Add files using upload-large-folder tool
Browse files- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-043437fb-fde6-4180-b927-bb5cfebf6b0e1759058799627-2025_09_28-13.27.24.771/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-04d29a91-f2bd-4b62-89da-678008b090861755160613254-2025_08_14-10.37.33.283/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-0538b58f-47c4-46c6-9641-bb5ac5ffb10f1759593994343-2025_10_04-18.07.24.830/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-1103382d-4338-4569-ad1a-fd99664b131a1756899628806-2025_09_03-13.40.59.36/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-33d627dc-2555-472b-a3d4-76d3b06878631756723263490-2025_09_01-12.41.38.380/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-50f57e52-47d1-4d1e-826b-d55baff9760f1752091399803-2025_07_09-22.05.03.836/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-5612308a-14dd-4f05-ae65-c6cd496f68351752499707410-2025_07_14-15.29.02.547/source.csv +127 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-619817ef-f0fa-4ff3-8f61-fe4646f7e2971752667688154-2025_07_16-14.09.00.461/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-69e8c2c1-3349-4fa0-950c-3777ceb9c3751758621585749-2025_09_23-12.01.04.60/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-7f667af1-e8dc-4534-8563-a4ef7be9de461754942097068-2025_08_11-21.55.24.740/source.csv +22 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-89289833-78d5-4138-b766-a3025aedfd811759399195678-2025_10_02-12.00.33.280/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-94286bd5-ed1d-4a7b-9c23-3104eeb712871757340853772-2025_09_08-16.14.51.932/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-9995c396-b28a-4817-9971-5d6e293ae32a1753180233918-2025_07_22-12.30.43.584/source.csv +2 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-a5aad456-d021-49a4-8368-8264b401bd3e1757237946327-2025_09_07-11.39.28.68/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-b85ae4be-e003-4b5f-b92d-85f73081a0e71757580020189-2025_09_11-10.41.19.400/source.csv +4 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c0873732-2064-4df8-be5c-5c4fe03049aa1751367864693-2025_07_01-13.05.12.330/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c190e976-4e73-4096-9dfa-66c3a0cce2061752157080469-2025_07_10-16.18.14.262/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c2f47c74-4812-499a-abea-eafb3b83db691752315370022-2025_07_12-12.17.39.814/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-e11dd94b-65ba-41d4-a0d1-6151e7930d741752738204303-2025_07_17-09.44.02.899/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-f514f25a-b687-4125-bb3e-747ec98309d21750930471988-2025_06_26-11.34.36.685/source.csv +4 -0
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-043437fb-fde6-4180-b927-bb5cfebf6b0e1759058799627-2025_09_28-13.27.24.771/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-04d29a91-f2bd-4b62-89da-678008b090861755160613254-2025_08_14-10.37.33.283/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-0538b58f-47c4-46c6-9641-bb5ac5ffb10f1759593994343-2025_10_04-18.07.24.830/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-1103382d-4338-4569-ad1a-fd99664b131a1756899628806-2025_09_03-13.40.59.36/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-33d627dc-2555-472b-a3d4-76d3b06878631756723263490-2025_09_01-12.41.38.380/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-50f57e52-47d1-4d1e-826b-d55baff9760f1752091399803-2025_07_09-22.05.03.836/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-5612308a-14dd-4f05-ae65-c6cd496f68351752499707410-2025_07_14-15.29.02.547/source.csv
ADDED
|
@@ -0,0 +1,127 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,206,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"3:29:02 PM [info] Activating crowd-code\n3:29:02 PM [info] Recording started\n3:29:02 PM [info] Initializing git provider using file system watchers...\n",Log,tab
|
| 3 |
+
3,365,"extension-output-pdoom-org.crowd-code-#1-crowd-code",150,0,"3:29:02 PM [info] Git repository found\n3:29:02 PM [info] Git provider initialized successfully\n3:29:02 PM [info] Initial git state: [object Object]\n",Log,content
|
| 4 |
+
4,2862,"TERMINAL",0,0,"/bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt",,terminal_command
|
| 5 |
+
5,2909,"TERMINAL",0,0,"]633;E;2025-07-14 15:29:05 /bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt;96649176-7412-48ad-98ea-7639505dc671]633;C",,terminal_output
|
| 6 |
+
6,2952,"TERMINAL",0,0,"]0;tum_cte0515@hkn1990:/hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash]633;D;0",,terminal_output
|
| 7 |
+
7,7671,"TERMINAL",0,0,"watch",,terminal_focus
|
| 8 |
+
8,24040,"TERMINAL",0,0,"salloc --time=10:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5",,terminal_command
|
| 9 |
+
9,24098,"TERMINAL",0,0,"]633;E;2025-07-14 15:29:26 salloc --time=10:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5;ba61a590-3751-4127-93b1-5716be1c7bd0]633;Csalloc: Pending job allocation 3344502\r\nsalloc: job 3344502 queued and waiting for resources\r\n",,terminal_output
|
| 10 |
+
10,25688,"TERMINAL",0,0,"^Csalloc: Job allocation 3344502 has been revoked.\r\nsalloc: Job aborted due to signal\r\n]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;1]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output
|
| 11 |
+
11,30368,"TERMINAL",0,0,"salloc --time=01:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5",,terminal_command
|
| 12 |
+
12,30400,"TERMINAL",0,0,"]633;E;2025-07-14 15:29:32 salloc --time=01:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5;ba61a590-3751-4127-93b1-5716be1c7bd0]633;Csalloc: Pending job allocation 3344503\r\nsalloc: job 3344503 queued and waiting for resources\r\n",,terminal_output
|
| 13 |
+
13,31244,"TERMINAL",0,0,"^Csalloc: Job allocation 3344503 has been revoked.\r\nsalloc: Job aborted due to signal\r\n]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;1",,terminal_output
|
| 14 |
+
14,35141,"TERMINAL",0,0,"salloc --time=01:00:00 --partition=dev_accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5",,terminal_command
|
| 15 |
+
15,35153,"TERMINAL",0,0,"]633;E;2025-07-14 15:29:37 salloc --time=01:00:00 --partition=dev_accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5;ba61a590-3751-4127-93b1-5716be1c7bd0]633;Csalloc: Granted job allocation 3344505\r\n",,terminal_output
|
| 16 |
+
16,35274,"TERMINAL",0,0,"salloc: Waiting for resource configuration\r\n",,terminal_output
|
| 17 |
+
17,37228,"TERMINAL",0,0,"bash",,terminal_focus
|
| 18 |
+
18,38854,"TERMINAL",0,0,"queue",,terminal_command
|
| 19 |
+
19,38926,"TERMINAL",0,0,"]633;E;2025-07-14 15:29:41 queue;450fbccd-48d1-432b-b44b-06394616d158]633;C[?1049h[22;0;0t[1;18r(B[m[4l[?7h[H[2JEvery 1.0s: squeue --me[1;162Hhkn1990.localdomain: Mon Jul 14 15:29:41 2025[3;14HJOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)[4;12H3341079 accelerat train_to tum_cte0 R 1-04:59:15\t 2 hkn[0628-0629][5;12H3341080 accelerat train_to tum_cte0 R 1-04:59:15\t 2 hkn[0631-0632][6;12H3344246 accelerat interact tum_cte0 R 1:13:32\t 2 hkn[0626,0630][7;12H3344505 dev_accel interact tum_cte0 R\t0:04\t 1 hkn0402[18;206H",,terminal_output
|
| 20 |
+
20,39972,"TERMINAL",0,0,"[1;201H2[4;60H6[5d6[6d3[7d5[18;206H",,terminal_output
|
| 21 |
+
21,41004,"TERMINAL",0,0,"[1;201H3[4;60H7[5d7[6d4[7d6[18;206H",,terminal_output
|
| 22 |
+
22,42041,"TERMINAL",0,0,"[1;201H4[4;60H8[5d8[6d5[7d7[18;206H",,terminal_output
|
| 23 |
+
23,43101,"TERMINAL",0,0,"[1;201H5[4;60H9[5d9[6d6[7d8[18;206H",,terminal_output
|
| 24 |
+
24,44132,"TERMINAL",0,0,"[1;201H6[4;59H20[5d20[6d7[7d9[18;206H",,terminal_output
|
| 25 |
+
25,45165,"TERMINAL",0,0,"[1;201H7[4;60H1[5d1[6d8[7d10[18;206H",,terminal_output
|
| 26 |
+
26,46236,"TERMINAL",0,0,"[1;201H8[4;60H2[5d2[6d9[7d1[18;206H",,terminal_output
|
| 27 |
+
27,46520,"TERMINAL",0,0,"[18;1H[?1049l[23;0;0t\r[?1l>]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;0",,terminal_output
|
| 28 |
+
28,49617,"TERMINAL",0,0,"idling",,terminal_command
|
| 29 |
+
29,49671,"TERMINAL",0,0,"]633;E;2025-07-14 15:29:52 idling;450fbccd-48d1-432b-b44b-06394616d158]633;C[?1049h[22;0;0t[1;18r(B[m[4l[?7h[H[2JEvery 1.0s: sinfo_t_idle[1;162Hhkn1990.localdomain: Mon Jul 14 15:29:52 2025[3;1HPartition dev_cpuonly[3;35H: 10 nodes idle\r[4dPartition cpuonly[4;35H: 42 nodes idle\r[5dPartition dev_accelerated[5;35H:\t 1 nodes idle\r[6dPartition accelerated[6;35H: 40 nodes idle\r[7dPartition dev_accelerated-h100 :\t 1 nodes idle\r[8dPartition accelerated-h100[8;35H:\t 0 nodes idle\r[9dPartition large[9;35H:\t 7 nodes idle[18;206H",,terminal_output
|
| 30 |
+
30,50700,"TERMINAL",0,0,"[1;201H3[18;206H",,terminal_output
|
| 31 |
+
31,51749,"TERMINAL",0,0,"[1;201H4[18;206H",,terminal_output
|
| 32 |
+
32,52802,"TERMINAL",0,0,"[1;201H5[18;206H",,terminal_output
|
| 33 |
+
33,53446,"TERMINAL",0,0,"[18;1H[?1049l[23;0;0t\r[?1l>]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;0",,terminal_output
|
| 34 |
+
34,62317,"TERMINAL",0,0,"salloc: Nodes hkn0402 are ready for job\r\n",,terminal_output
|
| 35 |
+
35,63118,"TERMINAL",0,0,"]0;tum_cte0515@hkn0402:~/Projects/jafar[?2004h[tum_cte0515@hkn0402 jafar]$ ",,terminal_output
|
| 36 |
+
36,84352,"TERMINAL",0,0,"srun",,terminal_focus
|
| 37 |
+
37,85382,"TERMINAL",0,0,"[?25ls[2mo[22m[45;32H[?25h[?25l[45;31Ho[45;32H[?25h",,terminal_output
|
| 38 |
+
38,85520,"TERMINAL",0,0,"[?25l[45;32Hu[45;33H[?25h",,terminal_output
|
| 39 |
+
39,85615,"TERMINAL",0,0,"[?25l[45;33Hr[45;34H[?25h",,terminal_output
|
| 40 |
+
40,85840,"TERMINAL",0,0,"[?25l[45;34Hc[45;35H[?25h",,terminal_output
|
| 41 |
+
41,85947,"TERMINAL",0,0,"[?25l[45;35He[45;36H[?25h",,terminal_output
|
| 42 |
+
42,86009,"TERMINAL",0,0,"[?25l[45;36H [45;37H[?25h",,terminal_output
|
| 43 |
+
43,86141,"TERMINAL",0,0,"[?25l[45;37H.[45;39H[?25h[?25l[45;38Hv[45;39H[?25h",,terminal_output
|
| 44 |
+
44,86295,"TERMINAL",0,0,"env/",,terminal_output
|
| 45 |
+
45,86587,"TERMINAL",0,0,"",,terminal_output
|
| 46 |
+
46,86945,"TERMINAL",0,0,"[?25l[45;43Hb[45;44H[?25h",,terminal_output
|
| 47 |
+
47,87054,"TERMINAL",0,0,"in/",,terminal_output
|
| 48 |
+
48,87522,"TERMINAL",0,0,"[?25l[45;47Ha[45;48H[?25h",,terminal_output
|
| 49 |
+
49,87637,"TERMINAL",0,0,"[?25l[45;48Hc[45;49H[?25h",,terminal_output
|
| 50 |
+
50,87784,"TERMINAL",0,0,"tivate",,terminal_output
|
| 51 |
+
51,88103,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0402:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0402 jafar]$ ",,terminal_output
|
| 52 |
+
52,94820,"sample.py",0,0,"from dataclasses import dataclass\nimport time\nimport os\n\nimport dm_pix as pix\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\nimport numpy as np\nfrom orbax.checkpoint import PyTreeCheckpointer\nfrom flax.training.train_state import TrainState\nimport grain\nimport orbax.checkpoint as ocp\nimport optax\nfrom PIL import Image, ImageDraw\nimport tyro\n\nfrom genie import Genie\nfrom utils.dataloader import get_dataloader\n\n\n@dataclass\nclass Args:\n # Experiment\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = ""data/coinrun_episodes""\n checkpoint: str = """"\n # Sampling\n batch_size: int = 1\n maskgit_steps: int = 25\n temperature: float = 1.0\n sample_argmax: bool = True\n start_frame: int = 0\n # Tokenizer checkpoint\n tokenizer_dim: int = 512\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 8\n tokenizer_num_heads: int = 8\n # LAM checkpoint\n lam_dim: int = 512\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 8\n lam_num_heads: int = 8\n # Dynamics checkpoint\n dyna_dim: int = 512\n dyna_num_blocks: int = 12\n dyna_num_heads: int = 8\n\n\nargs = tyro.cli(Args)\nrng = jax.random.PRNGKey(args.seed)\n\n# --- Load Genie checkpoint ---\ngenie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=True,\n # Dynamics\n dyna_dim=args.dyna_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n)\nrng, _rng = jax.random.split(rng)\nimage_shape = (args.image_height, args.image_width, args.image_channels)\ndummy_inputs = dict(\n videos=jnp.zeros((args.batch_size, args.seq_len, *image_shape), dtype=jnp.float32),\n mask_rng=_rng,\n)\nrng, _rng = jax.random.split(rng)\nparams = genie.init(_rng, dummy_inputs)\n\n# ckpt = PyTreeCheckpointer().restore(args.checkpoint)[""model""][""params""][""params""]\n# Create a dummy TrainState for checkpoint structure\nlr_schedule = optax.warmup_cosine_decay_schedule(\n 0.0, 3e-4, 1000, 300000\n)\ntx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4)\ndummy_train_state = TrainState.create(\n apply_fn=genie.apply,\n params=params, # or params if your model expects that\n tx=tx, # No optimizer needed for eval\n)\n\nhandler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\nhandler_registry.add('model_state', ocp.args.StandardSave, ocp.handlers.StandardCheckpointHandler)\nhandler_registry.add('model_state', ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler)\n\n\n# Set up Orbax CheckpointManager\ncheckpoint_manager = ocp.CheckpointManager(\n args.checkpoint, # Directory containing the checkpoint\n options=ocp.CheckpointManagerOptions(max_to_keep=1, step_format_fixed_length=6),\n handler_registry=handler_registry\n)\n\n# Prepare abstract state for restore\nabstract_train_state = jax.tree_util.tree_map(\n ocp.utils.to_shape_dtype_struct, dummy_train_state\n)\n\n# Restore the checkpoint (only model_state needed for sampling)\nrestored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.StandardRestore(abstract_train_state),\n ),\n)\nrestored_train_state = restored[""model_state""]\n\n# Extract model parameters for inference\nparams = restored_train_state.params\n\n\n# params[""params""].update(ckpt)\n\n\ndef _sampling_wrapper(module, batch):\n return module.sample(batch, args.seq_len, args.maskgit_steps, args.temperature, args.sample_argmax)\n\n# --- Define autoregressive sampling loop ---\ndef _autoreg_sample(rng, video_batch, action_batch):\n vid = video_batch[:, : args.start_frame + 1]\n sampling_fn = jax.jit(nn.apply(_sampling_wrapper, genie)) \n rng, _rng = jax.random.split(rng)\n batch = dict(videos=vid, latent_actions=action_batch, rng=_rng)\n generated_vid = sampling_fn(\n params,\n batch\n )\n return generated_vid\n\n# # --- Get video + latent actions ---\narray_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n]\ngrain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n)\ninitial_state = grain_dataloader._create_initial_state()\ngrain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\njax.debug.breakpoint()\nvideo_batch = next(grain_iterator)\njax.debug.breakpoint()\n# video_batch = np.load(""overfit_dir/single_sample_corner.npy"")\n# Get latent actions for all videos in the batch\nbatch = dict(videos=video_batch)\naction_batch = genie.apply(params, batch, False, method=Genie.vq_encode)\naction_batch = action_batch.reshape(video_batch.shape[0], args.seq_len - 1, 1)\n\n# --- Sample + evaluate video ---\nvid = _autoreg_sample(rng, video_batch, action_batch)\ngt = video_batch[:, : vid.shape[1]].clip(0, 1).reshape(-1, *video_batch.shape[2:])\nrecon = vid.clip(0, 1).reshape(-1, *vid.shape[2:])\nssim = pix.ssim(gt[:, args.start_frame + 1 :], recon[:, args.start_frame + 1 :]).mean()\nprint(f""SSIM: {ssim}"")\n\n# --- Construct video ---\ntrue_videos = (video_batch * 255).astype(np.uint8)\npred_videos = (vid * 255).astype(np.uint8)\nvideo_comparison = np.zeros((2, *vid.shape), dtype=np.uint8)\nvideo_comparison[0] = true_videos[:, :args.seq_len]\nvideo_comparison[1] = pred_videos\nframes = einops.rearrange(video_comparison, ""n b t h w c -> t (b h) (n w) c"")\n\n# --- Save video --- \nimgs = [Image.fromarray(img) for img in frames]\n# Write actions on each frame, on each row (i.e., for each video in the batch, on the GT row)\nfor t, img in enumerate(imgs[1:]):\n d = ImageDraw.Draw(img)\n for row in range(action_batch.shape[0]):\n action = action_batch[row, t, 0]\n y_offset = row * video_batch.shape[2] + 2\n d.text((2, y_offset), f""{action}"", fill=255)\nimgs[0].save(\n f""generation_{time.time()}.gif"",\n save_all=True,\n append_images=imgs[1:],\n duration=250,\n loop=0,\n)\n",python,tab
|
| 53 |
+
53,94915,"TERMINAL",0,0,"\r[K(jafar) [tum_cte0515@hkn0402 jafar]$ \r[K(jafar) [tum_cte0515@hkn0402 jafar]$ ",,terminal_output
|
| 54 |
+
54,148636,"TERMINAL",0,0,"source .venv/bin/activate",,terminal_output
|
| 55 |
+
55,148874,"TERMINAL",0,0,"idling[K",,terminal_output
|
| 56 |
+
56,149041,"TERMINAL",0,0,"python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output
|
| 57 |
+
57,160599,"TERMINAL",0,0,"bash",,terminal_focus
|
| 58 |
+
58,164468,"TERMINAL",0,0,"tmux a",,terminal_command
|
| 59 |
+
59,164479,"TERMINAL",0,0,"]633;E;2025-07-14 15:31:46 tmux a;450fbccd-48d1-432b-b44b-06394616d158]633;Cno sessions\r\n]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;1",,terminal_output
|
| 60 |
+
60,176523,"TERMINAL",0,0,"srun",,terminal_focus
|
| 61 |
+
61,181253,"TERMINAL",0,0,"[A[A[Aidling[K\r\n\r[K\r\n\r[K\r\n\r[K[A[A[A[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C",,terminal_output
|
| 62 |
+
62,182032,"TERMINAL",0,0,"source .venv/bin/activate",,terminal_output
|
| 63 |
+
63,182304,"TERMINAL",0,0,"[K",,terminal_output
|
| 64 |
+
64,182514,"TERMINAL",0,0,"",,terminal_output
|
| 65 |
+
65,183774,"TERMINAL",0,0,"\r(reverse-i-search)`': [K",,terminal_output
|
| 66 |
+
66,184082,"TERMINAL",0,0,"m': python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_[7mm[27minecraft_arrayrecords_chunked[A[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C",,terminal_output
|
| 67 |
+
67,184283,"TERMINAL",0,0,"[A[A\r[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[Co': python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_[7mmo[27mdelsize_scaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked[A[A[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C",,terminal_output
|
| 68 |
+
68,184445,"TERMINAL",0,0,"[A\r[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[Cd': python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_[7mmod[27melsize_scaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked[A[A[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C",,terminal_output
|
| 69 |
+
69,184620,"TERMINAL",0,0,"[?25l[43;85H[7;39;49mm[43;85H[0m[A\r[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[46Pu': [7mmodu[27mle unload devel/cuda/12.4\r\n\r[K\r\n\r[K\r\n\r[K[A[A[A[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[?25h",,terminal_output
|
| 70 |
+
70,184813,"TERMINAL",0,0,"[1@l': [7mmodul[27m",,terminal_output
|
| 71 |
+
71,184920,"TERMINAL",0,0,"[?25l[42;28H[7;39;49mm[42;28H[0m[1@e': [7mmodule[27m[?25h",,terminal_output
|
| 72 |
+
72,186275,"TERMINAL",0,0,"[?25l[42;29H\r[C[9@jafar) [tum_cte0515@hkn0402 jafar]$ module\r\n[?2004l\r[?25h",,terminal_output
|
| 73 |
+
73,186368,"TERMINAL",0,0,"[1;31mLmod Warning: [0m\r\n[1;31m----------------------------------------------------------------------------------------------------[0m\r\nThe following dependent module(s) are not currently loaded: devel/cuda/12.4 (required by:\r\nmpi/openmpi/5.0)\r\n[1;31m----------------------------------------------------------------------------------------------------[0m\r\n\r\n\r\n\r\n]0;tum_cte0515@hkn0402:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0402 jafar]$ ",,terminal_output
|
| 74 |
+
74,187458,"TERMINAL",0,0,"m",,terminal_output
|
| 75 |
+
75,188318,"TERMINAL",0,0,"\r(reverse-i-search)`': [K",,terminal_output
|
| 76 |
+
76,188537,"TERMINAL",0,0,"[?25l[45;23H[X[0mm': [7mm[27module unload devel/cuda/12.4\r[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[?25h",,terminal_output
|
| 77 |
+
77,188712,"TERMINAL",0,0,"[?25l[45;24H[7;39;49mm[45;24H[0m[1@o': [7mmo[27m[?25h",,terminal_output
|
| 78 |
+
78,188827,"TERMINAL",0,0,"[?25l[45;25H[7;39;49mm[45;25H[0m[1@d': [7mmod[27m[?25h",,terminal_output
|
| 79 |
+
79,188885,"TERMINAL",0,0,"[?25l[45;26H[7;39;49mm[45;26H[0me': python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_[7mmode[27mlsize_scaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked[A[A[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[?25h",,terminal_output
|
| 80 |
+
80,189124,"TERMINAL",0,0,"[A\r[Cfailed reverse-i-search)`modeu': python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize[Cscaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked[A[A[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C",,terminal_output
|
| 81 |
+
81,189331,"TERMINAL",0,0,"[?25l[43;94H[0mm[43;94H[0m[A\r[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[Cl': python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked[A[A[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[?25h",,terminal_output
|
| 82 |
+
82,190338,"TERMINAL",0,0,"^C\r\n\r\n\r[?2004l\r[?2004h[?2004l\r\r\n]0;tum_cte0515@hkn0402:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0402 jafar]$ ",,terminal_output
|
| 83 |
+
83,190778,"TERMINAL",0,0,"\r(reverse-i-search)`': [K",,terminal_output
|
| 84 |
+
84,190999,"TERMINAL",0,0,"m': [7mm[27module unload devel/cuda/12.4\r[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C",,terminal_output
|
| 85 |
+
85,191308,"TERMINAL",0,0,"[?25l[45;24H[7;39;49mm[45;24H[0m[1@o': [7mmo[27m[?25h",,terminal_output
|
| 86 |
+
86,192625,"TERMINAL",0,0,"python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_[7mmo[27mdelsize_scaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked[A[A[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C",,terminal_output
|
| 87 |
+
87,193357,"TERMINAL",0,0,"[7mmo[27mdelsize-scaling/train_dynamics_modelsize_scaling_3[A[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C",,terminal_output
|
| 88 |
+
88,193987,"TERMINAL",0,0,"modelsize-scaling/train_dynamics_[7mmo[27mdelsize_scaling_3[C[C[C[C[C[C[C[C[C[C087000 --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked[A[A[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C",,terminal_output
|
| 89 |
+
89,194447,"TERMINAL",0,0,"[7mmo[27mdelsize-scaling/train_dynamics_modelsize_scaling_3[A[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C",,terminal_output
|
| 90 |
+
90,194807,"TERMINAL",0,0,"[A\r[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[48P[7mmo[27mdule unload devel/cuda/12.4\r\n\r[K\r\n\r[K\r\n\r[K[A[A[A[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C",,terminal_output
|
| 91 |
+
91,195611,"TERMINAL",0,0,"[C[C[C[C[C[C[C[C[C[C[C[C[C[Cmpi/openmpi/5.0\r[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C",,terminal_output
|
| 92 |
+
92,196151,"TERMINAL",0,0,"[?25l[42;25H\r[C[13@jafar) [tum_cte0515@hkn0402 jafar]$ mo\r\n[?2004l\r[?25h",,terminal_output
|
| 93 |
+
93,196276,"TERMINAL",0,0,"]0;tum_cte0515@hkn0402:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0402 jafar]$ ",,terminal_output
|
| 94 |
+
94,197041,"TERMINAL",0,0,"module unload mpi/openmpi/5.0",,terminal_output
|
| 95 |
+
95,197244,"TERMINAL",0,0,"devel/cuda/12.4",,terminal_output
|
| 96 |
+
96,197809,"TERMINAL",0,0,"[?25l[?2004l\r[?25h\r\nNote: the module ""devel/cuda/12.4"" cannot be unloaded because it was not loaded.\r\n\r\n]0;tum_cte0515@hkn0402:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0402 jafar]$ ",,terminal_output
|
| 97 |
+
97,202765,"TERMINAL",0,0,"module unload devel/cuda/12.4",,terminal_output
|
| 98 |
+
98,202990,"TERMINAL",0,0,"mpi/openmpi/5.0",,terminal_output
|
| 99 |
+
99,203154,"TERMINAL",0,0,"devel/cuda/12.4",,terminal_output
|
| 100 |
+
100,203381,"TERMINAL",0,0,"[4Psource .venv/bin/activate",,terminal_output
|
| 101 |
+
101,203782,"TERMINAL",0,0,"idling[K",,terminal_output
|
| 102 |
+
102,204328,"TERMINAL",0,0,"python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output
|
| 103 |
+
103,205187,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output
|
| 104 |
+
104,206994,"sample.py",0,0,"",python,tab
|
| 105 |
+
105,206997,"sample.py",5254,0,"",python,selection_mouse
|
| 106 |
+
106,207025,"sample.py",5253,0,"",python,selection_command
|
| 107 |
+
107,218430,"TERMINAL",0,0,"2025-07-14 15:32:40.851942: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output
|
| 108 |
+
108,231780,"TERMINAL",0,0,"2025-07-14 15:32:54.216693: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output
|
| 109 |
+
109,248304,"TERMINAL",0,0,"2025-07-14 15:33:10.775327: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output
|
| 110 |
+
110,252736,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 91000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/091000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 89000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/089000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 90000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/090000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/080000/metrics/metrics not found.\r\n",,terminal_output
|
| 111 |
+
111,273123,"sample.py",5019,0,"",python,selection_mouse
|
| 112 |
+
112,273126,"sample.py",5018,0,"",python,selection_command
|
| 113 |
+
113,273749,"sample.py",5254,0,"",python,selection_mouse
|
| 114 |
+
114,273751,"sample.py",5253,0,"",python,selection_command
|
| 115 |
+
115,277876,"TERMINAL",0,0,"bash",,terminal_focus
|
| 116 |
+
116,283470,"TERMINAL",0,0,"cd $ws_dir",,terminal_command
|
| 117 |
+
117,284396,"TERMINAL",0,0,"cd logs/",,terminal_command
|
| 118 |
+
118,284816,"TERMINAL",0,0,"ls",,terminal_command
|
| 119 |
+
119,284828,"TERMINAL",0,0,"]633;E;2025-07-14 15:33:47 ls;450fbccd-48d1-432b-b44b-06394616d158]633;C[0m[01;34m3306965[0m [01;34mlogs_alfred[0m [01;34mlogs_franz[0m [01;34mlogs_mihir[0m\r\n]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs]633;D;0",,terminal_output
|
| 120 |
+
120,287852,"TERMINAL",0,0,"Entering jdb:\r\n(jdb) ",,terminal_output
|
| 121 |
+
121,287863,"TERMINAL",0,0,"cd logs_mihir/",,terminal_command
|
| 122 |
+
122,287879,"TERMINAL",0,0,"]633;E;2025-07-14 15:33:50 cd logs_mihir/;450fbccd-48d1-432b-b44b-06394616d158]633;C]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output
|
| 123 |
+
123,288184,"TERMINAL",0,0,"ls",,terminal_command
|
| 124 |
+
124,288235,"TERMINAL",0,0,"]633;E;2025-07-14 15:33:50 ls;450fbccd-48d1-432b-b44b-06394616d158]633;C",,terminal_output
|
| 125 |
+
125,288365,"TERMINAL",0,0,"[0m[01;34mbig_run[0m train_lam_action_space_scaling_20_3329788.log train_lam_model_size_scaling_38M_3317098.log train_tokenizer_model_size_scaling_140M_3316019.log\r\n[01;34mbig-runs[0m train_lam_action_space_scaling_20_3329803.log train_lam_model_size_scaling_38M_3317115.log train_tokenizer_model_size_scaling_200M_3313563.log\r\ntrain_dyn_yolorun_3333026.log train_lam_action_space_scaling_20_3331285.log train_lam_model_size_scaling_38M_3317231.log train_tokenizer_model_size_scaling_200M_3316020.log\r\ntrain_dyn_yolorun_3333448.log train_lam_action_space_scaling_50_3320180.log train_tokenizer_batch_size_scaling_16_node_3321526.log train_tokenizer_model_size_scaling_227M_3317234.log\r\ntrain_dyn_yolorun_3335345.log train_lam_action_space_scaling_50_3329789.log train_tokenizer_batch_size_scaling_1_node_3318551.log train_tokenizer_model_size_scaling_227M_3318555.log\r\ntrain_dyn_yolorun_3335362.log train_lam_action_space_scaling_50_3329804.log train_tokenizer_batch_size_scaling_2_node_3318552.log train_tokenizer_model_size_scaling_227M_3320173.log\r\ntrain_lam_action_space_scaling_10_3320179.log train_lam_action_space_scaling_50_3331286.log train_tokenizer_batch_size_scaling_2_node_3330806.log train_tokenizer_model_size_scaling_227M_3321523.log\r\ntrain_lam_action_space_scaling_10_3321529.log train_lam_action_space_scaling_6_3318549.log train_tokenizer_batch_size_scaling_2_node_3330848.log train_tokenizer_model_size_scaling_37M_3313565.log\r\ntrain_lam_action_space_scaling_10_3329786.log train_lam_action_space_scaling_6_3320178.log train_tokenizer_batch_size_scaling_2_node_3331282.log train_tokenizer_model_size_scaling_37M_3316022.log\r\ntrain_lam_action_space_scaling_10_3329801.log train_lam_action_space_scaling_6_3321528.log train_tokenizer_batch_size_scaling_4_node_3318553.log train_tokenizer_model_size_scaling_37M_3317232.log\r\ntrain_lam_action_space_scaling_10_3331283.log train_lam_action_space_scaling_6_3329790.log train_tokenizer_batch_size_scaling_4_node_3320175.log train_tokenizer_model_size_scaling_37M_3317239.log\r\ntrain_lam_action_space_scaling_12_3318546.log train_lam_action_space_scaling_6_3329805.log train_tokenizer_batch_size_scaling_4_node_3321524.log train_tokenizer_model_size_scaling_37M_3318556.log\r\ntrain_lam_action_space_scaling_12_3320177.log train_lam_action_space_scaling_6_3331287.log train_tokenizer_batch_size_scaling_8_node_3320176.log train_tokenizer_model_size_scaling_74M_3318557.log\r\ntrain_lam_action_space_scaling_12_3321527.log train_lam_action_space_scaling_8_3318550.log train_tokenizer_batch_size_scaling_8_node_3321525.log train_tokenizer_model_size_scaling_74M_3320174.log\r\ntrain_lam_action_space_scaling_12_3329787.log train_lam_action_space_scaling_8_3329791.log train_tokenizer_minecraft_overfit_sample_3309656.log train_tokenizer_model_size_scaling_74M_3321522.log\r\ntrain_lam_action_space_scaling_12_3329802.log train_lam_action_space_scaling_8_3329806.log train_tokenizer_model_size_scaling_127M_3317233.log train_tokenizer_model_size_scaling_80M_3313564.log\r\ntrain_lam_action_space_scaling_12_3331284.log train_lam_action_space_scaling_8_3331288.log train_tokenizer_model_size_scaling_127M_3318554.log train_tokenizer_model_size_scaling_80M_3316026.log\r\ntrain_lam_action_space_scaling_20_3318547.log train_lam_minecraft_overfit_sample_3309655.log train_tokenizer_model_size_scaling_140M_3313562.log\r\n]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output
|
| 126 |
+
126,291553,"TERMINAL",0,0,"cd big-runs/",,terminal_command
|
| 127 |
+
127,292418,"TERMINAL",0,0,"ls",,terminal_command
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-619817ef-f0fa-4ff3-8f61-fe4646f7e2971752667688154-2025_07_16-14.09.00.461/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-69e8c2c1-3349-4fa0-950c-3777ceb9c3751758621585749-2025_09_23-12.01.04.60/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-7f667af1-e8dc-4534-8563-a4ef7be9de461754942097068-2025_08_11-21.55.24.740/source.csv
ADDED
|
@@ -0,0 +1,22 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,461,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"9:55:24 PM [info] Activating crowd-code\n9:55:24 PM [info] Recording started\n9:55:24 PM [info] Initializing git provider using file system watchers...\n9:55:24 PM [info] Git repository found\n9:55:24 PM [info] Git provider initialized successfully\n",Log,tab
|
| 3 |
+
3,483,"TERMINAL",0,0,"git branch",,terminal_command
|
| 4 |
+
4,537,"TERMINAL",0,0,"]633;E;2025-08-11 21:55:24 git branch;686e78ec-8807-43f4-8f60-628f6dc2af90]633;C[?1h=\r",,terminal_output
|
| 5 |
+
5,732,"TERMINAL",0,0,"* [32m(HEAD detached at ab43d16)[m[m\r\n add-wandb-name-and-tags[m[m\r\n before-nnx[m[m\r\n causal-mem-reduce[m[m\r\n causal-spatiotemporal-kv-cache[m[m\r\n causal-st-transformer[m[m\r\n causal-transformer-dynamics-model[m[m\r\n causal-transformer-nnx-no-kv-cache[m[m\r\n convert-to-jax-array-in-iter[m[m\r\n correct-batched-sampling[m[m\r\n dev[m[m\r\n dont-let-tf-see-gpu[m[m\r\n feat/explicit-image-dims[m[m\r\n fix-action-padding-lam-future-information-access[m[m\r\n fix-sampling[m[m\r\n fix-transformer-forwardpass[m[m\r\n fix/spatiotemporal-pe-once-in-STTransformer[m[m\r\n grad-norm-log-and-clip[m[m\r\n grain-dataloader[m[m\r\n input_pipeline/add-npy2array_record[m[m\r\n logging-variants[m[m\r\n lr-schedules[m[m\r\n main[m[m\r\n maskgit-different-maskprob-per-sample[m[m\r\n:[K",,terminal_output
|
| 6 |
+
6,2249,"TERMINAL",0,0,"\r[K maskgit-sampling-iterative-unmasking-fix[m[m\r\n:[K",,terminal_output
|
| 7 |
+
7,2427,"TERMINAL",0,0,"\r[K metrics-logging-for-dynamics-model[m[m\r\n:[K",,terminal_output
|
| 8 |
+
8,2573,"TERMINAL",0,0,"\r[K monkey-patch[m[m\r\n:[K",,terminal_output
|
| 9 |
+
9,2811,"TERMINAL",0,0,"\r[K new-arch-sampling[m[m\r\n:[K",,terminal_output
|
| 10 |
+
10,2907,"TERMINAL",0,0,"\r[K preprocess_video[m[m\r\n:[K",,terminal_output
|
| 11 |
+
11,2986,"TERMINAL",0,0,"\r[K refactor-tmp[m[m\r\n:[K",,terminal_output
|
| 12 |
+
12,3470,"TERMINAL",0,0,"\r[K[HM causal-st-transformer[m[m\r\n[25;1H\r[K:[K",,terminal_output
|
| 13 |
+
13,4286,"TERMINAL",0,0,"\r[K[HM causal-spatiotemporal-kv-cache[m[m\r\n[25;1H\r[K:[K\r[K[HM causal-mem-reduce[m[m\r\n[25;1H\r[K:[K\r[K[HM before-nnx[m[m\r\n[25;1H\r[K:[K\r[K[HM add-wandb-name-and-tags[m[m\r\n[25;1H\r[K:[K\r[K[HM* [32m(HEAD detached at ab43d16)[m[m\r\n[25;1H\r[K:[K\r[K\r[K:[K\r[K\r[K:[K\r[K\r[K:[K\r[K\r[K:[K\r[K\r[K:[K\r[K\r[K:[K\r[K\r[K:[K",,terminal_output
|
| 14 |
+
14,4417,"TERMINAL",0,0,"\r[K\r[K:[K\r[K\r[K:[K\r[K\r[K:[K\r[K\r[K:[K\r[K\r[K:[K",,terminal_output
|
| 15 |
+
15,4470,"TERMINAL",0,0,"\r[K\r[K:[K\r[K\r[K:[K",,terminal_output
|
| 16 |
+
16,4524,"TERMINAL",0,0,"\r[K\r[K:[K",,terminal_output
|
| 17 |
+
17,4848,"TERMINAL",0,0,"\r[K[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output
|
| 18 |
+
18,7630,"TERMINAL",0,0,"git checkout main",,terminal_command
|
| 19 |
+
19,7730,"TERMINAL",0,0,"]633;E;2025-08-11 21:55:32 git checkout main;686e78ec-8807-43f4-8f60-628f6dc2af90]633;Cerror: Your local changes to the following files would be overwritten by checkout:\r\n\ttrain_dynamics.py\r\nPlease commit your changes or stash them before you switch branches.\r\nAborting\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;1",,terminal_output
|
| 20 |
+
20,11378,"TERMINAL",0,0,"git stash",,terminal_command
|
| 21 |
+
21,11436,"TERMINAL",0,0,"]633;E;2025-08-11 21:55:36 git stash;686e78ec-8807-43f4-8f60-628f6dc2af90]633;C",,terminal_output
|
| 22 |
+
22,11610,"TERMINAL",0,0,"Saved working directory and index state WIP on (no branch): ab43d16 fix: missing reshape and shape suffixes in `sample.py` (#134)\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-89289833-78d5-4138-b766-a3025aedfd811759399195678-2025_10_02-12.00.33.280/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-94286bd5-ed1d-4a7b-9c23-3104eeb712871757340853772-2025_09_08-16.14.51.932/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-9995c396-b28a-4817-9971-5d6e293ae32a1753180233918-2025_07_22-12.30.43.584/source.csv
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,58754,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"12:30:43 PM [info] Activating crowd-code\n12:30:43 PM [info] Recording started\n12:30:43 PM [info] Initializing git provider using file system watchers...\n12:30:45 PM [info] Retrying git provider initialization...\n12:30:46 PM [error] Not a git repository: EntryNotFound (FileSystemError): Error: ENOENT: no such file or directory, stat '/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/train_tokenizer_lr_sweep_1e-4/.git'\n12:30:47 PM [error] Not a git repository: EntryNotFound (FileSystemError): Error: ENOENT: no such file or directory, stat '/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/train_tokenizer_lr_sweep_1e-4/.git'\n",Log,tab
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-a5aad456-d021-49a4-8368-8264b401bd3e1757237946327-2025_09_07-11.39.28.68/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-b85ae4be-e003-4b5f-b92d-85f73081a0e71757580020189-2025_09_11-10.41.19.400/source.csv
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,15,"train_dynamics.py",0,0,"from dataclasses import dataclass, field\nimport os\nimport time\n\nimport einops\nfrom flax.training import orbax_utils\nfrom flax.training.train_state import TrainState\nimport optax\nimport orbax\nimport numpy as np\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\n\nfrom genie import Genie, restore_genie_components\nfrom utils.dataloader import get_dataloader\n\nts = int(time.time())\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_resolution: int = 64\n data_dir: str = ""data/coinrun_episodes""\n # Optimization\n batch_size: int = 36\n min_lr: float = 3e-6\n max_lr: float = 3e-5\n warmup_steps: int = 5000\n # Tokenizer\n tokenizer_dim: int = 512\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 8\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 8\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_dim: int = 512\n dyna_num_blocks: int = 12\n dyna_num_heads: int = 8\n dropout: float = 0.0\n mask_limit: float = 0.5\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_gradients: bool = False\n name: str = """"\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n wandb_id: str = """"\n\nargs = tyro.cli(Args)\n\n\ndef dynamics_loss_fn(params, state, inputs):\n """"""Compute masked dynamics loss""""""\n outputs = state.apply_fn(\n params, inputs, training=True, rngs={""dropout"": inputs[""dropout_rng""]}\n )\n mask = outputs[""mask""]\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask * ce_loss).sum() / mask.sum()\n acc = outputs[""token_logits""].argmax(-1) == outputs[""video_tokens""]\n acc = (mask * acc).sum() / mask.sum()\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_accuracy=acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n )\n return ce_loss, (outputs[""recon""], metrics)\n\n\n@jax.jit\ndef train_step(state, inputs):\n """"""Update state and compute metrics""""""\n grad_fn = jax.value_and_grad(dynamics_loss_fn, has_aux=True, allow_int=True)\n (loss, (recon, metrics)), grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return state, loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n rng = jax.random.PRNGKey(args.seed)\n if args.log:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n # --- Initialize model ---\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n # Dynamics\n dyna_dim=args.dyna_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n mask_limit=args.mask_limit,\n )\n rng, _rng = jax.random.split(rng)\n image_shape = (args.image_resolution, args.image_resolution, args.image_channels)\n dummy_inputs = dict(\n videos=jnp.zeros(\n (args.batch_size, args.seq_len, *image_shape), dtype=jnp.float32\n ),\n mask_rng=_rng,\n )\n rng, _rng = jax.random.split(rng)\n init_params = genie.init(_rng, dummy_inputs)\n init_params = restore_genie_components(\n init_params, args.tokenizer_checkpoint, args.lam_checkpoint\n )\n\n # --- Initialize optimizer ---\n lr_schedule = optax.warmup_cosine_decay_schedule(\n args.min_lr, args.max_lr, args.warmup_steps, args.num_steps\n )\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4)\n train_state = TrainState.create(apply_fn=genie.apply, params=init_params, tx=tx)\n\n # --- TRAIN LOOP ---\n dataloader = get_dataloader(args.data_dir, args.seq_len, args.batch_size)\n step = 0\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng, _mask_rng = jax.random.split(rng, 3)\n inputs = dict(\n videos=videos,\n action=jnp.zeros((args.batch_size, args.seq_len), dtype=jnp.float32),\n dropout_rng=_rng,\n mask_rng=_mask_rng,\n )\n train_state, loss, recon, metrics = train_step(train_state, inputs)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0:\n wandb.log({""loss"": loss, ""step"": step, **metrics})\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0]\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[15])),\n recon=wandb.Image(np.asarray(recon_seq[15])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n if step % args.log_checkpoint_interval == 0:\n ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""genie_{ts}_{step}""),\n ckpt,\n save_args=save_args,\n )\n if step >= args.num_steps:\n break\n",python,tab
|
| 3 |
+
2,910,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"10:41:19 AM [info] Activating crowd-code\n10:41:19 AM [info] Recording started\n10:41:19 AM [info] Initializing git provider using file system watchers...\n",Log,tab
|
| 4 |
+
3,1419,"extension-output-pdoom-org.crowd-code-#1-crowd-code",153,0,"10:41:19 AM [info] Git repository found\n10:41:19 AM [info] Git provider initialized successfully\n10:41:19 AM [info] Initial git state: [object Object]\n",Log,content
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c0873732-2064-4df8-be5c-5c4fe03049aa1751367864693-2025_07_01-13.05.12.330/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c190e976-4e73-4096-9dfa-66c3a0cce2061752157080469-2025_07_10-16.18.14.262/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c2f47c74-4812-499a-abea-eafb3b83db691752315370022-2025_07_12-12.17.39.814/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-e11dd94b-65ba-41d4-a0d1-6151e7930d741752738204303-2025_07_17-09.44.02.899/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-f514f25a-b687-4125-bb3e-747ec98309d21750930471988-2025_06_26-11.34.36.685/source.csv
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,377,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"11:34:36 AM [info] Activating crowd-code\n11:34:36 AM [info] Recording started\n11:34:36 AM [info] Initializing git provider using file system watchers...\n11:34:36 AM [error] Not a git repository: EntryNotFound (FileSystemError): Error: ENOENT: no such file or directory, stat '/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3291405/.git'\n",Log,tab
|
| 3 |
+
3,2083,"extension-output-pdoom-org.crowd-code-#1-crowd-code",358,0,"11:34:38 AM [info] Retrying git provider initialization...\n",Log,content
|
| 4 |
+
4,2162,"extension-output-pdoom-org.crowd-code-#1-crowd-code",417,0,"11:34:38 AM [error] Not a git repository: EntryNotFound (FileSystemError): Error: ENOENT: no such file or directory, stat '/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3291405/.git'\n",Log,content
|