Add files using upload-large-folder tool
Browse files- 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-b4fb6daf-58d3-49e8-97f2-6b7c628e60da1759594854455-2025_10_04-18.21.01.508/source.csv +0 -0
- 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-b994f4be-9bbc-4edd-a199-1b569abe5b6d1764494257169-2025_11_30-10.17.44.938/source.csv +0 -0
- 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-ef5ea013-ac2b-459c-8783-a7b025d58a391754900011518-2025_08_11-10.13.34.249/source.csv +4 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-014bca4d-a98d-462e-b6ae-e78a3d34654d1759675990290-2025_10_05-16.53.54.343/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-05e5774f-8d01-4366-931e-9e4275a288a01759222637213-2025_09_30-10.58.14.459/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-066da5ef-4ba0-4726-8184-5dc4ef5e75811751465994612-2025_07_02-16.20.42.177/source.csv +358 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-0f5513f7-8bc9-4c5d-856d-79d92f75113d1751284706913-2025_06_30-14.24.04.501/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-3fb0e2a5-88e1-4992-bce0-2a2c4a35a7161758449976442-2025_09_21-12.20.21.273/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-5c146b3b-a208-4bdf-96e7-7e0722fd3fa01751383718572-2025_07_01-18.25.45.514/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-5e1c58f1-93d2-473f-9eaf-a2de01442cff1758800954786-2025_09_25-13.49.57.480/source.csv +268 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-6791460b-ec38-4da2-872f-193943c12d601753274780799-2025_07_23-17.17.23.114/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-75084265-2b4d-4d6f-86ae-c0ab064f62491758992086879-2025_09_27-18.55.18.494/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-7b06cc31-85b7-4591-87e0-b26b0ddee2111758710564601-2025_09_24-12.43.30.174/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-92ab1593-f937-4cc4-a174-544581a6ac991751909174142-2025_07_07-19.26.40.736/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-9ab67804-e8b6-44ba-9ee4-ddec1e42461f1757968391960-2025_09_15-22.33.57.229/source.csv +37 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-9eb2e164-5989-4db7-8f31-6e8db1a38df41757236520211-2025_09_07-11.16.00.62/source.csv +5 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-b150d533-89a6-42d8-b7b7-d5a004d568971759420118221-2025_10_02-17.49.22.758/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c52c974d-6c8a-40d7-8b6d-40ee3b3624c21759325522612-2025_10_01-15.32.35.344/source.csv +0 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-e9c5a28e-55ca-497d-97b5-e0c37d2af2781751878142668-2025_07_07-10.49.22.270/source.csv +213 -0
- 927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-f1a23455-555b-44b7-b7f2-5fb8550d75021753722279799-2025_07_28-19.04.51.762/source.csv +3 -0
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-b4fb6daf-58d3-49e8-97f2-6b7c628e60da1759594854455-2025_10_04-18.21.01.508/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-b994f4be-9bbc-4edd-a199-1b569abe5b6d1764494257169-2025_11_30-10.17.44.938/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-ef5ea013-ac2b-459c-8783-a7b025d58a391754900011518-2025_08_11-10.13.34.249/source.csv
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
3503,453188525,"Untitled-1",0,0,"",plaintext,tab
|
| 2 |
+
3504,453189769,"train_dynamics.py",0,0,"",python,tab
|
| 3 |
+
3505,453190506,"Untitled-1",0,0,"",plaintext,tab
|
| 4 |
+
3506,453190928,"train_dynamics.py",0,0,"",python,tab
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-014bca4d-a98d-462e-b6ae-e78a3d34654d1759675990290-2025_10_05-16.53.54.343/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-05e5774f-8d01-4366-931e-9e4275a288a01759222637213-2025_09_30-10.58.14.459/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-066da5ef-4ba0-4726-8184-5dc4ef5e75811751465994612-2025_07_02-16.20.42.177/source.csv
ADDED
|
@@ -0,0 +1,358 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,241,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"4:20:42 PM [info] Activating crowd-code\n4:20:42 PM [info] Recording started\n4:20:42 PM [info] Initializing git provider using file system watchers...\n",Log,tab
|
| 3 |
+
3,825,"extension-output-pdoom-org.crowd-code-#1-crowd-code",150,0,"4:20:42 PM [info] Git repository found\n4:20:42 PM [info] Git provider initialized successfully\n",Log,content
|
| 4 |
+
4,1121,"extension-output-pdoom-org.crowd-code-#1-crowd-code",245,0,"4:20:43 PM [info] Initial git state: [object Object]\n",Log,content
|
| 5 |
+
5,3223,"TERMINAL",0,0,"idling",,terminal_command
|
| 6 |
+
6,3312,"TERMINAL",0,0,"]633;E;2025-07-02 16:20:45 idling;26cd839c-476e-4913-967a-1422bf7b3816]633;C[?1049h[22;0;0t[1;39r(B[m[4l[?7h[H[2JEvery 1.0s: sinfo_t_idle[1;158Hhkn1993.localdomain: Wed Jul 2 16:20:45 2025[3;1HPartition dev_cpuonly[3;35H:\t 9 nodes idle\r[4dPartition cpuonly[4;35H: 69 nodes idle\r[5dPartition dev_accelerated[5;35H:\t 0 nodes idle\r[6dPartition accelerated[6;35H:\t 0 nodes idle\r[7dPartition dev_accelerated-h100 :\t 0 nodes idle\r[8dPartition accelerated-h100[8;35H:\t 0 nodes idle\r[9dPartition large[9;35H:\t 6 nodes idle[39;202H",,terminal_output
|
| 7 |
+
7,3863,"TERMINAL",0,0,"/bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt",,terminal_command
|
| 8 |
+
8,3899,"TERMINAL",0,0,"]633;E;2025-07-02 16:20:45 /bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt;14446405-24c7-42a5-a6ff-f36b2c7c204e]633;C",,terminal_output
|
| 9 |
+
9,3923,"TERMINAL",0,0,"]0;tum_cte0515@hkn1993:/hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash]633;D;0",,terminal_output
|
| 10 |
+
10,4353,"TERMINAL",0,0,"[1;197H6[3;41H10[6d1[39;202H",,terminal_output
|
| 11 |
+
11,5398,"TERMINAL",0,0,"[1;197H7[39d\t ",,terminal_output
|
| 12 |
+
12,5843,"TERMINAL",0,0,"[39;1H[?1049l[23;0;0t\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output
|
| 13 |
+
13,11230,"TERMINAL",0,0,"source .venv/bin/activate",,terminal_command
|
| 14 |
+
14,15010,"TERMINAL",0,0,"queue",,terminal_command
|
| 15 |
+
15,15052,"TERMINAL",0,0,"]633;E;2025-07-02 16:20:57 queue;26cd839c-476e-4913-967a-1422bf7b3816]633;C",,terminal_output
|
| 16 |
+
16,15145,"TERMINAL",0,0,"[?1049h[22;0;0t[1;39r(B[m[4l[?7h[H[2JEvery 1.0s: squeue --me[1;158Hhkn1993.localdomain: Wed Jul 2 16:20:57 2025[3;14HJOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)[4;12H3311671 accelerat train_to tum_cte0 R 1:56:41\t 1 hkn0717[39;202H",,terminal_output
|
| 17 |
+
17,16168,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 18 |
+
18,17176,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 19 |
+
19,18229,"TERMINAL",0,0,"[1;194H1:00[4;60H4[39;202H",,terminal_output
|
| 20 |
+
20,19280,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 21 |
+
21,20325,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 22 |
+
22,21375,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 23 |
+
23,22381,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 24 |
+
24,23415,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 25 |
+
25,24458,"TERMINAL",0,0,"[1;197H6[4;59H50[39;202H",,terminal_output
|
| 26 |
+
26,25500,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 27 |
+
27,26554,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 28 |
+
28,27638,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 29 |
+
29,28771,"TERMINAL",0,0,"[1;196H10[4;60H4[39;202H",,terminal_output
|
| 30 |
+
30,29687,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 31 |
+
31,30801,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 32 |
+
32,31799,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 33 |
+
33,32842,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 34 |
+
34,33915,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 35 |
+
35,34938,"TERMINAL",0,0,"[1;197H6[4;57H7:01[39;202H",,terminal_output
|
| 36 |
+
36,35991,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 37 |
+
37,37044,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 38 |
+
38,38084,"TERMINAL",0,0,"[1;196H20[4;60H4[39;202H",,terminal_output
|
| 39 |
+
39,39136,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 40 |
+
40,40175,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 41 |
+
41,41228,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 42 |
+
42,42316,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 43 |
+
43,43362,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 44 |
+
44,44399,"TERMINAL",0,0,"[1;197H6[4;59H10[39;202H",,terminal_output
|
| 45 |
+
45,45439,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 46 |
+
46,46484,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 47 |
+
47,47524,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 48 |
+
48,48558,"TERMINAL",0,0,"[1;196H30[4;60H4[39;202H",,terminal_output
|
| 49 |
+
49,49609,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 50 |
+
50,50628,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 51 |
+
51,51687,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 52 |
+
52,52733,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 53 |
+
53,53771,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 54 |
+
54,54813,"TERMINAL",0,0,"[1;197H6[4;59H20[39;202H",,terminal_output
|
| 55 |
+
55,55873,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 56 |
+
56,56914,"TERMINAL",0,0,"[1;197H8[4;60H3[39;202H",,terminal_output
|
| 57 |
+
57,57963,"TERMINAL",0,0,"[1;196H40[4;60H4[39;202H",,terminal_output
|
| 58 |
+
58,59001,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 59 |
+
59,60059,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 60 |
+
60,61088,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 61 |
+
61,62175,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 62 |
+
62,63190,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 63 |
+
63,64223,"TERMINAL",0,0,"[1;197H6[4;59H30[39;202H",,terminal_output
|
| 64 |
+
64,65267,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 65 |
+
65,66304,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 66 |
+
66,67359,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 67 |
+
67,68400,"TERMINAL",0,0,"[1;196H50[4;60H4[39;202H",,terminal_output
|
| 68 |
+
68,69439,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 69 |
+
69,70458,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 70 |
+
70,71507,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 71 |
+
71,72552,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 72 |
+
72,73601,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 73 |
+
73,74642,"TERMINAL",0,0,"[1;197H6[4;59H40[39;202H",,terminal_output
|
| 74 |
+
74,75776,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 75 |
+
75,76720,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 76 |
+
76,77825,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 77 |
+
77,78850,"TERMINAL",0,0,"[1;194H2:00[4;60H4[39;202H",,terminal_output
|
| 78 |
+
78,79867,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 79 |
+
79,80916,"TERMINAL",0,0,"[1;197H2[4;60H7[39;202H",,terminal_output
|
| 80 |
+
80,82024,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 81 |
+
81,83048,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 82 |
+
82,84062,"TERMINAL",0,0,"[1;197H6[4;59H50[39;202H",,terminal_output
|
| 83 |
+
83,85199,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 84 |
+
84,86186,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 85 |
+
85,87220,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 86 |
+
86,88273,"TERMINAL",0,0,"[1;196H10[4;60H4[39;202H",,terminal_output
|
| 87 |
+
87,89320,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 88 |
+
88,90408,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 89 |
+
89,91444,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 90 |
+
90,92572,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 91 |
+
91,93550,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 92 |
+
92,94557,"TERMINAL",0,0,"[1;197H6[4;57H8:00[39;202H",,terminal_output
|
| 93 |
+
93,95645,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 94 |
+
94,96668,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 95 |
+
95,97691,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 96 |
+
96,98818,"TERMINAL",0,0,"[1;196H20[4;60H4[39;202H",,terminal_output
|
| 97 |
+
97,99755,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 98 |
+
98,100869,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 99 |
+
99,101827,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 100 |
+
100,102871,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 101 |
+
101,103910,"TERMINAL",0,0,"[1;197H5[4;59H10[39;202H",,terminal_output
|
| 102 |
+
102,104953,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 103 |
+
103,106086,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 104 |
+
104,107056,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 105 |
+
105,108157,"TERMINAL",0,0,"[1;196H30[4;60H4[39;202H",,terminal_output
|
| 106 |
+
106,109213,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 107 |
+
107,110257,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 108 |
+
108,111288,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 109 |
+
109,112341,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 110 |
+
110,113392,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 111 |
+
111,114483,"TERMINAL",0,0,"[1;197H6[4;59H20[39;202H",,terminal_output
|
| 112 |
+
112,115480,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 113 |
+
113,116502,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 114 |
+
114,117589,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 115 |
+
115,118589,"TERMINAL",0,0,"[1;196H40[4;60H4[39;202H",,terminal_output
|
| 116 |
+
116,119706,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 117 |
+
117,120737,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 118 |
+
118,121857,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 119 |
+
119,122882,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 120 |
+
120,123879,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 121 |
+
121,125032,"TERMINAL",0,0,"[1;197H6[4;59H31[39;202H",,terminal_output
|
| 122 |
+
122,126058,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 123 |
+
123,127078,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 124 |
+
124,128105,"TERMINAL",0,0,"[1;196H50[4;60H4[39;202H",,terminal_output
|
| 125 |
+
125,129144,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 126 |
+
126,130159,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 127 |
+
127,131201,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 128 |
+
128,132261,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 129 |
+
129,133300,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 130 |
+
130,134348,"TERMINAL",0,0,"[1;197H6[4;59H40[39;202H",,terminal_output
|
| 131 |
+
131,135480,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 132 |
+
132,136516,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 133 |
+
133,137528,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 134 |
+
134,138545,"TERMINAL",0,0,"[1;194H3:00[4;60H4[39;202H",,terminal_output
|
| 135 |
+
135,139569,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 136 |
+
136,140617,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 137 |
+
137,141723,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 138 |
+
138,142750,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 139 |
+
139,143875,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 140 |
+
140,144920,"TERMINAL",0,0,"[1;197H6[4;59H50[39;202H",,terminal_output
|
| 141 |
+
141,145925,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 142 |
+
142,146945,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 143 |
+
143,147970,"TERMINAL",0,0,"[1;197H9[4;60H4[39;202H",,terminal_output
|
| 144 |
+
144,149098,"TERMINAL",0,0,"[1;196H11[4;60H5[39;202H",,terminal_output
|
| 145 |
+
145,150048,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 146 |
+
146,151144,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 147 |
+
147,152130,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 148 |
+
148,153197,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 149 |
+
149,154233,"TERMINAL",0,0,"[1;197H6[4;57H9:00[39;202H",,terminal_output
|
| 150 |
+
150,155287,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 151 |
+
151,156368,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 152 |
+
152,157391,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 153 |
+
153,158444,"TERMINAL",0,0,"[1;196H20[4;60H4[39;202H",,terminal_output
|
| 154 |
+
154,159544,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 155 |
+
155,160568,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 156 |
+
156,161586,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 157 |
+
157,162636,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 158 |
+
158,163687,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 159 |
+
159,164735,"TERMINAL",0,0,"[1;197H6[4;59H10[39;202H",,terminal_output
|
| 160 |
+
160,165787,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 161 |
+
161,166805,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 162 |
+
162,167938,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 163 |
+
163,168870,"TERMINAL",0,0,"[1;196H30[4;60H4[39;202H",,terminal_output
|
| 164 |
+
164,169988,"TERMINAL",0,0,"[1;197H1[4;60H6[39;202H",,terminal_output
|
| 165 |
+
165,171008,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 166 |
+
166,172736,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 167 |
+
167,173818,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 168 |
+
168,174823,"TERMINAL",0,0,"[1;197H6[4;59H20[39;202H",,terminal_output
|
| 169 |
+
169,175879,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 170 |
+
170,176929,"TERMINAL",0,0,"[1;197H8[4;60H3[39;202H",,terminal_output
|
| 171 |
+
171,177980,"TERMINAL",0,0,"[1;196H40[4;60H4[39;202H",,terminal_output
|
| 172 |
+
172,179020,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 173 |
+
173,180047,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 174 |
+
174,181098,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 175 |
+
175,182144,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 176 |
+
176,183199,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 177 |
+
177,184252,"TERMINAL",0,0,"[1;197H6[4;59H30[39;202H",,terminal_output
|
| 178 |
+
178,185293,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 179 |
+
179,186346,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 180 |
+
180,187386,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 181 |
+
181,188438,"TERMINAL",0,0,"[1;196H50[4;60H4[39;202H",,terminal_output
|
| 182 |
+
182,189489,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 183 |
+
183,190534,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 184 |
+
184,191589,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 185 |
+
185,192642,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 186 |
+
186,193686,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 187 |
+
187,194730,"TERMINAL",0,0,"[1;197H6[4;59H40[39;202H",,terminal_output
|
| 188 |
+
188,195783,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 189 |
+
189,196832,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 190 |
+
190,197945,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 191 |
+
191,198909,"TERMINAL",0,0,"[1;194H4:00[4;60H5[39;202H",,terminal_output
|
| 192 |
+
192,199955,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 193 |
+
193,201013,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 194 |
+
194,202045,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 195 |
+
195,203085,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 196 |
+
196,204132,"TERMINAL",0,0,"[1;197H6[4;59H50[39;202H",,terminal_output
|
| 197 |
+
197,205180,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 198 |
+
198,206228,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 199 |
+
199,207263,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 200 |
+
200,208305,"TERMINAL",0,0,"[1;196H10[4;60H4[39;202H",,terminal_output
|
| 201 |
+
201,209343,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 202 |
+
202,210433,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 203 |
+
203,211401,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 204 |
+
204,212440,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 205 |
+
205,213533,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 206 |
+
206,214534,"TERMINAL",0,0,"[1;197H6[4;54H2:00:00[39;202H",,terminal_output
|
| 207 |
+
207,215556,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 208 |
+
208,216683,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 209 |
+
209,217704,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 210 |
+
210,218676,"TERMINAL",0,0,"[1;196H20[4;60H4[39;202H",,terminal_output
|
| 211 |
+
211,219713,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 212 |
+
212,220882,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 213 |
+
213,221902,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 214 |
+
214,222853,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 215 |
+
215,223913,"TERMINAL",0,0,"[1;197H5[4;59H10[39;202H",,terminal_output
|
| 216 |
+
216,224993,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 217 |
+
217,225981,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 218 |
+
218,227095,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 219 |
+
219,228148,"TERMINAL",0,0,"[1;196H30[4;60H4[39;202H",,terminal_output
|
| 220 |
+
220,229183,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 221 |
+
221,230174,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 222 |
+
222,231210,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 223 |
+
223,232274,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 224 |
+
224,233322,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 225 |
+
225,234362,"TERMINAL",0,0,"[1;197H6[4;59H20[39;202H",,terminal_output
|
| 226 |
+
226,235404,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 227 |
+
227,236454,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 228 |
+
228,237503,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 229 |
+
229,238542,"TERMINAL",0,0,"[1;196H40[4;60H4[39;202H",,terminal_output
|
| 230 |
+
230,239598,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 231 |
+
231,240645,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 232 |
+
232,241692,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 233 |
+
233,242731,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 234 |
+
234,243808,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 235 |
+
235,244792,"TERMINAL",0,0,"[1;197H6[4;59H30[39;202H",,terminal_output
|
| 236 |
+
236,245835,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 237 |
+
237,246903,"TERMINAL",0,0,"[1;197H8[4;60H3[39;202H",,terminal_output
|
| 238 |
+
238,248016,"TERMINAL",0,0,"[1;196H50[4;60H4[39;202H",,terminal_output
|
| 239 |
+
239,249001,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 240 |
+
240,250082,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 241 |
+
241,251096,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 242 |
+
242,252230,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 243 |
+
243,253213,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 244 |
+
244,254265,"TERMINAL",0,0,"[1;197H6[4;59H40[39;202H",,terminal_output
|
| 245 |
+
245,255317,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 246 |
+
246,256445,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 247 |
+
247,257415,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 248 |
+
248,258467,"TERMINAL",0,0,"[1;194H5:00[4;60H4[39;202H",,terminal_output
|
| 249 |
+
249,259491,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 250 |
+
250,260541,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 251 |
+
251,261597,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 252 |
+
252,262644,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 253 |
+
253,263683,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 254 |
+
254,265423,"TERMINAL",0,0,"[1;197H6[4;59H51[39;202H",,terminal_output
|
| 255 |
+
255,266549,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 256 |
+
256,267573,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 257 |
+
257,268536,"TERMINAL",0,0,"[1;196H10[4;60H4[39;202H",,terminal_output
|
| 258 |
+
258,269659,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 259 |
+
259,270603,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 260 |
+
260,271673,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 261 |
+
261,272686,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 262 |
+
262,273739,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 263 |
+
263,274843,"TERMINAL",0,0,"[1;197H6[4;57H1:00[39;202H",,terminal_output
|
| 264 |
+
264,275869,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 265 |
+
265,276857,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 266 |
+
266,277915,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 267 |
+
267,278940,"TERMINAL",0,0,"[1;196H20[4;60H5[39;202H",,terminal_output
|
| 268 |
+
268,280068,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 269 |
+
269,281093,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 270 |
+
270,282114,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 271 |
+
271,283129,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 272 |
+
272,284264,"TERMINAL",0,0,"[1;197H6[4;59H10[39;202H",,terminal_output
|
| 273 |
+
273,285295,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 274 |
+
274,286261,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 275 |
+
275,287311,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 276 |
+
276,288349,"TERMINAL",0,0,"[1;196H30[4;60H4[39;202H",,terminal_output
|
| 277 |
+
277,289487,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 278 |
+
278,290438,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 279 |
+
279,291535,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 280 |
+
280,292563,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 281 |
+
281,293569,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 282 |
+
282,294644,"TERMINAL",0,0,"[1;197H6[4;59H20[39;202H",,terminal_output
|
| 283 |
+
283,295655,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 284 |
+
284,296695,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 285 |
+
285,297741,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 286 |
+
286,298794,"TERMINAL",0,0,"[1;196H40[4;60H4[39;202H",,terminal_output
|
| 287 |
+
287,299838,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 288 |
+
288,300882,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 289 |
+
289,301983,"TERMINAL",0,0,"[1;197H3[4;60H8[39;202H",,terminal_output
|
| 290 |
+
290,302951,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 291 |
+
291,303999,"TERMINAL",0,0,"[1;197H6[4;59H30[39;202H",,terminal_output
|
| 292 |
+
292,305051,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 293 |
+
293,306178,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 294 |
+
294,307202,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 295 |
+
295,308176,"TERMINAL",0,0,"[1;196H50[4;60H4[39;202H",,terminal_output
|
| 296 |
+
296,309221,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 297 |
+
297,310276,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 298 |
+
298,311314,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 299 |
+
299,312428,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 300 |
+
300,313448,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 301 |
+
301,314473,"TERMINAL",0,0,"[1;197H6[4;59H40[39;202H",,terminal_output
|
| 302 |
+
302,315495,"TERMINAL",0,0,"[1;197H7[4;60H1[39;202H",,terminal_output
|
| 303 |
+
303,316521,"TERMINAL",0,0,"[1;197H8[4;60H2[39;202H",,terminal_output
|
| 304 |
+
304,317570,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 305 |
+
305,318624,"TERMINAL",0,0,"[1;194H6:00[4;60H4[39;202H",,terminal_output
|
| 306 |
+
306,319697,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 307 |
+
307,320678,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 308 |
+
308,321736,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 309 |
+
309,322777,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 310 |
+
310,323894,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 311 |
+
311,324922,"TERMINAL",0,0,"[1;197H6[4;59H50[39;202H",,terminal_output
|
| 312 |
+
312,325942,"TERMINAL",0,0,"[1;197H7[4;60H2[39;202H",,terminal_output
|
| 313 |
+
313,326965,"TERMINAL",0,0,"[1;197H9[4;60H3[39;202H",,terminal_output
|
| 314 |
+
314,327991,"TERMINAL",0,0,"[1;196H10[4;60H4[39;202H",,terminal_output
|
| 315 |
+
315,329116,"TERMINAL",0,0,"[1;197H1[4;60H5[39;202H",,terminal_output
|
| 316 |
+
316,330243,"TERMINAL",0,0,"[1;197H2[4;60H6[39;202H",,terminal_output
|
| 317 |
+
317,331133,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 318 |
+
318,332188,"TERMINAL",0,0,"[1;197H4[4;60H8[39;202H",,terminal_output
|
| 319 |
+
319,333240,"TERMINAL",0,0,"[1;197H5[4;60H9[39;202H",,terminal_output
|
| 320 |
+
320,334286,"TERMINAL",0,0,"[1;197H6[4;57H2:00[39;202H",,terminal_output
|
| 321 |
+
321,334771,"TERMINAL",0,0,"[39;1H[?1049l[23;0;0t\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output
|
| 322 |
+
322,4560228,"TERMINAL",0,0,"queue",,terminal_command
|
| 323 |
+
323,4560272,"TERMINAL",0,0,"]633;E;2025-07-02 17:36:42 queue;26cd839c-476e-4913-967a-1422bf7b3816]633;C",,terminal_output
|
| 324 |
+
324,4560343,"TERMINAL",0,0,"[?1049h[22;0;0t[1;39r(B[m[4l[?7h[H[2JEvery 1.0s: squeue --me[1;158Hhkn1993.localdomain: Wed Jul 2 17:36:42 2025[3;14HJOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)[4;12H3311671 accelerat train_to tum_cte0 R 3:12:26\t 1 hkn0717[39;202H",,terminal_output
|
| 325 |
+
325,4561383,"TERMINAL",0,0,"[1;197H3[4;60H7[39;202H",,terminal_output
|
| 326 |
+
326,4561489,"TERMINAL",0,0,"[39;1H[?1049l[23;0;0t\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output
|
| 327 |
+
327,4562886,"TERMINAL",0,0,"idling",,terminal_command
|
| 328 |
+
328,4562983,"TERMINAL",0,0,"]633;E;2025-07-02 17:36:44 idling;26cd839c-476e-4913-967a-1422bf7b3816]633;C[?1049h[22;0;0t[1;39r(B[m[4l[?7h[H[2JEvery 1.0s: sinfo_t_idle[1;158Hhkn1993.localdomain: Wed Jul 2 17:36:45 2025[3;1HPartition dev_cpuonly[3;35H: 12 nodes idle\r[4dPartition cpuonly[4;35H:\t 4 nodes idle\r[5dPartition dev_accelerated[5;35H:\t 0 nodes idle\r[6dPartition accelerated[6;35H:\t 0 nodes idle\r[7dPartition dev_accelerated-h100 :\t 0 nodes idle\r[8dPartition accelerated-h100[8;35H:\t 0 nodes idle\r[9dPartition large[9;35H:\t 7 nodes idle[39;202H",,terminal_output
|
| 329 |
+
329,4564005,"TERMINAL",0,0,"[1;197H6[39d\t ",,terminal_output
|
| 330 |
+
330,4565051,"TERMINAL",0,0,"[1;197H7[39d\t ",,terminal_output
|
| 331 |
+
331,4566099,"TERMINAL",0,0,"[1;197H8[39d\t ",,terminal_output
|
| 332 |
+
332,4566873,"TERMINAL",0,0,"[39;1H[?1049l[23;0;0t\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output
|
| 333 |
+
333,4569190,"TERMINAL",0,0,"",,terminal_focus
|
| 334 |
+
334,4575053,"scripts_horeka/overfit_batch_tiny/sample.sh",0,0,"#!/usr/bin/env bash\n\n# Unload modules that may interfere\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\n\n# Activate virtual environment\nsource .venv/bin/activate\n\n# Set workspace and checkpoint directory (update slurm_job_id as needed)\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared'\n# Replace the following with the actual job id/checkpoint you want to sample from\nslurm_job_id=3301029\n\n# job_name=train_dynamics_minecraft_overfit_sample_tiny\nCHECKPOINT_DIR=$ws_dir/checkpoints/${slurm_job_id}\n\n# Example: If you want to use a specific checkpoint, set it here\n# CHECKPOINT_PATH=$ws_dir/checkpoints/3299272/dynamics-tiny-overfit-big-lr-3299272_50000/\n# Or use the latest in the directory\n# CHECKPOINT_PATH=$(ls -d $CHECKPOINT_DIR/*/ | sort | tail -n 1)\nCHECKPOINT_PATH=$CHECKPOINT_DIR/genie_1751067601_200000/\n# CHECKPOINT_PATH=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/0000/genie_1751301068_2000/\n# CHECKPOINT_PATH=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/../checkpoints/3307618/genie_1751322003_15500/\n# CHECKPOINT_PATH=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3307619/genie_1751322003_200000/\nCHECKPOINT_PATH=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3309699/genie_1751384516_200000/\n\n\necho ""Sampling from checkpoint: $CHECKPOINT_PATH""\n\npython sample.py \\n --checkpoint ""$CHECKPOINT_PATH"" \\n --tokenizer_dim=384 \\n --latent_patch_dim=32 \\n --num_patch_latents=1024 \\n --patch_size=4 \\n --tokenizer_num_blocks=8 \\n --tokenizer_num_heads=8 \\n --lam_dim=384 \\n --latent_action_dim=32 \\n --lam_patch_size=16 \\n --lam_num_blocks=8 \\n --lam_num_heads=8 \\n --dyna_dim=128 \\n --dyna_num_blocks=2 \\n --dyna_num_heads=4 \\n --maskgit_steps=1000 \\n --num_latent_actions=6 \\n --seq_len=16 \\n --start_frame=0\n\n# python sample.py \\n # --checkpoint ""$CHECKPOINT_PATH"" \\n # --data_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/coinrun_episodes\n",shellscript,tab
|
| 335 |
+
335,4582430,"TERMINAL",0,0,"bash",,terminal_focus
|
| 336 |
+
336,4584712,"TERMINAL",0,0,"bash",,terminal_focus
|
| 337 |
+
337,4584714,"scripts_horeka/overfit_batch_tiny/sample.sh",0,0,"",shellscript,tab
|
| 338 |
+
338,4585497,"scripts_horeka/modelsize_scaling/dynamics/model_sizes.md",0,0,"# Genie 1 - Model Sizes and their configs\n\n## Tokenizer model: sizes\n\ndefault: \n| Model | model_dim | num_blocks | num_heads | latent_dim | num_latents | Est. Params |\n|-------|-----------|------------|-----------|------------|-------------|-------------|\n| default | 512 | 8 | 8 | 32 | 1024 | ~38M |\n\n### scaling up \n#### (not tested yet - TODO @mihir)\n\n| Model | model_dim | num_blocks | num_heads | latent_dim | num_latents | Est. Params |\n|-------|-----------|------------|-----------|------------|-------------|-------------|\n| L1 | 768 | 12 | 12 | 64 | 2048 | ~80M |\n| L2 | 1024 | 12 | 16 | 128 | 2048 | ~140M |\n| L3 | 1152 | 16 | 16 | 128 | 4096 | ~200M |\n| L4 | 896 | 16 | 14 | 96 | 4096 | ~120M |\n| L5 | 1536 | 12 | 24 | 256 | 2048 | ~190M |\n\n\n### tiny models\n| Model | model_dim | num_blocks | num_heads | latent_dim | num_latents | Est. Params |\n|-------|-----------|------------|-----------|------------|-------------|-------------|\n| S1 | 128 | 2 | 2 | 8 | 128 | ~0.6M |\n| S2 | 192 | 2 | 3 | 16 | 128 | ~1.3M |\n| S3 | 256 | 3 | 4 | 16 | 256 | ~3.6M |\n| S4 | 320 | 4 | 5 | 24 | 256 | ~7.4M |\n| S5 | 384 | 4 | 6 | 32 | 512 | ~10M |\n\n\n## Latent Action model: sizes\ndefault: \n| Model | model_dim | num_blocks | num_heads | latent_dim | num_latents | Est. Params |\n|-------|-----------|------------|-----------|------------|-------------|-------------|\n| default | 512 | 8 | 8 | 32 | 6 | ~39M |\n\n### scaling up \n#### (not tested yet - TODO @mihir)\n\n| Name | model_dim | num_blocks | num_heads | latent_dim | num_latents | Est. Params |\n|--------------|-----------|------------|-----------|------------|-------------|-------------|\n| XL | 1024 | 12 | 16 | 64 | 12 | ~200M |\n| L | 896 | 12 | 14 | 48 | 8 | ~150M |\n| M+ | 768 | 10 | 12 | 48 | 8 | ~100M |\n| M | 640 | 10 | 10 | 32 | 8 | ~70M |\n| Base+ | 512 | 12 | 8 | 32 | 8 | ~55M |\n\n\n### tiny models\n| Name | model_dim | num_blocks | num_heads | latent_dim | num_latents | Est. Params |\n|--------------|-----------|------------|-----------|------------|-------------|-------------|\n| XS | 128 | 2 | 2 | 8 | 4 | ~0.9M |\n| S | 160 | 2 | 2 | 8 | 4 | ~1.3M |\n| S+ | 192 | 3 | 3 | 8 | 4 | ~2.4M |\n| M- | 256 | 4 | 4 | 16 | 6 | ~5.4M |\n| M | 320 | 6 | 4 | 16 | 6 | ~12M |\n\n\n## Dynamics model: sizes \n\n| Config | dyna_dim | dyna_num_blocks | dyna_num_heads | Approx. Params |\n|--------|----------|-----------------|---------------|----------------|\n| 1 | 512 | 12 | 8 | ~36M |\n| 2 | 768 | 16 | 12 | ~110M |\n| 3 | 1024 | 16 | 16 | ~180M |\n| 4 | 1024 | 24 | 16 | ~270M |\n| 5 | 1536 | 24 | 24 | ~500M |\n\n\n### tiny models\n| Config | dyna_dim | dyna_num_blocks | dyna_num_heads | Approx. Params |\n|--------|----------|-----------------|---------------|----------------|\n| A | 128 | 2 | 4 | ~1.5M |\n| B | 256 | 2 | 4 | ~3.5M |\n| C | 256 | 4 | 4 | ~6M |\n| D | 384 | 4 | 6 | ~12M |\n| E | 512 | 4 | 8 | ~18M |",markdown,tab
|
| 339 |
+
339,4873342,"TERMINAL",0,0,"bash",,terminal_focus
|
| 340 |
+
340,5485002,"TERMINAL",0,0,"salloc --time=00:30:00 --partition=dev_accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5 --mem=50G",,terminal_command
|
| 341 |
+
341,5485037,"TERMINAL",0,0,"]633;E;2025-07-02 17:52:07 salloc --time=00:30:00 --partition=dev_accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5 --mem=50G;26cd839c-476e-4913-967a-1422bf7b3816]633;Csalloc: Pending job allocation 3312853\r\nsalloc: job 3312853 queued and waiting for resources\r\n",,terminal_output
|
| 342 |
+
342,5486062,"TERMINAL",0,0,"^Csalloc: Job allocation 3312853 has been revoked.\r\nsalloc: Job aborted due to signal\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;1",,terminal_output
|
| 343 |
+
343,5488299,"TERMINAL",0,0,"salloc --time=00:30:00 --partition=dev_accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5 --mem=50G^C",,terminal_command
|
| 344 |
+
344,5488347,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]633;E;;26cd839c-476e-4913-967a-1422bf7b3816]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D",,terminal_output
|
| 345 |
+
345,5507657,"TERMINAL",0,0,"salloc --time=01:30:00 --partition=dev_accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5 --mem=50G",,terminal_command
|
| 346 |
+
346,5507713,"TERMINAL",0,0,"]633;E;2025-07-02 17:52:29 salloc --time=01:30:00 --partition=dev_accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5 --mem=50G;26cd839c-476e-4913-967a-1422bf7b3816]633;Csalloc: error: Job submit/allocate failed: Requested time limit is invalid (missing or exceeds some limit)\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;1",,terminal_output
|
| 347 |
+
347,5522205,"TERMINAL",0,0,"salloc --time=01:30:00 --partition=accelerated --nodes=1 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5 --mem=50G",,terminal_command
|
| 348 |
+
348,5522257,"TERMINAL",0,0,"]633;E;2025-07-02 17:52:44 salloc --time=01:30:00 --partition=accelerated --nodes=1 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5 --mem=50G;26cd839c-476e-4913-967a-1422bf7b3816]633;Csalloc: Pending job allocation 3312854\r\nsalloc: job 3312854 queued and waiting for resources\r\n",,terminal_output
|
| 349 |
+
349,5528628,"TERMINAL",0,0,"bash",,terminal_focus
|
| 350 |
+
350,5529561,"TERMINAL",0,0,"queue",,terminal_command
|
| 351 |
+
351,5529625,"TERMINAL",0,0,"\r\n[?2004l\r]633;E;2025-07-02 17:52:51 queue;0598f850-442d-4019-9770-f648eaf5abbd]633;C[?1049h[22;0;0t[1;15r(B[m[4l[?7h[H[2JEvery 1.0s: squeue --me[1;158Hhkn1993.localdomain: Wed Jul 2 17:52:51 2025[3;14HJOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)[4;12H3312854 accelerat interact tum_cte0 PD\t0:00\t 1 (Priority)[5;12H3311671 accelerat train_to tum_cte0 R 3:28:35\t 1 hkn0717[15;202H",,terminal_output
|
| 352 |
+
352,5530673,"TERMINAL",0,0,"[1;197H2[5;60H6[15;202H",,terminal_output
|
| 353 |
+
353,5531169,"TERMINAL",0,0,"[15;1H[?1049l[23;0;0t\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar[?2004h",,terminal_output
|
| 354 |
+
354,5532122,"TERMINAL",0,0,"idling",,terminal_command
|
| 355 |
+
355,5532190,"TERMINAL",0,0,"]633;E;2025-07-02 17:52:54 idling;0598f850-442d-4019-9770-f648eaf5abbd]633;C[?1049h[22;0;0t[1;15r(B[m[4l[?7h[H[2JEvery 1.0s: sinfo_t_idle[1;158Hhkn1993.localdomain: Wed Jul 2 17:52:54 2025[3;1HPartition dev_cpuonly[3;35H: 11 nodes idle\r[4dPartition cpuonly[4;35H: 115 nodes idle\r[5dPartition dev_accelerated[5;35H:\t 0 nodes idle\r[6dPartition accelerated[6;35H:\t 0 nodes idle\r[7dPartition dev_accelerated-h100 :\t 0 nodes idle\r[8dPartition accelerated-h100[8;35H:\t 0 nodes idle\r[9dPartition large[9;35H:\t 7 nodes idle[15;202H",,terminal_output
|
| 356 |
+
356,5533235,"TERMINAL",0,0,"[1;197H5[15d\t ",,terminal_output
|
| 357 |
+
357,5533929,"TERMINAL",0,0,"[15;1H[?1049l[23;0;0t\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output
|
| 358 |
+
358,5771719,"TERMINAL",0,0,"salloc",,terminal_focus
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-0f5513f7-8bc9-4c5d-856d-79d92f75113d1751284706913-2025_06_30-14.24.04.501/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-3fb0e2a5-88e1-4992-bce0-2a2c4a35a7161758449976442-2025_09_21-12.20.21.273/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-5c146b3b-a208-4bdf-96e7-7e0722fd3fa01751383718572-2025_07_01-18.25.45.514/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-5e1c58f1-93d2-473f-9eaf-a2de01442cff1758800954786-2025_09_25-13.49.57.480/source.csv
ADDED
|
@@ -0,0 +1,268 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,4,"jasmine/models/dynamics.py",0,0,"from typing import Dict\n\nimport jax\nimport jax.numpy as jnp\nimport flax.nnx as nnx\n\nfrom utils.nn import STTransformer, Transformer\n\n\nclass DynamicsMaskGIT(nnx.Module):\n """"""\n MaskGIT dynamics model\n\n Dimension keys:\n B: batch size\n T: sequence length\n N: number of patches per frame\n L: latent dimension\n V: vocabulary size (number of latents)\n """"""\n\n def __init__(\n self,\n model_dim: int,\n ffn_dim: int,\n num_latents: int,\n latent_action_dim: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n mask_limit: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n rngs: nnx.Rngs,\n ):\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.num_latents = num_latents\n self.latent_action_dim = latent_action_dim\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.mask_limit = mask_limit\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n\n self.transformer = STTransformer(\n self.model_dim,\n self.model_dim,\n self.ffn_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.patch_embed = nnx.Embed(self.num_latents, self.model_dim, rngs=rngs)\n self.mask_token = nnx.Param(\n nnx.initializers.lecun_uniform()(rngs.params(), (1, 1, 1, self.model_dim))\n )\n self.action_up = nnx.Linear(\n self.latent_action_dim,\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n def __call__(\n self,\n batch: Dict[str, jax.Array],\n ) -> tuple[jax.Array, jax.Array]:\n # --- Mask videos ---\n video_tokens_BTN = batch[""video_tokens""]\n latent_actions_BTm11L = batch[""latent_actions""]\n vid_embed_BTNM = self.patch_embed(video_tokens_BTN)\n\n batch_size = vid_embed_BTNM.shape[0]\n _rng_prob, *_rngs_mask = jax.random.split(batch[""mask_rng""], batch_size + 1)\n mask_prob = jax.random.uniform(\n _rng_prob, shape=(batch_size,), minval=self.mask_limit\n )\n per_sample_shape = vid_embed_BTNM.shape[1:-1]\n mask = jax.vmap(\n lambda rng, prob: jax.random.bernoulli(rng, prob, per_sample_shape),\n in_axes=(0, 0),\n )(jnp.asarray(_rngs_mask), mask_prob)\n mask = mask.at[:, 0].set(False)\n vid_embed_BTNM = jnp.where(\n jnp.expand_dims(mask, -1), self.mask_token.value, vid_embed_BTNM\n )\n\n # --- Predict transition ---\n act_embed_BTm11M = self.action_up(latent_actions_BTm11L)\n padded_act_embed_BT1M = jnp.pad(\n act_embed_BTm11M, ((0, 0), (1, 0), (0, 0), (0, 0))\n )\n padded_act_embed_BTNM = jnp.broadcast_to(\n padded_act_embed_BT1M, vid_embed_BTNM.shape\n )\n vid_embed_BTNM += padded_act_embed_BTNM\n logits_BTNV = self.transformer(vid_embed_BTNM)\n return logits_BTNV, mask\n\n\nclass DynamicsCausal(nnx.Module):\n """"""Causal dynamics model""""""\n\n def __init__(\n self,\n model_dim: int,\n ffn_dim: int,\n num_latents: int,\n latent_action_dim: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n ):\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.num_latents = num_latents\n self.latent_action_dim = latent_action_dim\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.decode = decode\n\n self.transformer = Transformer(\n self.model_dim,\n self.model_dim,\n self.ffn_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=self.decode,\n rngs=rngs,\n )\n self.patch_embed = nnx.Embed(self.num_latents, self.model_dim, rngs=rngs)\n self.action_up = nnx.Linear(\n self.latent_action_dim,\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n def __call__(\n self,\n batch: Dict[str, jax.Array],\n ) -> tuple[jax.Array, jax.Array]:\n video_tokens_BTN = batch[""video_tokens""]\n latent_actions_BTm11L = batch[""latent_actions""]\n vid_embed_BTNM = self.patch_embed(video_tokens_BTN)\n act_embed_BTm11M = self.action_up(latent_actions_BTm11L)\n padded_act_embed_BT1M = jnp.pad(\n act_embed_BTm11M, ((0, 0), (1, 0), (0, 0), (0, 0))\n )\n vid_embed_BTNp1M = jnp.concatenate(\n [padded_act_embed_BT1M, vid_embed_BTNM], axis=2\n )\n logits_BTNp1V = self.transformer(vid_embed_BTNp1M)\n logits_BTNV = logits_BTNp1V[:, :, :-1]\n return logits_BTNV, jnp.ones_like(video_tokens_BTN)\n",python,tab
|
| 3 |
+
2,732,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"1:49:57 PM [info] Activating crowd-code\n1:49:57 PM [info] Recording started\n1:49:57 PM [info] Initializing git provider using file system watchers...\n",Log,tab
|
| 4 |
+
3,1087,"extension-output-pdoom-org.crowd-code-#1-crowd-code",150,0,"1:49:57 PM [info] Git repository found\n1:49:57 PM [info] Git provider initialized successfully\n1:49:57 PM [info] Initial git state: [object Object]\n",Log,content
|
| 5 |
+
4,6981,"TERMINAL",0,0,"bash",,terminal_focus
|
| 6 |
+
5,11991,"TERMINAL",0,0,"queue",,terminal_command
|
| 7 |
+
6,12012,"TERMINAL",0,0,"]633;C[?1049h[22;0;0t[1;62r(B[m[4l[?7h[H[2JEvery 1.0s: squeue --me[1;87Hhkn1991.localdomain: Thu Sep 25 13:50:09 2025[3;14HJOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)[62;131H",,terminal_output
|
| 8 |
+
7,12913,"TERMINAL",0,0,"[62;1H[?1049l[23;0;0t\r[?1l>]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output
|
| 9 |
+
8,21218,"TERMINAL",0,0,"salloc --time=05:00:00 --partition=accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8",,terminal_command
|
| 10 |
+
9,21287,"TERMINAL",0,0,"]633;Csalloc: Pending job allocation 3520566\r\nsalloc: job 3520566 queued and waiting for resources\r\n",,terminal_output
|
| 11 |
+
10,26666,"TERMINAL",0,0,"salloc: job 3520566 has been allocated resources\r\nsalloc: Granted job allocation 3520566\r\nsalloc: Waiting for resource configuration\r\n",,terminal_output
|
| 12 |
+
11,53262,"TERMINAL",0,0,"salloc: Nodes hkn0801 are ready for job\r\n",,terminal_output
|
| 13 |
+
12,53922,"TERMINAL",0,0,"]0;tum_cte0515@hkn0801:~/Projects/jasmine[?2004h[tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 14 |
+
13,80884,"TERMINAL",0,0,"so",,terminal_output
|
| 15 |
+
14,80963,"TERMINAL",0,0,"u",,terminal_output
|
| 16 |
+
15,81017,"TERMINAL",0,0,"r",,terminal_output
|
| 17 |
+
16,81231,"TERMINAL",0,0,"c",,terminal_output
|
| 18 |
+
17,81353,"TERMINAL",0,0,"e",,terminal_output
|
| 19 |
+
18,81418,"TERMINAL",0,0," ",,terminal_output
|
| 20 |
+
19,81615,"TERMINAL",0,0,".",,terminal_output
|
| 21 |
+
20,81736,"TERMINAL",0,0,"",,terminal_output
|
| 22 |
+
21,82521,"TERMINAL",0,0,"v",,terminal_output
|
| 23 |
+
22,82753,"TERMINAL",0,0,"e",,terminal_output
|
| 24 |
+
23,82846,"TERMINAL",0,0,"n",,terminal_output
|
| 25 |
+
24,82953,"TERMINAL",0,0,"v/",,terminal_output
|
| 26 |
+
25,83176,"TERMINAL",0,0,"b",,terminal_output
|
| 27 |
+
26,83303,"TERMINAL",0,0,"a",,terminal_output
|
| 28 |
+
27,83664,"TERMINAL",0,0,"[K",,terminal_output
|
| 29 |
+
28,83813,"TERMINAL",0,0,"i",,terminal_output
|
| 30 |
+
29,83923,"TERMINAL",0,0,"n/",,terminal_output
|
| 31 |
+
30,84413,"TERMINAL",0,0,"ac",,terminal_output
|
| 32 |
+
31,84549,"TERMINAL",0,0,"tivate",,terminal_output
|
| 33 |
+
32,85001,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0801:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 34 |
+
33,85218,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0801:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 35 |
+
34,86479,"TERMINAL",0,0,"q",,terminal_output
|
| 36 |
+
35,86549,"TERMINAL",0,0,"u",,terminal_output
|
| 37 |
+
36,86664,"TERMINAL",0,0,"eu",,terminal_output
|
| 38 |
+
37,86813,"TERMINAL",0,0,"e",,terminal_output
|
| 39 |
+
38,86879,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output
|
| 40 |
+
39,86948,"TERMINAL",0,0,"[?1049h[22;0;0t[1;62r(B[m[4l[?7h[H[2JEvery 1.0s: squeue --me[1;87Hhkn0801.localdomain: Thu Sep 25 13:51:24 2025[3;14HJOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)[4;12H3520566 accelerat interact tum_cte0 R\t1:01\t 1 hkn0801[62;131H",,terminal_output
|
| 41 |
+
40,87731,"TERMINAL",0,0,"[62;1H[?1049l[23;0;0t\r[?1l>]0;tum_cte0515@hkn0801:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 42 |
+
41,88231,"TERMINAL",0,0,"c",,terminal_output
|
| 43 |
+
42,88284,"TERMINAL",0,0,"l",,terminal_output
|
| 44 |
+
43,88350,"TERMINAL",0,0,"e",,terminal_output
|
| 45 |
+
44,88515,"TERMINAL",0,0,"a",,terminal_output
|
| 46 |
+
45,88684,"TERMINAL",0,0,"r",,terminal_output
|
| 47 |
+
46,88766,"TERMINAL",0,0,"\r\n[?2004l\r[H[2J[3J]0;tum_cte0515@hkn0801:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 48 |
+
47,91357,"TERMINAL",0,0,"\r[K(jasmine) [tum_cte0515@hkn0801 jasmine]$ \r[K(jasmine) [tum_cte0515@hkn0801 jasmine]$ \r[K(jasmine) [tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 49 |
+
48,91969,"jasmine/models/dynamics.py",0,0,"",python,tab
|
| 50 |
+
49,91972,"jasmine/models/dynamics.py",2538,0,"",python,selection_mouse
|
| 51 |
+
50,92574,"jasmine/models/dynamics.py",2927,0,"",python,selection_mouse
|
| 52 |
+
51,93139,"jasmine/models/dynamics.py",2860,0,"",python,selection_mouse
|
| 53 |
+
52,93960,"jasmine/models/dynamics.py",2926,0,"",python,selection_mouse
|
| 54 |
+
53,93975,"jasmine/models/dynamics.py",2925,0,"",python,selection_command
|
| 55 |
+
54,144844,"TERMINAL",0,0,"q",,terminal_output
|
| 56 |
+
55,144954,"TERMINAL",0,0,"u",,terminal_output
|
| 57 |
+
56,145025,"TERMINAL",0,0,"e",,terminal_output
|
| 58 |
+
57,145132,"TERMINAL",0,0,"u",,terminal_output
|
| 59 |
+
58,145197,"TERMINAL",0,0,"e",,terminal_output
|
| 60 |
+
59,145368,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h[22;0;0t[1;62r(B[m[4l[?7h[H[2JEvery 1.0s: squeue --me[1;78Hhkn0801.localdomain: Thu Sep 25 13:52:22 2025[3;14HJOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)[4;12H3520566 accelerat interact tum_cte0 R\t1:59\t 1 hkn0801[62;122H",,terminal_output
|
| 61 |
+
60,146193,"TERMINAL",0,0,"[62;1H[?1049l[23;0;0t\r[?1l>]0;tum_cte0515@hkn0801:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 62 |
+
61,147631,"TERMINAL",0,0,"g",,terminal_output
|
| 63 |
+
62,147695,"TERMINAL",0,0,"i",,terminal_output
|
| 64 |
+
63,147769,"TERMINAL",0,0,"t",,terminal_output
|
| 65 |
+
64,147833,"TERMINAL",0,0," ",,terminal_output
|
| 66 |
+
65,147977,"TERMINAL",0,0,"b",,terminal_output
|
| 67 |
+
66,148073,"TERMINAL",0,0,"r",,terminal_output
|
| 68 |
+
67,148298,"TERMINAL",0,0,"a",,terminal_output
|
| 69 |
+
68,148385,"TERMINAL",0,0,"c",,terminal_output
|
| 70 |
+
69,148727,"TERMINAL",0,0,"[K",,terminal_output
|
| 71 |
+
70,149401,"TERMINAL",0,0,"n",,terminal_output
|
| 72 |
+
71,149512,"TERMINAL",0,0,"c",,terminal_output
|
| 73 |
+
72,149885,"TERMINAL",0,0,"h",,terminal_output
|
| 74 |
+
73,150141,"TERMINAL",0,0,"\r\n[?2004l\r[?1h=\r",,terminal_output
|
| 75 |
+
74,150381,"TERMINAL",0,0," action-mapper[m[m\r\n add-noise-to-combat-exposure-bias[m[m\r\n add-wandb-name-and-tags[m[m\r\n before-nnx[m[m\r\n causal-mem-reduce[m[m\r\n causal-spatiotemporal-kv-cache[m[m\r\n causal-st-transformer[m[m\r\n causal-transformer-dynamics-model[m[m\r\n causal-transformer-nnx-no-kv-cache[m[m\r\n coinrun-gt-actions[m[m\r\n convert-to-jax-array-in-iter[m[m\r\n correct-batched-sampling[m[m\r\n dev[m[m\r\n dont-let-tf-see-gpu[m[m\r\n feat/darkness-filter[m[m\r\n feat/explicit-image-dims[m[m\r\n fix-action-padding-lam-future-information-access[m[m\r\n fix-sampling[m[m\r\n fix-transformer-forwardpass[m[m\r\n fix/dyn-restore-after-nnx-upgrade[m[m\r\n fix/spatiotemporal-pe-once-in-STTransformer[m[m\r\n generate-minatar-breakout-dataset[m[m\r\n grad-norm-log-and-clip[m[m\r\n grain-dataloader[m[m\r\n gt-actions[m[m\r\n hotfix/eval-full-frame-fix[m[m\r\n hotfix/fix-val-loss-maskgit-masking[m[m\r\n hotfix/full-frame-eval-only-calculate-last-frame-metrics[m[m\r\n hotfix/sampling-shapes-error[m[m\r\n input_pipeline/add-npy2array_record[m[m\r\n logging-variants[m[m\r\n lr-schedules[m[m\r\n* [32mmain[m[m\r\n maskgit-different-maskprob-per-sample[m[m\r\n maskgit-sampling-iterative-unmasking-fix[m[m\r\n metrics-logging-for-dynamics-model[m[m\r\n monkey-patch[m[m\r\n new-arch-sampling[m[m\r\n preprocess_video[m[m\r\n refactor-full-frame-val-loss[m[m\r\n refactor-tmp[m[m\r\n remove-restore-branching[m[m\r\n revised-dataloader[m[m\r\n runner[m[m\r\n runner-grain[m[m\r\n sample-ali-branch[m[m\r\n sample-from-different-topologies[m[m\r\n sampling-script-add-metrics[m[m\r\n sampling-startframe-indexing-fix[m[m\r\n speedup-tfrecord-preprocessing[m[m\r\n train_lam_coinrun_ablation_wsd_3e-6_28747[m[m\r\n val-loss[m[m\r\n\r[K[?1l>]0;tum_cte0515@hkn0801:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 76 |
+
75,152630,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=01:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/sampling/maskgit/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/sampling/maskgit/%x_%j.log\n#SBATCH --job-name=coinrun_sample_maskgit\n\n# Unload modules that may interfere\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\n\n# Activate virtual environment\n# source .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\n\nCHECKPOINT_PATH=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn-gt-actions/train_dyn_default_gt_actions_breakout_longer_smaller_lr/3519530\n\n\necho ""Sampling from checkpoint: $CHECKPOINT_PATH""\n\nsrun python jasmine/sample.py \\n --checkpoint ""$CHECKPOINT_PATH"" \\n --data_dir=$array_records_dir \\n --seq_len=16 \\n --batch_size=12 \\n --start_frame=4 \\n --image_height=10 \\n --image_width=10 \\n --dyna_type=maskgit \\n --lam_patch_size=4 \\n --no-print-action-indices \\n --use_gt_actions \\n --output_dir ""gifs/50k/gt-actions""",shellscript,tab
|
| 77 |
+
76,153858,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",770,0,"",shellscript,selection_mouse
|
| 78 |
+
77,154737,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",770,0,"\n",shellscript,content
|
| 79 |
+
78,155739,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",770,0,"",shellscript,selection_mouse
|
| 80 |
+
79,156088,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",770,0,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn-gt-actions/train_dyn_default_gt_actions_breakout_longer/3519698",shellscript,content
|
| 81 |
+
80,156887,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1071,0,"",shellscript,selection_mouse
|
| 82 |
+
81,157929,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",915,0,"",shellscript,selection_command
|
| 83 |
+
82,158165,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",914,0,"\n",shellscript,content
|
| 84 |
+
83,159368,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",916,0,"",shellscript,selection_command
|
| 85 |
+
84,160343,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",916,0,"#",shellscript,content
|
| 86 |
+
85,160345,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",917,0,"",shellscript,selection_keyboard
|
| 87 |
+
86,160830,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",917,0," ",shellscript,content
|
| 88 |
+
87,160831,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",918,0,"",shellscript,selection_keyboard
|
| 89 |
+
88,160967,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",917,0,"",shellscript,selection_command
|
| 90 |
+
89,245053,"TERMINAL",0,0,"g",,terminal_output
|
| 91 |
+
90,245177,"TERMINAL",0,0,"t",,terminal_output
|
| 92 |
+
91,245888,"TERMINAL",0,0,"[K",,terminal_output
|
| 93 |
+
92,246100,"TERMINAL",0,0,"i",,terminal_output
|
| 94 |
+
93,246188,"TERMINAL",0,0,"t",,terminal_output
|
| 95 |
+
94,246349,"TERMINAL",0,0," ",,terminal_output
|
| 96 |
+
95,246424,"TERMINAL",0,0,"c",,terminal_output
|
| 97 |
+
96,246555,"TERMINAL",0,0,"he",,terminal_output
|
| 98 |
+
97,246692,"TERMINAL",0,0,"c",,terminal_output
|
| 99 |
+
98,246745,"TERMINAL",0,0,"k",,terminal_output
|
| 100 |
+
99,246921,"TERMINAL",0,0,"o",,terminal_output
|
| 101 |
+
100,246985,"TERMINAL",0,0,"u",,terminal_output
|
| 102 |
+
101,247049,"TERMINAL",0,0,"t",,terminal_output
|
| 103 |
+
102,247114,"TERMINAL",0,0," ",,terminal_output
|
| 104 |
+
103,247497,"TERMINAL",0,0,"[7madd-noise-to-combat-exposure-bias[27m",,terminal_output
|
| 105 |
+
104,247717,"TERMINAL",0,0,"add-noise-to-combat-exposure-bias\r\n[?2004l\r",,terminal_output
|
| 106 |
+
105,248450,"TERMINAL",0,0,"Switched to branch 'add-noise-to-combat-exposure-bias'\r\nYour branch is ahead of 'origin/add-noise-to-combat-exposure-bias' by 3 commits.\r\n (use ""git push"" to publish your local commits)\r\n]0;tum_cte0515@hkn0801:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 107 |
+
106,250505,"",0,0,"Switched from branch 'main' to 'add-noise-to-combat-exposure-bias'",,git_branch_checkout
|
| 108 |
+
107,275721,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=01:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/sampling/maskgit/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/sampling/maskgit/%x_%j.log\n#SBATCH --job-name=coinrun_sample_maskgit\n\n# Unload modules that may interfere\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\n\n# Activate virtual environment\n# source .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\n\nCHECKPOINT_PATH=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn-gt-actions/train_dyn_default_gt_actions_breakout_longer/3519698\n\n# /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn-gt-actions/train_dyn_default_gt_actions_breakout_longer_smaller_lr/3519530\n\n\necho ""Sampling from checkpoint: $CHECKPOINT_PATH""\n\nsrun python jasmine/sample.py \\n --checkpoint ""$CHECKPOINT_PATH"" \\n --data_dir=$array_records_dir \\n --seq_len=16 \\n --batch_size=12 \\n --start_frame=4 \\n --image_height=10 \\n --image_width=10 \\n --dyna_type=maskgit \\n --lam_patch_size=4 \\n --no-print-action-indices \\n --use_gt_actions \\n --output_dir ""gifs/50k/gt-actions""",shellscript,tab
|
| 109 |
+
108,275722,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1487,0,"",shellscript,selection_mouse
|
| 110 |
+
109,275723,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1486,0,"",shellscript,selection_command
|
| 111 |
+
110,275729,"TERMINAL",0,0,"bash",,terminal_focus
|
| 112 |
+
111,275787,"TERMINAL",0,0,"git git^C[?2004l\r[?2004h[?2004l\r\r\n]0;tum_cte0515@hkn0801:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 113 |
+
112,279744,"TERMINAL",0,0,"srun",,terminal_focus
|
| 114 |
+
113,281646,"TERMINAL",0,0,"g",,terminal_output
|
| 115 |
+
114,281705,"TERMINAL",0,0,"i",,terminal_output
|
| 116 |
+
115,281777,"TERMINAL",0,0,"t",,terminal_output
|
| 117 |
+
116,281834,"TERMINAL",0,0," ",,terminal_output
|
| 118 |
+
117,282004,"TERMINAL",0,0,"st",,terminal_output
|
| 119 |
+
118,282157,"TERMINAL",0,0,"a",,terminal_output
|
| 120 |
+
119,282210,"TERMINAL",0,0,"t",,terminal_output
|
| 121 |
+
120,282274,"TERMINAL",0,0,"u",,terminal_output
|
| 122 |
+
121,282538,"TERMINAL",0,0,"s",,terminal_output
|
| 123 |
+
122,282611,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output
|
| 124 |
+
123,282782,"TERMINAL",0,0,"On branch add-noise-to-combat-exposure-bias\r\nYour branch is ahead of 'origin/add-noise-to-combat-exposure-bias' by 3 commits.\r\n (use ""git push"" to publish your local commits)\r\n\r\n",,terminal_output
|
| 125 |
+
124,282859,"TERMINAL",0,0,"Last commands done (2 commands done):\r\n pick ba37453 feat: generate coinrun dataset with val split\r\n pick faadd10 feat: implemented validation loss for all three models\r\nNext commands to do (26 remaining commands):\r\n pick 9a17dbb fix: pass val data path to dataloader\r\n pick 6e69cdb fix typo in image logging\r\n (use ""git rebase --edit-todo"" to view and edit)\r\nYou are currently editing a commit while rebasing branch 'gt-actions' on 'c7522f2'.\r\n (use ""git commit --amend"" to amend the current commit)\r\n (use ""git rebase --continue"" once you are satisfied with your changes)\r\n\r\nUntracked files:\r\n (use ""git add <file>..."" to include in what will be committed)\r\n\t[31mdiff.diff[m\r\n\t[31mdiff2.diff[m\r\n\t[31minput_pipeline/[m\r\n\t[31mkiller.sh[m\r\n\t[31mkiller_partition.sh[m\r\n\t[31mlog.log[m\r\n\t[31moverfit_dir.zip[m\r\n\t[31mrequirements-franz.txt[m\r\n\t[31msamples/[m\r\n\t[31mscripts_cremers/[m\r\n\t[31mslurm/[m\r\n\t[31mtest.py[m\r\n\t[31mutils/[m\r\n\t[31muv.lock[m\r\n\r\nnothing added to commit but untracked files present (use ""git add"" to track)\r\n]0;tum_cte0515@hkn0801:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 126 |
+
125,294827,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",0,0,"",shellscript,tab
|
| 127 |
+
126,294828,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1475,0,"",shellscript,selection_mouse
|
| 128 |
+
127,295978,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1475,0,"-",shellscript,content
|
| 129 |
+
128,295980,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1476,0,"",shellscript,selection_keyboard
|
| 130 |
+
129,296561,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1476,0,"w",shellscript,content
|
| 131 |
+
130,296563,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1477,0,"",shellscript,selection_keyboard
|
| 132 |
+
131,296654,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1477,0,"-",shellscript,content
|
| 133 |
+
132,296655,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1478,0,"",shellscript,selection_keyboard
|
| 134 |
+
133,297206,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1478,0,"n",shellscript,content
|
| 135 |
+
134,297207,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1479,0,"",shellscript,selection_keyboard
|
| 136 |
+
135,297496,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1479,0,"i",shellscript,content
|
| 137 |
+
136,297497,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1480,0,"",shellscript,selection_keyboard
|
| 138 |
+
137,297797,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1479,1,"",shellscript,content
|
| 139 |
+
138,298013,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1479,0,"p",shellscript,content
|
| 140 |
+
139,298015,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1480,0,"",shellscript,selection_keyboard
|
| 141 |
+
140,298115,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1480,0,"i",shellscript,content
|
| 142 |
+
141,298117,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1481,0,"",shellscript,selection_keyboard
|
| 143 |
+
142,298415,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1480,1,"",shellscript,content
|
| 144 |
+
143,298540,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1479,1,"",shellscript,content
|
| 145 |
+
144,298713,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1479,0,"o",shellscript,content
|
| 146 |
+
145,298715,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1480,0,"",shellscript,selection_keyboard
|
| 147 |
+
146,298844,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1480,0,"i",shellscript,content
|
| 148 |
+
147,298846,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1481,0,"",shellscript,selection_keyboard
|
| 149 |
+
148,298877,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1481,0,"s",shellscript,content
|
| 150 |
+
149,298878,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1482,0,"",shellscript,selection_keyboard
|
| 151 |
+
150,299003,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1482,0,"e",shellscript,content
|
| 152 |
+
151,299004,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1483,0,"",shellscript,selection_keyboard
|
| 153 |
+
152,308200,"TERMINAL",0,0,"q",,terminal_output
|
| 154 |
+
153,308306,"TERMINAL",0,0,"u",,terminal_output
|
| 155 |
+
154,308385,"TERMINAL",0,0,"e",,terminal_output
|
| 156 |
+
155,308467,"TERMINAL",0,0,"ue",,terminal_output
|
| 157 |
+
156,308698,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h[22;0;0t[1;62r(B[m[4l[?7h[H[2JEvery 1.0s: squeue --me[1;78Hhkn0801.localdomain: Thu Sep 25 13:55:06 2025[3;14HJOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)[4;12H3520566 accelerat interact tum_cte0 R\t4:43\t 1 hkn0801[62;122H",,terminal_output
|
| 158 |
+
157,309613,"TERMINAL",0,0,"[62;1H[?1049l[23;0;0t\r[?1l>]0;tum_cte0515@hkn0801:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 159 |
+
158,310080,"TERMINAL",0,0,"c",,terminal_output
|
| 160 |
+
159,310144,"TERMINAL",0,0,"l",,terminal_output
|
| 161 |
+
160,310315,"TERMINAL",0,0,"e",,terminal_output
|
| 162 |
+
161,310520,"TERMINAL",0,0,"a",,terminal_output
|
| 163 |
+
162,310634,"TERMINAL",0,0,"r",,terminal_output
|
| 164 |
+
163,310784,"TERMINAL",0,0,"\r\n[?2004l\r[H[2J[3J]0;tum_cte0515@hkn0801:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 165 |
+
164,316438,"TERMINAL",0,0,"s",,terminal_output
|
| 166 |
+
165,316543,"TERMINAL",0,0,"h",,terminal_output
|
| 167 |
+
166,316636,"TERMINAL",0,0," ",,terminal_output
|
| 168 |
+
167,317338,"TERMINAL",0,0,"[7mslurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch[27m",,terminal_output
|
| 169 |
+
168,317731,"TERMINAL",0,0,"\r[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[Cslurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch\r\n[?2004l\r",,terminal_output
|
| 170 |
+
169,317863,"TERMINAL",0,0,"Sampling from checkpoint: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn-gt-actions/train_dyn_default_gt_actions_breakout_longer/3519698\r\n",,terminal_output
|
| 171 |
+
170,317994,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output
|
| 172 |
+
171,333523,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output
|
| 173 |
+
172,363785,"TERMINAL",0,0,"Per-frame SSIM:\r\n [0.30596483 0.414882 0.5163348 0.51084375 0.5155197 0.5155652\r\n 0.51382315 0.4914374 0.45811233 0.4443463 0.44947392 0.4618884 ]\r\nPer-frame PSNR:\r\n [18.745987 18.47467 18.217417 18.06475 18.032337 17.890877 17.707855\r\n 17.542704 17.384125 17.292942 17.300459 17.316257]\r\nSSIM: 0.4665159583091736\r\nPSNR: 17.83086585998535\r\n",,terminal_output
|
| 174 |
+
173,364404,"TERMINAL",0,0,"W0925 13:56:01.732225 2006205 pjrt_client.cc:1469] WatchJobStateAsync failed for task goo.gle/debugproto job_name: ""jax_worker"": UNAVAILABLE: failed to connect to all addresses; last error: UNKNOWN: ipv4:10.0.1.177:63542: Failed to connect to remote host: Connection refused\r\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/WatchJobState:\r\n:UNKNOWN:Error received from peer {grpc_message:""failed to connect to all addresses; last error: UNKNOWN: ipv4:10.0.1.177:63542: Failed to connect to remote host: Connection refused"", grpc_status:14}\r\n",,terminal_output
|
| 175 |
+
174,365078,"TERMINAL",0,0,"]0;tum_cte0515@hkn0801:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 176 |
+
175,372087,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",0,0,"",shellscript,tab
|
| 177 |
+
176,372088,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1425,0,"",shellscript,selection_mouse
|
| 178 |
+
177,372588,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1487,0,"",shellscript,selection_mouse
|
| 179 |
+
178,373273,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1425,0,"",shellscript,selection_mouse
|
| 180 |
+
179,373893,"slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",1448,0,"",shellscript,selection_mouse
|
| 181 |
+
180,442282,"jasmine/sample.py",0,0,"from dataclasses import dataclass\nimport time\nimport os\nimport optax\n\nimport dm_pix as pix\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\nimport numpy as np\nimport orbax.checkpoint as ocp\nfrom PIL import Image, ImageDraw\nimport tyro\nfrom flax import nnx\n\nfrom genie import Genie\nfrom utils.dataloader import get_dataloader\n\n\n@dataclass\nclass Args:\n # Experiment\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = ""data/coinrun_episodes""\n checkpoint: str = """"\n print_action_indices: bool = True\n output_dir: str = ""gifs/""\n # Sampling\n batch_size: int = 1\n maskgit_steps: int = 25\n temperature: float = 1.0\n sample_argmax: bool = True\n start_frame: int = 1\n noise_level: float = 0.0\n noise_buckets: int = 10\n # Tokenizer checkpoint\n tokenizer_dim: int = 512\n tokenizer_ffn_dim: int = 2048\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 4\n tokenizer_num_heads: int = 8\n # LAM checkpoint\n lam_dim: int = 512\n lam_ffn_dim: int = 2048\n latent_action_dim: int = 32\n num_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 4\n lam_num_heads: int = 8\n use_gt_actions: bool = False\n # Dynamics checkpoint\n dyna_type: str = ""maskgit""\n dyna_dim: int = 512\n dyna_ffn_dim: int = 2048\n dyna_num_blocks: int = 6\n dyna_num_heads: int = 8\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n\n\nargs = tyro.cli(Args)\n\nif __name__ == ""__main__"":\n """"""\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n """"""\n jax.distributed.initialize()\n\n rng = jax.random.key(args.seed)\n\n # --- Load Genie checkpoint ---\n rngs = nnx.Rngs(rng)\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n lam_ffn_dim=args.lam_ffn_dim,\n latent_action_dim=args.latent_action_dim,\n num_actions=args.num_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=False,\n max_noise_level=0.0,\n noise_buckets=args.noise_buckets,\n use_gt_actions=args.use_gt_actions,\n # Dynamics\n dyna_type=args.dyna_type,\n dyna_dim=args.dyna_dim,\n dyna_ffn_dim=args.dyna_ffn_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n # FIXME (f.srambical): implement spatiotemporal KV caching and set decode=True\n decode=False,\n rngs=rngs,\n )\n\n # Need to delete lam decoder for checkpoint loading\n if not args.use_gt_actions:\n assert genie.lam is not None\n del genie.lam.decoder\n\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n step_format_fixed_length=6,\n )\n checkpoint_manager = ocp.CheckpointManager(\n args.checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n dummy_tx = optax.adamw(\n learning_rate=optax.linear_schedule(0.0001, 0.0001, 10000),\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n dummy_optimizer = nnx.ModelAndOptimizer(genie, dummy_tx)\n\n abstract_optimizer = nnx.eval_shape(lambda: dummy_optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(dummy_optimizer, restored_optimizer_state)\n\n # --- Define sampling function ---\n def _sampling_fn(model: Genie, batch: dict) -> jax.Array:\n """"""Runs Genie.sample with pre-defined generation hyper-parameters.""""""\n assert args.dyna_type in [\n ""maskgit"",\n ""causal"",\n ], f""Invalid dynamics type: {args.dyna_type}""\n frames, _ = model.sample(\n batch,\n args.seq_len,\n args.temperature,\n args.sample_argmax,\n args.maskgit_steps,\n )\n return frames\n\n # --- Define autoregressive sampling loop ---\n def _autoreg_sample(genie, rng, batch):\n batch[""videos""] = batch[""videos""][:, : args.start_frame]\n batch[""rng""] = rng\n generated_vid_BSHWC = _sampling_fn(genie, batch)\n return generated_vid_BSHWC\n\n # --- Get video + latent actions ---\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n args.batch_size,\n args.image_height,\n args.image_width,\n args.image_channels,\n # We don't use workers in order to avoid grain shutdown issues (https://github.com/google/grain/issues/398)\n num_workers=0,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n dataloader = iter(dataloader)\n batch = next(dataloader)\n gt_video = jnp.asarray(batch[""videos""], dtype=jnp.float32) / 255.0\n batch[""videos""] = gt_video.astype(args.dtype)\n # Get latent actions for all videos in the batch\n action_batch_E = None\n if not args.use_gt_actions:\n action_batch_E = genie.vq_encode(batch, training=False)\n batch[""latent_actions""] = action_batch_E\n\n # --- Sample + evaluate video ---\n recon_video_BSHWC = _autoreg_sample(genie, rng, batch)\n recon_video_BSHWC = recon_video_BSHWC.astype(jnp.float32)\n\n gt = gt_video.clip(0, 1)[:, args.start_frame :]\n recon = recon_video_BSHWC.clip(0, 1)[:, args.start_frame :]\n\n ssim_vmap = jax.vmap(pix.ssim, in_axes=(0, 0))\n psnr_vmap = jax.vmap(pix.psnr, in_axes=(0, 0))\n ssim = ssim_vmap(gt, recon)\n psnr = psnr_vmap(gt, recon)\n per_frame_ssim = ssim.mean(0)\n per_frame_psnr = psnr.mean(0)\n avg_ssim = ssim.mean()\n avg_psnr = psnr.mean()\n\n print(""Per-frame SSIM:\n"", per_frame_ssim)\n print(""Per-frame PSNR:\n"", per_frame_psnr)\n\n print(f""SSIM: {avg_ssim}"")\n print(f""PSNR: {avg_psnr}"")\n\n # --- Construct video ---\n true_videos = (gt_video * 255).astype(np.uint8)\n pred_videos = (recon_video_BSHWC * 255).astype(np.uint8)\n video_comparison = np.zeros((2, *recon_video_BSHWC.shape), dtype=np.uint8)\n video_comparison[0] = true_videos[:, : args.seq_len]\n video_comparison[1] = pred_videos\n frames = einops.rearrange(video_comparison, ""n b t h w c -> t (b h) (n w) c"")\n\n # --- Save video ---\n imgs = [Image.fromarray(img) for img in frames]\n # Write actions on each frame, on each row (i.e., for each video in the batch, on the GT row)\n B = batch[""videos""].shape[0]\n if action_batch_E is not None:\n action_batch_BSm11 = jnp.reshape(action_batch_E, (B, args.seq_len - 1, 1))\n else:\n action_batch_BSm11 = jnp.reshape(\n batch[""actions""][:, :-1], (B, args.seq_len - 1, 1)\n )\n for t, img in enumerate(imgs[1:]):\n d = ImageDraw.Draw(img)\n for row in range(B):\n if args.print_action_indices:\n action = action_batch_BSm11[row, t, 0]\n y_offset = row * batch[""videos""].shape[2] + 2\n d.text((2, y_offset), f""{action}"", fill=255)\n\n os.makedirs(args.output_dir, exist_ok=True)\n imgs[0].save(\n os.path.join(args.output_dir, f""generation_{time.time()}.gif""),\n save_all=True,\n append_images=imgs[1:],\n duration=250,\n loop=0,\n )\n",python,tab
|
| 182 |
+
181,454590,"jasmine/sample.py",5519,0,"",python,selection_mouse
|
| 183 |
+
182,454728,"jasmine/sample.py",4823,0,"",python,selection_command
|
| 184 |
+
183,457080,"jasmine/sample.py",5118,0,"",python,selection_mouse
|
| 185 |
+
184,457562,"jasmine/genie.py",0,0,"from typing import Dict\n\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.nnx as nnx\nimport orbax.checkpoint as ocp\n\nfrom models.dynamics import DynamicsMaskGIT, DynamicsCausal\nfrom models.lam import LatentActionModel\nfrom models.tokenizer import TokenizerVQVAE\n\n\nclass Genie(nnx.Module):\n """"""Genie model""""""\n\n def __init__(\n self,\n in_dim: int,\n tokenizer_dim: int,\n tokenizer_ffn_dim: int,\n latent_patch_dim: int,\n num_patch_latents: int,\n patch_size: int,\n tokenizer_num_blocks: int,\n tokenizer_num_heads: int,\n lam_dim: int,\n lam_ffn_dim: int,\n latent_action_dim: int,\n num_actions: int,\n lam_patch_size: int,\n lam_num_blocks: int,\n lam_num_heads: int,\n lam_co_train: bool,\n use_gt_actions: bool,\n dyna_type: str,\n dyna_dim: int,\n dyna_ffn_dim: int,\n dyna_num_blocks: int,\n dyna_num_heads: int,\n max_noise_level: float,\n noise_buckets: int,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n dropout: float = 0.0,\n mask_limit: float = 0.0,\n ):\n # --- Tokenizer ---\n self.in_dim = in_dim\n self.tokenizer_dim = tokenizer_dim\n self.tokenizer_ffn_dim = tokenizer_ffn_dim\n self.latent_patch_dim = latent_patch_dim\n self.num_patch_latents = num_patch_latents\n self.patch_size = patch_size\n self.tokenizer_num_blocks = tokenizer_num_blocks\n self.tokenizer_num_heads = tokenizer_num_heads\n # --- LAM ---\n self.lam_dim = lam_dim\n self.lam_ffn_dim = lam_ffn_dim\n self.latent_action_dim = latent_action_dim\n self.num_actions = num_actions\n self.lam_patch_size = lam_patch_size\n self.lam_num_blocks = lam_num_blocks\n self.lam_num_heads = lam_num_heads\n self.lam_co_train = lam_co_train\n self.use_gt_actions = use_gt_actions\n # --- Dynamics ---\n self.dyna_type = dyna_type\n self.dyna_dim = dyna_dim\n self.dyna_ffn_dim = dyna_ffn_dim\n self.dyna_num_blocks = dyna_num_blocks\n self.dyna_num_heads = dyna_num_heads\n self.max_noise_level = max_noise_level\n self.noise_buckets = noise_buckets\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.dropout = dropout\n self.mask_limit = mask_limit\n self.decode = decode\n\n self.tokenizer = TokenizerVQVAE(\n in_dim=self.in_dim,\n model_dim=self.tokenizer_dim,\n ffn_dim=self.tokenizer_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_patch_latents,\n patch_size=self.patch_size,\n num_blocks=self.tokenizer_num_blocks,\n num_heads=self.tokenizer_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n if self.use_gt_actions:\n self.action_embed = nnx.Embed(\n self.num_actions, self.latent_action_dim, rngs=rngs\n )\n self.lam = None\n else:\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n ffn_dim=self.lam_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_actions,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.action_embed = None\n if self.dyna_type == ""maskgit"":\n self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n max_noise_level=self.max_noise_level,\n noise_buckets=self.noise_buckets,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n elif self.dyna_type == ""causal"":\n self.dynamics = DynamicsCausal(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n max_noise_level=self.max_noise_level,\n noise_buckets=self.noise_buckets,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=decode,\n rngs=rngs,\n )\n else:\n raise ValueError(f""Invalid dynamics type: {self.dyna_type}"")\n\n def __call__(\n self,\n batch: Dict[str, jax.Array],\n ) -> Dict[str, jax.Array]:\n videos_BTHWC = batch[""videos""]\n tokenizer_outputs = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_indices_BTN = tokenizer_outputs[""indices""]\n latent_actions_BTm11L = None\n action_embeddings_BTm11L = None\n if self.use_gt_actions:\n assert self.action_embed is not None\n action_indices_E = None\n action_embeddings_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n action_embeddings_BTm11L = action_embeddings_BT1L[:, :-1]\n else:\n assert self.lam is not None\n lam_outputs = self.lam.vq_encode(videos_BTHWC, training=False)\n z_q_BTm11L = lam_outputs[""z_q""]\n action_indices_E = lam_outputs[""indices""]\n latent_actions_BTm11L = jax.lax.cond(\n self.lam_co_train,\n lambda: z_q_BTm11L,\n lambda: jax.lax.stop_gradient(z_q_BTm11L),\n )\n outputs = dict(\n video_tokens=jax.lax.stop_gradient(token_indices_BTN),\n latent_actions=(\n action_embeddings_BTm11L\n if self.use_gt_actions\n else latent_actions_BTm11L\n ),\n )\n outputs[""mask_rng""] = batch[""rng""]\n dyna_logits_BTNV, dyna_mask = self.dynamics(outputs)\n outputs[""token_logits""] = dyna_logits_BTNV\n outputs[""mask""] = dyna_mask\n mle_indices_BTN = jnp.argmax(outputs[""token_logits""], axis=-1)\n H, W = batch[""videos""].shape[2:4]\n outputs[""recon""] = self.tokenizer.decode(mle_indices_BTN, (H, W))\n if action_indices_E is not None:\n outputs[""lam_indices""] = action_indices_E\n return outputs\n\n def sample(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n maskgit_steps: int = 25,\n ) -> tuple[jax.Array, jax.Array]:\n if self.dyna_type == ""maskgit"":\n return self.sample_maskgit(\n batch, seq_len, 0.0, maskgit_steps, temperature, sample_argmax\n )\n elif self.dyna_type == ""causal"":\n return self.sample_causal(batch, seq_len, temperature, sample_argmax)\n else:\n raise ValueError(f""Dynamics model type unknown: {self.dyna_type}"")\n\n def sample_maskgit(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n noise_level: float,\n steps: int = 25,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> tuple[jax.Array, jax.Array]:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n P: S * N\n """"""\n assert isinstance(self.dynamics, DynamicsMaskGIT)\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n init_logits_BSNV = jnp.zeros(\n shape=(*token_idxs_BSN.shape, self.num_patch_latents)\n )\n noise_level = jnp.array(noise_level)\n if self.use_gt_actions:\n assert self.action_embed is not None\n latent_actions_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n latent_actions_BTm11L = latent_actions_BT1L[:, :-1]\n action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\n else:\n assert self.lam is not None\n latent_actions_E = batch[""latent_actions""]\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n\n # --- Extract submodule state ---\n dynamics_state = nnx.state(self.dynamics)\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def maskgit_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array],\n step: jax.Array,\n ) -> tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array]:\n rng, token_idxs_BSN, logits_BSNV, mask_BSN, action_tokens_EL = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # We need to reconstruct the submodule inside scan body to prevent trace context mismatches\n dynamics_maskgit = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n max_noise_level=self.max_noise_level,\n noise_buckets=self.noise_buckets,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=nnx.Rngs(0),\n )\n nnx.update(dynamics_maskgit, dynamics_state)\n\n # --- Construct + encode video ---\n vid_embed_BSNM = dynamics_maskgit.patch_embed(token_idxs_BSN)\n mask_token_111M = dynamics_maskgit.mask_token.value\n mask_expanded_BSN1 = mask_BSN[..., None]\n vid_embed_BSNM = jnp.where(\n mask_expanded_BSN1, mask_token_111M, vid_embed_BSNM\n )\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = dynamics_maskgit.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(\n act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\n )\n # TODO mihir\n\n rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 / self.max_noise_level) * self.noise_buckets\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n vid_embed_BSNM += jnp.expand_dims(noise_level_B11, -1)\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )\n unmasked_ratio = jnp.cos(jnp.pi * (step + 1) / (steps * 2))\n step_temp = temperature * (1.0 - unmasked_ratio)\n final_logits_BSNp2V = (\n dynamics_maskgit.transformer(vid_embed_BSNp2M) / step_temp\n )\n final_logits_BSNV = final_logits_BSNp2V[:, :, 2:]\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_BSN = jnp.argmax(final_logits_BSNV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_BSN = jax.random.categorical(_rng, final_logits_BSNV)\n gather_fn = jax.vmap(jax.vmap(jax.vmap(lambda x, y: x[y])))\n final_token_probs_BSN = gather_fn(\n jax.nn.softmax(final_logits_BSNV), sampled_token_idxs_BSN\n )\n final_token_probs_BSN += ~mask_BSN\n # Update masked tokens and logits only\n token_idxs_BSN = jnp.where(mask_BSN, sampled_token_idxs_BSN, token_idxs_BSN)\n logits_BSNV = jnp.where(\n jnp.expand_dims(mask_BSN, -1), final_logits_BSNV, logits_BSNV\n )\n\n # --- Update mask ---\n num_unmasked_tokens = jnp.round(N * (1.0 - unmasked_ratio)).astype(int)\n final_token_probs_flat_BP = einops.rearrange(\n final_token_probs_BSN, ""b s n -> b (s n)""\n )\n idx_mask_P = (\n jnp.arange(final_token_probs_flat_BP.shape[-1])\n <= N - num_unmasked_tokens\n )\n sorted_idxs_BP = jnp.argsort(final_token_probs_flat_BP, axis=-1)\n mask_update_fn = jax.vmap(lambda msk, ids: msk.at[ids].set(idx_mask_P))\n mask_flat_BP = einops.rearrange(mask_BSN, ""b s n -> b (s n)"")\n new_mask_flat_BP = mask_update_fn(mask_flat_BP, sorted_idxs_BP)\n new_mask_BSN = einops.rearrange(new_mask_flat_BP, ""b (s n) -> b s n"", n=N)\n\n new_carry = (\n rng,\n token_idxs_BSN,\n logits_BSNV,\n new_mask_BSN,\n action_tokens_EL,\n )\n return new_carry\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[jax.Array, jax.Array, jax.Array]:\n rng, current_token_idxs_BSN, current_logits_BSNV = carry\n rng, step_rng = jax.random.split(rng)\n\n # Mask current frame (i.e., t == step_t)\n mask_S = jnp.arange(seq_len) == step_t\n mask_BSN = jnp.broadcast_to(mask_S[None, :, None], (B, seq_len, N)).astype(\n bool\n )\n masked_token_idxs_BSN = current_token_idxs_BSN * ~mask_BSN\n masked_logits_BSNV = current_logits_BSNV * jnp.expand_dims(~mask_BSN, -1)\n\n # --- Initialize and run MaskGIT loop ---\n init_carry_maskgit = (\n step_rng,\n masked_token_idxs_BSN,\n masked_logits_BSNV,\n mask_BSN,\n action_tokens_EL,\n )\n final_carry_maskgit = maskgit_step_fn(init_carry_maskgit, jnp.arange(steps))\n updated_token_idxs_BSN = final_carry_maskgit[1]\n updated_logits_BSNV = final_carry_maskgit[2]\n new_carry = (rng, updated_token_idxs_BSN, updated_logits_BSNV)\n return new_carry\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN, init_logits_BSNV)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry = generation_step_fn(initial_carry, timesteps_to_scan)\n final_token_idxs_BSN = final_carry[1]\n final_logits_BSNV = final_carry[2]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC, final_logits_BSNV\n\n def sample_causal(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> tuple[jax.Array, jax.Array]:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n """"""\n assert isinstance(self.dynamics, DynamicsCausal)\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n logits_BSNV = jnp.zeros((*token_idxs_BSN.shape, self.num_patch_latents))\n dynamics_state = nnx.state(self.dynamics)\n\n if self.use_gt_actions:\n assert self.action_embed is not None\n latent_actions_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n latent_actions_BTm11L = latent_actions_BT1L[:, :-1]\n action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\n else:\n assert self.lam is not None\n latent_actions_E = batch[""latent_actions""]\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def causal_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array],\n step_n: jax.Array,\n ) -> tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array]:\n rng, token_idxs_BSN, logits_BSNV, action_tokens_EL, step_t = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # We need to reconstruct the submodule inside scan body to prevent trace context mismatches\n dynamics_causal = DynamicsCausal(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=self.decode,\n rngs=nnx.Rngs(0),\n )\n nnx.update(dynamics_causal, dynamics_state)\n\n # --- Construct + encode video ---\n vid_embed_BSNM = dynamics_causal.patch_embed(token_idxs_BSN)\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = dynamics_causal.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(\n act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\n )\n vid_embed_BSNp1M = jnp.concatenate([act_embed_BS1M, vid_embed_BSNM], axis=2)\n final_logits_BTNp1V = (\n dynamics_causal.transformer(vid_embed_BSNp1M, (step_t, step_n))\n / temperature\n )\n final_logits_BV = final_logits_BTNp1V[:, step_t, step_n, :]\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_B = jnp.argmax(final_logits_BV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_B = jax.random.categorical(_rng, final_logits_BV)\n # Update next tokens only\n token_idxs_BSN = token_idxs_BSN.at[:, step_t, step_n].set(\n sampled_token_idxs_B\n )\n logits_BSNV = logits_BSNV.at[:, step_t, step_n].set(final_logits_BV)\n\n new_carry = (rng, token_idxs_BSN, logits_BSNV, action_tokens_EL, step_t)\n return new_carry\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[jax.Array, jax.Array, jax.Array]:\n rng, current_token_idxs_BSN, current_logits_BSNV = carry\n rng, step_rng = jax.random.split(rng)\n\n # --- Initialize and run causal loop ---\n init_carry_causal = (\n step_rng,\n current_token_idxs_BSN,\n current_logits_BSNV,\n action_tokens_EL,\n step_t,\n )\n final_carry_causal = causal_step_fn(init_carry_causal, jnp.arange(N))\n updated_token_idxs_BSN = final_carry_causal[1]\n updated_logits_BSNV = final_carry_causal[2]\n new_carry = (rng, updated_token_idxs_BSN, updated_logits_BSNV)\n return new_carry\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN, logits_BSNV)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry = generation_step_fn(initial_carry, timesteps_to_scan)\n final_token_idxs_BSN = final_carry[1]\n final_logits_BSNV = final_carry[2]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC, final_logits_BSNV\n\n def vq_encode(self, batch: Dict[str, jax.Array], training: bool) -> jax.Array:\n # --- Preprocess videos ---\n assert self.lam is not None\n video_BTHWC = batch[""videos""]\n lam_output: Dict[str, jax.Array] = self.lam.vq_encode(\n video_BTHWC, training=training\n )\n lam_indices_E = lam_output[""indices""]\n return lam_indices_E\n\n\n# FIXME (f.srambical): add conversion script for old checkpoints\ndef restore_genie_components(\n optimizer: nnx.ModelAndOptimizer,\n sharding: jax.sharding.NamedSharding,\n rng: jax.Array,\n args,\n) -> nnx.ModelAndOptimizer:\n """"""Restore pre-trained Genie components""""""\n rng_tokenizer, rng_lam = jax.random.split(rng)\n rngs_tokenizer = nnx.Rngs(rng_tokenizer)\n rngs_lam = nnx.Rngs(rng_lam)\n\n tx = optimizer.tx\n model = optimizer.model\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n step_format_fixed_length=6,\n )\n tokenizer_checkpoint_manager = ocp.CheckpointManager(\n directory=args.tokenizer_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.tokenizer_dim,\n ffn_dim=args.tokenizer_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n num_blocks=args.tokenizer_num_blocks,\n num_heads=args.tokenizer_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs_tokenizer,\n )\n dummy_tokenizer_optimizer = nnx.ModelAndOptimizer(dummy_tokenizer, tx)\n dummy_tokenizer_optimizer_state = nnx.state(dummy_tokenizer_optimizer)\n abstract_sharded_tokenizer_optimizer_state = _create_abstract_sharded_pytree(\n dummy_tokenizer_optimizer_state, sharding\n )\n restored_tokenizer = tokenizer_checkpoint_manager.restore(\n step=tokenizer_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_tokenizer_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_tokenizer_optimizer.model, restored_tokenizer.model)\n model.tokenizer = dummy_tokenizer_optimizer.model\n tokenizer_checkpoint_manager.close()\n\n if args.lam_checkpoint:\n lam_checkpoint_manager = ocp.CheckpointManager(\n directory=args.lam_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.lam_dim,\n ffn_dim=args.lam_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_actions,\n patch_size=args.lam_patch_size,\n num_blocks=args.lam_num_blocks,\n num_heads=args.lam_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs_lam,\n )\n dummy_lam_optimizer = nnx.ModelAndOptimizer(dummy_lam, tx)\n dummy_lam_optimizer_state = nnx.state(dummy_lam_optimizer)\n abstract_sharded_lam_optimizer_state = _create_abstract_sharded_pytree(\n dummy_lam_optimizer_state, sharding\n )\n restored_lam_optimizer = lam_checkpoint_manager.restore(\n step=lam_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_lam_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_lam_optimizer.model, restored_lam_optimizer.model)\n model.lam = dummy_lam_optimizer.model\n # Remove the LAM decoder to save memory and avoid unnecessary computation.\n del model.lam.decoder\n lam_checkpoint_manager.close()\n\n # Reinitialize the optimizer states\n optimizer = nnx.ModelAndOptimizer(model, tx)\n return optimizer\n\n\ndef _create_abstract_sharded_pytree(\n pytree_template: nnx.GraphState, sharding_spec: jax.sharding.NamedSharding\n) -> jax.Array:\n """"""Replaces arrays in a pytree with ShapeDtypeStructs having the given sharding.""""""\n\n def map_fn(leaf_template):\n if hasattr(leaf_template, ""shape"") and hasattr(leaf_template, ""dtype""):\n return jax.ShapeDtypeStruct(\n leaf_template.shape, leaf_template.dtype, sharding=sharding_spec\n )\n return leaf_template\n\n return jax.tree_util.tree_map(map_fn, pytree_template)\n",python,tab
|
| 186 |
+
185,457564,"jasmine/genie.py",7548,0,"",python,selection_command
|
| 187 |
+
186,460084,"jasmine/genie.py",7881,0,"",python,selection_mouse
|
| 188 |
+
187,460257,"jasmine/genie.py",7881,1,"0",python,selection_mouse
|
| 189 |
+
188,460258,"jasmine/genie.py",7881,2,"0.",python,selection_mouse
|
| 190 |
+
189,460304,"jasmine/genie.py",7881,3,"0.0",python,selection_mouse
|
| 191 |
+
190,460682,"jasmine/genie.py",7884,0,"",python,selection_mouse
|
| 192 |
+
191,461284,"jasmine/genie.py",7881,0,"",python,selection_mouse
|
| 193 |
+
192,461418,"jasmine/genie.py",7881,1,"0",python,selection_mouse
|
| 194 |
+
193,461475,"jasmine/genie.py",7881,2,"0.",python,selection_mouse
|
| 195 |
+
194,461475,"jasmine/genie.py",7881,3,"0.0",python,selection_mouse
|
| 196 |
+
195,461861,"jasmine/genie.py",7884,0,"",python,selection_mouse
|
| 197 |
+
196,468902,"jasmine/genie.py",7883,1,"",python,content
|
| 198 |
+
197,469934,"jasmine/genie.py",7883,0,"3",python,content
|
| 199 |
+
198,469935,"jasmine/genie.py",7884,0,"",python,selection_keyboard
|
| 200 |
+
199,471302,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",,terminal_output
|
| 201 |
+
200,472807,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output
|
| 202 |
+
201,472919,"TERMINAL",0,0,"Sampling from checkpoint: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn-gt-actions/train_dyn_default_gt_actions_breakout_longer/3519698\r\n",,terminal_output
|
| 203 |
+
202,473043,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output
|
| 204 |
+
203,481577,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output
|
| 205 |
+
204,511471,"TERMINAL",0,0,"Per-frame SSIM:\r\n [0.30596483 0.414882 0.5163348 0.51084375 0.5155197 0.5155652\r\n 0.51382315 0.4914374 0.45811233 0.4443463 0.44947392 0.4618884 ]\r\nPer-frame PSNR:\r\n [18.745987 18.47467 18.217417 18.06475 18.032337 17.890877 17.707855\r\n 17.542704 17.384125 17.292942 17.300459 17.316257]\r\nSSIM: 0.4665159583091736\r\nPSNR: 17.83086585998535\r\n",,terminal_output
|
| 206 |
+
205,511990,"TERMINAL",0,0,"W0925 13:58:29.386010 2007888 pjrt_client.cc:1469] WatchJobStateAsync failed for task goo.gle/debugstr job_name: ""jax_worker"": CANCELLED: CANCELLED\r\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/WatchJobState:\r\n:UNKNOWN:Error received from peer {grpc_message:""CANCELLED"", grpc_status:1} [type.googleapis.com/tensorflow.DerivedStatus='']\r\n",,terminal_output
|
| 207 |
+
206,512521,"TERMINAL",0,0,"]0;tum_cte0515@hkn0801:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 208 |
+
207,541605,"jasmine/genie.py",0,0,"",python,tab
|
| 209 |
+
208,541607,"jasmine/genie.py",7883,0,"",python,selection_mouse
|
| 210 |
+
209,542863,"jasmine/genie.py",7883,1,"",python,content
|
| 211 |
+
210,542996,"jasmine/genie.py",7883,0,"7",python,content
|
| 212 |
+
211,542998,"jasmine/genie.py",7884,0,"",python,selection_keyboard
|
| 213 |
+
212,545050,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch",,terminal_output
|
| 214 |
+
213,545231,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output
|
| 215 |
+
214,545329,"TERMINAL",0,0,"Sampling from checkpoint: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn-gt-actions/train_dyn_default_gt_actions_breakout_longer/3519698\r\n",,terminal_output
|
| 216 |
+
215,545463,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output
|
| 217 |
+
216,548159,"jasmine/genie.py",0,0,"",python,tab
|
| 218 |
+
217,548160,"jasmine/genie.py",7841,0,"",python,selection_mouse
|
| 219 |
+
218,548757,"jasmine/genie.py",8167,0,"",python,selection_command
|
| 220 |
+
219,553802,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output
|
| 221 |
+
220,555890,"jasmine/genie.py",8269,0,"",python,selection_mouse
|
| 222 |
+
221,556022,"jasmine/genie.py",8264,11,"noise_level",python,selection_mouse
|
| 223 |
+
222,583516,"TERMINAL",0,0,"Per-frame SSIM:\r\n [0.30596483 0.414882 0.5163348 0.51084375 0.5155197 0.5155652\r\n 0.51382315 0.4914374 0.45811233 0.4443463 0.44947392 0.4618884 ]\r\nPer-frame PSNR:\r\n [18.745987 18.47467 18.217417 18.06475 18.032337 17.890877 17.707855\r\n 17.542704 17.384125 17.292942 17.300459 17.316257]\r\nSSIM: 0.4665159583091736\r\nPSNR: 17.83086585998535\r\n",,terminal_output
|
| 224 |
+
223,584098,"TERMINAL",0,0,"W0925 13:59:41.451424 2009277 pjrt_client.cc:1469] WatchJobStateAsync failed for task goo.gle/debugproto job_name: ""jax_worker"": UNAVAILABLE: failed to connect to all addresses; last error: UNKNOWN: ipv4:10.0.1.177:63542: Failed to connect to remote host: Connection refused\r\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/WatchJobState:\r\n:UNKNOWN:Error received from peer {grpc_message:""failed to connect to all addresses; last error: UNKNOWN: ipv4:10.0.1.177:63542: Failed to connect to remote host: Connection refused"", grpc_status:14}\r\n",,terminal_output
|
| 225 |
+
224,584621,"TERMINAL",0,0,"]0;tum_cte0515@hkn0801:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 226 |
+
225,759719,"jasmine/genie.py",13893,0,"",python,selection_mouse
|
| 227 |
+
226,814183,"jasmine/genie.py",0,0,"",python,tab
|
| 228 |
+
227,814184,"jasmine/genie.py",13869,0,"",python,selection_mouse
|
| 229 |
+
228,823620,"jasmine/genie.py",14116,0,"",python,selection_mouse
|
| 230 |
+
229,893246,"TERMINAL",0,0,"q",,terminal_output
|
| 231 |
+
230,893311,"TERMINAL",0,0,"u",,terminal_output
|
| 232 |
+
231,893434,"TERMINAL",0,0,"e",,terminal_output
|
| 233 |
+
232,893487,"TERMINAL",0,0,"u",,terminal_output
|
| 234 |
+
233,893627,"TERMINAL",0,0,"e",,terminal_output
|
| 235 |
+
234,893813,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h[22;0;0t[1;62r(B[m[4l[?7h[H[2JEvery 1.0s: squeue --me[1;78Hhkn0801.localdomain: Thu Sep 25 14:04:51 2025[3;14HJOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)[4;12H3520566 accelerat interact tum_cte0 R[56G14:28\t 1 hkn0801[62;122H",,terminal_output
|
| 236 |
+
235,894871,"TERMINAL",0,0,"[1;117H2[4;60H9[62;122H",,terminal_output
|
| 237 |
+
236,894980,"TERMINAL",0,0,"[62;1H[?1049l[23;0;0t\r[?1l>]0;tum_cte0515@hkn0801:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0801 jasmine]$ ",,terminal_output
|
| 238 |
+
237,900148,"jasmine/genie.py",0,0,"",python,tab
|
| 239 |
+
238,900149,"jasmine/genie.py",14080,0,"",python,selection_mouse
|
| 240 |
+
239,900647,"jasmine/genie.py",14079,0,"",python,selection_command
|
| 241 |
+
240,900965,"jasmine/genie.py",14115,0,"",python,selection_command
|
| 242 |
+
241,901139,"jasmine/genie.py",14165,0,"",python,selection_command
|
| 243 |
+
242,901286,"jasmine/genie.py",14227,0,"",python,selection_command
|
| 244 |
+
243,901585,"jasmine/genie.py",14316,0,"",python,selection_command
|
| 245 |
+
244,901908,"jasmine/genie.py",14373,0,"",python,selection_command
|
| 246 |
+
245,902094,"jasmine/genie.py",14435,0,"",python,selection_command
|
| 247 |
+
246,902232,"jasmine/genie.py",14461,0,"",python,selection_command
|
| 248 |
+
247,902395,"jasmine/genie.py",14508,0,"",python,selection_command
|
| 249 |
+
248,902561,"jasmine/genie.py",14559,0,"",python,selection_command
|
| 250 |
+
249,902711,"jasmine/genie.py",14621,0,"",python,selection_command
|
| 251 |
+
250,902867,"jasmine/genie.py",14685,0,"",python,selection_command
|
| 252 |
+
251,903018,"jasmine/genie.py",14747,0,"",python,selection_command
|
| 253 |
+
252,903161,"jasmine/genie.py",14777,0,"",python,selection_command
|
| 254 |
+
253,903344,"jasmine/genie.py",14779,0,"",python,selection_command
|
| 255 |
+
254,903485,"jasmine/genie.py",14812,0,"",python,selection_command
|
| 256 |
+
255,903626,"jasmine/genie.py",14874,0,"",python,selection_command
|
| 257 |
+
256,903785,"jasmine/genie.py",14954,0,"",python,selection_command
|
| 258 |
+
257,904314,"jasmine/genie.py",15012,0,"",python,selection_command
|
| 259 |
+
258,904340,"jasmine/genie.py",15026,0,"",python,selection_command
|
| 260 |
+
259,904414,"jasmine/genie.py",15053,0,"",python,selection_command
|
| 261 |
+
260,904645,"jasmine/genie.py",15115,0,"",python,selection_command
|
| 262 |
+
261,904840,"jasmine/genie.py",15160,0,"",python,selection_command
|
| 263 |
+
262,905342,"jasmine/genie.py",15174,0,"",python,selection_command
|
| 264 |
+
263,905374,"jasmine/genie.py",15236,0,"",python,selection_command
|
| 265 |
+
264,905440,"jasmine/genie.py",15313,0,"",python,selection_command
|
| 266 |
+
265,905488,"jasmine/genie.py",15397,0,"",python,selection_command
|
| 267 |
+
266,905520,"jasmine/genie.py",15471,0,"",python,selection_command
|
| 268 |
+
267,905521,"jasmine/genie.py",15547,0,"",python,selection_command
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-6791460b-ec38-4da2-872f-193943c12d601753274780799-2025_07_23-17.17.23.114/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-75084265-2b4d-4d6f-86ae-c0ab064f62491758992086879-2025_09_27-18.55.18.494/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-7b06cc31-85b7-4591-87e0-b26b0ddee2111758710564601-2025_09_24-12.43.30.174/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-92ab1593-f937-4cc4-a174-544581a6ac991751909174142-2025_07_07-19.26.40.736/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-9ab67804-e8b6-44ba-9ee4-ddec1e42461f1757968391960-2025_09_15-22.33.57.229/source.csv
ADDED
|
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,1209,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"10:33:57 PM [info] Activating crowd-code\n10:33:57 PM [info] Recording started\n10:33:57 PM [info] Initializing git provider using file system watchers...\n10:33:57 PM [info] Git repository found\n10:33:57 PM [info] Git provider initialized successfully\n10:33:57 PM [info] Initial git state: [object Object]\n",Log,tab
|
| 3 |
+
3,2243,"TERMINAL",0,0,"bash",,terminal_focus
|
| 4 |
+
4,10241,"TERMINAL",0,0,"git status",,terminal_command
|
| 5 |
+
5,10311,"TERMINAL",0,0,"]633;COn branch val-loss\r\nYour branch is up to date with 'origin/val-loss'.\r\n\r\nChanges not staged for commit:\r\n (use ""git add <file>..."" to update what will be committed)\r\n (use ""git restore <file>..."" to discard changes in working directory)\r\n\t[31mmodified: models/dynamics.py[m\r\n\r\nUntracked files:\r\n (use ""git add <file>..."" to include in what will be committed)\r\n\t[31mdata/[m\r\n\t[31mdiff.diff[m\r\n\t[31mkiller.sh[m\r\n\t[31mkiller_partition.sh[m\r\n\t[31mlog.log[m\r\n\t[31mlogs/[m\r\n\t[31moverfit_dir.zip[m\r\n\t[31mrequirements-franz.txt[m\r\n\t[31msamples/[m\r\n\t[31mscripts_cremers/[m\r\n\t[31mslurm/[m\r\n\t[31mutils/visualizer.py[m\r\n\r\nno changes added to commit (use ""git add"" and/or ""git commit -a"")\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output
|
| 6 |
+
6,14147,"TERMINAL",0,0,"git diff",,terminal_command
|
| 7 |
+
7,14189,"TERMINAL",0,0,"]633;C[?1h=\r[1mdiff --git a/models/dynamics.py b/models/dynamics.py[m[m\r\n[1mindex 74fde10..1ad90cf 100644[m[m\r\n[1m--- a/models/dynamics.py[m[m\r\n[1m+++ b/models/dynamics.py[m[m\r\n[36m@@ -72,9 +72,14 @@[m [mclass DynamicsMaskGIT(nnx.Module):[m[m\r\n )[m[m\r\n [m[m\r\n def __call__([m[m\r\n[31m- self, batch: Dict[str, jax.Array], training: bool = True, pred_full_frame: bool = False,[m[m\r\n[32m+[m[32m self,[m[m\r\n[32m+[m[32m batch: Dict[str, jax.Array],[m[m\r\n[32m+[m[32m training: bool = True,[m[m\r\n[32m+[m[32m pred_full_frame: bool = False,[m[m\r\n ) -> tuple[jax.Array, jax.Array | None]:[m[m\r\n[31m- assert not (training and pred_full_frame), ""Cannot evaluate full frame prediction during training.""[m[m\r\n[32m+[m[32m assert not ([m[m\r\n[32m+[m[32m training and pred_full_frame[m[m\r\n[32m+[m[32m ), ""Cannot evaluate full frame prediction during training.""[m[m\r\n # --- Mask videos ---[m[m\r\n video_tokens_BTN = batch[""video_tokens""][m[m\r\n latent_actions_BTm11L = batch[""latent_actions""][m[m\r\n[36m@@ -170,9 +175,14 @@[m [mclass DynamicsCausal(nnx.Module):[m[m\r\n )[m[m\r\n [m[m\r\n def __call__([m[m\r\n[31m- self, batch: Dict[str, jax.Array], training: bool = True, pred_full_frame: bool = False,[m[m\r\n[32m+[m[32m self,[m[m\r\n[32m+[m[32m batch: Dict[str, jax.Array],[m[m\r\n[32m+[m[32m training: bool = True,[m[m\r\n[32m+[m[32m pred_full_frame: bool = False,[m[m\r\n ) -> tuple[jax.Array, jax.Array | None]:[m[m\r\n:[K",,terminal_output
|
| 8 |
+
8,15345,"TERMINAL",0,0,"",,terminal_command
|
| 9 |
+
9,16216,"TERMINAL",0,0,"\r[K[31m- assert not (training and pred_full_frame), ""Cannot evaluate full frame prediction during training.""[m[m\r\n:[K",,terminal_output
|
| 10 |
+
10,17045,"TERMINAL",0,0,"\r[K[32m+[m[32m assert not ([m[m\r\n:[K\r[K[32m+[m[32m training and pred_full_frame[m[m\r\n:[K\r[K[32m+[m[32m ), ""Cannot evaluate full frame prediction during training.""[m[m\r\n:[K\r[K video_tokens_BTN = batch[""video_tokens""][m[m\r\n:[K\r[K latent_actions_BTm11L = batch[""latent_actions""][m[m\r\n:[K\r[K if pred_full_frame:[m[m\r\n:[K\r[K[36m@@ -184,16 +194,31 @@[m [mclass DynamicsCausal(nnx.Module):[m[m\r\n:[K\r[K def _pred_full_frame(carry, step_n):[m[m\r\n:[K\r[K video_tokens_BTN, final_logits_BTNV = carry[m[m\r\n:[K\r[K # We need to reconstruct submodules inside scan body to prevent trace context mismatches[m[m\r\n:[K\r[K[31m- patch_embed = nnx.Embed(self.num_latents, self.model_dim, rngs=nnx.Rngs(0))[m[m\r\n:[K\r[K[32m+[m[32m patch_embed = nnx.Embed([m[m\r\n:[K\r[K[32m+[m[32m self.num_latents, self.model_dim, rngs=nnx.Rngs(0)[m[m\r\n:[K",,terminal_output
|
| 11 |
+
11,17400,"TERMINAL",0,0,"\r[K[32m+[m[32m )[m[m\r\n:[K\r[K nnx.update(patch_embed, patch_embed_state)[m[m\r\n:[K\r[K action_up = nnx.Linear([m[m\r\n:[K\r[K[31m- self.latent_action_dim, self.model_dim, param_dtype=self.param_dtype, dtype=self.dtype, rngs=nnx.Rn[m :[K\r[K[31mgs(0)[m[m\r\n:[K\r[K[32m+[m[32m self.latent_action_dim,[m[m\r\n:[K\r[K[32m+[m[32m self.model_dim,[m[m\r\n:[K\r[K[32m+[m[32m param_dtype=self.param_dtype,[m[m\r\n:[K\r[K[32m+[m[32m dtype=self.dtype,[m[m\r\n:[K\r[K[32m+[m[32m rngs=nnx.Rngs(0),[m[m\r\n:[K\r[K )[m[m\r\n:[K\r[K nnx.update(action_up, action_up_state)[m[m\r\n:[K",,terminal_output
|
| 12 |
+
12,17643,"TERMINAL",0,0,"\r[K transformer = Transformer([m[m\r\n:[K\r[K[31m- self.model_dim, self.model_dim, self.ffn_dim, self.num_latents, self.num_blocks, self.num_heads,[m[m\r\n:[K\r[K[31m- self.dropout, self.param_dtype, self.dtype, use_flash_attention=self.use_flash_attention,[m[m\r\n:[K\r[K[31m- decode=self.decode, rngs=nnx.Rngs(0)[m[m\r\n:[K\r[K[32m+[m[32m self.model_dim,[m[m\r\n:[K\r[K[32m+[m[32m self.model_dim,[m[m\r\n:[K\r[K[32m+[m[32m self.ffn_dim,[m[m\r\n:[K\r[K[32m+[m[32m self.num_latents,[m[m\r\n:[K",,terminal_output
|
| 13 |
+
13,17762,"TERMINAL",0,0,"\r[K[32m+[m[32m self.num_blocks,[m[m\r\n:[K\r[K[32m+[m[32m self.num_heads,[m[m\r\n:[K",,terminal_output
|
| 14 |
+
14,17883,"TERMINAL",0,0,"\r[K[32m+[m[32m self.dropout,[m[m\r\n:[K\r[K[32m+[m[32m self.param_dtype,[m[m\r\n:[K",,terminal_output
|
| 15 |
+
15,18897,"TERMINAL",0,0,"\r[K[32m+[m[32m self.dtype,[m[m\r\n:[K",,terminal_output
|
| 16 |
+
16,19279,"TERMINAL",0,0,"\r[K[32m+[m[32m use_flash_attention=self.use_flash_attention,[m[m\r\n:[K",,terminal_output
|
| 17 |
+
17,19756,"TERMINAL",0,0,"\r[K[32m+[m[32m decode=self.decode,[m[m\r\n:[K\r[K[32m+[m[32m rngs=nnx.Rngs(0),[m[m\r\n:[K",,terminal_output
|
| 18 |
+
18,19862,"TERMINAL",0,0,"\r[K )[m[m\r\n:[K\r[K nnx.update(transformer, transformer_state)[m[m\r\n:[K\r[K [m[m\r\n:[K",,terminal_output
|
| 19 |
+
19,19966,"TERMINAL",0,0,"\r[K[36m@@ -207,7 +232,9 @@[m [mclass DynamicsCausal(nnx.Module):[m[m\r\n:[K\r[K )[m[m\r\n:[K\r[K step_logits_BTNp1V = transformer(vid_embed_BTNp1M)[m[m\r\n:[K",,terminal_output
|
| 20 |
+
20,20114,"TERMINAL",0,0,"\r[K step_logits_BV = step_logits_BTNp1V[:, -1, step_n, :][m[m\r\n:[K\r[K[31m- final_logits_BTNV = final_logits_BTNV.at[:, -1, step_n].set(step_logits_BV)[m[m\r\n:[K\r[K[32m+[m[32m final_logits_BTNV = final_logits_BTNV.at[:, -1, step_n].set([m[m\r\n:[K\r[K[32m+[m[32m step_logits_BV[m[m\r\n:[K\r[K[32m+[m[32m )[m[m\r\n:[K\r[K sampled_token_idxs_B = jnp.argmax(step_logits_BV, axis=-1)[m[m\r\n:[K",,terminal_output
|
| 21 |
+
21,20303,"TERMINAL",0,0,"\r[K video_tokens_BTN = video_tokens_BTN.at[:, -1, step_n].set([m[m\r\n:[K\r[K sampled_token_idxs_B[m[m\r\n:[K\r[K[36m@@ -216,10 +243,11 @@[m [mclass DynamicsCausal(nnx.Module):[m[m\r\n:[K\r[K [m[m\r\n:[K\r[K (_, final_logits_BTNV), _ = jax.lax.scan([m[m\r\n:[K",,terminal_output
|
| 22 |
+
22,20421,"TERMINAL",0,0,"\r[K _pred_full_frame,[m[m\r\n:[K\r[K[31m- (video_tokens_BTN, jnp.zeros(([m[m\r\n:[K",,terminal_output
|
| 23 |
+
23,20492,"TERMINAL",0,0,"\r[K[31m- **video_tokens_BTN.shape,[m[m\r\n:[K\r[K[31m- self.num_latents))),[m[m\r\n:[K\r[K[31m- jnp.arange(video_tokens_BTN.shape[2])[m[m\r\n:[K\r[K[32m+[m[32m ([m[m\r\n:[K",,terminal_output
|
| 24 |
+
24,20800,"TERMINAL",0,0,"\r[K[32m+[m[32m video_tokens_BTN,[m[m\r\n:[K\r[K[32m+[m[32m jnp.zeros((*video_tokens_BTN.shape, self.num_latents)),[m[m\r\n:[K\r[K[32m+[m[32m ),[m[m\r\n:[K\r[K[32m+[m[32m jnp.arange(video_tokens_BTN.shape[2]),[m[m\r\n:[K\r[K )[m[m\r\n:[K\r[K mask_out = jnp.zeros_like(video_tokens_BTN)[m[m\r\n:[K\r[K mask_out = mask_out.at[:, -1].set(True)[m[m\r\n:[K\r[K\r[K[7m(END)[27m[K\r[K\r[K[7m(END)[27m[K\r[K\r[K[7m(END)[27m[K\r[K\r[K[7m(END)[27m[K\r[K\r[K[7m(END)[27m[K",,terminal_output
|
| 25 |
+
25,21746,"TERMINAL",0,0,"\r[K[?1l>]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output
|
| 26 |
+
26,22679,"TERMINAL",0,0,"",,terminal_command
|
| 27 |
+
27,31747,"TERMINAL",0,0,"git commit -am ""run pre-commit""",,terminal_command
|
| 28 |
+
28,31829,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 29 |
+
29,34352,"TERMINAL",0,0,"[INFO][m Installing environment for https://github.com/psf/black.\r\n[INFO][m Once installed this environment will be reused.\r\n[INFO][m This may take a few minutes...\r\n",,terminal_output
|
| 30 |
+
30,50819,"TERMINAL",0,0,"black....................................................................",,terminal_output
|
| 31 |
+
31,50994,"TERMINAL",0,0,"[42mPassed[m\r\n",,terminal_output
|
| 32 |
+
32,51145,"TERMINAL",0,0,"[val-loss 263a0c0] run pre-commit\r\n 1 file changed, 42 insertions(+), 14 deletions(-)\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output
|
| 33 |
+
33,60713,"TERMINAL",0,0,"git push",,terminal_command
|
| 34 |
+
34,60808,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 35 |
+
35,62167,"TERMINAL",0,0,"Enumerating objects: 7, done.\r\nCounting objects: 14% (1/7)\rCounting objects: 28% (2/7)\rCounting objects: 42% (3/7)\rCounting objects: 57% (4/7)\rCounting objects: 71% (5/7)\rCounting objects: 85% (6/7)\rCounting objects: 100% (7/7)\rCounting objects: 100% (7/7), done.\r\nDelta compression using up to 152 threads\r\nCompressing objects: 25% (1/4)\rCompressing objects: 50% (2/4)\rCompressing objects: 75% (3/4)\rCompressing objects: 100% (4/4)\rCompressing objects: 100% (4/4), done.\r\nWriting objects: 25% (1/4)\rWriting objects: 50% (2/4)\rWriting objects: 75% (3/4)\rWriting objects: 100% (4/4)\rWriting objects: 100% (4/4), 714 bytes | 714.00 KiB/s, done.\r\nTotal 4 (delta 2), reused 0 (delta 0), pack-reused 0\r\nremote: Resolving deltas: 0% (0/2)[K\rremote: Resolving deltas: 50% (1/2)[K\rremote: Resolving deltas: 100% (2/2)[K\rremote: Resolving deltas: 100% (2/2), completed with 2 local objects.[K\r\n",,terminal_output
|
| 36 |
+
36,62453,"TERMINAL",0,0,"To github.com:p-doom/jasmine.git\r\n a9f9ec1..263a0c0 val-loss -> val-loss\r\n",,terminal_output
|
| 37 |
+
37,62482,"TERMINAL",0,0,"]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-9eb2e164-5989-4db7-8f31-6e8db1a38df41757236520211-2025_09_07-11.16.00.62/source.csv
ADDED
|
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,2178,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"11:16:00 AM [info] Activating crowd-code\n11:16:00 AM [info] Recording started\n11:16:00 AM [info] Initializing git provider using file system watchers...\n11:16:00 AM [info] Git repository found\n11:16:00 AM [info] Git provider initialized successfully\n11:16:01 AM [info] Initial git state: [object Object]\n",Log,tab
|
| 3 |
+
3,108194,"TERMINAL",0,0,"bash",,terminal_focus
|
| 4 |
+
4,215106,"TERMINAL",0,0,"bash",,terminal_focus
|
| 5 |
+
5,218728,"TERMINAL",0,0,"bash",,terminal_focus
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-b150d533-89a6-42d8-b7b7-d5a004d568971759420118221-2025_10_02-17.49.22.758/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c52c974d-6c8a-40d7-8b6d-40ee3b3624c21759325522612-2025_10_01-15.32.35.344/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-e9c5a28e-55ca-497d-97b5-e0c37d2af2781751878142668-2025_07_07-10.49.22.270/source.csv
ADDED
|
@@ -0,0 +1,213 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,764,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"10:49:22 AM [info] Activating crowd-code\n10:49:22 AM [info] Recording started\n10:49:22 AM [info] Initializing git provider using file system watchers...\n10:49:22 AM [info] Git repository found\n10:49:22 AM [info] Git provider initialized successfully\n10:49:22 AM [info] Initial git state: [object Object]\n",Log,tab
|
| 3 |
+
3,4308,"TERMINAL",0,0,"/bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt",,terminal_command
|
| 4 |
+
4,4359,"TERMINAL",0,0,"]633;E;2025-07-07 10:49:26 /bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt;bf94d117-cd0c-449b-b408-59a02a05b60e]633;C]0;tum_cte0515@hkn1991:/hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash]633;D;0",,terminal_output
|
| 5 |
+
5,7832,"TERMINAL",0,0,"git branch",,terminal_command
|
| 6 |
+
6,7884,"TERMINAL",0,0,"]633;E;2025-07-07 10:49:30 git branch;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C[?1h=\r add-wandb-name-and-tags[m[m\r\n convert-to-jax-array-in-iter[m[m\r\n dont-let-tf-see-gpu[m[m\r\n feat/explicit-image-dims[m[m\r\n fix-sampling[m[m\r\n main[m[m\r\n preprocess_video[m[m\r\n revised-dataloader[m[m\r\n* [32mrunner[m[m\r\n tmp[m[m\r\n",,terminal_output
|
| 7 |
+
7,7960,"TERMINAL",0,0,"\r[K[?1l>]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output
|
| 8 |
+
8,77190,"TERMINAL",0,0,"git status",,terminal_command
|
| 9 |
+
9,77237,"TERMINAL",0,0,"]633;E;2025-07-07 10:50:39 git status;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C",,terminal_output
|
| 10 |
+
10,77308,"TERMINAL",0,0,"On branch runner\r\nAll conflicts fixed but you are still merging.\r\n (use ""git commit"" to conclude merge)\r\n\r\nChanges to be committed:\r\n\t[32mmodified: train_dynamics.py[m\r\n\t[32mmodified: train_lam.py[m\r\n\t[32mmodified: train_tokenizer.py[m\r\n\r\nChanges not staged for commit:\r\n (use ""git add <file>..."" to update what will be committed)\r\n (use ""git restore <file>..."" to discard changes in working directory)\r\n\t[31mmodified: train_lam.py[m\r\n\t[31mmodified: train_tokenizer.py[m\r\n\r\nUntracked files:\r\n (use ""git add <file>..."" to include in what will be committed)\r\n\t[31mdata_tfrecord_duplicated/[m\r\n\t[31mdata_tfrecords/[m\r\n\t[31mlogs/[m\r\n\t[31mread_tf_record.py[m\r\n\t[31mrequirements-franz.txt[m\r\n\t[31mscripts_cremers/[m\r\n\t[31mscripts_horeka/[m\r\n\t[31mslurm-3309772.out[m\r\n\t[31mslurm/[m\r\n\t[31mutils/visualizer.py[m\r\n\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output
|
| 11 |
+
11,85704,"TERMINAL",0,0,"git diff train_lam.py",,terminal_command
|
| 12 |
+
12,85768,"TERMINAL",0,0,"]633;E;2025-07-07 10:50:47 git diff train_lam.py;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C[?1h=\r[1mdiff --git a/train_lam.py b/train_lam.py[m[m\r\n[1mindex 858990e..540a464 100644[m[m\r\n[1m--- a/train_lam.py[m[m\r\n[1m+++ b/train_lam.py[m[m\r\n[36m@@ -59,7 +59,6 @@[m [mclass Args:[m[m\r\n log_interval: int = 5[m[m\r\n log_image_interval: int = 250[m[m\r\n ckpt_dir: str = """"[m[m\r\n[31m- tmp_ckpt_dir: str = ""/tmp/checkpoints/""[m[m\r\n log_checkpoint_interval: int = 10000[m[m\r\n [m[m\r\n [m[m\r\n\r[K[?1l>]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output
|
| 13 |
+
13,97852,"TERMINAL",0,0,"git commit -am ""removed tmp""",,terminal_command
|
| 14 |
+
14,97890,"TERMINAL",0,0,"]633;E;2025-07-07 10:51:00 git commit -am ""removed tmp"";f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C",,terminal_output
|
| 15 |
+
15,98308,"TERMINAL",0,0,"[runner 316eae6] removed tmp\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output
|
| 16 |
+
16,109638,"TERMINAL",0,0,"git checkout revised-dataloader",,terminal_command
|
| 17 |
+
17,109689,"TERMINAL",0,0,"]633;E;2025-07-07 10:51:11 git checkout revised-dataloader;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C",,terminal_output
|
| 18 |
+
18,110012,"TERMINAL",0,0,"Switched to branch 'revised-dataloader'\r\nYour branch is up to date with 'origin/revised-dataloader'.\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output
|
| 19 |
+
19,110376,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"Switched from branch 'runner' to 'revised-dataloader'",Log,git_branch_checkout
|
| 20 |
+
20,110484,"extension-output-pdoom-org.crowd-code-#1-crowd-code",304,0,"10:51:12 AM [info] Branch checkout detected: runner -> revised-dataloader\n10:51:12 AM [info] Recording git checkout: Switched from branch 'runner' to 'revised-dataloader'\n10:51:12 AM [info] Resetting file cache due to branch checkout\n",Log,content
|
| 21 |
+
21,112104,"TERMINAL",0,0,"git status",,terminal_command
|
| 22 |
+
22,112207,"TERMINAL",0,0,"]633;E;2025-07-07 10:51:14 git status;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;COn branch revised-dataloader\r\nYour branch is up to date with 'origin/revised-dataloader'.\r\n\r\nUntracked files:\r\n (use ""git add <file>..."" to include in what will be committed)\r\n\t[31mdata_tfrecord_duplicated/[m\r\n\t[31mdata_tfrecords/[m\r\n\t[31mlogs/[m\r\n\t[31mread_tf_record.py[m\r\n\t[31mrequirements-franz.txt[m\r\n\t[31mscripts_cremers/[m\r\n\t[31mscripts_horeka/[m\r\n\t[31mslurm-3309772.out[m\r\n\t[31mslurm/[m\r\n\t[31mutils/visualizer.py[m\r\n\r\nnothing added to commit but untracked files present (use ""git add"" to track)\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output
|
| 23 |
+
23,128400,"TERMINAL",0,0,"git pull",,terminal_command
|
| 24 |
+
24,128448,"TERMINAL",0,0,"]633;E;2025-07-07 10:51:30 git pull;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C",,terminal_output
|
| 25 |
+
25,130043,"TERMINAL",0,0,"remote: Enumerating objects: 60, done.[K\r\nremote: Counting objects: 1% (1/54)[K\rremote: Counting objects: 3% (2/54)[K\rremote: Counting objects: 5% (3/54)[K\rremote: Counting objects: 7% (4/54)[K\rremote: Counting objects: 9% (5/54)[K\rremote: Counting objects: 11% (6/54)[K\rremote: Counting objects: 12% (7/54)[K\rremote: Counting objects: 14% (8/54)[K\rremote: Counting objects: 16% (9/54)[K\rremote: Counting objects: 18% (10/54)[K\rremote: Counting objects: 20% (11/54)[K\rremote: Counting objects: 22% (12/54)[K\rremote: Counting objects: 24% (13/54)[K\rremote: Counting objects: 25% (14/54)[K\rremote: Counting objects: 27% (15/54)[K\rremote: Counting objects: 29% (16/54)[K\rremote: Counting objects: 31% (17/54)[K\rremote: Counting objects: 33% (18/54)[K\rremote: Counting objects: 35% (19/54)[K\rremote: Counting objects: 37% (20/54)[K\rremote: Counting objects: 38% (21/54)[K\rremote: Counting objects: 40% (22/54)[K\rremote: Counting objects: 42% (23/54)[K\rremote: Counting objects: 44% (24/54)[K\rremote: Counting objects: 46% (25/54)[K\rremote: Counting objects: 48% (26/54)[K\rremote: Counting objects: 50% (27/54)[K\rremote: Counting objects: 51% (28/54)[K\rremote: Counting objects: 53% (29/54)[K\rremote: Counting objects: 55% (30/54)[K\rremote: Counting objects: 57% (31/54)[K\rremote: Counting objects: 59% (32/54)[K\rremote: Counting objects: 61% (33/54)[K\rremote: Counting objects: 62% (34/54)[K\rremote: Counting objects: 64% (35/54)[K\rremote: Counting objects: 66% (36/54)[K\rremote: Counting objects: 68% (37/54)[K\rremote: Counting objects: 70% (38/54)[K\rremote: Counting objects: 72% (39/54)[K\rremote: Counting objects: 74% (40/54)[K\rremote: Counting objects: 75% (41/54)[K\rremote: Counting objects: 77% (42/54)[K\rremote: Counting objects: 79% (43/54)[K\rremote: Counting objects: 81% (44/54)[K\rremote: Counting objects: 83% (45/54)[K\rremote: Counting objects: 85% (46/54)[K\rremote: Counting objects: 87% (47/54)[K\rremote: Counting objects: 88% (48/54)[K\rremote: Counting objects: 90% (49/54)[K\rremote: Counting objects: 92% (50/54)[K\rremote: Counting objects: 94% (51/54)[K\rremote: Counting objects: 96% (52/54)[K\rremote: Counting objects: 98% (53/54)[K\rremote: Counting objects: 100% (54/54)[K\rremote: Counting objects: 100% (54/54), done.[K\r\nremote: Compressing objects: 5% (1/20)[K\rremote: Compressing objects: 10% (2/20)[K\rremote: Compressing objects: 15% (3/20)[K\rremote: Compressing objects: 20% (4/20)[K\rremote: Compressing objects: 25% (5/20)[K\rremote: Compressing objects: 30% (6/20)[K\rremote: Compressing objects: 35% (7/20)[K\rremote: Compressing objects: 40% (8/20)[K\rremote: Compressing objects: 45% (9/20)[K\rremote: Compressing objects: 50% (10/20)[K\rremote: Compressing objects: 55% (11/20)[K\rremote: Compressing objects: 60% (12/20)[K\rremote: Compressing objects: 65% (13/20)[K\rremote: Compressing objects: 70% (14/20)[K\rremote: Compressing objects: 75% (15/20)[K\rremote: Compressing objects: 80% (16/20)[K\rremote: Compressing objects: 85% (17/20)[K\rremote: Compressing objects: 90% (18/20)[K\rremote: Compressing objects: 95% (19/20)[K\rremote: Compressing objects: 100% (20/20)[K\rremote: Compressing objects: 100% (20/20), done.[K\r\nremote: Total 36 (delta 26), reused 22 (delta 16), pack-reused 0 (from 0)[K\r\n",,terminal_output
|
| 26 |
+
26,130161,"TERMINAL",0,0,"Unpacking objects: 2% (1/36)\rUnpacking objects: 5% (2/36)\rUnpacking objects: 8% (3/36)\rUnpacking objects: 11% (4/36)\rUnpacking objects: 13% (5/36)\rUnpacking objects: 16% (6/36)\rUnpacking objects: 19% (7/36)\r",,terminal_output
|
| 27 |
+
27,130215,"TERMINAL",0,0,"Unpacking objects: 22% (8/36)\rUnpacking objects: 25% (9/36)\r",,terminal_output
|
| 28 |
+
28,130280,"TERMINAL",0,0,"Unpacking objects: 27% (10/36)\rUnpacking objects: 30% (11/36)\r",,terminal_output
|
| 29 |
+
29,130404,"TERMINAL",0,0,"Unpacking objects: 33% (12/36)\rUnpacking objects: 36% (13/36)\rUnpacking objects: 38% (14/36)\rUnpacking objects: 41% (15/36)\rUnpacking objects: 44% (16/36)\rUnpacking objects: 47% (17/36)\rUnpacking objects: 50% (18/36)\rUnpacking objects: 52% (19/36)\rUnpacking objects: 55% (20/36)\rUnpacking objects: 58% (21/36)\rUnpacking objects: 61% (22/36)\rUnpacking objects: 63% (23/36)\rUnpacking objects: 66% (24/36)\r",,terminal_output
|
| 30 |
+
30,130473,"TERMINAL",0,0,"Unpacking objects: 69% (25/36)\rUnpacking objects: 72% (26/36)\rUnpacking objects: 75% (27/36)\rUnpacking objects: 77% (28/36)\rUnpacking objects: 80% (29/36)\rUnpacking objects: 83% (30/36)\r",,terminal_output
|
| 31 |
+
31,130587,"TERMINAL",0,0,"Unpacking objects: 86% (31/36)\rUnpacking objects: 88% (32/36)\rUnpacking objects: 91% (33/36)\rUnpacking objects: 94% (34/36)\rUnpacking objects: 97% (35/36)\rUnpacking objects: 100% (36/36)\rUnpacking objects: 100% (36/36), 6.91 KiB | 14.00 KiB/s, done.\r\n",,terminal_output
|
| 32 |
+
32,130774,"TERMINAL",0,0,"From github.com:p-doom/jafar\r\n * [new branch] correct-batched-sampling -> origin/correct-batched-sampling\r\n 4ec9ebe..9edd0c1 dynamics-lam-co-training -> origin/dynamics-lam-co-training\r\n 32b3f04..3176718 feat/restore_train_state -> origin/feat/restore_train_state\r\n ae9451f..6e623c6 fix-sampling -> origin/fix-sampling\r\n c8dd7ea..9fb362e main -> origin/main\r\n * [new branch] make-warmup-default -> origin/make-warmup-default\r\n",,terminal_output
|
| 33 |
+
33,130837,"TERMINAL",0,0,"Already up to date.\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output
|
| 34 |
+
34,133144,"TERMINAL",0,0,"git status",,terminal_command
|
| 35 |
+
35,133186,"TERMINAL",0,0,"]633;E;2025-07-07 10:51:35 git status;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;COn branch revised-dataloader\r\nYour branch is up to date with 'origin/revised-dataloader'.\r\n\r\nUntracked files:\r\n (use ""git add <file>..."" to include in what will be committed)\r\n\t[31mdata_tfrecord_duplicated/[m\r\n\t[31mdata_tfrecords/[m\r\n\t[31mlogs/[m\r\n\t[31mread_tf_record.py[m\r\n\t[31mrequirements-franz.txt[m\r\n\t[31mscripts_cremers/[m\r\n\t[31mscripts_horeka/[m\r\n\t[31mslurm-3309772.out[m\r\n\t[31mslurm/[m\r\n\t[31mutils/visualizer.py[m\r\n\r\nnothing added to commit but untracked files present (use ""git add"" to track)\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output
|
| 36 |
+
36,148140,"utils/dataloader.py",0,0,"import functools\nimport jax\n\nimport tensorflow as tf\n\n# reserve GPU memory for JAX only if tensorflow is built with GPU support\ntf.config.experimental.set_visible_devices([], ""GPU"")\n\n\n# --- TensorFlow function for processing: slicing, normalization ---\ndef _tf_process_episode(episode_tensor, seq_len, image_h, image_w, image_c):\n """"""\n Processes a raw episode tensor in TensorFlow.\n Takes a full episode, extracts a random sequence, and normalizes it.\n Args:\n episode_tensor: A TensorFlow tensor representing a full video episode.\n Expected shape: (dynamic_length, image_h, image_w, image_c)\n Expected dtype: e.g., tf.uint8 (raw pixel values)\n seq_len: The desired length of the sub-sequence to extract.\n image_h: The height of each frame.\n image_w: The width of each frame.\n image_c: The number of channels in each frame.\n Returns:\n A TensorFlow tensor representing the processed video sequence.\n Shape: (seq_len, image_h, image_w, image_c)\n Dtype: tf.float32 (normalized pixel values)\n """"""\n current_episode_len = tf.shape(episode_tensor)[0]\n\n max_start_idx = current_episode_len - seq_len\n\n start_idx = tf.random.uniform(\n shape=(), minval=0, maxval=max_start_idx + 1, dtype=tf.int32\n )\n\n seq = episode_tensor[start_idx : start_idx + seq_len]\n\n seq = tf.cast(seq, tf.float32) / 255.0\n\n # Ensure the final shape is statically known for batching.\n # tf.reshape is robust, but tf.ensure_shape or set_shape can also be used if confident.\n processed_sequence = tf.reshape(seq, [seq_len, image_h, image_w, image_c])\n\n return processed_sequence\n\n\ndef _parse_tfrecord_fn(example_proto, image_h, image_w, image_c):\n feature_description = {\n ""height"": tf.io.FixedLenFeature([], tf.int64),\n ""width"": tf.io.FixedLenFeature([], tf.int64),\n ""channels"": tf.io.FixedLenFeature([], tf.int64),\n ""sequence_length"": tf.io.FixedLenFeature([], tf.int64),\n ""raw_video"": tf.io.FixedLenFeature([], tf.string),\n }\n example = tf.io.parse_single_example(example_proto, feature_description)\n\n video_shape = (example[""sequence_length""], image_h, image_w, image_c)\n\n episode_tensor = tf.io.decode_raw(example[""raw_video""], out_type=tf.uint8)\n episode_tensor = tf.reshape(episode_tensor, video_shape)\n\n episode_tensor = tf.ensure_shape(episode_tensor, [None, image_h, image_w, image_c])\n return episode_tensor\n\n\ndef _create_processed_dataset_from_file(file_path, image_h, image_w, image_c, seq_len, num_parallel_calls):\n """"""Creates a fully processed dataset from a single TFRecord file.""""""\n dataset = tf.data.TFRecordDataset([file_path])\n \n parse_fn = functools.partial(\n _parse_tfrecord_fn, image_h=image_h, image_w=image_w, image_c=image_c\n )\n dataset = dataset.map(parse_fn, num_parallel_calls=num_parallel_calls)\n\n tf_process_fn = functools.partial(\n _tf_process_episode,\n seq_len=seq_len,\n image_h=image_h,\n image_w=image_w,\n image_c=image_c,\n )\n dataset = dataset.map(tf_process_fn, num_parallel_calls=num_parallel_calls)\n \n return dataset\n\n\ndef get_dataloader(\n tfrecord_paths: list[str],\n seq_len: int,\n global_batch_size: int,\n image_h: int,\n image_w: int,\n image_c: int,\n shuffle_buffer_size: int = 1000,\n num_parallel_calls: int = tf.data.AUTOTUNE,\n seed: int = 42,\n cycle_length: int = 4,\n block_length: int = 1,\n):\n """"""\n Creates a tf.data.Dataset pipeline from TFRecord files.\n """"""\n if not tfrecord_paths:\n raise ValueError(""tfrecord_paths list cannot be empty."")\n\n process_id = jax.process_index()\n num_processes = jax.process_count()\n\n assert (\n global_batch_size % num_processes == 0\n ), f""Global batch size {global_batch_size} \\n must be divisible by the number of JAX processes {num_processes} for proper sharding.""\n per_process_batch_size = global_batch_size // num_processes\n\n def dataset_fn(file_path):\n return _create_processed_dataset_from_file(\n file_path, image_h, image_w, image_c, seq_len, num_parallel_calls\n )\n \n dataset = tf.data.Dataset.from_tensor_slices(tfrecord_paths)\n dataset = dataset.shard(num_shards=num_processes, index=process_id)\n \n dataset = dataset.interleave(\n dataset_fn,\n cycle_length=cycle_length,\n block_length=block_length,\n num_parallel_calls=num_parallel_calls,\n deterministic=False\n )\n \n if shuffle_buffer_size > 0:\n dataset = dataset.shuffle(\n buffer_size=shuffle_buffer_size, seed=seed, reshuffle_each_iteration=True\n )\n\n dataset = dataset.repeat(None)\n dataset = dataset.batch(per_process_batch_size, drop_remainder=True)\n dataset = dataset.prefetch(tf.data.AUTOTUNE)\n\n return dataset.as_numpy_iterator()\n",python,tab
|
| 37 |
+
37,153941,"utils/dataloader.py",3046,0,"",python,selection_mouse
|
| 38 |
+
38,153957,"utils/dataloader.py",3045,0,"",python,selection_command
|
| 39 |
+
39,167620,"TERMINAL",0,0,"git checkout runner",,terminal_command
|
| 40 |
+
40,167670,"TERMINAL",0,0,"]633;E;2025-07-07 10:52:09 git checkout runner;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C",,terminal_output
|
| 41 |
+
41,167734,"TERMINAL",0,0,"Switched to branch 'runner'\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output
|
| 42 |
+
42,169889,"utils/dataloader.py",0,0,"",python,tab
|
| 43 |
+
43,170107,"utils/dataloader.py",253,3946,"def _tf_process_episode(episode_tensor, seq_len, image_h, image_w, image_c, seed):\n """"""\n Processes a raw episode tensor in TensorFlow.\n Takes a full episode, extracts a random sequence, and normalizes it.\n Args:\n episode_tensor: A TensorFlow tensor representing a full video episode.\n Expected shape: (dynamic_length, image_h, image_w, image_c)\n Expected dtype: e.g., tf.uint8 (raw pixel values)\n seq_len: The desired length of the sub-sequence to extract.\n image_h: The height of each frame.\n image_w: The width of each frame.\n image_c: The number of channels in each frame.\n seed: The seed for the random number generator.\n Returns:\n A TensorFlow tensor representing the processed video sequence.\n Shape: (seq_len, image_h, image_w, image_c)\n Dtype: tf.float32 (normalized pixel values)\n """"""\n current_episode_len = tf.shape(episode_tensor)[0]\n\n max_start_idx = current_episode_len - seq_len\n\n start_idx = tf.random.uniform(\n shape=(), minval=0, maxval=max_start_idx + 1, dtype=tf.int32, seed=seed\n )\n\n seq = episode_tensor[start_idx : start_idx + seq_len]\n\n seq = tf.cast(seq, tf.float32) / 255.0\n\n # Ensure the final shape is statically known for batching.\n # tf.reshape is robust, but tf.ensure_shape or set_shape can also be used if confident.\n processed_sequence = tf.reshape(seq, [seq_len, image_h, image_w, image_c])\n\n return processed_sequence\n\n\ndef _parse_tfrecord_fn(example_proto, image_h, image_w, image_c):\n feature_description = {\n ""height"": tf.io.FixedLenFeature([], tf.int64),\n ""width"": tf.io.FixedLenFeature([], tf.int64),\n ""channels"": tf.io.FixedLenFeature([], tf.int64),\n ""sequence_length"": tf.io.FixedLenFeature([], tf.int64),\n ""raw_video"": tf.io.FixedLenFeature([], tf.string),\n }\n example = tf.io.parse_single_example(example_proto, feature_description)\n\n video_shape = (example[""sequence_length""], image_h, image_w, image_c)\n\n episode_tensor = tf.io.decode_raw(example[""raw_video""], out_type=tf.uint8)\n episode_tensor = tf.reshape(episode_tensor, video_shape)\n\n episode_tensor = tf.ensure_shape(episode_tensor, [None, image_h, image_w, image_c])\n return episode_tensor\n\n\ndef _create_processed_dataset_from_file(file_path, image_h, image_w, image_c, seq_len, num_parallel_calls, seed):\n """"""Creates a fully processed dataset from a single TFRecord file.""""""\n dataset = tf.data.TFRecordDataset([file_path])\n \n parse_fn = functools.partial(\n _parse_tfrecord_fn, image_h=image_h, image_w=image_w, image_c=image_c\n )\n dataset = dataset.map(parse_fn, num_parallel_calls=num_parallel_calls)\n\n # Filter out episodes that are too short\n def filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)\n\n tf_process_fn = functools.partial(\n _tf_process_episode,\n seq_len=seq_len,\n image_h=image_h,\n image_w=image_w,\n image_c=image_c,\n seed=seed,\n )\n dataset = dataset.map(tf_process_fn, num_parallel_calls=num_parallel_calls)\n \n return dataset\n\n\ndef get_dataloader(\n tfrecord_paths: list[str],\n seq_len: int,\n global_batch_size: int,\n image_h: int,\n image_w: int,\n image_c: int,\n shuffle_buffer_size: int = 10,\n num_parallel_calls: int = tf.data.AUTOTUNE,\n seed: int = 42,\n cycle_length: int = 4,\n block_length: int = 1,\n):\n """"""\n Creates a tf.data.Dataset pipeline from TFRecord files.\n """"""\n if not tfrecord_paths:\n raise ValueError(""tfrecord_paths list cannot be empty."")\n\n process_id = jax.process_index()\n num_processes = jax.process_count()\n\n assert (\n global_batch_size % num_processes == 0\n ), f""Global batch size {global_batch_size} \\n must be divisible by the number of JAX processes {num_processes} for proper sharding.""\n per_process_batch_size = global_batch_size // num_processes\n\n def dataset_fn(file_path):\n return _create_processed_dataset_from_file(\n file_path, image_h, image_w, image_c, seq_len, num_parallel_calls, seed\n",python,content
|
| 44 |
+
44,170389,"utils/dataloader.py",0,0,"Switched from branch 'revised-dataloader' to 'runner'",python,git_branch_checkout
|
| 45 |
+
45,172501,"utils/dataloader.py",3153,0,"",python,selection_mouse
|
| 46 |
+
46,172512,"utils/dataloader.py",3152,0,"",python,selection_command
|
| 47 |
+
47,172733,"utils/dataloader.py",3152,1,"n",python,selection_mouse
|
| 48 |
+
48,172734,"utils/dataloader.py",3152,6,"n\n ",python,selection_mouse
|
| 49 |
+
49,172735,"utils/dataloader.py",3135,17,"sor)[0] >= seq_le",python,selection_mouse
|
| 50 |
+
50,172735,"utils/dataloader.py",3079,73,"des(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_le",python,selection_mouse
|
| 51 |
+
51,172751,"utils/dataloader.py",3153,0,"",python,selection_command
|
| 52 |
+
52,172824,"utils/dataloader.py",3027,126,"isodes that are too short\n def filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len",python,selection_mouse
|
| 53 |
+
53,172825,"utils/dataloader.py",3021,132,"out episodes that are too short\n def filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len",python,selection_mouse
|
| 54 |
+
54,172826,"utils/dataloader.py",3018,135,"er out episodes that are too short\n def filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len",python,selection_mouse
|
| 55 |
+
55,172842,"utils/dataloader.py",3014,139,"Filter out episodes that are too short\n def filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len",python,selection_mouse
|
| 56 |
+
56,172874,"utils/dataloader.py",3007,146,"\n # Filter out episodes that are too short\n def filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len",python,selection_mouse
|
| 57 |
+
57,173126,"utils/dataloader.py",2932,221," dataset = dataset.map(parse_fn, num_parallel_calls=num_parallel_calls)\n\n # Filter out episodes that are too short\n def filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len",python,selection_mouse
|
| 58 |
+
58,177555,"utils/dataloader.py",3210,0,"",python,selection_mouse
|
| 59 |
+
59,177618,"utils/dataloader.py",3209,0,"",python,selection_command
|
| 60 |
+
60,177705,"utils/dataloader.py",3209,1,")",python,selection_mouse
|
| 61 |
+
61,177723,"utils/dataloader.py",3210,0,"",python,selection_command
|
| 62 |
+
62,177807,"utils/dataloader.py",3207,3,"es)",python,selection_mouse
|
| 63 |
+
63,177808,"utils/dataloader.py",3158,52,"\n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 64 |
+
64,177815,"utils/dataloader.py",3131,79,"_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 65 |
+
65,177818,"utils/dataloader.py",3127,83,"sode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 66 |
+
66,177843,"utils/dataloader.py",3124,86,"episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 67 |
+
67,177843,"utils/dataloader.py",3123,87,"(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 68 |
+
68,177910,"utils/dataloader.py",3121,89,"pe(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 69 |
+
69,177912,"utils/dataloader.py",3119,91,"hape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 70 |
+
70,177920,"utils/dataloader.py",3116,94,"f.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 71 |
+
71,177943,"utils/dataloader.py",3114,96," tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 72 |
+
72,178059,"utils/dataloader.py",3112,98,"rn tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 73 |
+
73,178060,"utils/dataloader.py",3111,99,"urn tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 74 |
+
74,178060,"utils/dataloader.py",3109,101,"eturn tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 75 |
+
75,178062,"utils/dataloader.py",3108,102,"return tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 76 |
+
76,178062,"utils/dataloader.py",3060,150," filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 77 |
+
77,178083,"utils/dataloader.py",3059,151,"f filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 78 |
+
78,178145,"utils/dataloader.py",3058,152,"ef filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 79 |
+
79,178216,"utils/dataloader.py",3057,153,"def filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 80 |
+
80,178360,"utils/dataloader.py",3056,154," def filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 81 |
+
81,178361,"utils/dataloader.py",3055,155," def filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 82 |
+
82,178504,"utils/dataloader.py",3010,200," # Filter out episodes that are too short\n def filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 83 |
+
83,178505,"utils/dataloader.py",3055,155," def filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 84 |
+
84,178505,"utils/dataloader.py",3009,201," # Filter out episodes that are too short\n def filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 85 |
+
85,178827,"utils/dataloader.py",3008,202," # Filter out episodes that are too short\n def filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,selection_mouse
|
| 86 |
+
86,189453,"TERMINAL",0,0,"git checkout revised-dataloader",,terminal_command
|
| 87 |
+
87,189506,"TERMINAL",0,0,"]633;E;2025-07-07 10:52:31 git checkout revised-dataloader;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;CSwitched to branch 'revised-dataloader'\r\nYour branch is up to date with 'origin/revised-dataloader'.\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output
|
| 88 |
+
88,190380,"",0,0,"Switched from branch 'runner' to 'revised-dataloader'",,git_branch_checkout
|
| 89 |
+
89,190846,"utils/dataloader.py",0,0,"import functools\nimport jax\n\nimport tensorflow as tf\n\n# reserve GPU memory for JAX only if tensorflow is built with GPU support\ntf.config.experimental.set_visible_devices([], ""GPU"")\n\n\n# --- TensorFlow function for processing: slicing, normalization ---\ndef _tf_process_episode(episode_tensor, seq_len, image_h, image_w, image_c, seed):\n """"""\n Processes a raw episode tensor in TensorFlow.\n Takes a full episode, extracts a random sequence, and normalizes it.\n Args:\n episode_tensor: A TensorFlow tensor representing a full video episode.\n Expected shape: (dynamic_length, image_h, image_w, image_c)\n Expected dtype: e.g., tf.uint8 (raw pixel values)\n seq_len: The desired length of the sub-sequence to extract.\n image_h: The height of each frame.\n image_w: The width of each frame.\n image_c: The number of channels in each frame.\n seed: The seed for the random number generator.\n Returns:\n A TensorFlow tensor representing the processed video sequence.\n Shape: (seq_len, image_h, image_w, image_c)\n Dtype: tf.float32 (normalized pixel values)\n """"""\n current_episode_len = tf.shape(episode_tensor)[0]\n\n max_start_idx = current_episode_len - seq_len\n\n start_idx = tf.random.uniform(\n shape=(), minval=0, maxval=max_start_idx + 1, dtype=tf.int32, seed=seed\n )\n\n seq = episode_tensor[start_idx : start_idx + seq_len]\n\n seq = tf.cast(seq, tf.float32) / 255.0\n\n # Ensure the final shape is statically known for batching.\n # tf.reshape is robust, but tf.ensure_shape or set_shape can also be used if confident.\n processed_sequence = tf.reshape(seq, [seq_len, image_h, image_w, image_c])\n\n return processed_sequence\n\n\ndef _parse_tfrecord_fn(example_proto, image_h, image_w, image_c):\n feature_description = {\n ""height"": tf.io.FixedLenFeature([], tf.int64),\n ""width"": tf.io.FixedLenFeature([], tf.int64),\n ""channels"": tf.io.FixedLenFeature([], tf.int64),\n ""sequence_length"": tf.io.FixedLenFeature([], tf.int64),\n ""raw_video"": tf.io.FixedLenFeature([], tf.string),\n }\n example = tf.io.parse_single_example(example_proto, feature_description)\n\n video_shape = (example[""sequence_length""], image_h, image_w, image_c)\n\n episode_tensor = tf.io.decode_raw(example[""raw_video""], out_type=tf.uint8)\n episode_tensor = tf.reshape(episode_tensor, video_shape)\n\n episode_tensor = tf.ensure_shape(episode_tensor, [None, image_h, image_w, image_c])\n return episode_tensor\n\n\ndef _create_processed_dataset_from_file(file_path, image_h, image_w, image_c, seq_len, num_parallel_calls, seed):\n """"""Creates a fully processed dataset from a single TFRecord file.""""""\n dataset = tf.data.TFRecordDataset([file_path])\n \n parse_fn = functools.partial(\n _parse_tfrecord_fn, image_h=image_h, image_w=image_w, image_c=image_c\n )\n dataset = dataset.map(parse_fn, num_parallel_calls=num_parallel_calls)\n\n # Filter out episodes that are too short\n def filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)\n\n tf_process_fn = functools.partial(\n _tf_process_episode,\n seq_len=seq_len,\n image_h=image_h,\n image_w=image_w,\n image_c=image_c,\n seed=seed,\n )\n dataset = dataset.map(tf_process_fn, num_parallel_calls=num_parallel_calls)\n \n return dataset\n\n\ndef get_dataloader(\n tfrecord_paths: list[str],\n seq_len: int,\n global_batch_size: int,\n image_h: int,\n image_w: int,\n image_c: int,\n shuffle_buffer_size: int = 10,\n num_parallel_calls: int = tf.data.AUTOTUNE,\n seed: int = 42,\n cycle_length: int = 4,\n block_length: int = 1,\n):\n """"""\n Creates a tf.data.Dataset pipeline from TFRecord files.\n """"""\n if not tfrecord_paths:\n raise ValueError(""tfrecord_paths list cannot be empty."")\n\n process_id = jax.process_index()\n num_processes = jax.process_count()\n\n assert (\n global_batch_size % num_processes == 0\n ), f""Global batch size {global_batch_size} \\n must be divisible by the number of JAX processes {num_processes} for proper sharding.""\n per_process_batch_size = global_batch_size // num_processes\n\n def dataset_fn(file_path):\n return _create_processed_dataset_from_file(\n file_path, image_h, image_w, image_c, seq_len, num_parallel_calls, seed\n )\n \n dataset = tf.data.Dataset.from_tensor_slices(tfrecord_paths)\n dataset = dataset.shard(num_shards=num_processes, index=process_id)\n \n dataset = dataset.interleave(\n dataset_fn,\n cycle_length=cycle_length,\n block_length=block_length,\n num_parallel_calls=num_parallel_calls,\n deterministic=False\n )\n \n if shuffle_buffer_size > 0:\n dataset = dataset.shuffle(\n buffer_size=shuffle_buffer_size, seed=seed, reshuffle_each_iteration=True\n )\n\n dataset = dataset.repeat(None)\n dataset = dataset.batch(per_process_batch_size, drop_remainder=True)\n dataset = dataset.prefetch(tf.data.AUTOTUNE)\n\n return dataset.as_numpy_iterator()\n",python,tab
|
| 90 |
+
90,191033,"utils/dataloader.py",253,4252,"def _tf_process_episode(episode_tensor, seq_len, image_h, image_w, image_c):\n """"""\n Processes a raw episode tensor in TensorFlow.\n Takes a full episode, extracts a random sequence, and normalizes it.\n Args:\n episode_tensor: A TensorFlow tensor representing a full video episode.\n Expected shape: (dynamic_length, image_h, image_w, image_c)\n Expected dtype: e.g., tf.uint8 (raw pixel values)\n seq_len: The desired length of the sub-sequence to extract.\n image_h: The height of each frame.\n image_w: The width of each frame.\n image_c: The number of channels in each frame.\n Returns:\n A TensorFlow tensor representing the processed video sequence.\n Shape: (seq_len, image_h, image_w, image_c)\n Dtype: tf.float32 (normalized pixel values)\n """"""\n current_episode_len = tf.shape(episode_tensor)[0]\n\n max_start_idx = current_episode_len - seq_len\n\n start_idx = tf.random.uniform(\n shape=(), minval=0, maxval=max_start_idx + 1, dtype=tf.int32\n )\n\n seq = episode_tensor[start_idx : start_idx + seq_len]\n\n seq = tf.cast(seq, tf.float32) / 255.0\n\n # Ensure the final shape is statically known for batching.\n # tf.reshape is robust, but tf.ensure_shape or set_shape can also be used if confident.\n processed_sequence = tf.reshape(seq, [seq_len, image_h, image_w, image_c])\n\n return processed_sequence\n\n\ndef _parse_tfrecord_fn(example_proto, image_h, image_w, image_c):\n feature_description = {\n ""height"": tf.io.FixedLenFeature([], tf.int64),\n ""width"": tf.io.FixedLenFeature([], tf.int64),\n ""channels"": tf.io.FixedLenFeature([], tf.int64),\n ""sequence_length"": tf.io.FixedLenFeature([], tf.int64),\n ""raw_video"": tf.io.FixedLenFeature([], tf.string),\n }\n example = tf.io.parse_single_example(example_proto, feature_description)\n\n video_shape = (example[""sequence_length""], image_h, image_w, image_c)\n\n episode_tensor = tf.io.decode_raw(example[""raw_video""], out_type=tf.uint8)\n episode_tensor = tf.reshape(episode_tensor, video_shape)\n\n episode_tensor = tf.ensure_shape(episode_tensor, [None, image_h, image_w, image_c])\n return episode_tensor\n\n\ndef _create_processed_dataset_from_file(file_path, image_h, image_w, image_c, seq_len, num_parallel_calls):\n """"""Creates a fully processed dataset from a single TFRecord file.""""""\n dataset = tf.data.TFRecordDataset([file_path])\n \n parse_fn = functools.partial(\n _parse_tfrecord_fn, image_h=image_h, image_w=image_w, image_c=image_c\n )\n dataset = dataset.map(parse_fn, num_parallel_calls=num_parallel_calls)\n\n tf_process_fn = functools.partial(\n _tf_process_episode,\n seq_len=seq_len,\n image_h=image_h,\n image_w=image_w,\n image_c=image_c,\n )\n dataset = dataset.map(tf_process_fn, num_parallel_calls=num_parallel_calls)\n \n return dataset\n\n\ndef get_dataloader(\n tfrecord_paths: list[str],\n seq_len: int,\n global_batch_size: int,\n image_h: int,\n image_w: int,\n image_c: int,\n shuffle_buffer_size: int = 1000,\n num_parallel_calls: int = tf.data.AUTOTUNE,\n seed: int = 42,\n cycle_length: int = 4,\n block_length: int = 1,\n):\n """"""\n Creates a tf.data.Dataset pipeline from TFRecord files.\n """"""\n if not tfrecord_paths:\n raise ValueError(""tfrecord_paths list cannot be empty."")\n\n process_id = jax.process_index()\n num_processes = jax.process_count()\n\n assert (\n global_batch_size % num_processes == 0\n ), f""Global batch size {global_batch_size} \\n must be divisible by the number of JAX processes {num_processes} for proper sharding.""\n per_process_batch_size = global_batch_size // num_processes\n\n def dataset_fn(file_path):\n return _create_processed_dataset_from_file(\n file_path, image_h, image_w, image_c, seq_len, num_parallel_calls\n",python,content
|
| 91 |
+
91,192106,"utils/dataloader.py",2928,0,"",python,selection_mouse
|
| 92 |
+
92,193410,"utils/dataloader.py",2928,0,"\n",python,content
|
| 93 |
+
93,193787,"utils/dataloader.py",2929,0," # Filter out episodes that are too short\n def filter_short_episodes(episode_tensor):\n return tf.shape(episode_tensor)[0] >= seq_len\n \n dataset = dataset.filter(filter_short_episodes)",python,content
|
| 94 |
+
94,194713,"utils/dataloader.py",3131,0,"\n ",python,content
|
| 95 |
+
95,195064,"utils/dataloader.py",3132,4,"",python,content
|
| 96 |
+
96,195607,"utils/dataloader.py",3080,0,"",python,selection_command
|
| 97 |
+
97,195836,"utils/dataloader.py",3075,0,"",python,selection_command
|
| 98 |
+
98,195995,"utils/dataloader.py",3021,0,"",python,selection_command
|
| 99 |
+
99,196145,"utils/dataloader.py",2974,0,"",python,selection_command
|
| 100 |
+
100,196311,"utils/dataloader.py",2929,0,"",python,selection_command
|
| 101 |
+
101,202514,"TERMINAL",0,0,"git status",,terminal_command
|
| 102 |
+
102,202573,"TERMINAL",0,0,"]633;E;2025-07-07 10:52:44 git status;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;COn branch revised-dataloader\r\nYour branch is up to date with 'origin/revised-dataloader'.\r\n\r\nChanges not staged for commit:\r\n (use ""git add <file>..."" to update what will be committed)\r\n (use ""git restore <file>..."" to discard changes in working directory)\r\n\t[31mmodified: utils/dataloader.py[m\r\n\r\nUntracked files:\r\n (use ""git add <file>..."" to include in what will be committed)\r\n\t[31mdata_tfrecord_duplicated/[m\r\n\t[31mdata_tfrecords/[m\r\n\t[31mlogs/[m\r\n\t[31mread_tf_record.py[m\r\n\t[31mrequirements-franz.txt[m\r\n\t[31mscripts_cremers/[m\r\n\t[31mscripts_horeka/[m\r\n\t[31mslurm-3309772.out[m\r\n\t[31mslurm/[m\r\n\t[31mutils/visualizer.py[m\r\n\r\nno changes added to commit (use ""git add"" and/or ""git commit -a"")\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output
|
| 103 |
+
103,206815,"TERMINAL",0,0,"git add utils/dataloader.py",,terminal_command
|
| 104 |
+
104,206843,"TERMINAL",0,0,"]633;E;2025-07-07 10:52:49 git add utils/dataloader.py ;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output
|
| 105 |
+
105,217694,"TERMINAL",0,0,"git commit -m ""added filter for too short episodes in dataloader""",,terminal_command
|
| 106 |
+
106,217741,"TERMINAL",0,0,"]633;E;2025-07-07 10:52:59 git commit -m ""added filter for too short episodes in dataloader"";f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C",,terminal_output
|
| 107 |
+
107,217921,"TERMINAL",0,0,"[revised-dataloader 1e306ff] added filter for too short episodes in dataloader\r\n 1 file changed, 6 insertions(+)\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output
|
| 108 |
+
108,219165,"TERMINAL",0,0,"git push",,terminal_command
|
| 109 |
+
109,219215,"TERMINAL",0,0,"]633;E;2025-07-07 10:53:01 git push;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C",,terminal_output
|
| 110 |
+
110,220680,"TERMINAL",0,0,"Enumerating objects: 7, done.\r\nCounting objects: 14% (1/7)\rCounting objects: 28% (2/7)\rCounting objects: 42% (3/7)\rCounting objects: 57% (4/7)\rCounting objects: 71% (5/7)\rCounting objects: 85% (6/7)\rCounting objects: 100% (7/7)\rCounting objects: 100% (7/7), done.\r\nDelta compression using up to 152 threads\r\nCompressing objects: 25% (1/4)\rCompressing objects: 50% (2/4)\rCompressing objects: 75% (3/4)\rCompressing objects: 100% (4/4)\rCompressing objects: 100% (4/4), done.\r\nWriting objects: 25% (1/4)\rWriting objects: 50% (2/4)\rWriting objects: 75% (3/4)\rWriting objects: 100% (4/4)\rWriting objects: 100% (4/4), 489 bytes | 244.00 KiB/s, done.\r\nTotal 4 (delta 3), reused 0 (delta 0), pack-reused 0\r\nremote: Resolving deltas: 0% (0/3)[K\rremote: Resolving deltas: 33% (1/3)[K\rremote: Resolving deltas: 66% (2/3)[K\rremote: Resolving deltas: 100% (3/3)[K\rremote: Resolving deltas: 100% (3/3), completed with 3 local objects.[K\r\n",,terminal_output
|
| 111 |
+
111,220984,"TERMINAL",0,0,"To github.com:p-doom/jafar.git\r\n 1eac634..1e306ff revised-dataloader -> revised-dataloader\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output
|
| 112 |
+
112,704052,"TERMINAL",0,0,"queue",,terminal_command
|
| 113 |
+
113,704136,"TERMINAL",0,0,"]633;E;2025-07-07 11:01:06 queue;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C[?1049h[22;0;0t[1;63r(B[m[4l[?7h[H[2JEvery 1.0s: squeue --me[1;358Hhkn1991.localdomain: Mon Jul 7 11:01:06 2025[3;14HJOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)[4;12H3320180 accelerat train_la tum_cte0 CG\t0:00\t 1 hkn0405[63;402H",,terminal_output
|
| 114 |
+
114,705241,"TERMINAL",0,0,"[1;397H7[63d\t ",,terminal_output
|
| 115 |
+
115,706271,"TERMINAL",0,0,"[1;397H8[63d\t ",,terminal_output
|
| 116 |
+
116,707322,"TERMINAL",0,0,"[1;397H9[63d\t ",,terminal_output
|
| 117 |
+
117,708343,"TERMINAL",0,0,"[1;396H10[63d\t ",,terminal_output
|
| 118 |
+
118,708459,"TERMINAL",0,0,"[63;1H[?1049l[23;0;0t\r[?1l>]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output
|
| 119 |
+
119,722511,"TERMINAL",0,0,"cd $ws_dir",,terminal_command
|
| 120 |
+
120,724566,"TERMINAL",0,0,"cd ..",,terminal_command
|
| 121 |
+
121,726017,"TERMINAL",0,0,"cd logs/",,terminal_command
|
| 122 |
+
122,726381,"TERMINAL",0,0,"ls",,terminal_command
|
| 123 |
+
123,726418,"TERMINAL",0,0,"]633;E;2025-07-07 11:01:28 ls;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C[0m[01;34m3306965[0m [01;34mlogs_alfred[0m [01;34mlogs_franz[0m [01;34mlogs_mihir[0m\r\n]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs]633;D;0",,terminal_output
|
| 124 |
+
124,728472,"TERMINAL",0,0,"cd logs_mihir/",,terminal_command
|
| 125 |
+
125,728753,"TERMINAL",0,0,"ls",,terminal_command
|
| 126 |
+
126,728804,"TERMINAL",0,0,"]633;E;2025-07-07 11:01:30 ls;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C",,terminal_output
|
| 127 |
+
127,729035,"TERMINAL",0,0,"train_lam_action_space_scaling_10_3320179.log train_lam_action_space_scaling_6_3318549.log train_lam_model_size_scaling_38M_3317231.log train_tokenizer_batch_size_scaling_8_node_3320176.log train_tokenizer_model_size_scaling_200M_3313563.log train_tokenizer_model_size_scaling_37M_3316022.log train_tokenizer_model_size_scaling_80M_3313564.log\r\ntrain_lam_action_space_scaling_10_3321529.log train_lam_action_space_scaling_6_3320178.log train_tokenizer_batch_size_scaling_16_node_3321526.log train_tokenizer_batch_size_scaling_8_node_3321525.log train_tokenizer_model_size_scaling_200M_3316020.log train_tokenizer_model_size_scaling_37M_3317232.log train_tokenizer_model_size_scaling_80M_3316026.log\r\ntrain_lam_action_space_scaling_12_3318546.log train_lam_action_space_scaling_6_3321528.log train_tokenizer_batch_size_scaling_1_node_3318551.log train_tokenizer_minecraft_overfit_sample_3309656.log train_tokenizer_model_size_scaling_227M_3317234.log train_tokenizer_model_size_scaling_37M_3317239.log\r\ntrain_lam_action_space_scaling_12_3320177.log train_lam_action_space_scaling_8_3318550.log train_tokenizer_batch_size_scaling_2_node_3318552.log train_tokenizer_model_size_scaling_127M_3317233.log train_tokenizer_model_size_scaling_227M_3318555.log train_tokenizer_model_size_scaling_37M_3318556.log\r\ntrain_lam_action_space_scaling_12_3321527.log train_lam_minecraft_overfit_sample_3309655.log train_tokenizer_batch_size_scaling_4_node_3318553.log train_tokenizer_model_size_scaling_127M_3318554.log train_tokenizer_model_size_scaling_227M_3320173.log train_tokenizer_model_size_scaling_74M_3318557.log\r\ntrain_lam_action_space_scaling_20_3318547.log train_lam_model_size_scaling_38M_3317098.log train_tokenizer_batch_size_scaling_4_node_3320175.log train_tokenizer_model_size_scaling_140M_3313562.log train_tokenizer_model_size_scaling_227M_3321523.log train_tokenizer_model_size_scaling_74M_3320174.log\r\ntrain_lam_action_space_scaling_50_3320180.log train_lam_model_size_scaling_38M_3317115.log train_tokenizer_batch_size_scaling_4_node_3321524.log train_tokenizer_model_size_scaling_140M_3316019.log train_tokenizer_model_size_scaling_37M_3313565.log train_tokenizer_model_size_scaling_74M_3321522.log\r\n]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output
|
| 128 |
+
128,756754,"TERMINAL",0,0,"echo $(pwd)/train_tokenizer_batch_size_scaling_16_node_3321526.log",,terminal_command
|
| 129 |
+
129,756798,"TERMINAL",0,0,"]633;E;2025-07-07 11:01:58 echo $(pwd)/train_tokenizer_batch_size_scaling_16_node_3321526.log;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_tokenizer_batch_size_scaling_16_node_3321526.log\r\n]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output
|
| 130 |
+
130,1152416,"utils/dataloader.py",0,0,"",python,tab
|
| 131 |
+
131,1177005,"TERMINAL",0,0,"git checkout runner",,terminal_command
|
| 132 |
+
132,1177092,"TERMINAL",0,0,"]633;E;2025-07-07 11:08:59 git checkout runner;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;Cfatal: not a git repository (or any parent up to mount point /hkfs)\r\nStopping at filesystem boundary (GIT_DISCOVERY_ACROSS_FILESYSTEM not set).\r\n]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;128",,terminal_output
|
| 133 |
+
133,1190795,"utils/dataloader.py",0,0,"",python,tab
|
| 134 |
+
134,1384534,"utils/dataloader.py",4648,0,"",python,selection_mouse
|
| 135 |
+
135,1384551,"utils/dataloader.py",4647,0,"",python,selection_command
|
| 136 |
+
136,1388072,"TERMINAL",0,0,"queue",,terminal_command
|
| 137 |
+
137,1388130,"TERMINAL",0,0,"]633;E;2025-07-07 11:12:30 queue;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C[?1049h[22;0;0t[1;63r(B[m[4l[?7h[H[2JEvery 1.0s: squeue --me[1;358Hhkn1991.localdomain: Mon Jul 7 11:12:30 2025[3;14HJOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)[4;12H3320180 accelerat train_la tum_cte0 CG\t0:00\t 1 hkn0405[63;402H",,terminal_output
|
| 138 |
+
138,1389126,"TERMINAL",0,0,"[1;397H1[63d\t ",,terminal_output
|
| 139 |
+
139,1390163,"TERMINAL",0,0,"[63;1H[?1049l[23;0;0t\r[?1l>]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output
|
| 140 |
+
140,1393075,"TERMINAL",0,0,"idling",,terminal_command
|
| 141 |
+
141,1393136,"TERMINAL",0,0,"]633;E;2025-07-07 11:12:35 idling;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C[?1049h[22;0;0t[1;63r(B[m[4l[?7h[H[2JEvery 1.0s: sinfo_t_idle[1;358Hhkn1991.localdomain: Mon Jul 7 11:12:35 2025[3;1HPartition dev_cpuonly[3;35H: 11 nodes idle\r[4dPartition cpuonly[4;35H: 325 nodes idle\r[5dPartition dev_accelerated[5;35H:\t 3 nodes idle\r[6dPartition accelerated[6;35H: 39 nodes idle\r[7dPartition dev_accelerated-h100 :\t 0 nodes idle\r[8dPartition accelerated-h100[8;35H:\t 1 nodes idle\r[9dPartition large[9;35H:\t 7 nodes idle[63;402H",,terminal_output
|
| 142 |
+
142,1394285,"TERMINAL",0,0,"[1;397H6[63d\t [63;1H[?1049l[23;0;0t\r[?1l>]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output
|
| 143 |
+
143,1397165,"TERMINAL",0,0,"cd",,terminal_command
|
| 144 |
+
144,1397180,"TERMINAL",0,0,"]633;E;2025-07-07 11:12:39 cd;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;C]0;tum_cte0515@hkn1991:~]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515",,terminal_output
|
| 145 |
+
145,1404501,"TERMINAL",0,0,"cd Projects/jafar",,terminal_command
|
| 146 |
+
146,1445110,"TERMINAL",0,0,"salloc --time=10:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5",,terminal_command
|
| 147 |
+
147,1445175,"TERMINAL",0,0,"]633;E;2025-07-07 11:13:27 salloc --time=10:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5 ;f27e7af0-1f60-464b-afac-68c0fe98f46d]633;Csalloc: Granted job allocation 3326035\r\n",,terminal_output
|
| 148 |
+
148,1445309,"TERMINAL",0,0,"salloc: Waiting for resource configuration\r\n",,terminal_output
|
| 149 |
+
149,1472365,"TERMINAL",0,0,"salloc: Nodes hkn0734 are ready for job\r\n",,terminal_output
|
| 150 |
+
150,1473217,"TERMINAL",0,0,"]0;tum_cte0515@hkn0734:~/Projects/jafar[?2004h[tum_cte0515@hkn0734 jafar]$ ",,terminal_output
|
| 151 |
+
151,1699781,"TERMINAL",0,0,"s",,terminal_output
|
| 152 |
+
152,1699852,"TERMINAL",0,0,"[?25l[63;31Ho[63;32H[?25h",,terminal_output
|
| 153 |
+
153,1699960,"TERMINAL",0,0,"[?25l[63;32Hu[63;33H[?25h",,terminal_output
|
| 154 |
+
154,1700026,"TERMINAL",0,0,"[?25l[63;33Hr[63;34H[?25h",,terminal_output
|
| 155 |
+
155,1700174,"TERMINAL",0,0,"[?25l[63;34Hc[63;35H[?25h",,terminal_output
|
| 156 |
+
156,1700311,"TERMINAL",0,0,"[?25l[63;35He[63;36H[?25h",,terminal_output
|
| 157 |
+
157,1700451,"TERMINAL",0,0,"[?25l[63;36H [63;37H[?25h",,terminal_output
|
| 158 |
+
158,1700517,"TERMINAL",0,0,"[?25l[63;37H.[63;38H[?25h",,terminal_output
|
| 159 |
+
159,1700959,"TERMINAL",0,0,"[?25l[63;38Hv[63;40H[?25h[?25l[63;39He[63;40H[?25h",,terminal_output
|
| 160 |
+
160,1701278,"TERMINAL",0,0,"nv/",,terminal_output
|
| 161 |
+
161,1701883,"TERMINAL",0,0,"[?25l[63;43Hb[63;44H[?25h",,terminal_output
|
| 162 |
+
162,1702072,"TERMINAL",0,0,"in/",,terminal_output
|
| 163 |
+
163,1702317,"TERMINAL",0,0,"[?25l[63;47Ha[63;48H[?25h",,terminal_output
|
| 164 |
+
164,1702396,"TERMINAL",0,0,"[?25l[63;48Hc[63;49H[?25h",,terminal_output
|
| 165 |
+
165,1702648,"TERMINAL",0,0,"tivate",,terminal_output
|
| 166 |
+
166,1702996,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0734:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0734 jafar]$ ",,terminal_output
|
| 167 |
+
167,1705224,"TERMINAL",0,0,"[?25lg[2mi[22m[63;40H[?25h[?25l[63;39Hi[63;40H[?25h",,terminal_output
|
| 168 |
+
168,1705388,"TERMINAL",0,0,"[?25l[63;40Ht[63;42H[?25h",,terminal_output
|
| 169 |
+
169,1705441,"TERMINAL",0,0,"[?25l[63;41H [63;42H[?25h",,terminal_output
|
| 170 |
+
170,1705641,"TERMINAL",0,0,"[?25l[63;42He[63;43H[?25h",,terminal_output
|
| 171 |
+
171,1707051,"TERMINAL",0,0,"[?25l[63;42Hc[63;43H[?25h",,terminal_output
|
| 172 |
+
172,1707116,"TERMINAL",0,0,"[?25l[63;43Hh[63;44H[?25h",,terminal_output
|
| 173 |
+
173,1707261,"TERMINAL",0,0,"[?25l[63;44He[63;45H[?25h",,terminal_output
|
| 174 |
+
174,1707413,"TERMINAL",0,0,"[?25l[63;45Hc[63;47H[?25h[?25l[63;46Hk[63;47H[?25h",,terminal_output
|
| 175 |
+
175,1707656,"TERMINAL",0,0,"[?25l[63;47Ho[63;48H[?25h",,terminal_output
|
| 176 |
+
176,1707736,"TERMINAL",0,0,"[?25l[63;48Hu[63;49H[?25h",,terminal_output
|
| 177 |
+
177,1707856,"TERMINAL",0,0,"[?25l[63;49Ht[63;51H[?25h",,terminal_output
|
| 178 |
+
178,1707922,"TERMINAL",0,0,"[?25l[63;50H [63;51H[?25h",,terminal_output
|
| 179 |
+
179,1708772,"TERMINAL",0,0,"[?25l[63;51Hf[63;52H[?25h",,terminal_output
|
| 180 |
+
180,1708895,"TERMINAL",0,0,"[?25l[63;52Hi[63;53H[?25h",,terminal_output
|
| 181 |
+
181,1709131,"TERMINAL",0,0,"[?25l[63;53Hx[63;54H[?25h",,terminal_output
|
| 182 |
+
182,1709618,"TERMINAL",0,0,"[?25l[63;54H-[63;55H[?25h",,terminal_output
|
| 183 |
+
183,1709774,"TERMINAL",0,0,"[?25l[63;55Hs[63;56H[?25h",,terminal_output
|
| 184 |
+
184,1709934,"TERMINAL",0,0,"[?25l[63;56Ha[63;57H[?25h",,terminal_output
|
| 185 |
+
185,1710041,"TERMINAL",0,0,"[?25l[63;57Hm[63;58H[?25h",,terminal_output
|
| 186 |
+
186,1710274,"TERMINAL",0,0,"[?25l[63;58Hp[63;60H[?25h[?25l[63;59Hl[63;60H[?25h",,terminal_output
|
| 187 |
+
187,1710539,"TERMINAL",0,0,"[?25l[63;60Hi[63;62H[?25h[?25l[63;61Hn[63;62H[?25h",,terminal_output
|
| 188 |
+
188,1710608,"TERMINAL",0,0,"[?25l[63;62Hg[63;63H[?25h",,terminal_output
|
| 189 |
+
189,1710830,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output
|
| 190 |
+
190,1711585,"TERMINAL",0,0,"Switched to branch 'fix-sampling'\r\nYour branch is behind 'origin/fix-sampling' by 2 commits, and can be fast-forwarded.\r\n (use ""git pull"" to update your local branch)\r\n]0;tum_cte0515@hkn0734:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0734 jafar]$ ",,terminal_output
|
| 191 |
+
191,1712699,"TERMINAL",0,0,"g",,terminal_output
|
| 192 |
+
192,1712812,"TERMINAL",0,0,"[?25l[63;39Hi[63;40H[?25h",,terminal_output
|
| 193 |
+
193,1712904,"TERMINAL",0,0,"[?25l[63;40Ht[63;42H[?25h[?25l[63;41H [63;42H[?25h",,terminal_output
|
| 194 |
+
194,1713002,"TERMINAL",0,0,"[?25l[63;42Hp[63;43H[?25h",,terminal_output
|
| 195 |
+
195,1713211,"TERMINAL",0,0,"[?25l[63;43Hu[63;44H[?25h",,terminal_output
|
| 196 |
+
196,1713391,"TERMINAL",0,0,"[?25l[63;44Hl[63;45H[?25h",,terminal_output
|
| 197 |
+
197,1713523,"TERMINAL",0,0,"[?25l[63;45Hl[63;46H[?25h",,terminal_output
|
| 198 |
+
198,1713632,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output
|
| 199 |
+
199,1715291,"TERMINAL",0,0,"Updating ae9451f..6e623c6\r\nFast-forward\r\n genie.py | 16 [32m++++++++[m[31m--------[m\r\n 1 file changed, 8 insertions(+), 8 deletions(-)\r\n]0;tum_cte0515@hkn0734:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0734 jafar]$ ",,terminal_output
|
| 200 |
+
200,1715555,"",0,0,"Switched from branch 'revised-dataloader' to 'fix-sampling'",,git_branch_checkout
|
| 201 |
+
201,1722688,"genie.py",0,0,"from typing import Dict, Any\n\nimport optax\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\nfrom jax import NamedSharding\nfrom flax.training.train_state import TrainState\nfrom flax.training import orbax_utils\nfrom orbax.checkpoint import PyTreeCheckpointer\n\nfrom models.dynamics import DynamicsMaskGIT\nfrom models.lam import LatentActionModel\nfrom models.tokenizer import TokenizerVQVAE\n\n\nclass Genie(nn.Module):\n """"""Genie model""""""\n\n # --- Tokenizer ---\n in_dim: int\n tokenizer_dim: int\n latent_patch_dim: int\n num_patch_latents: int\n patch_size: int\n tokenizer_num_blocks: int\n tokenizer_num_heads: int\n # --- LAM ---\n lam_dim: int\n latent_action_dim: int\n num_latent_actions: int\n lam_patch_size: int\n lam_num_blocks: int\n lam_num_heads: int\n # --- Dynamics ---\n dyna_dim: int\n dyna_num_blocks: int\n dyna_num_heads: int\n dropout: float = 0.0\n mask_limit: float = 0.0\n\n def setup(self):\n self.tokenizer = TokenizerVQVAE(\n in_dim=self.in_dim,\n model_dim=self.tokenizer_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_patch_latents,\n patch_size=self.patch_size,\n num_blocks=self.tokenizer_num_blocks,\n num_heads=self.tokenizer_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n )\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_latent_actions,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n )\n self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n num_latents=self.num_patch_latents,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n )\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n tokenizer_outputs = self.tokenizer.vq_encode(batch[""videos""], training=False)\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\n outputs = dict(\n video_tokens=jax.lax.stop_gradient(tokenizer_outputs[""indices""]),\n latent_actions=jax.lax.stop_gradient(lam_outputs[""z_q""]),\n )\n outputs[""mask_rng""] = batch[""mask_rng""]\n dyna_outputs = self.dynamics(outputs, training)\n outputs.update(dyna_outputs)\n mle_indices = jnp.argmax(outputs[""token_logits""], axis=-1)\n outputs[""recon""] = self.tokenizer.decode(\n mle_indices, batch[""videos""].shape[2:4]\n )\n return outputs\n\n @nn.compact\n def sample(\n self,\n batch: Dict[str, Any],\n seq_len: int,\n steps: int = 25,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> Any:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by \n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size \n T: number of input (conditioning) frames \n N: patches per frame \n S: sequence length \n A: action space \n D: model latent dimension\n """"""\n # --- Encode videos and actions ---\n tokenizer_out = self.tokenizer.vq_encode(batch[""videos""], training=False)\n token_idxs = tokenizer_out[""indices""] # (B, T, N)\n B, T, N = token_idxs.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs.dtype)\n token_idxs = jnp.concatenate([token_idxs, pad], axis=1) # (B, S, N)\n action_tokens = self.lam.vq.get_codes(batch[""latent_actions""])\n\n for step_t in range(T, seq_len):\n print(f""Sampling Frame {step_t}..."")\n # mask current and future frames (i.e., t >= step_t)\n mask = jnp.arange(seq_len) >= step_t # (S,)\n mask = jnp.broadcast_to(mask[None, :, None], (B, seq_len, N)) # (B, S, N)\n mask = mask.astype(bool)\n token_idxs *= ~mask\n\n # --- Initialize MaskGIT ---\n init_carry = (\n batch[""rng""],\n token_idxs,\n mask,\n action_tokens,\n )\n\n MaskGITLoop = nn.scan(\n MaskGITStep,\n variable_broadcast=""params"",\n split_rngs={""params"": False},\n in_axes=0,\n out_axes=0,\n length=steps,\n )\n\n # --- Run MaskGIT loop ---\n loop_fn = MaskGITLoop(\n dynamics=self.dynamics,\n tokenizer=self.tokenizer,\n temperature=temperature,\n sample_argmax=sample_argmax,\n steps=steps,\n )\n final_carry, _ = loop_fn(init_carry, jnp.arange(steps))\n token_idxs = final_carry[1]\n\n final_frames = self.tokenizer.decode(\n token_idxs,\n video_hw=batch[""videos""].shape[2:4],\n )\n return final_frames \n\n def vq_encode(self, batch, training) -> Dict[str, Any]:\n # --- Preprocess videos ---\n lam_output = self.lam.vq_encode(batch[""videos""], training=training)\n return lam_output[""indices""]\n\n\nclass MaskGITStep(nn.Module):\n dynamics: nn.Module\n tokenizer: nn.Module\n temperature: float\n sample_argmax: bool\n steps: int\n\n @nn.compact\n def __call__(self, carry, x):\n rng, token_idxs, mask, action_tokens = carry\n step = x\n N = token_idxs.shape[2]\n\n # --- Construct + encode video ---\n vid_embed = self.dynamics.patch_embed(token_idxs) # (B, S, N, D)\n mask_token = self.dynamics.mask_token # (1, 1, 1, D,)\n mask_expanded = mask[..., None] # (B, S, N, 1) \n vid_embed = jnp.where(mask_expanded, mask_token, vid_embed)\n\n # --- Predict transition ---\n act_embed = self.dynamics.action_up(action_tokens)\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n unmasked_ratio = jnp.cos(jnp.pi * (step + 1) / (self.steps * 2))\n step_temp = self.temperature * (1.0 - unmasked_ratio)\n final_logits = self.dynamics.dynamics(vid_embed) / step_temp\n\n # --- Sample new tokens for final frame ---\n if self.sample_argmax:\n sampled_token_idxs = jnp.argmax(final_logits, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs = jnp.where(\n step == self.steps - 1,\n jnp.argmax(final_logits, axis=-1),\n jax.random.categorical(_rng, final_logits),\n )\n gather_fn = jax.vmap(jax.vmap(jax.vmap(lambda x, y: x[y])))\n final_token_probs = gather_fn(jax.nn.softmax(final_logits), sampled_token_idxs)\n final_token_probs += ~mask\n # Update masked tokens only\n token_idxs = jnp.where(mask, sampled_token_idxs, token_idxs)\n\n # --- Update mask ---\n num_unmasked_tokens = jnp.round(N * (1.0 - unmasked_ratio)).astype(int)\n idx_mask = jnp.arange(final_token_probs.shape[-1]) > num_unmasked_tokens\n sorted_idxs = jnp.argsort(final_token_probs, axis=-1, descending=True)\n mask_update_fn = jax.vmap(lambda msk, ids: msk.at[ids].set(idx_mask))\n new_mask = mask_update_fn(mask, sorted_idxs)\n\n new_carry = (rng, token_idxs, new_mask, action_tokens)\n return new_carry, None\n\ndef restore_genie_components(\n train_state: TrainState,\n sharding: NamedSharding,\n inputs: Dict[str, jax.Array],\n rng: jax.Array,\n args,\n):\n """"""Restore pre-trained Genie components""""""\n rng, _rng = jax.random.split(rng)\n\n dummy_tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.tokenizer_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n num_blocks=args.tokenizer_num_blocks,\n num_heads=args.tokenizer_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n )\n dummy_lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.lam_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_latent_actions,\n patch_size=args.lam_patch_size,\n num_blocks=args.lam_num_blocks,\n num_heads=args.lam_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n )\n tokenizer_init_params = dummy_tokenizer.init(_rng, inputs)\n lam_init_params = dummy_lam.init(_rng, inputs)\n\n # dummy values since we only use tx to initialize the dummy train states\n dummy_tx = optax.adamw(\n learning_rate=optax.constant_schedule(args.max_lr),\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n )\n\n dummy_tokenizer_train_state = TrainState.create(\n apply_fn=dummy_tokenizer.apply, params=tokenizer_init_params, tx=dummy_tx\n )\n dummy_lam_train_state = TrainState.create(\n apply_fn=dummy_lam.apply, params=lam_init_params, tx=dummy_tx\n )\n\n def create_abstract_sharded_pytree(pytree_template, sharding_spec):\n """"""Replaces arrays in a pytree with ShapeDtypeStructs having the given sharding.""""""\n\n def map_fn(leaf_template):\n if hasattr(leaf_template, ""shape"") and hasattr(leaf_template, ""dtype""):\n return jax.ShapeDtypeStruct(\n leaf_template.shape, leaf_template.dtype, sharding=sharding_spec\n )\n return leaf_template\n\n return jax.tree_util.tree_map(map_fn, pytree_template)\n\n abstract_sharded_tokenizer_state = create_abstract_sharded_pytree(\n dummy_tokenizer_train_state, sharding\n )\n abstract_sharded_lam_state = create_abstract_sharded_pytree(\n dummy_lam_train_state, sharding\n )\n\n tokenizer_restore_target = {""model"": abstract_sharded_tokenizer_state}\n lam_restore_target = {""model"": abstract_sharded_lam_state}\n\n tokenizer_restore_args = orbax_utils.restore_args_from_target(\n tokenizer_restore_target\n )\n lam_restore_args = orbax_utils.restore_args_from_target(lam_restore_target)\n\n restored_tokenizer_params = (\n PyTreeCheckpointer()\n .restore(\n args.tokenizer_checkpoint,\n item=tokenizer_restore_target,\n restore_args=tokenizer_restore_args,\n )[""model""]\n .params[""params""]\n )\n restored_lam_params = (\n PyTreeCheckpointer()\n .restore(\n args.lam_checkpoint, item=lam_restore_target, restore_args=lam_restore_args\n )[""model""]\n .params[""params""]\n )\n # Genie does not initialize all LAM modules, thus we omit those extra modules during restoration\n # (f.srambical) FIXME: Currently, this is a small HBM memory crunch since the LAM's decoder is loaded into HBM and immediately dicarded.\n # A workaround would be to restore to host memory first, and only move the weights to HBM after pruning the decoder\n restored_lam_params = {\n k: v\n for k, v in restored_lam_params.items()\n if k in train_state.params[""params""][""lam""]\n }\n\n train_state.params[""params""][""tokenizer""].update(restored_tokenizer_params)\n train_state.params[""params""][""lam""].update(restored_lam_params)\n\n return train_state\n",python,tab
|
| 202 |
+
202,1731948,"sample.py",0,0,"from dataclasses import dataclass\nimport time\nimport os\n\nimport dm_pix as pix\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\nimport numpy as np\nfrom orbax.checkpoint import PyTreeCheckpointer\nfrom PIL import Image, ImageDraw\nimport tyro\n\nfrom genie import Genie\nfrom utils.dataloader import get_dataloader\n\n\n@dataclass\nclass Args:\n # Experiment\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = ""data/coinrun_episodes""\n checkpoint: str = """"\n # Sampling\n batch_size: int = 1\n maskgit_steps: int = 25\n temperature: float = 1.0\n sample_argmax: bool = True\n start_frame: int = 0\n # Tokenizer checkpoint\n tokenizer_dim: int = 512\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 8\n tokenizer_num_heads: int = 8\n # LAM checkpoint\n lam_dim: int = 512\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 8\n lam_num_heads: int = 8\n # Dynamics checkpoint\n dyna_dim: int = 512\n dyna_num_blocks: int = 12\n dyna_num_heads: int = 8\n\n\nargs = tyro.cli(Args)\nrng = jax.random.PRNGKey(args.seed)\n\n# --- Load Genie checkpoint ---\ngenie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n # Dynamics\n dyna_dim=args.dyna_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n)\nrng, _rng = jax.random.split(rng)\nimage_shape = (args.image_height, args.image_width, args.image_channels)\ndummy_inputs = dict(\n videos=jnp.zeros((args.batch_size, args.seq_len, *image_shape), dtype=jnp.float32),\n mask_rng=_rng,\n)\nrng, _rng = jax.random.split(rng)\nparams = genie.init(_rng, dummy_inputs)\nckpt = PyTreeCheckpointer().restore(args.checkpoint)[""model""][""params""][""params""]\nparams[""params""].update(ckpt)\n\n\ndef _sampling_wrapper(module, batch):\n return module.sample(batch, args.seq_len, args.maskgit_steps, args.temperature, args.sample_argmax)\n\n# --- Define autoregressive sampling loop ---\ndef _autoreg_sample(rng, video_batch, action_batch):\n vid = video_batch[:, : args.start_frame + 1]\n sampling_fn = jax.jit(nn.apply(_sampling_wrapper, genie)) \n rng, _rng = jax.random.split(rng)\n batch = dict(videos=vid, latent_actions=action_batch, rng=_rng)\n generated_vid = sampling_fn(\n params,\n batch\n )\n return generated_vid\n\n# --- Get video + latent actions ---\ntfrecord_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".tfrecord"")\n]\ndataloader = get_dataloader(\n tfrecord_files,\n args.seq_len,\n args.batch_size,\n args.image_height,\n args.image_width,\n args.image_channels,\n seed=args.seed,\n)\nvideo_batch = next(iter(dataloader))\n# Get latent actions from first video only; clip them down to the specified seq_len\nfirst_video = video_batch[:1, :args.seq_len]\nbatch = dict(videos=first_video)\naction_batch = genie.apply(params, batch, False, method=Genie.vq_encode)\naction_batch = action_batch.reshape(1, args.seq_len - 1, 1)\n# Use actions from first video for all videos\naction_batch = jnp.repeat(action_batch, video_batch.shape[0], axis=0)\n\n# --- Sample + evaluate video ---\nvid = _autoreg_sample(rng, video_batch, action_batch)\ngt = video_batch[:, : vid.shape[1]].clip(0, 1).reshape(-1, *video_batch.shape[2:])\nrecon = vid.clip(0, 1).reshape(-1, *vid.shape[2:])\nssim = pix.ssim(gt[:, args.start_frame + 1 :], recon[:, args.start_frame + 1 :]).mean()\nprint(f""SSIM: {ssim}"")\n\n# --- Construct video ---\nfirst_true = (video_batch[0:1] * 255).astype(np.uint8)\nfirst_pred = (vid[0:1] * 255).astype(np.uint8)\nfirst_video_comparison = np.zeros((2, *vid.shape[1:5]), dtype=np.uint8)\nfirst_video_comparison[0] = first_true[:, : vid.shape[1]]\nfirst_video_comparison[1] = first_pred\n# For other videos, only show generated video\nother_preds = (vid[1:] * 255).astype(np.uint8)\nall_frames = np.concatenate([first_video_comparison, other_preds], axis=0)\nflat_vid = einops.rearrange(all_frames, ""n t h w c -> t h (n w) c"")\n\n# --- Save video ---\nimgs = [Image.fromarray(img) for img in flat_vid]\n# Write actions on each frame\nfor img, action in zip(imgs[1:], action_batch[0, :, 0]):\n d = ImageDraw.Draw(img)\n d.text((2, 2), f""{action}"", fill=255)\nimgs[0].save(\n f""generation_{time.time()}.gif"",\n save_all=True,\n append_images=imgs[1:],\n duration=250,\n loop=0,\n)\n",python,tab
|
| 203 |
+
203,1733276,"sample.py",3057,0,"",python,selection_mouse
|
| 204 |
+
204,1733276,"sample.py",3056,0,"",python,selection_command
|
| 205 |
+
205,1734341,"genie.py",0,0,"",python,tab
|
| 206 |
+
206,1758226,"genie.py",4621,0,"",python,selection_mouse
|
| 207 |
+
207,1761801,"genie.py",4527,0,"",python,selection_mouse
|
| 208 |
+
208,1773944,"genie.py",4950,0,"",python,selection_mouse
|
| 209 |
+
209,1774091,"genie.py",4946,5,"batch",python,selection_mouse
|
| 210 |
+
210,1782472,"genie.py",4954,0,"",python,selection_mouse
|
| 211 |
+
211,1782652,"genie.py",4953,3,"rng",python,selection_mouse
|
| 212 |
+
212,10441352,"genie.py",0,0,"Switched from branch 'fix-sampling' to 'correct-batched-sampling'",python,git_branch_checkout
|
| 213 |
+
213,15726982,"genie.py",0,0,"Switched from branch 'correct-batched-sampling' to 'main'",python,git_branch_checkout
|
927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-f1a23455-555b-44b7-b7f2-5fb8550d75021753722279799-2025_07_28-19.04.51.762/source.csv
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,5,"utils/nn.py",0,0,"import math\nfrom typing import Tuple\n\nfrom flax import linen as nn\nimport jax\nimport jax.numpy as jnp\nimport einops\n\n\nclass PositionalEncoding(nn.Module):\n """"""https://uvadlc-notebooks.readthedocs.io/en/latest/tutorial_notebooks/JAX/tutorial6/Transformers_and_MHAttention.html""""""\n\n d_model: int # Hidden dimensionality of the input.\n max_len: int = 5000 # Maximum length of a sequence to expect.\n\n def setup(self):\n # Create matrix of [SeqLen, HiddenDim] representing the positional encoding for max_len inputs\n self.pe = jnp.zeros((self.max_len, self.d_model))\n position = jnp.arange(0, self.max_len, dtype=jnp.float32)[:, None]\n div_term = jnp.exp(\n jnp.arange(0, self.d_model, 2) * (-math.log(10000.0) / self.d_model)\n )\n self.pe = self.pe.at[:, 0::2].set(jnp.sin(position * div_term))\n self.pe = self.pe.at[:, 1::2].set(jnp.cos(position * div_term))\n\n def __call__(self, x):\n x = x + self.pe[: x.shape[2]]\n return x\n\n\nclass STBlock(nn.Module):\n dim: int\n ffn_dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n use_flash_attention: bool\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n attention_fn=_create_flash_attention_fn(\n self.use_flash_attention, is_causal=False\n ),\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n attention_fn=_create_flash_attention_fn(\n self.use_flash_attention, is_causal=True\n ),\n # FIXME (f.srambical): check whether we should still pass the mask if we set is_causal=True\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n z = nn.Dense(\n self.ffn_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n return x\n\n\nclass STTransformer(nn.Module):\n model_dim: int\n ffn_dim: int\n out_dim: int\n num_blocks: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n use_flash_attention: bool\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n x = nn.Sequential(\n [\n nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n ]\n )(x)\n for _ in range(self.num_blocks):\n x = STBlock(\n dim=self.model_dim,\n ffn_dim=self.ffn_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n )(x)\n x = nn.Dense(\n self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n return x # (B, T, E)\n\n\ndef normalize(x):\n return x / (jnp.linalg.norm(x, ord=2, axis=-1, keepdims=True) + 1e-8)\n\n\nclass VectorQuantizer(nn.Module):\n latent_dim: int\n num_latents: int\n dropout: float\n\n def setup(self):\n self.codebook = normalize(\n self.param(\n ""codebook"",\n nn.initializers.lecun_uniform(),\n (self.num_latents, self.latent_dim),\n )\n )\n self.drop = nn.Dropout(self.dropout, deterministic=False)\n\n def __call__(\n self, x: jax.Array, training: bool\n ) -> Tuple[jax.Array, jax.Array, jax.Array, jax.Array]:\n # --- Compute distances ---\n x = normalize(x)\n codebook = normalize(self.codebook)\n distance = -jnp.matmul(x, codebook.T)\n if training:\n dropout_key = self.make_rng(""dropout"")\n distance = self.drop(distance, rng=dropout_key)\n\n # --- Get indices and embeddings ---\n indices = jnp.argmin(distance, axis=-1)\n z = self.codebook[indices]\n\n # --- Straight through estimator ---\n z_q = x + jax.lax.stop_gradient(z - x)\n return z_q, z, x, indices\n\n def get_codes(self, indices: jax.Array):\n return self.codebook[indices]\n\n\ndef _create_flash_attention_fn(use_flash_attention: bool, is_causal: bool):\n """"""\n Create an attention function that uses flash attention if enabled.\n\n Flax MultiHeadAttention provides tensors with shape (batch..., length, num_heads, head_dim)\n jax.nn.dot_product_attention expects (batch, length, num_heads, head_dim).\n\n We need to reshape to ensure compatibility. cuDNN's flash attention additionally\n requires a sequence length that is a multiple of 4. We pad the sequence length to the nearest\n multiple of 4 and mask accordingly.\n """"""\n\n def attention_fn(query, key, value, bias=None, mask=None, **kwargs):\n implementation = ""cudnn"" if use_flash_attention else None\n\n def _rearrange(x):\n return einops.rearrange(x, ""... l h d -> (...) l h d"")\n\n def _pad(x):\n return jnp.pad(x, ((0, 0), (0, pad_size), (0, 0), (0, 0)))\n\n def _fuse_masks(mask: jax.Array, attention_mask: jax.Array) -> jax.Array:\n mask_bool = mask.astype(jnp.bool_)\n expanded_mask = jnp.pad(\n mask_bool, ((0, pad_size), (0, pad_size)), constant_values=False\n )\n return jnp.logical_and(attention_mask, expanded_mask)\n\n original_shape = query.shape\n original_seq_len = query.shape[-3]\n\n # Pad to nearest multiple of 4\n target_seq_len = ((original_seq_len + 3) // 4) * 4\n pad_size = target_seq_len - original_seq_len\n\n query_4d = _pad(_rearrange(query))\n key_4d = _pad(_rearrange(key))\n value_4d = _pad(_rearrange(value))\n\n attention_mask = jnp.ones((target_seq_len, target_seq_len), dtype=jnp.bool_)\n attention_mask = attention_mask.at[original_seq_len:, :].set(False)\n attention_mask = attention_mask.at[:, original_seq_len:].set(False)\n\n mask_4d = (\n _fuse_masks(mask, attention_mask) if mask is not None else attention_mask\n )\n mask_4d = mask_4d[jnp.newaxis, jnp.newaxis, :, :] # (1, 1, seq_len, seq_len)\n\n bias_4d = _pad(_rearrange(bias)) if bias is not None else None\n\n output_4d = jax.nn.dot_product_attention(\n query=query_4d,\n key=key_4d,\n value=value_4d,\n bias=bias_4d,\n mask=mask_4d,\n implementation=implementation,\n is_causal=is_causal,\n **kwargs\n )\n return output_4d[..., :original_seq_len, :, :].reshape(original_shape)\n\n return attention_fn\n",python,tab
|
| 3 |
+
2,511,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"7:04:51 PM [info] Activating crowd-code\n7:04:51 PM [info] Recording started\n7:04:51 PM [info] Initializing git provider using file system watchers...\n7:04:52 PM [info] Git repository found\n7:04:52 PM [info] Git provider initialized successfully\n7:04:52 PM [info] Initial git state: [object Object]\n",Log,tab
|