Add b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-0162b81a-7b28-4ef5-b0e9-c56355c44e161765287461281-2025_12_09-14.41.45.741/source.csv +28 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-048f6756-c59b-4be5-887a-8f28003e7da91761121031395-2025_10_22-10.17.37.67/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-04d765c9-bb6f-4faa-b6ea-8904418254d61760945425208-2025_10_20-09.30.54.363/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-0811166c-19e5-47a9-bbfb-5deb433c29821760527863000-2025_10_15-13.31.33.167/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-090f467e-141d-440a-88f0-fb9757f142d31761307268455-2025_10_24-14.01.33.483/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-09e95c4f-d571-45a4-a506-a80e6e398a3f1761597703208-2025_10_27-21.42.06.986/source.csv +3 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-0ebb31ee-57f5-41bb-b048-9c99802b107c1764329048669-2025_11_28-12.24.33.567/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-0f122a13-52f0-458a-b70b-8adddf08d7fc1764349079194-2025_11_28-17.58.34.412/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-0f1ad218-d2d6-4ace-9150-2ff8d2109f6d1762606110433-2025_11_08-13.49.11.795/source.csv +27 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-11074bd2-09df-4856-bd07-fbc4dd6df28c1761329974242-2025_10_24-20.20.11.685/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-117eaf48-658e-44af-bdb3-ac16b413b55d1764499426748-2025_11_30-11.44.10.289/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-11aa9396-05bd-4153-b391-316c69fcb7011764002704562-2025_11_24-17.45.28.743/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-17597979-3a78-4d91-bff1-68d4cc98cb621761146980297-2025_10_22-17.30.03.925/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-1d012a4c-25e5-4957-be82-a3e8ebc2f90b1760629588479-2025_10_16-17.46.51.747/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-1dbf94d5-e80c-4f76-8c82-ec481c7acea31762603480406-2025_11_08-13.05.09.810/source.csv +5 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-20676a2c-2459-41f4-8896-bf5b1bc9afe71764494282718-2025_11_30-10.18.52.300/source.csv +5 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-21579c2d-8ebc-43ea-af3d-20ff8c9f5eb51761135573580-2025_10_22-14.19.54.427/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-2165ae8d-ebb7-4134-a4e8-9043ba00c6551762612659284-2025_11_08-15.38.17.91/source.csv +25 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-23300773-7107-48dc-95de-2ef6fdc0e97c1760710518542-2025_10_17-16.15.55.319/source.csv +16 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-25df760a-b46a-4185-aa09-ab9e63322f9f1764066566259-2025_11_25-11.29.59.139/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-2b611b25-766c-40fe-9c56-c37495c49abf1764154622152-2025_11_26-11.57.21.424/source.csv +78 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-2f756e36-5a16-427a-9b47-60002a9d13571761321558034-2025_10_24-17.59.44.602/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-32bb221d-d676-4384-9ca4-18916339b62b1764092860950-2025_11_25-18.48.03.180/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-34b6dec0-9dde-4479-9af4-cd0656225b821761477933551-2025_10_26-12.25.57.396/source.csv +39 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-35174d2b-ac4a-4076-abc0-3de665efb30d1764851634497-2025_12_04-13.34.18.756/source.csv +78 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-386afd79-6aa8-4e11-939b-d2205ae947e71760884674230-2025_10_19-16.38.53.787/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-3b2f337f-0bf2-4eb1-81d0-1152f026aa5c1764086742729-2025_11_25-17.06.01.787/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-3e34f5c4-13f2-481c-a23c-176fc713ea5e1761657198697-2025_10_28-14.13.44.256/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-3ee1c8d9-397d-4186-890e-abbebcb677cd1760442276619-2025_10_14-13.49.47.394/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-41e637ef-086e-4e89-977d-363e70d7d3151764500068578-2025_11_30-11.55.34.47/source.csv +74 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-427a4d3b-1eee-4e1b-ad62-0a837e5c21da1764859633492-2025_12_04-15.47.37.05/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-452e8a0c-f3ce-4668-9b2f-7561bcdbf12a1760538538230-2025_10_15-16.29.25.680/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-4e503427-8b75-4417-944d-0f8b64a7ff841761599689551-2025_10_27-22.15.16.579/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-4e6bed5e-6b9e-447e-859a-ff87eb2aa7061761311558837-2025_10_24-15.13.03.416/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-4ffb36c2-9d46-4459-9b1b-f93f144ad2f61765193679229-2025_12_08-12.35.03.82/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-529e4bba-7d8a-4661-afaa-55c4c5c035291764770291321-2025_12_03-14.58.42.143/source.csv +6 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-59804d2c-df27-4b41-9fe6-ffdfef8825021760703869419-2025_10_17-14.25.51.972/source.csv +3 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-5a5585eb-19d8-428e-a477-32e8773045fc1761049920108-2025_10_21-14.32.24.316/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-5f1cd0c2-8bc8-41c4-bed3-625c4b4adb111761657862721-2025_10_28-14.24.45.589/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-5f8ae991-2efe-46c5-b76d-9518b238b91d1761295762673-2025_10_24-10.49.49.33/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-6048c8e1-2439-424f-8372-d4ccd17c2d5f1760522924995-2025_10_15-12.09.21.860/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-6116c825-ee17-41e6-b50b-e2d4d748cd831761555737545-2025_10_27-10.02.40.737/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-66805e21-c18a-477d-a6fd-7ee8129b9a211760898484058-2025_10_19-20.28.33.558/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-6877ad19-9807-438f-8b51-95c34f7ec7121764157755585-2025_11_26-12.49.36.76/source.csv +4 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-698eb099-f494-4ec6-a868-8174b5764b0d1765283687620-2025_12_09-13.35.26.529/source.csv +7 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-73f8939e-63ce-4a1f-ae80-e8b595f3c9321764500864397-2025_11_30-12.08.15.208/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-75fcc8f0-691c-4551-9c8c-7a9bcc824a7e1760553685394-2025_10_15-20.41.46.66/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-76116d19-106d-4bbc-8d56-8e736ef694361765124265000-2025_12_07-17.18.05.133/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-7652521e-e335-40e6-8416-914c425804a11765271690051-2025_12_09-10.15.15.804/source.csv +0 -0
- b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-772b9e7f-7206-4184-bf66-5ac053c11d621761648011932-2025_10_28-11.40.35.184/source.csv +0 -0
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-0162b81a-7b28-4ef5-b0e9-c56355c44e161765287461281-2025_12_09-14.41.45.741/source.csv
ADDED
|
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,4,"scripts/run-qwen3-0.6B-base-sft.sh",0,0,"#!/bin/bash\n\n# for rerun the task\npkill -9 sglang\nsleep 3\nray stop --force\npkill -9 ray\npkill -9 python\nsleep 3\npkill -9 ray\npkill -9 python\n\nset -ex\n\n# will prevent ray from buffering stdout/stderr\nexport PYTHONBUFFERED=16\n\nNVLINK_COUNT=$(nvidia-smi topo -m 2>/dev/null | grep -o 'NV[0-9][0-9]*' | wc -l)\nif [ ""$NVLINK_COUNT"" -gt 0 ]; then\n HAS_NVLINK=1\nelse\n HAS_NVLINK=0\nfi\necho ""HAS_NVLINK: $HAS_NVLINK (detected $NVLINK_COUNT NVLink references)""\n\nSCRIPT_DIR=""$(cd -- ""$(dirname -- ""${BASH_SOURCE[0]}"")"" &>/dev/null && pwd)""\necho ""SCRIPT_DIR: $SCRIPT_DIR""\nsource ""${SCRIPT_DIR}/models/qwen3-0.6B.sh""\n\nCKPT_ARGS=(\n --hf-checkpoint /fast/project/HFMI_SynergyUnit/tab_model/huggingface/Qwen3-0.6B/\n --ref-load /fast/project/HFMI_SynergyUnit/tab_model/huggingface/Qwen3-0.6B_torch_dist\n --load /fast/project/HFMI_SynergyUnit/tab_model/huggingface/Qwen3-0.6B_miles/\n --save /fast/project/HFMI_SynergyUnit/tab_model/huggingface/Qwen3-0.6B_miles/\n --save-interval 1000\n)\n\nSFT_ARGS=(\n --rollout-function-path miles.rollout.sft_rollout.generate_rollout\n --prompt-data /fast/project/HFMI_SynergyUnit/tab_model/huggingface/openhermes2_5.parquet\n --apply-chat-template\n --input-key messages\n --rollout-shuffle\n --num-epoch 3\n --rollout-batch-size 1\n --global-batch-size 1\n\n --loss-type sft_loss\n --calculate-per-token-loss\n --disable-compute-advantages-and-returns\n --debug-train-only\n)\n\nPERF_ARGS=(\n --tensor-model-parallel-size 1\n --sequence-parallel\n --pipeline-model-parallel-size 1\n --context-parallel-size 1\n --expert-model-parallel-size 1\n --expert-tensor-parallel-size 1\n\n --recompute-granularity full\n --recompute-method uniform\n --recompute-num-layers 1\n\n # --micro-batch-size 1\n --use-dynamic-batch-size\n --max-tokens-per-gpu 9216\n)\n\nOPTIMIZER_ARGS=(\n --optimizer adam\n --lr 1e-5\n --lr-decay-style cosine\n --min-lr 1e-6\n --lr-warmup-fraction 0.1\n --weight-decay 0.1\n --adam-beta1 0.9\n --adam-beta2 0.95\n)\n\nWANDB_ARGS=(\n # --use-wandb\n # --wandb-project miles-dev\n # --wandb-group qwen3-4B-base-sft\n # --wandb-key ${WANDB_KEY}\n)\n\nMISC_ARGS=(\n # default dropout in megatron is 0.1\n --attention-dropout 0.0\n --hidden-dropout 0.0\n # should be good for model performance\n --accumulate-allreduce-grads-in-fp32\n --attention-softmax-in-fp32\n # need to comment this when using model with MLA\n --attention-backend flash\n)\n\n# launch the master node of ray in container\nexport MASTER_ADDR=${MASTER_ADDR:-""10.86.2.50""}\nexport no_proxy=""10.86.2.50,${MASTER_ADDR}""\nray start --head --node-ip-address ${MASTER_ADDR} --num-gpus 2 --disable-usage-stats --dashboard-host=0.0.0.0 --dashboard-port=8265\n\n\n# Build the runtime environment JSON with proper variable substitution\nRUNTIME_ENV_JSON=""{\n \""env_vars\"": {\n \""PYTHONPATH\"": \""/fast/project/HFMI_SynergyUnit/tab_model/Megatron-LM/\"",\n \""CUDA_DEVICE_MAX_CONNECTIONS\"": \""1\"",\n \""NCCL_NVLS_ENABLE\"": \""${HAS_NVLINK}\"",\n \""PYTORCH_CUDA_ALLOC_CONF\"": \""expandable_segments:True\""\n }\n}""\n\nray job submit --address=""http://10.86.2.50:8265"" \\n --runtime-env-json=""${RUNTIME_ENV_JSON}"" \\n -- python3 train_async.py \\n --actor-num-nodes 1 \\n --actor-num-gpus-per-node 2 \\n ${MODEL_ARGS[@]} \\n ${CKPT_ARGS[@]} \\n ${SFT_ARGS[@]} \\n ${OPTIMIZER_ARGS[@]} \\n ${WANDB_ARGS[@]} \\n ${PERF_ARGS[@]} \\n ${EVAL_ARGS[@]} \\n ${MISC_ARGS[@]}\n",shellscript,tab
|
| 3 |
+
2,28880,"scripts/run-qwen3-0.6B-base-sft.sh",1553,0,"",shellscript,selection_mouse
|
| 4 |
+
3,32556,"scripts/run-qwen3-0.6B-base-sft.sh",1552,0,"",shellscript,selection_command
|
| 5 |
+
4,54151,"TERMINAL",0,0,"",,terminal_focus
|
| 6 |
+
5,109805,"scripts/run-qwen3-0.6B-base-sft.sh",1623,0,"",shellscript,selection_mouse
|
| 7 |
+
6,144315,"TERMINAL",0,0,"queue",,terminal_command
|
| 8 |
+
7,145344,"TERMINAL",0,0,"]633;C[?1049h[22;0;0t[1;11r(B[m[4l[?7h[H[2JEvery 1.0s: squeue --me[1;64Hhai-login1.haicore.berlin: Tue Dec 9 14:43:59 2025[3;3HJOBID[15GUSER PARTITION NODES CPUS ST[3;53HSUBMIT_TIME[3;74HSTART_TIME TIME TIME_LIMIT NODELIST(R[4;1HEASON)[11;114H",,terminal_output
|
| 9 |
+
8,146052,"TERMINAL",0,0,"[1;106H4:00[11d\t ",,terminal_output
|
| 10 |
+
9,147444,"TERMINAL",0,0,"[1;109H1[11d\t ",,terminal_output
|
| 11 |
+
10,148382,"TERMINAL",0,0,"[1;109H2[11d\t ",,terminal_output
|
| 12 |
+
11,148682,"TERMINAL",0,0,"[1;109H3[11d\t ",,terminal_output
|
| 13 |
+
12,149104,"TERMINAL",0,0,"[1;109H4[11d\t [1;109H5[11d\t [1;109H6[11d\t [1;109H7[11d\t [H[2JEvery 1.0s: squeue --me[1;64Hhai-login1.haicore.berlin: Tue Dec 9 14:44:07 2025[3;3HJOBID[15GUSER PARTITION NODES CPUS ST[3;53HSUBMIT_TIME[3;74HSTART_TIME TIME TIME_LIMIT NODELIST(R[4;1HEASON)[31;114H[1;109H8[31d\t ",,terminal_output
|
| 14 |
+
13,149394,"TERMINAL",0,0,"[1;109H9[31d\t [1;108H10[31d\t [1;109H1[31d\t [1;109H2[31d\t ",,terminal_output
|
| 15 |
+
14,149474,"TERMINAL",0,0,"[1;109H3[31d\t ",,terminal_output
|
| 16 |
+
15,149922,"TERMINAL",0,0,"[1;109H4[31d\t ",,terminal_output
|
| 17 |
+
16,165760,"TERMINAL",0,0,"[1;109H5[31d\t ",,terminal_output
|
| 18 |
+
17,165891,"TERMINAL",0,0,"[1;109H6[31d\t [1;109H7[31d\t ",,terminal_output
|
| 19 |
+
18,166638,"TERMINAL",0,0,"[1;109H8[31d\t ",,terminal_output
|
| 20 |
+
19,166802,"TERMINAL",0,0,"[1;109H9[31d\t [1;108H20[31d\t ",,terminal_output
|
| 21 |
+
20,166978,"TERMINAL",0,0,"[1;109H1[31d\t ",,terminal_output
|
| 22 |
+
21,176592,"TERMINAL",0,0,"[1;109H2[31d\t [1;109H3[31d\t [1;109H4[31d\t [1;109H5[31d\t [1;109H7[31d\t [1;109H8[31d\t [1;109H9[31d\t [1;108H30[31d\t [31;1H[?1049l[23;0;0t\r[?1l>]0;mihir.mahajan@hai-login1:~/Projects/miles",,terminal_output
|
| 23 |
+
22,372641,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"2:41:45 PM [info] Activating crowd-code\n2:41:45 PM [info] Recording started\n2:41:45 PM [info] Initializing git provider using file system watchers...\n2:41:47 PM [info] Retrying git provider initialization...\n2:42:21 PM [info] Git repository found\n2:42:21 PM [info] Git provider initialized successfully\n2:42:21 PM [info] Initial git state: [object Object]\n2:42:29 PM [info] Git repository found\n2:42:29 PM [info] Git provider initialized successfully\n2:42:29 PM [info] Initial git state: [object Object]\n",Log,tab
|
| 24 |
+
23,445141,"scripts/run-qwen3-0.6B-base-sft.sh",0,0,"",shellscript,tab
|
| 25 |
+
24,453380,"TERMINAL",0,0,"source /home/mihir.mahajan/Projects/miles/.venv/bin/activate",,terminal_command
|
| 26 |
+
25,454711,"TERMINAL",0,0,"]633;C]0;mihir.mahajan@hai-login1:~/Projects/miles",,terminal_output
|
| 27 |
+
26,558063,"scripts/run-qwen3-0.6B-base-sft.sh",1524,0,"",shellscript,selection_mouse
|
| 28 |
+
27,562108,"scripts/run-qwen3-0.6B-base-sft.sh",1523,0,"",shellscript,selection_command
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-048f6756-c59b-4be5-887a-8f28003e7da91761121031395-2025_10_22-10.17.37.67/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-04d765c9-bb6f-4faa-b6ea-8904418254d61760945425208-2025_10_20-09.30.54.363/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-0811166c-19e5-47a9-bbfb-5deb433c29821760527863000-2025_10_15-13.31.33.167/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-090f467e-141d-440a-88f0-fb9757f142d31761307268455-2025_10_24-14.01.33.483/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-09e95c4f-d571-45a4-a506-a80e6e398a3f1761597703208-2025_10_27-21.42.06.986/source.csv
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,3,"/fast/home/mihir.mahajan/Projects/jasmine/jasmine/baselines/diffusion/train_tokenizer_mae.py",0,0,"import os\n\nos.environ.setdefault(""XLA_PYTHON_CLIENT_MEM_FRACTION"", ""0.98"")\n\nfrom dataclasses import dataclass, field\nfrom typing import cast, Optional\n\nimport einops\nimport itertools\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom jasmine.models.tokenizer import TokenizerMAE\nfrom jasmine.utils.dataloader import get_dataloader\nfrom jasmine.utils.train_utils import (\n get_lr_schedule,\n count_parameters_by_component,\n print_mem_stats,\n print_compiled_memory_stats,\n print_compiled_cost_analysis,\n)\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 300_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 64\n image_width: int = 64\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 48\n init_lr: float = 0.0\n max_lr: float = 3e-4\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 30_000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n warmup_steps: int = 10000\n # Tokenizer\n model_dim: int = 512\n ffn_dim: int = 2048\n latent_dim: int = 32\n num_latents: int = 1024\n patch_size: int = 16\n num_blocks: int = 4\n num_heads: int = 8\n dropout: float = 0.0\n max_mask_ratio: float = 0.9\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n # Logging\n log: bool = True\n entity: str = """"\n project: str = """"\n name: str = ""train_tokenizer_mae""\n tags: list[str] = field(default_factory=lambda: [""tokenizer"", ""mae""])\n log_interval: int = 50\n log_image_interval: int = 1000\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 1000\n log_checkpoint_keep_period: int = 20_000\n log_gradients: bool = False\n val_data_dir: str = """"\n val_interval: int = 20_000\n val_steps: int = 50\n wandb_id: str = """"\n\n\ndef build_model(args: Args, rng: jax.Array) -> tuple[TokenizerMAE, jax.Array]:\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n tokenizer = TokenizerMAE(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n ffn_dim=args.ffn_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n max_mask_ratio=args.max_mask_ratio,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n )\n return tokenizer, rng\n\n\ndef build_optimizer(model: TokenizerMAE, args: Args) -> nnx.ModelAndOptimizer:\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.param_dtype, # moments in full precision\n )\n optimizer = nnx.ModelAndOptimizer(model, tx)\n return optimizer\n\n\ndef build_mesh_and_sharding(\n num_devices: int,\n) -> tuple[Mesh, NamedSharding, NamedSharding]:\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n return mesh, replicated_sharding, videos_sharding\n\n\ndef shard_optimizer_states(\n optimizer: nnx.ModelAndOptimizer, replicated_sharding: NamedSharding\n) -> None:\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n\ndef build_dataloader(args: Args, data_dir: str) -> grain.DataLoaderIterator:\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(data_dir, x)\n for x in os.listdir(data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n return grain_iterator\n\n\ndef build_checkpoint_manager(args: Args) -> Optional[ocp.CheckpointManager]:\n if args.restore_ckpt or args.save_ckpt:\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n if args.val_data_dir:\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(\n ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler\n ),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(\n ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler\n ),\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n return checkpoint_manager\n else:\n return None\n\n\ndef restore_checkpoint_if_needed(\n args: Args,\n checkpoint_manager: Optional[ocp.CheckpointManager],\n optimizer: nnx.ModelAndOptimizer,\n train_iterator: grain.DataLoaderIterator,\n val_iterator: Optional[grain.DataLoaderIterator],\n restore_step: Optional[int] = None,\n) -> tuple[\n int, nnx.ModelAndOptimizer, grain.DataLoaderIterator, grain.DataLoaderIterator\n]:\n step = 0\n if checkpoint_manager and restore_step is None:\n restore_step = checkpoint_manager.latest_step()\n if args.restore_ckpt:\n assert checkpoint_manager is not None\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n if val_iterator:\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n val_dataloader_state=grain.checkpoint.CheckpointRestore(val_iterator), # type: ignore\n )\n else:\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n )\n restored = checkpoint_manager.restore(restore_step, args=restore_args)\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n train_iterator = restored[""train_dataloader_state""]\n if val_iterator:\n val_iterator = restored[""val_dataloader_state""]\n step = restore_step or 0\n print(f""Restored dataloader and model state from step {step}"")\n return step, optimizer, train_iterator, val_iterator\n\n\ndef main(args: Args) -> None:\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n tokenizer, rng = build_model(args, rng)\n\n _, params, _ = nnx.split(tokenizer, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n optimizer = build_optimizer(tokenizer, args)\n del tokenizer\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n _, replicated_sharding, videos_sharding = build_mesh_and_sharding(num_devices)\n\n shard_optimizer_states(optimizer, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n checkpoint_manager = build_checkpoint_manager(args)\n\n # --- Create DataLoaderIterator from dataloader ---\n train_iterator = build_dataloader(args, args.data_dir)\n val_iterator = None\n if args.val_data_dir:\n val_iterator = build_dataloader(args, args.val_data_dir)\n\n # --- Restore checkpoint ---\n step, optimizer, train_iterator, val_iterator = restore_checkpoint_if_needed(\n args, checkpoint_manager, optimizer, train_iterator, val_iterator\n )\n\n # --- Define loss and train step (close over args) ---\n def tokenizer_loss_fn(\n model: TokenizerMAE, inputs: dict, training: bool = False\n ) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n mse = jnp.square(gt - outputs[""recon""]).mean()\n\n gt_clipped = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt_clipped, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt_clipped, recon)).mean()\n\n metrics = dict(\n mse=mse,\n psnr=psnr,\n ssim=ssim,\n loss=mse,\n )\n\n return mse, (outputs[""recon""], metrics)\n\n @nnx.jit(donate_argnums=0)\n def train_step(\n optimizer: nnx.ModelAndOptimizer, inputs: dict\n ) -> tuple[jax.Array, jax.Array, dict]:\n def loss_fn(\n model: TokenizerMAE,\n ) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n model.train()\n return tokenizer_loss_fn(model, inputs, training=True)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\n optimizer.model\n )\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""encoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""encoder""]\n )\n metrics[""decoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""decoder""]\n )\n return loss, recon, metrics\n\n @nnx.jit\n def val_step(\n tokenizer: TokenizerMAE, inputs: dict\n ) -> tuple[jax.Array, jax.Array, dict]:\n tokenizer.eval()\n (loss, (recon, metrics)) = tokenizer_loss_fn(tokenizer, inputs, training=False)\n return loss, recon, metrics\n\n def calculate_validation_metrics(val_dataloader, tokenizer, rng):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n batch = None\n recon = None\n for batch in val_dataloader:\n rng, _rng_mask = jax.random.split(rng, 2)\n batch[""rng""] = _rng_mask\n loss, recon, metrics = val_step(tokenizer, batch)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(\n f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}""\n )\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([float(m[key]) for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n val_metrics[""val_loss""] = val_loss\n return val_metrics, batch, recon\n\n # --- TRAIN LOOP ---\n dataloader_train = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n }\n for elem in train_iterator\n )\n dataloader_val = None\n if val_iterator:\n dataloader_val = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n }\n for elem in val_iterator\n )\n if jax.process_index() == 0:\n rng, _rng = jax.random.split(rng)\n first_batch = next(dataloader_train)\n first_batch[""rng""] = _rng\n compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training\n dataloader_train = itertools.chain([first_batch], dataloader_train)\n print(f""Starting training from step {step}..."")\n first_step = step\n while step < args.num_steps:\n for batch in dataloader_train:\n # --- Train step ---\n rng, _rng = jax.random.split(rng)\n batch[""rng""] = _rng\n loss, recon, metrics = train_step(optimizer, batch)\n if step == first_step:\n print_mem_stats(""After params initialized"")\n step += 1\n\n # --- Validation loss ---\n val_results = {}\n if dataloader_val and step % args.val_interval == 0:\n print(""Calculating validation metrics..."")\n rng, _rng_mask_val = jax.random.split(rng, 2)\n val_metrics, val_gt_batch, val_recon = calculate_validation_metrics(\n dataloader_val, optimizer.model, _rng_mask_val\n )\n print(f""Step {step}, validation loss: {val_metrics['val_loss']}"")\n val_results = {\n ""metrics"": val_metrics,\n ""gt_batch"": val_gt_batch,\n ""recon"": val_recon,\n }\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n log_dict = {""loss"": loss, ""step"": step, **metrics}\n if val_results:\n log_dict.update(val_results[""metrics""])\n wandb.log(log_dict)\n if step % args.log_image_interval == 0:\n gt_seq = batch[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if val_results and step % args.val_interval == 0:\n val_results[""gt_seq_val""] = (\n val_results[""gt_batch""][""videos""][0].astype(jnp.float32)\n / 255.0\n )\n val_results[""recon_seq_val""] = val_results[""recon""][0].clip(\n 0, 1\n )\n val_results[""val_comparison_seq""] = jnp.concatenate(\n (val_results[""gt_seq_val""], val_results[""recon_seq_val""]),\n axis=1,\n )\n val_results[""val_comparison_seq""] = einops.rearrange(\n val_results[""val_comparison_seq""] * 255,\n ""t h w c -> h (t w) c"",\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n if val_results and step % args.val_interval == 0:\n log_images.update(\n dict(\n val_image=wandb.Image(\n np.asarray(val_results[""gt_seq_val""][0])\n ),\n val_recon=wandb.Image(\n np.asarray(val_results[""recon_seq_val""][0])\n ),\n val_true_vs_recon=wandb.Image(\n np.asarray(\n val_results[""val_comparison_seq""].astype(\n np.uint8\n )\n )\n ),\n )\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n assert checkpoint_manager is not None\n optimizer_state = nnx.state(optimizer)\n if val_iterator:\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n ),\n val_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n val_iterator # type: ignore\n ),\n )\n else:\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n ),\n )\n checkpoint_manager.save(step, args=ckpt_manager_args)\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n if checkpoint_manager:\n checkpoint_manager.close()\n\n\nif __name__ == ""__main__"":\n args = tyro.cli(Args)\n main(args)\n",python,tab
|
| 3 |
+
2,924197,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"9:42:06 PM [info] Activating crowd-code\n9:42:06 PM [info] Recording started\n9:42:06 PM [info] Initializing git provider using file system watchers...\n9:42:07 PM [info] Git repository found\n9:42:07 PM [info] Git provider initialized successfully\n9:42:07 PM [info] Initial git state: [object Object]\n",Log,tab
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-0ebb31ee-57f5-41bb-b048-9c99802b107c1764329048669-2025_11_28-12.24.33.567/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-0f122a13-52f0-458a-b70b-8adddf08d7fc1764349079194-2025_11_28-17.58.34.412/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-0f1ad218-d2d6-4ace-9150-2ff8d2109f6d1762606110433-2025_11_08-13.49.11.795/source.csv
ADDED
|
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,9721,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"1:49:11 PM [info] Activating crowd-code\n1:49:11 PM [info] Recording started\n1:49:11 PM [info] Initializing git provider using file system watchers...\n1:49:13 PM [info] Retrying git provider initialization...\n1:49:15 PM [info] Git repository found\n1:49:15 PM [info] Git provider initialized successfully\n1:49:15 PM [info] Initial git state: [object Object]\n1:49:15 PM [info] Git repository found\n1:49:15 PM [info] Git provider initialized successfully\n1:49:15 PM [info] Initial git state: [object Object]\n",Log,tab
|
| 3 |
+
3,56495,"TERMINAL",0,0,"",,terminal_focus
|
| 4 |
+
4,96627,"TERMINAL",0,0,"",,terminal_focus
|
| 5 |
+
5,167415,"TERMINAL",0,0,"git clone https://github.com/p-doom/slurm",,terminal_command
|
| 6 |
+
6,167477,"TERMINAL",0,0,"]633;CCloning into 'slurm'...\r\n",,terminal_output
|
| 7 |
+
7,172593,"TERMINAL",0,0,"^[^[",,terminal_output
|
| 8 |
+
8,175222,"TERMINAL",0,0,"remote: Repository not found.\r\nfatal: Authentication failed for 'https://github.com/p-doom/slurm/'\r\n]0;mihir.mahajan@hai-login2:~/Projects/crowd-pilot",,terminal_output
|
| 9 |
+
9,192444,"TERMINAL",0,0,"git clone git@github.com:p-doom/slurm.git",,terminal_command
|
| 10 |
+
10,192494,"TERMINAL",0,0,"]633;CCloning into 'slurm'...\r\n",,terminal_output
|
| 11 |
+
11,194019,"TERMINAL",0,0,"remote: Enumerating objects: 5247, done.[K\r\nremote: Counting objects: 0% (1/443)[K\rremote: Counting objects: 1% (5/443)[K\rremote: Counting objects: 2% (9/443)[K\rremote: Counting objects: 3% (14/443)[K\rremote: Counting objects: 4% (18/443)[K\rremote: Counting objects: 5% (23/443)[K\rremote: Counting objects: 6% (27/443)[K\rremote: Counting objects: 7% (32/443)[K\rremote: Counting objects: 8% (36/443)[K\rremote: Counting objects: 9% (40/443)[K\rremote: Counting objects: 10% (45/443)[K\rremote: Counting objects: 11% (49/443)[K\rremote: Counting objects: 12% (54/443)[K\rremote: Counting objects: 13% (58/443)[K\rremote: Counting objects: 14% (63/443)[K\rremote: Counting objects: 15% (67/443)[K\rremote: Counting objects: 16% (71/443)[K\rremote: Counting objects: 17% (76/443)[K\rremote: Counting objects: 18% (80/443)[K\rremote: Counting objects: 19% (85/443)[K\rremote: Counting objects: 20% (89/443)[K\rremote: Counting objects: 21% (94/443)[K\rremote: Counting objects: 22% (98/443)[K\rremote: Counting objects: 23% (102/443)[K\rremote: Counting objects: 24% (107/443)[K\rremote: Counting objects: 25% (111/443)[K\rremote: Counting objects: 26% (116/443)[K\rremote: Counting objects: 27% (120/443)[K\rremote: Counting objects: 28% (125/443)[K\rremote: Counting objects: 29% (129/443)[K\rremote: Counting objects: 30% (133/443)[K\rremote: Counting objects: 31% (138/443)[K\rremote: Counting objects: 32% (142/443)[K\rremote: Counting objects: 33% (147/443)[K\rremote: Counting objects: 34% (151/443)[K\rremote: Counting objects: 35% (156/443)[K\rremote: Counting objects: 36% (160/443)[K\rremote: Counting objects: 37% (164/443)[K\rremote: Counting objects: 38% (169/443)[K\rremote: Counting objects: 39% (173/443)[K\r",,terminal_output
|
| 12 |
+
12,194213,"TERMINAL",0,0,"remote: Counting objects: 40% (178/443)[K\rremote: Counting objects: 41% (182/443)[K\rremote: Counting objects: 42% (187/443)[K\rremote: Counting objects: 43% (191/443)[K\rremote: Counting objects: 44% (195/443)[K\rremote: Counting objects: 45% (200/443)[K\rremote: Counting objects: 46% (204/443)[K\rremote: Counting objects: 47% (209/443)[K\rremote: Counting objects: 48% (213/443)[K\rremote: Counting objects: 49% (218/443)[K\rremote: Counting objects: 50% (222/443)[K\rremote: Counting objects: 51% (226/443)[K\rremote: Counting objects: 52% (231/443)[K\rremote: Counting objects: 53% (235/443)[K\rremote: Counting objects: 54% (240/443)[K\rremote: Counting objects: 55% (244/443)[K\rremote: Counting objects: 56% (249/443)[K\rremote: Counting objects: 57% (253/443)[K\rremote: Counting objects: 58% (257/443)[K\rremote: Counting objects: 59% (262/443)[K\rremote: Counting objects: 60% (266/443)[K\rremote: Counting objects: 61% (271/443)[K\rremote: Counting objects: 62% (275/443)[K\rremote: Counting objects: 63% (280/443)[K\rremote: Counting objects: 64% (284/443)[K\rremote: Counting objects: 65% (288/443)[K\rremote: Counting objects: 66% (293/443)[K\rremote: Counting objects: 67% (297/443)[K\rremote: Counting objects: 68% (302/443)[K\rremote: Counting objects: 69% (306/443)[K\rremote: Counting objects: 70% (311/443)[K\rremote: Counting objects: 71% (315/443)[K\rremote: Counting objects: 72% (319/443)[K\rremote: Counting objects: 73% (324/443)[K\rremote: Counting objects: 74% (328/443)[K\rremote: Counting objects: 75% (333/443)[K\rremote: Counting objects: 76% (337/443)[K\rremote: Counting objects: 77% (342/443)[K\rremote: Counting objects: 78% (346/443)[K\rremote: Counting objects: 79% (350/443)[K\rremote: Counting objects: 80% (355/443)[K\rremote: Counting objects: 81% (359/443)[K\rremote: Counting objects: 82% (364/443)[K\rremote: Counting objects: 83% (368/443)[K\rremote: Counting objects: 84% (373/443)[K\rremote: Counting objects: 85% (377/443)[K\rremote: Counting objects: 86% (381/443)[K\rremote: Counting objects: 87% (386/443)[K\rremote: Counting objects: 88% (390/443)[K\rremote: Counting objects: 89% (395/443)[K\rremote: Counting objects: 90% (399/443)[K\rremote: Counting objects: 91% (404/443)[K\rremote: Counting objects: 92% (408/443)[K\rremote: Counting objects: 93% (412/443)[K\rremote: Counting objects: 94% (417/443)[K\rremote: Counting objects: 95% (421/443)[K\rremote: Counting objects: 96% (426/443)[K\rremote: Counting objects: 97% (430/443)[K\rremote: Counting objects: 98% (435/443)[K\rremote: Counting objects: 99% (439/443)[K\rremote: Counting objects: 100% (443/443)[K\rremote: Counting objects: 100% (443/443), done.[K\r\nremote: Compressing objects: 0% (1/310)[K\rremote: Compressing objects: 1% (4/310)[K\rremote: Compressing objects: 2% (7/310)[K\rremote: Compressing objects: 3% (10/310)[K\rremote: Compressing objects: 4% (13/310)[K\rremote: Compressing objects: 5% (16/310)[K\rremote: Compressing objects: 6% (19/310)[K\rremote: Compressing objects: 7% (22/310)[K\rremote: Compressing objects: 8% (25/310)[K\rremote: Compressing objects: 9% (28/310)[K\rremote: Compressing objects: 10% (31/310)[K\rremote: Compressing objects: 11% (35/310)[K\rremote: Compressing objects: 12% (38/310)[K\rremote: Compressing objects: 13% (41/310)[K\rremote: Compressing objects: 14% (44/310)[K\rremote: Compressing objects: 15% (47/310)[K\rremote: Compressing objects: 16% (50/310)[K\rremote: Compressing objects: 17% (53/310)[K\rremote: Compressing objects: 18% (56/310)[K\rremote: Compressing objects: 19% (59/310)[K\rremote: Compressing objects: 20% (62/310)[K\rremote: Compressing objects: 21% (66/310)[K\rremote: Compressing objects: 22% (69/310)[K\rremote: Compressing objects: 23% (72/310)[K\rremote: Compressing objects: 24% (75/310)[K\rremote: Compressing objects: 25% (78/310)[K\rremote: Compressing objects: 26% (81/310)[K\rremote: Compressing objects: 27% (84/310)[K\rremote: Compressing objects: 28% (87/310)[K\rremote: Compressing objects: 29% (90/310)[K\rremote: Compressing objects: 30% (93/310)[K\rremote: Compressing objects: 31% (97/310)[K\rremote: Compressing objects: 32% (100/310)[K\rremote: Compressing objects: 33% (103/310)[K\rremote: Compressing objects: 34% (106/310)[K\rremote: Compressing objects: 35% (109/310)[K\rremote: Compressing objects: 36% (112/310)[K\rremote: Compressing objects: 37% (115/310)[K\rremote: Compressing objects: 38% (118/310)[K\rremote: Compressing objects: 39% (121/310)[K\rremote: Compressing objects: 40% (124/310)[K\rremote: Compressing objects: 41% (128/310)[K\rremote: Compressing objects: 42% (131/310)[K\rremote: Compressing objects: 43% (134/310)[K\rremote: Compressing objects: 44% (137/310)[K\rremote: Compressing objects: 45% (140/310)[K\rremote: Compressing objects: 46% (143/310)[K\rremote: Compressing objects: 47% (146/310)[K\rremote: Compressing objects: 48% (149/310)[K\rremote: Compressing objects: 49% (152/310)[K\rremote: Compressing objects: 50% (155/310)[K\rremote: Compressing objects: 51% (159/310)[K\rremote: Compressing objects: 52% (162/310)[K\rremote: Compressing objects: 53% (165/310)[K\rremote: Compressing objects: 54% (168/310)[K\rremote: Compressing objects: 55% (171/310)[K\rremote: Compressing objects: 56% (174/310)[K\rremote: Compressing objects: 57% (177/310)[K\rremote: Compressing objects: 58% (180/310)[K\rremote: Compressing objects: 59% (183/310)[K\rremote: Compressing objects: 60% (186/310)[K\rremote: Compressing objects: 61% (190/310)[K\rremote: Compressing objects: 62% (193/310)[K\rremote: Compressing objects: 63% (196/310)[K\rremote: Compressing objects: 64% (199/310)[K\rremote: Compressing objects: 65% (202/310)[K\rremote: Compressing objects: 66% (205/310)[K\rremote: Compressing objects: 67% (208/310)[K\rremote: Compressing objects: 68% (211/310)[K\rremote: Compressing objects: 69% (214/310)[K\rremote: Compressing objects: 70% (217/310)[K\rremote: Compressing objects: 71% (221/310)[K\rremote: Compressing objects: 72% (224/310)[K\rremote: Compressing objects: 73% (227/310)[K\rremote: Compressing objects: 74% (230/310)[K\rremote: Compressing objects: 75% (233/310)[K\rremote: Compressing objects: 76% (236/310)[K\rremote: Compressing objects: 77% (239/310)[K\rremote: Compressing objects: 78% (242/310)[K\rremote: Compressing objects: 79% (245/310)[K\rremote: Compressing objects: 80% (248/310)[K\rremote: Compressing objects: 81% (252/310)[K\rremote: Compressing objects: 82% (255/310)[K\rremote: Compressing objects: 83% (258/310)[K\rremote: Compressing objects: 84% (261/310)[K\rremote: Compressing objects: 85% (264/310)[K\rremote: Compressing objects: 86% (267/310)[K\rremote: Compressing objects: 87% (270/310)[K\rremote: Compressing objects: 88% (273/310)[K\rremote: Compressing objects: 89% (276/310)[K\rremote: Compressing objects: 90% (279/310)[K\rremote: Compressing objects: 91% (283/310)[K\rremote: Compressing objects: 92% (286/310)[K\rremote: Compressing objects: 93% (289/310)[K\rremote: Compressing objects: 94% (292/310)[K\rremote: Compressing objects: 95% (295/310)[K\rremote: Compressing objects: 96% (298/310)[K\rremote: Compressing objects: 97% (301/310)[K\rremote: Compressing objects: 98% (304/310)[K\rremote: Compressing objects: 99% (307/310)[K\rremote: Compressing objects: 100% (310/310)[K\rremote: Compressing objects: 100% (310/310), done.[K\r\nReceiving objects: 0% (1/5247)\rReceiving objects: 1% (53/5247)\rReceiving objects: 2% (105/5247)\rReceiving objects: 3% (158/5247)\rReceiving objects: 4% (210/5247)\rReceiving objects: 5% (263/5247)\rReceiving objects: 6% (315/5247)\rReceiving objects: 7% (368/5247)\rReceiving objects: 8% (420/5247)\r",,terminal_output
|
| 13 |
+
13,194432,"TERMINAL",0,0,"Receiving objects: 9% (473/5247)\rReceiving objects: 10% (525/5247)\rReceiving objects: 11% (578/5247)\rReceiving objects: 12% (630/5247)\rReceiving objects: 13% (683/5247)\rReceiving objects: 14% (735/5247)\rReceiving objects: 15% (788/5247)\rReceiving objects: 16% (840/5247)\rReceiving objects: 17% (892/5247)\rReceiving objects: 18% (945/5247)\rReceiving objects: 19% (997/5247)\rReceiving objects: 20% (1050/5247)\rReceiving objects: 21% (1102/5247)\rReceiving objects: 22% (1155/5247)\rReceiving objects: 23% (1207/5247)\rReceiving objects: 24% (1260/5247)\rReceiving objects: 25% (1312/5247)\rReceiving objects: 26% (1365/5247)\rReceiving objects: 27% (1417/5247)\rReceiving objects: 28% (1470/5247)\rReceiving objects: 29% (1522/5247)\rReceiving objects: 30% (1575/5247)\rReceiving objects: 31% (1627/5247)\rReceiving objects: 32% (1680/5247)\rReceiving objects: 33% (1732/5247)\rReceiving objects: 34% (1784/5247)\rReceiving objects: 35% (1837/5247)\rReceiving objects: 36% (1889/5247)\rReceiving objects: 37% (1942/5247)\rReceiving objects: 38% (1994/5247)\rReceiving objects: 39% (2047/5247)\rReceiving objects: 40% (2099/5247)\rReceiving objects: 41% (2152/5247)\rReceiving objects: 42% (2204/5247)\rReceiving objects: 43% (2257/5247)\rReceiving objects: 44% (2309/5247)\r",,terminal_output
|
| 14 |
+
14,194711,"TERMINAL",0,0,"Receiving objects: 45% (2362/5247)\rReceiving objects: 46% (2414/5247)\rReceiving objects: 47% (2467/5247)\rReceiving objects: 48% (2519/5247)\rReceiving objects: 49% (2572/5247)\rReceiving objects: 50% (2624/5247)\rReceiving objects: 51% (2676/5247)\rReceiving objects: 52% (2729/5247)\rReceiving objects: 53% (2781/5247)\rReceiving objects: 54% (2834/5247)\rReceiving objects: 55% (2886/5247)\rReceiving objects: 56% (2939/5247)\rReceiving objects: 57% (2991/5247)\rReceiving objects: 58% (3044/5247)\rReceiving objects: 59% (3096/5247)\rReceiving objects: 60% (3149/5247)\rReceiving objects: 61% (3201/5247)\rReceiving objects: 62% (3254/5247)\rReceiving objects: 63% (3306/5247)\rReceiving objects: 64% (3359/5247)\rReceiving objects: 65% (3411/5247)\rReceiving objects: 66% (3464/5247)\rReceiving objects: 67% (3516/5247)\rReceiving objects: 68% (3568/5247)\rReceiving objects: 69% (3621/5247)\rReceiving objects: 70% (3673/5247)\rReceiving objects: 71% (3726/5247)\rReceiving objects: 72% (3778/5247)\rReceiving objects: 73% (3831/5247)\rReceiving objects: 74% (3883/5247)\rReceiving objects: 75% (3936/5247)\rReceiving objects: 76% (3988/5247)\rReceiving objects: 77% (4041/5247)\rReceiving objects: 78% (4093/5247)\rReceiving objects: 79% (4146/5247)\rReceiving objects: 80% (4198/5247)\rReceiving objects: 81% (4251/5247)\rReceiving objects: 82% (4303/5247)\rReceiving objects: 83% (4356/5247)\rReceiving objects: 84% (4408/5247)\rReceiving objects: 85% (4460/5247)\rReceiving objects: 86% (4513/5247)\rReceiving objects: 87% (4565/5247)\rReceiving objects: 88% (4618/5247)\rReceiving objects: 89% (4670/5247)\rReceiving objects: 90% (4723/5247)\rremote: Total 5247 (delta 217), reused 334 (delta 125), pack-reused 4804 (from 1)[K\r\nReceiving objects: 91% (4775/5247)\rReceiving objects: 92% (4828/5247)\rReceiving objects: 93% (4880/5247)\rReceiving objects: 94% (4933/5247)\rReceiving objects: 95% (4985/5247)\rReceiving objects: 96% (5038/5247)\rReceiving objects: 97% (5090/5247)\rReceiving objects: 98% (5143/5247)\rReceiving objects: 99% (5195/5247)\rReceiving objects: 100% (5247/5247)\rReceiving objects: 100% (5247/5247), 659.26 KiB | 1.30 MiB/s, done.\r\nResolving deltas: 0% (0/3291)\rResolving deltas: 1% (33/3291)\rResolving deltas: 2% (68/3291)\rResolving deltas: 3% (100/3291)\rResolving deltas: 4% (132/3291)\rResolving deltas: 5% (165/3291)\rResolving deltas: 6% (199/3291)\rResolving deltas: 7% (232/3291)\rResolving deltas: 8% (264/3291)\rResolving deltas: 9% (299/3291)\rResolving deltas: 10% (330/3291)\rResolving deltas: 11% (363/3291)\rResolving deltas: 12% (395/3291)\rResolving deltas: 13% (429/3291)\rResolving deltas: 14% (461/3291)\rResolving deltas: 15% (494/3291)\rResolving deltas: 16% (528/3291)\rResolving deltas: 17% (560/3291)\rResolving deltas: 18% (594/3291)\rResolving deltas: 19% (626/3291)\rResolving deltas: 20% (661/3291)\rResolving deltas: 21% (693/3291)\rResolving deltas: 22% (726/3291)\rResolving deltas: 23% (757/3291)\rResolving deltas: 24% (790/3291)\rResolving deltas: 25% (823/3291)\rResolving deltas: 26% (856/3291)\rResolving deltas: 27% (889/3291)\rResolving deltas: 28% (922/3291)\rResolving deltas: 29% (955/3291)\rResolving deltas: 30% (988/3291)\rResolving deltas: 31% (1021/3291)\rResolving deltas: 32% (1054/3291)\rResolving deltas: 33% (1087/3291)\rResolving deltas: 34% (1120/3291)\rResolving deltas: 35% (1152/3291)\rResolving deltas: 36% (1185/3291)\rResolving deltas: 37% (1218/3291)\rResolving deltas: 38% (1251/3291)\rResolving deltas: 39% (1284/3291)\rResolving deltas: 40% (1319/3291)\rResolving deltas: 41% (1350/3291)\rResolving deltas: 42% (1383/3291)\rResolving deltas: 43% (1417/3291)\rResolving deltas: 44% (1450/3291)\rResolving deltas: 45% (1481/3291)\rResolving deltas: 46% (1515/3291)\rResolving deltas: 47% (1547/3291)\rResolving deltas: 48% (1580/3291)\rResolving deltas: 49% (1613/3291)\rResolving deltas: 50% (1646/3291)\rResolving deltas: 51% (1679/3291)\rResolving deltas: 52% (1713/3291)\rResolving deltas: 53% (1745/3291)\rResolving deltas: 54% (1778/3291)\rResolving deltas: 55% (1811/3291)\rResolving deltas: 56% (1843/3291)\rResolving deltas: 57% (1876/3291)\rResolving deltas: 58% (1909/3291)\rResolving deltas: 59% (1942/3291)\rResolving deltas: 60% (1975/3291)\rResolving deltas: 61% (2008/3291)\rResolving deltas: 62% (2041/3291)\rResolving deltas: 63% (2075/3291)\rResolving deltas: 64% (2107/3291)\rResolving deltas: 65% (2140/3291)\rResolving deltas: 66% (2173/3291)\rResolving deltas: 67% (2205/3291)\rResolving deltas: 68% (2238/3291)\rResolving deltas: 69% (2272/3291)\rResolving deltas: 70% (2304/3291)\rResolving deltas: 71% (2337/3291)\rResolving deltas: 72% (2370/3291)\rResolving deltas: 73% (2403/3291)\rResolving deltas: 74% (2436/3291)\rResolving deltas: 75% (2469/3291)\rResolving deltas: 76% (2502/3291)\rResolving deltas: 77% (2535/3291)\rResolving deltas: 78% (2567/3291)\rResolving deltas: 79% (2600/3291)\rResolving deltas: 80% (2636/3291)\rResolving deltas: 81% (2666/3291)\rResolving deltas: 82% (2699/3291)\rResolving deltas: 83% (2732/3291)\rResolving deltas: 84% (2765/3291)\rResolving deltas: 85% (2798/3291)\rResolving deltas: 86% (2831/3291)\rResolving deltas: 87% (2864/3291)\rResolving deltas: 88% (2897/3291)\rResolving deltas: 89% (2929/3291)\rResolving deltas: 90% (2963/3291)\rResolving deltas: 91% (2995/3291)\rResolving deltas: 92% (3031/3291)\rResolving deltas: 93% (3061/3291)\rResolving deltas: 94% (3095/3291)\rResolving deltas: 95% (3127/3291)\rResolving deltas: 96% (3160/3291)\rResolving deltas: 97% (3194/3291)\rResolving deltas: 98% (3227/3291)\rResolving deltas: 99% (3259/3291)\rResolving deltas: 100% (3291/3291)\rResolving deltas: 100% (3291/3291), done.\r\n",,terminal_output
|
| 15 |
+
15,195903,"TERMINAL",0,0,"Updating files: 30% (408/1342)\rUpdating files: 31% (417/1342)\rUpdating files: 32% (430/1342)\r",,terminal_output
|
| 16 |
+
16,196274,"TERMINAL",0,0,"Updating files: 33% (443/1342)\rUpdating files: 34% (457/1342)\rUpdating files: 35% (470/1342)\rUpdating files: 36% (484/1342)\rUpdating files: 37% (497/1342)\rUpdating files: 38% (510/1342)\rUpdating files: 39% (524/1342)\rUpdating files: 40% (537/1342)\r",,terminal_output
|
| 17 |
+
17,196440,"TERMINAL",0,0,"Updating files: 41% (551/1342)\rUpdating files: 42% (564/1342)\rUpdating files: 43% (578/1342)\rUpdating files: 44% (591/1342)\r",,terminal_output
|
| 18 |
+
18,196716,"TERMINAL",0,0,"Updating files: 45% (604/1342)\rUpdating files: 46% (618/1342)\rUpdating files: 47% (631/1342)\rUpdating files: 48% (645/1342)\rUpdating files: 49% (658/1342)\rUpdating files: 50% (671/1342)\rUpdating files: 51% (685/1342)\rUpdating files: 52% (698/1342)\rUpdating files: 53% (712/1342)\rUpdating files: 54% (725/1342)\rUpdating files: 55% (739/1342)\rUpdating files: 56% (752/1342)\rUpdating files: 57% (765/1342)\rUpdating files: 58% (779/1342)\rUpdating files: 59% (792/1342)\rUpdating files: 60% (806/1342)\rUpdating files: 61% (819/1342)\rUpdating files: 62% (833/1342)\rUpdating files: 63% (846/1342)\r",,terminal_output
|
| 19 |
+
19,197013,"TERMINAL",0,0,"Updating files: 64% (859/1342)\rUpdating files: 65% (873/1342)\rUpdating files: 65% (874/1342)\r",,terminal_output
|
| 20 |
+
20,197234,"TERMINAL",0,0,"Updating files: 66% (886/1342)\rUpdating files: 67% (900/1342)\rUpdating files: 68% (913/1342)\rUpdating files: 69% (926/1342)\rUpdating files: 70% (940/1342)\rUpdating files: 71% (953/1342)\rUpdating files: 72% (967/1342)\rUpdating files: 73% (980/1342)\rUpdating files: 74% (994/1342)\rUpdating files: 75% (1007/1342)\rUpdating files: 76% (1020/1342)\rUpdating files: 77% (1034/1342)\rUpdating files: 78% (1047/1342)\rUpdating files: 79% (1061/1342)\rUpdating files: 80% (1074/1342)\rUpdating files: 81% (1088/1342)\rUpdating files: 82% (1101/1342)\r",,terminal_output
|
| 21 |
+
21,197682,"TERMINAL",0,0,"Updating files: 83% (1114/1342)\rUpdating files: 84% (1128/1342)\rUpdating files: 85% (1141/1342)\rUpdating files: 86% (1155/1342)\rUpdating files: 87% (1168/1342)\rUpdating files: 88% (1181/1342)\rUpdating files: 89% (1195/1342)\rUpdating files: 90% (1208/1342)\rUpdating files: 91% (1222/1342)\rUpdating files: 92% (1235/1342)\rUpdating files: 93% (1249/1342)\rUpdating files: 94% (1262/1342)\rUpdating files: 95% (1275/1342)\rUpdating files: 96% (1289/1342)\rUpdating files: 97% (1302/1342)\rUpdating files: 98% (1316/1342)\rUpdating files: 99% (1329/1342)\rUpdating files: 100% (1342/1342)\rUpdating files: 100% (1342/1342), done.\r\n]0;mihir.mahajan@hai-login2:~/Projects/crowd-pilot",,terminal_output
|
| 22 |
+
22,219692,"TERMINAL",0,0,"cat slurm/dev/alfred/berlin/crowd_pilot/install_sglang.sh",,terminal_command
|
| 23 |
+
23,219694,"TERMINAL",0,0,"]633;C#! /bin/bash\r\n\r\n# NOTE: Make sure that the project is set correctly using jutil\r\nmkdir $PROJECT/alfred/\r\nexport CARGO_HOME=$PROJECT/alfred/.cargo\r\nexport RUSTUP_HOME=$PROJECT/alfred/.rustup\r\n\r\nsource .venv/bin/activate\r\n\r\n# NOTE: A local rust installation already exists in the project root\r\ncurl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh -s -- -y\r\nuv pip install ""sglang"" --prerelease=allow\r\n]0;mihir.mahajan@hai-login2:~/Projects/crowd-pilot",,terminal_output
|
| 24 |
+
24,234818,"TERMINAL",0,0,"cat slurm/dev/alfred/berlin/crowd_pilot/launch_sglang.sh",,terminal_command
|
| 25 |
+
25,234836,"TERMINAL",0,0,"]633;CFLASHINFER_CACHE_DIR=$PROJECT/.cache/flashinfer\r\n\r\nsource .venv/bin/activate\r\n\r\npython3 -m sglang.launch_server --model-path qwen/qwen2.5-0.5b-instruct --host 0.0.0.0\r\n]0;mihir.mahajan@hai-login2:~/Projects/crowd-pilot",,terminal_output
|
| 26 |
+
26,275503,"TERMINAL",0,0,"cat slurm/dev/alfred/berlin/crowd_pilot/prompt_engineering/chat.py",,terminal_command
|
| 27 |
+
27,275506,"TERMINAL",0,0,"]633;Cimport requests\r\n\r\n# SERVER_URL = ""http://hai005:30000/generate""\r\nSERVER_URL = ""http://hai005:30000/v1/chat/completions""\r\n\r\n\r\nSYSTEM_PROMPT = """"""\r\n\t\t'Output ONLY a JSON array. No prose, no code fences.',\r\n\t\t'Allowed actions (TypeScript-like schema):',\r\n\t\t""{ kind: 'showTextDocument' }"",\r\n\t\t""{ kind: 'setSelections', selections: Array<{ start: [number, number], end: [number, number] }> }"",\r\n\t\t""{ kind: 'editInsert', position: [number, number], text: string }"",\r\n\t\t""{ kind: 'terminalShow' }"",\r\n\t\t""{ kind: 'terminalSendText', text: string }"",\r\n\t\t'Coordinates are zero-based [line, column].'\r\n""""""\r\n\r\ndemoGoal = """"""\r\n\t\t'Create a concise demo plan that:',y\r\n\t\t'- focuses the active text document',y\r\n\t\t'- moves the cursor to (0, 0)',y\r\n\t\t""- inserts the line \""hello from model\\n\"" at (0, 0)"",y\r\n\t\t'- focuses the terminal',y\r\n\t\t'- runs the command ""echo model run""'\r\n\r\n""""""\r\n\r\n\r\n\r\ndef chat(message):\r\n\r\n # payload = {\r\n # ""text"": message, # <- use ""text"" instead of ""prompt""\r\n # }\r\n\r\n payload = {\r\n ""model"": ""qwen/qwen2.5-0.5b-instruct"",\r\n ""messages"": [\r\n {""role"": ""system"", ""content"": SYSTEM_PROMPT},\r\n {""role"": ""user"", ""content"": message}\r\n ],\r\n }\r\n\r\n response = requests.post(SERVER_URL, json=payload)\r\n if response.status_code == 200:\r\n return response.json().get(""choices"", [{}])[0].get(""message"", {}).get(""content"", """")\r\n else:\r\n return f""Error: {response.status_code} - {response.text}""\r\n\r\n\r\nif __name__ == ""__main__"":\r\n while True:\r\n msg = input(""You: "")\r\n if msg.lower() in [""exit"", ""quit""]:\r\n break\r\n reply = chat(msg)\r\n print(""Bot:"", reply)\r\n]0;mihir.mahajan@hai-login2:~/Projects/crowd-pilot",,terminal_output
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-11074bd2-09df-4856-bd07-fbc4dd6df28c1761329974242-2025_10_24-20.20.11.685/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-117eaf48-658e-44af-bdb3-ac16b413b55d1764499426748-2025_11_30-11.44.10.289/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-11aa9396-05bd-4153-b391-316c69fcb7011764002704562-2025_11_24-17.45.28.743/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-17597979-3a78-4d91-bff1-68d4cc98cb621761146980297-2025_10_22-17.30.03.925/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-1d012a4c-25e5-4957-be82-a3e8ebc2f90b1760629588479-2025_10_16-17.46.51.747/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-1dbf94d5-e80c-4f76-8c82-ec481c7acea31762603480406-2025_11_08-13.05.09.810/source.csv
ADDED
|
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,1838,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"1:05:09 PM [info] Activating crowd-code\n1:05:09 PM [info] Recording started\n1:05:09 PM [info] Initializing git provider using file system watchers...\n1:05:09 PM [info] No workspace folder found\n",Log,tab
|
| 3 |
+
3,2285,"extension-output-pdoom-org.crowd-code-#1-crowd-code",194,0,"1:05:11 PM [info] Retrying git provider initialization...\n1:05:11 PM [info] No workspace folder found\n",Log,content
|
| 4 |
+
4,7180,"TERMINAL",0,0,"",,terminal_command
|
| 5 |
+
5,14300,"TERMINAL",0,0,"",,terminal_command
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-20676a2c-2459-41f4-8896-bf5b1bc9afe71764494282718-2025_11_30-10.18.52.300/source.csv
ADDED
|
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,3,"data/eval/handcrafted/add_import_easy.md",0,0,"# Assistant\n```bash\npython src/input_pipeline/validation_to_testcases.py --input_file data/temp/validation.jsonl --output_file data/temp/validation_testcases.jsonl\n```\n\n# User\n<stdout>\nTraceback (most recent call last):\n File ""/fast/home/mihir.mahajan/Projects/tab-model-eval/src/input_pipeline/validation_to_testcases.py"", line 57, in <module>\n convert_to_incremental_jsonl(line, args.output_file, task_name)\n ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File ""/fast/home/mihir.mahajan/Projects/tab-model-eval/src/input_pipeline/validation_to_testcases.py"", line 14, in convert_to_incremental_jsonl\n data = json.loads(line)\n ^^^^\nNameError: name 'json' is not defined. Did you forget to import 'json'?\n</stdout>\n\n\n# Assistant\n```bash\ncat -n src/input_pipeline/validation_to_testcases.py\n```\n\n# User\n<stdout>\n 1 import tyro\n 2 import os\n 3 from dataclasses import dataclass\n 4\n 5\n 6 @dataclass\n 7 class Args:\n 8 input_file: str = ""data/eval/val/validation.jsonl""\n 9 output_file: str = ""data/eval/val/validation_testcases.jsonl""\n 10 task_name: str = ""validation_set""\n 11\n 12\n 13 def convert_to_incremental_jsonl(line, output_file, task_name):\n 14 data = json.loads(line)\n 15\n 16 system_prompt = data.get(""system_prompt"", """")\n 17 conversations = data.get(""conversations"", [])\n 18\n 19 output_lines = []\n 20 context = []\n 21\n 22 for i, conv in enumerate(conversations):\n 23 # Convert ""Assistant""/""User"" to ""assistant""/""user""\n 24 role = conv[""from""].lower()\n 25 content = conv[""value""]\n 26\n 27 # Add current message to context\n 28 context.append({""role"": role, ""content"": content})\n 29\n 30 if i > 0 and role == ""assistant"":\n 31 task_entry = {\n 32 ""system_prompt"": system_prompt,\n 33 ""task_id"": f""{task_name}/{i-1}"",\n 34 ""context"": context[:-1].copy(),\n 35 ""expected_final_response"": content,\n 36 }\n 37 output_lines.append(task_entry)\n 38\n 39 with open(output_file, ""a"") as f:\n 40 for line in output_lines:\n 41 f.write(json.dumps(line) + ""\n"")\n 42\n 43 print(f""Converted {len(output_lines)} entries to {output_file}"")\n 44\n 45\n 46 if __name__ == ""__main__"":\n 47 args = tyro.cli(Args)\n 48\n 49 if os.path.exists(args.output_file):\n 50 os.remove(args.output_file)\n 51\n 52 # iterate over jsonl file and convert to incremental jsonl\n 53 conversation_id = 0\n 54 with open(args.input_file, ""r"") as f:\n 55 for line in f:\n 56 task_name = f""conversation_{conversation_id}/{args.task_name}""\n 57 convert_to_incremental_jsonl(line, args.output_file, task_name)\n 58 conversation_id += 1\n</stdout>\n\n# Assistant\n```bash\nsed -i '1i\import json' src/input_pipeline/validation_to_testcases.py\n```\n\n# User\n<stdout>\n</stdout>\n\n# Assistant\n```bash\npython src/input_pipeline/validation_to_testcases.py --input_file data/temp/validation.jsonl --output_file data/temp/validation_testcases.jsonl\n```\n\n# User\n<stdout>\nConverted 7 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 4 entries to data/temp/validation_testcases.jsonl\nConverted 7 entries to data/temp/validation_testcases.jsonl\nConverted 4 entries to data/temp/validation_testcases.jsonl\nConverted 8 entries to data/temp/validation_testcases.jsonl\nConverted 8 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 5 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 2 entries to data/temp/validation_testcases.jsonl\nConverted 4 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 5 entries to data/temp/validation_testcases.jsonl\nConverted 6 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\n</stdout>\n",markdown,tab
|
| 3 |
+
2,14760,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"10:18:52 AM [info] Activating crowd-code\n10:18:52 AM [info] Recording started\n10:18:52 AM [info] Initializing git provider using file system watchers...\n10:18:54 AM [info] Retrying git provider initialization...\n10:18:56 AM [info] Git repository found\n10:18:56 AM [info] Git provider initialized successfully\n10:18:56 AM [info] Initial git state: [object Object]\n10:18:56 AM [info] Git repository found\n10:18:56 AM [info] Git provider initialized successfully\n10:18:56 AM [info] Initial git state: [object Object]\n",Log,tab
|
| 4 |
+
3,67181,"data/eval/handcrafted/add_import_easy.md",0,0,"",markdown,tab
|
| 5 |
+
4,77316,"TERMINAL",0,0,"",,terminal_command
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-21579c2d-8ebc-43ea-af3d-20ff8c9f5eb51761135573580-2025_10_22-14.19.54.427/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-2165ae8d-ebb7-4134-a4e8-9043ba00c6551762612659284-2025_11_08-15.38.17.91/source.csv
ADDED
|
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,3,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",0,0,"import requests\n\n# SERVER_URL = ""http://hai005:30000/generate""\nSERVER_URL = ""http://hai006:30000/v1/chat/completions""\n\n\nSYSTEM_PROMPT = """"""\n\t\t'Output ONLY a JSON array. No prose, no code fences.',\n\t\t'Allowed actions (TypeScript-like schema):',\n\t\t""{ kind: 'showTextDocument' }"",\n\t\t""{ kind: 'setSelections', selections: Array<{ start: [number, number], end: [number, number] }> }"",\n\t\t""{ kind: 'editInsert', position: [number, number], text: string }"",\n\t\t""{ kind: 'terminalShow' }"",\n\t\t""{ kind: 'terminalSendText', text: string }"",\n\t\t'Coordinates are zero-based [line, column].'\n""""""\n\ndemoGoal = """"""\n\t\t'Create a concise demo plan that:',y\n\t\t'- focuses the active text document',y\n\t\t'- moves the cursor to (0, 0)',y\n\t\t""- inserts the line \""hello from model\\n\"" at (0, 0)"",y\n\t\t'- focuses the terminal',y\n\t\t'- runs the command ""echo model run""'\n\n""""""\n\n\n\ndef chat():\n\n payload = {\n ""model"": ""Qwen/Qwen3-0.6B"",\n ""messages"": [\n {""role"": ""system"", ""content"": SYSTEM_PROMPT},\n {""role"": ""user"", ""content"": demoGoal}\n ],\n }\n\n response = requests.post(SERVER_URL, json=payload)\n if response.status_code == 200:\n return response.json().get(""choices"", [{}])[0].get(""message"", {}).get(""content"", """")\n else:\n return f""Error: {response.status_code} - {response.text}""\n\n\nif __name__ == ""__main__"":\n reply = chat()\n print(reply)\n",python,tab
|
| 3 |
+
2,3540,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"3:38:17 PM [info] Activating crowd-code\n3:38:17 PM [info] Recording started\n3:38:17 PM [info] Initializing git provider using file system watchers...\n3:38:17 PM [info] Git repository found\n3:38:17 PM [info] Git provider initialized successfully\n3:38:17 PM [info] Initial git state: [object Object]\n",Log,tab
|
| 4 |
+
3,9553,"TERMINAL",0,0,"",,terminal_command
|
| 5 |
+
4,20322,"TERMINAL",0,0,"",,terminal_command
|
| 6 |
+
5,92697,"TERMINAL",0,0,"",,terminal_command
|
| 7 |
+
6,186522,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",0,0,"",python,tab
|
| 8 |
+
7,295932,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
|
| 9 |
+
8,298717,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",0,0,"",python,tab
|
| 10 |
+
9,357980,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",842,0,"",python,selection_mouse
|
| 11 |
+
10,358012,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",841,0,"",python,selection_command
|
| 12 |
+
11,401294,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",665,0,"",python,selection_mouse
|
| 13 |
+
12,401731,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",627,0,"",python,selection_mouse
|
| 14 |
+
13,402461,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",595,38,"\t\t'Create a concise demo plan that:',y",python,selection_command
|
| 15 |
+
14,402631,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",595,79,"\t\t'Create a concise demo plan that:',y\n\t\t'- focuses the active text document',y",python,selection_command
|
| 16 |
+
15,402781,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",595,114,"\t\t'Create a concise demo plan that:',y\n\t\t'- focuses the active text document',y\n\t\t'- moves the cursor to (0, 0)',y",python,selection_command
|
| 17 |
+
16,402891,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",595,173,"\t\t'Create a concise demo plan that:',y\n\t\t'- focuses the active text document',y\n\t\t'- moves the cursor to (0, 0)',y\n\t\t""- inserts the line \""hello from model\\n\"" at (0, 0)"",y",python,selection_command
|
| 18 |
+
17,403546,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",595,202,"\t\t'Create a concise demo plan that:',y\n\t\t'- focuses the active text document',y\n\t\t'- moves the cursor to (0, 0)',y\n\t\t""- inserts the line \""hello from model\\n\"" at (0, 0)"",y\n\t\t'- focuses the terminal',y",python,selection_command
|
| 19 |
+
18,403780,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",633,0,"",python,selection_command
|
| 20 |
+
19,404441,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",796,1,"",python,content
|
| 21 |
+
20,404441,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",767,1,"",python,content
|
| 22 |
+
21,404441,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",708,1,"",python,content
|
| 23 |
+
22,404441,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",673,1,"",python,content
|
| 24 |
+
23,404441,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",632,1,"",python,content
|
| 25 |
+
24,404811,"slurm/dev/mihir/berlin/crowd_pilot/prompt_engineering/chat.py",631,0,"",python,selection_command
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-23300773-7107-48dc-95de-2ef6fdc0e97c1760710518542-2025_10_17-16.15.55.319/source.csv
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,2828,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"4:15:55 PM [info] Activating crowd-code\n4:15:55 PM [info] Recording started\n4:15:55 PM [info] Initializing git provider using file system watchers...\n4:15:56 PM [info] Git repository found\n4:15:56 PM [info] Git provider initialized successfully\n4:15:56 PM [info] Initial git state: [object Object]\n",Log,tab
|
| 3 |
+
3,8847,"TERMINAL",0,0,"",,terminal_command
|
| 4 |
+
4,15626,"TERMINAL",0,0,"",,terminal_command
|
| 5 |
+
5,89719,"TERMINAL",0,0,"",,terminal_command
|
| 6 |
+
6,277515,"slurm/utils/mihir/ali_runs/merge_and_plot.sh",0,0,"python slurm/utils/mihir/run-renamer.py slurm/utils/mihir/jasmine.csv slurm/utils/mihir/jasmine_renamed.csv\n\npython slurm/utils/mihir/merge_wandb_runs.py -i slurm/utils/mihir/ali_runs/og*.csv -o slurm/utils/mihir/ali_runs/merged.csv\npython slurm/utils/mihir/run-renamer.py slurm/utils/mihir/ali_runs/merged.csv slurm/utils/mihir/ali_runs/merged_renamed.csv\n\npython slurm/utils/mihir/wandb_csv_to_matplotlib_minimal.py slurm/utils/mihir/jasmine_renamed.csv slurm/utils/mihir/ali_runs/merged_renamed.csv --x ""time"" --metric ""Train loss"" --save slurm/utils/mihir/ali_runs/train-loss.pdf",shellscript,tab
|
| 7 |
+
7,282537,"slurm/utils/mihir/jafar_full_run/merge_and_plot.sh",0,0,"python slurm/utils/mihir/run-renamer.py slurm/utils/mihir/jasmine.csv slurm/utils/mihir/jasmine_renamed.csv\n\npython slurm/utils/mihir/merge_wandb_runs.py -i slurm/utils/mihir/ali_runs/og*.csv -o slurm/utils/mihir/ali_runs/merged.csv\npython slurm/utils/mihir/run-renamer.py slurm/utils/mihir/ali_runs/merged.csv slurm/utils/mihir/ali_runs/merged_renamed.csv\n\npython slurm/utils/mihir/wandb_csv_to_matplotlib_minimal.py slurm/utils/mihir/jasmine_renamed.csv slurm/utils/mihir/ali_runs/merged_renamed.csv --x ""time"" --metric ""Train loss"" --save slurm/utils/mihir/ali_runs/train-loss.pdf",shellscript,tab
|
| 8 |
+
8,283659,"slurm/utils/mihir/jafar_full_run/merge_and_plot.sh",108,0,"",shellscript,selection_mouse
|
| 9 |
+
9,287332,"slurm/utils/mihir/jafar_full_run/merge_and_plot.sh",141,1,"a",shellscript,selection_command
|
| 10 |
+
10,287333,"slurm/utils/mihir/jafar_full_run/merge_and_plot.sh",175,2,"al",shellscript,selection_command
|
| 11 |
+
11,287512,"slurm/utils/mihir/jafar_full_run/merge_and_plot.sh",175,3,"ali",shellscript,selection_command
|
| 12 |
+
12,288427,"slurm/utils/mihir/jafar_full_run/merge_and_plot.sh",175,4,"ali_",shellscript,selection_command
|
| 13 |
+
13,288844,"slurm/utils/mihir/jafar_full_run/merge_and_plot.sh",175,5,"ali_r",shellscript,selection_command
|
| 14 |
+
14,288920,"slurm/utils/mihir/jafar_full_run/merge_and_plot.sh",175,6,"ali_ru",shellscript,selection_command
|
| 15 |
+
15,288972,"slurm/utils/mihir/jafar_full_run/merge_and_plot.sh",175,7,"ali_run",shellscript,selection_command
|
| 16 |
+
16,289111,"slurm/utils/mihir/jafar_full_run/merge_and_plot.sh",175,8,"ali_runs",shellscript,selection_command
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-25df760a-b46a-4185-aa09-ab9e63322f9f1764066566259-2025_11_25-11.29.59.139/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-2b611b25-766c-40fe-9c56-c37495c49abf1764154622152-2025_11_26-11.57.21.424/source.csv
ADDED
|
@@ -0,0 +1,78 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,3267,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"11:57:21 AM [info] Activating crowd-code\n11:57:21 AM [info] Recording started\n11:57:21 AM [info] Initializing git provider using file system watchers...\n11:57:23 AM [info] Retrying git provider initialization...\n11:57:23 AM [info] Git repository found\n11:57:23 AM [info] Git provider initialized successfully\n11:57:23 AM [info] Initial git state: [object Object]\n11:57:24 AM [info] Git repository found\n11:57:24 AM [info] Git provider initialized successfully\n11:57:24 AM [info] Initial git state: [object Object]\n",Log,tab
|
| 3 |
+
3,10154,"TERMINAL",0,0,"",,terminal_command
|
| 4 |
+
4,17051,"TERMINAL",0,0,"",,terminal_focus
|
| 5 |
+
5,17172,"TERMINAL",0,0,"",,terminal_command
|
| 6 |
+
6,17552,"TERMINAL",0,0,"source /home/mihir.mahajan/Projects/crowd-pilot/.uv_venv/bin/activate",,terminal_command
|
| 7 |
+
7,17563,"TERMINAL",0,0,"]633;C]0;mihir.mahajan@hai-login2:~/Projects/crowd-pilot",,terminal_output
|
| 8 |
+
8,19079,"TERMINAL",0,0,"watch",,terminal_focus
|
| 9 |
+
9,46947,"TERMINAL",0,0,"salloc --time=02:00:00 --gres=gpu:1 --nodes=1 --ntasks-per-node=1 --mem=200G --qos=low",,terminal_command
|
| 10 |
+
10,47003,"TERMINAL",0,0,"]633;Csalloc: Granted job allocation 34851\r\n",,terminal_output
|
| 11 |
+
11,47089,"TERMINAL",0,0,"salloc: Nodes hai005 are ready for job\r\n",,terminal_output
|
| 12 |
+
12,47370,"TERMINAL",0,0,"Running inside SLURM, Job ID 34851.\r\n",,terminal_output
|
| 13 |
+
13,47459,"TERMINAL",0,0,"]0;mihir.mahajan@hai-login2:~/Projects/crowd-pilot[?2004h[mihir.mahajan@hai005.haicore.berlin:~/Projects/crowd-pilot] $ ",,terminal_output
|
| 14 |
+
14,48513,"TERMINAL",0,0,"s",,terminal_output
|
| 15 |
+
15,48694,"TERMINAL",0,0,"o",,terminal_output
|
| 16 |
+
16,48824,"TERMINAL",0,0,"u",,terminal_output
|
| 17 |
+
17,48895,"TERMINAL",0,0,"r",,terminal_output
|
| 18 |
+
18,49077,"TERMINAL",0,0,"c",,terminal_output
|
| 19 |
+
19,49266,"TERMINAL",0,0,"e",,terminal_output
|
| 20 |
+
20,49382,"TERMINAL",0,0," .",,terminal_output
|
| 21 |
+
21,49559,"TERMINAL",0,0,"v",,terminal_output
|
| 22 |
+
22,49736,"TERMINAL",0,0,"env/",,terminal_output
|
| 23 |
+
23,49833,"TERMINAL",0,0,"b",,terminal_output
|
| 24 |
+
24,50284,"TERMINAL",0,0,"in/",,terminal_output
|
| 25 |
+
25,50410,"TERMINAL",0,0,"ac",,terminal_output
|
| 26 |
+
26,50766,"TERMINAL",0,0,"tivate",,terminal_output
|
| 27 |
+
27,50900,"TERMINAL",0,0,"\r\n[?2004l\r]0;mihir.mahajan@hai-login2:~/Projects/crowd-pilot[?2004h(crowd-pilot) [mihir.mahajan@hai005.haicore.berlin:~/Projects/crowd-pilot] $ ",,terminal_output
|
| 28 |
+
28,57713,"TERMINAL",0,0,"sc",,terminal_output
|
| 29 |
+
29,57911,"TERMINAL",0,0,"a",,terminal_output
|
| 30 |
+
30,58054,"TERMINAL",0,0,"n",,terminal_output
|
| 31 |
+
31,58198,"TERMINAL",0,0,"c",,terminal_output
|
| 32 |
+
32,58323,"TERMINAL",0,0,"e",,terminal_output
|
| 33 |
+
33,58419,"TERMINAL",0,0,"l",,terminal_output
|
| 34 |
+
34,58525,"TERMINAL",0,0," ",,terminal_output
|
| 35 |
+
35,58640,"TERMINAL",0,0,"[7m34[27m[7m8[27m[7m49[27m",,terminal_output
|
| 36 |
+
36,59051,"TERMINAL",0,0,"[A[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C34849\r\n[?2004l\r]0;mihir.mahajan@hai-login2:~/Projects/crowd-pilot[?2004h(crowd-pilot) [mihir.mahajan@hai005.haicore.berlin:~/Projects/crowd-pilot] $ ",,terminal_output
|
| 37 |
+
37,82286,"slurm/dev/mihir/berlin/crowd_pilot/nemo/iterating_on_stuff/test_full_finetune_4k.py",0,0,"import os\n\nimport nemo_run as run\nfrom nemo.collections import llm\nfrom lightning.pytorch.loggers import WandbLogger\n\nos.environ[""NEMORUN_HOME""] = ""/fast/project/HFMI_SynergyUnit/tab_model/franz/nemorun_home""\nos.environ[""NEMO_CACHE_DIR""] = ""/fast/project/HFMI_SynergyUnit/tab_model/franz/nemo_cache""\nslurm_job_id = os.environ.get(""SLURM_JOB_ID"", ""local"")\n\nNAME = f""qwen3_600m_full_4k_tokens_debug_{slurm_job_id}""\nDATASET_DIR = ""/fast/project/HFMI_SynergyUnit/tab_model/data/nemo_hf_part_jsonl_4k_tokens/"" \n\ndef recipe_to_dict(recipe, depth=4):\n if depth == 0:\n return dict()\n cfg_dict = dict()\n for k, v in vars(recipe)['__arguments__'].items():\n if isinstance(v, run.Config) or isinstance(v, dict):\n cfg_dict[k] = recipe_to_dict(v, depth-1)\n elif isinstance(v, (int, float, str, bool)):\n cfg_dict[k] = v\n else:\n cfg_dict[k] = str(v)\n return cfg_dict\n \nrecipe = llm.qwen3_600m.finetune_recipe(\n name=NAME,\n dir=f""/fast/project/HFMI_SynergyUnit/mihir/logs"",\n num_nodes=1,\n num_gpus_per_node=1,\n peft_scheme=None,\n # FIXME(f.srambical): check whether this is supported by ChatDataModule\n packed_sequence=False,\n)\n\ndef local_executor_torchrun(nodes: int, devices: int) -> run.LocalExecutor:\n # Env vars for jobs are configured here\n env_vars = {\n ""TORCH_NCCL_AVOID_RECORD_STREAMS"": ""1"",\n ""NCCL_NVLS_ENABLE"": ""0"",\n ""NVTE_DP_AMAX_REDUCE_INTERVAL"": ""0"",\n ""NVTE_ASYNC_AMAX_REDUCTION"": ""1"",\n }\n\n executor = run.LocalExecutor(ntasks_per_node=devices, launcher=""torchrun"", env_vars=env_vars)\n\n return executor\n\ndef run_finetuning():\n recipe.resume.restore_config.path = ""/fast/project/HFMI_SynergyUnit/tab_model/data/checkpoints/nemo_converted_weights_qwen3-600m/""\n recipe.data = run.Config(\n llm.ChatDataModule,\n dataset_root=DATASET_DIR,\n seq_length=4096,\n global_batch_size=32,\n micro_batch_size=4,\n num_workers=0,\n )\n\n MAX_STEPS = 11\n VAL_INTERVAL = 2\n NUM_CHECKPOINTS = 10\n recipe.trainer.max_steps = MAX_STEPS\n recipe.log.ckpt.every_n_train_steps = MAX_STEPS // NUM_CHECKPOINTS\n recipe.log.ckpt.save_top_k = NUM_CHECKPOINTS\n recipe.trainer.val_check_interval = VAL_INTERVAL\n\n cfg_dict = recipe_to_dict(recipe)\n recipe.log.wandb = run.Config(\n WandbLogger,\n name=NAME,\n project=""crowd-pilot-nemo"",\n save_dir=""/fast/project/HFMI_SynergyUnit/mihir/logs/wandb"",\n entity=""instant-uv"",\n config=cfg_dict,\n )\n\n executor = local_executor_torchrun(nodes=recipe.trainer.num_nodes, devices=recipe.trainer.devices)\n\n run.run(recipe, executor=executor, name=NAME)\n\n# This condition is necessary for the script to be compatible with Python's multiprocessing module.\nif __name__ == ""__main__"":\n run_finetuning()",python,tab
|
| 38 |
+
38,82287,"slurm/dev/mihir/berlin/crowd_pilot/nemo/iterating_on_stuff/test_full_finetune_4k.py",2083,0,"",python,selection_mouse
|
| 39 |
+
39,82361,"slurm/dev/mihir/berlin/crowd_pilot/nemo/iterating_on_stuff/test_full_finetune_4k.py",2082,0,"",python,selection_command
|
| 40 |
+
40,83172,"TERMINAL",0,0,"srun",,terminal_focus
|
| 41 |
+
41,88676,"TERMINAL",0,0,"ls",,terminal_output
|
| 42 |
+
42,89998,"TERMINAL",0,0,"\r\n[?2004l\rLICENSE [0m[01;34mcrowd-pilot-extension[0m [01;34mmaxtext[0m [01;34mslurm[0m uv.lock\r\nREADME.md [01;34mcrowd_pilot[0m pyproject.toml [01;34mtests[0m\r\n]0;mihir.mahajan@hai-login2:~/Projects/crowd-pilot[?2004h(crowd-pilot) [mihir.mahajan@hai005.haicore.berlin:~/Projects/crowd-pilot] $ ",,terminal_output
|
| 43 |
+
43,91865,"TERMINAL",0,0,"l",,terminal_output
|
| 44 |
+
44,91984,"TERMINAL",0,0,"o",,terminal_output
|
| 45 |
+
45,92163,"TERMINAL",0,0,"g",,terminal_output
|
| 46 |
+
46,92341,"TERMINAL",0,0,"s\r\n[?2004l\r]0;mihir.mahajan@hai-login2:/fast/project/HFMI_SynergyUnit/mihir/logs[?2004h(crowd-pilot) [mihir.mahajan@hai005.haicore.berlin:/fast/project/HFMI_SynergyUnit/mihir/logs] $ ",,terminal_output
|
| 47 |
+
47,92474,"TERMINAL",0,0,"l",,terminal_output
|
| 48 |
+
48,92652,"TERMINAL",0,0,"s\r\n[?2004l\rfinetune_lora_4k.log finetune_lora_4k_34587.log [0m[01;34mtb_logs[0m\r\nfinetune_lora_4k_34569.log finetune_lora_4k_34588.log [01;34mtrain[0m\r\nfinetune_lora_4k_34572.log [01;34mnemo[0m [01;34mwandb[0m\r\nfinetune_lora_4k_34584.log [01;34mqwen3_600m_full_4k_tokens_34783[0m\r\nfinetune_lora_4k_34585.log [01;34mqwen3_600m_full_4k_tokens_debug_34849[0m\r\n]0;mihir.mahajan@hai-login2:/fast/project/HFMI_SynergyUnit/mihir/logs[?2004h(crowd-pilot) [mihir.mahajan@hai005.haicore.berlin:/fast/project/HFMI_SynergyUnit/mihir/logs] $ ",,terminal_output
|
| 49 |
+
49,94698,"TERMINAL",0,0,"c",,terminal_output
|
| 50 |
+
50,94752,"TERMINAL",0,0,"d",,terminal_output
|
| 51 |
+
51,94892,"TERMINAL",0,0," ",,terminal_output
|
| 52 |
+
52,95161,"TERMINAL",0,0,"n",,terminal_output
|
| 53 |
+
53,95846,"TERMINAL",0,0,"[K",,terminal_output
|
| 54 |
+
54,95980,"TERMINAL",0,0,"[K",,terminal_output
|
| 55 |
+
55,100656,"TERMINAL",0,0," ",,terminal_output
|
| 56 |
+
56,104023,"TERMINAL",0,0,"[7mqwen3_600m_full_4k_tokens_debug_34849[27m",,terminal_output
|
| 57 |
+
57,104466,"TERMINAL",0,0,"\r[C[C[C[C[C[C[C[C[C[C[C[Cqwen3_600m_full_4k_tokens_debug_34849\r\n[?2004l\r]0;mihir.mahajan@hai-login2:/fast/project/HFMI_SynergyUnit/mihir/logs/qwen3_600m_full_4k_tokens_debug_34849[?2004h(crowd-pilot) [mihir.mahajan@hai005.haicore.berlin:/fast/project/HFMI_SynergyUnit/mihir/logs/qwen3_600m_full_4k_tokens_debug_34849] $ ",,terminal_output
|
| 58 |
+
58,104781,"TERMINAL",0,0,"ls",,terminal_output
|
| 59 |
+
59,104958,"TERMINAL",0,0,"\r\n[?2004l\r[0m[01;34m2025-11-26_11-33-01[0m\r\n]0;mihir.mahajan@hai-login2:/fast/project/HFMI_SynergyUnit/mihir/logs/qwen3_600m_full_4k_tokens_debug_34849[?2004h(crowd-pilot) [mihir.mahajan@hai005.haicore.berlin:/fast/project/HFMI_SynergyUnit/mihir/logs/qwen3_600m_full_4k_tokens_debug_34849] $ ",,terminal_output
|
| 60 |
+
60,107010,"TERMINAL",0,0,"cd",,terminal_output
|
| 61 |
+
61,107076,"TERMINAL",0,0," ",,terminal_output
|
| 62 |
+
62,107160,"TERMINAL",0,0,"2025-11-26_11-33-01/",,terminal_output
|
| 63 |
+
63,107227,"TERMINAL",0,0,"",,terminal_command
|
| 64 |
+
64,107484,"TERMINAL",0,0,"\r\n[?2004l\r]0;mihir.mahajan@hai-login2:/fast/project/HFMI_SynergyUnit/mihir/logs/qwen3_600m_full_4k_tokens_debug_34849/2025-11-26_11-33-01[?2004h(crowd-pilot) [mihir.mahajan@hai005.haicore.berlin:/fast/project/HFMI_SynergyUnit/mihir/logs/qwen3_600m_full_4k_tokens_debug_34849/2025-11-26_11-33-01] $ ",,terminal_output
|
| 65 |
+
65,107666,"TERMINAL",0,0,"l",,terminal_output
|
| 66 |
+
66,107902,"TERMINAL",0,0,"s",,terminal_output
|
| 67 |
+
67,108200,"TERMINAL",0,0,"\r\n[?2004l\r[0m[01;34mcheckpoints[0m git-info.log nemo_error_log.txt\r\ncmd-args.log lightning_logs.txt nemo_log_globalrank-0_localrank-0.txt\r\n]0;mihir.mahajan@hai-login2:/fast/project/HFMI_SynergyUnit/mihir/logs/qwen3_600m_full_4k_tokens_debug_34849/2025-11-26_11-33-01[?2004h(crowd-pilot) [mihir.mahajan@hai005.haicore.berlin:/fast/project/HFMI_SynergyUnit/mihir/logs/qwen3_600m_full_4k_tokens_debug_34849/2025-11-26_11-33-01] $ ",,terminal_output
|
| 68 |
+
68,142974,"TERMINAL",0,0,"l",,terminal_output
|
| 69 |
+
69,143053,"TERMINAL",0,0,"s",,terminal_output
|
| 70 |
+
70,143118,"TERMINAL",0,0," ",,terminal_output
|
| 71 |
+
71,143327,"TERMINAL",0,0,"c",,terminal_output
|
| 72 |
+
72,143493,"TERMINAL",0,0,"h",,terminal_output
|
| 73 |
+
73,143619,"TERMINAL",0,0,"e",,terminal_output
|
| 74 |
+
74,143768,"TERMINAL",0,0,"c",,terminal_output
|
| 75 |
+
75,143906,"TERMINAL",0,0,"kpoints/",,terminal_output
|
| 76 |
+
76,144336,"TERMINAL",0,0,"\r\n[?2004l\r[0m[01;34m'model_name=0--val_loss=1.09-step=10-consumed_samples=352.0-last'[0m\r\n[01;34m'model_name=0--val_loss=1.10-step=10-consumed_samples=352.0'[0m\r\n[01;34m'model_name=0--val_loss=1.14-step=8-consumed_samples=288.0'[0m\r\n[01;34m'model_name=0--val_loss=1.14-step=9-consumed_samples=320.0'[0m\r\n[01;34m'model_name=0--val_loss=1.15-step=6-consumed_samples=224.0'[0m\r\n[01;34m'model_name=0--val_loss=1.15-step=7-consumed_samples=256.0'[0m\r\n[01;34m'model_name=0--val_loss=1.16-step=2-consumed_samples=96.0'[0m\r\n[01;34m'model_name=0--val_loss=1.16-step=3-consumed_samples=128.0'[0m\r\n[01;34m'model_name=0--val_loss=1.16-step=4-consumed_samples=160.0'[0m\r\n[01;34m'model_name=0--val_loss=1.16-step=5-consumed_samples=192.0'[0m\r\n]0;mihir.mahajan@hai-login2:/fast/project/HFMI_SynergyUnit/mihir/logs/qwen3_600m_full_4k_tokens_debug_34849/2025-11-26_11-33-01[?2004h(crowd-pilot) [mihir.mahajan@hai005.haicore.berlin:/fast/project/HFMI_SynergyUnit/mihir/logs/qwen3_600m_full_4k_tokens_debug_34849/2025-11-26_11-33-01] $ ",,terminal_output
|
| 77 |
+
77,279191,"TERMINAL",0,0,"\r[K[A(crowd-pilot) [mihir.mahajan@hai005.haicore.berlin:/fast/project/HFMI_SynergyUnit/mihir\r\n\r/logs/qwen3_600m_full_4k_tokens_debug_34849/2025-11-26_11-33-01] $ \r[K[A(crowd-pilot) [mihir.mahajan@hai005.haicore.berlin:/fast/project/HFMI_SynergyUnit/mihir\r\n\r/logs/qwen3_600m_full_4k_tokens_debug_34849/2025-11-26_11-33-01] $ ",,terminal_output
|
| 78 |
+
78,608800,"TERMINAL",0,0,"\r[K[A(crowd-pilot) [mihir.mahajan@hai005.haicore.berlin:/fast/project/HFMI_SynergyUnit/mihir\r\n\r/logs/qwen3_600m_full_4k_tokens_debug_34849/2025-11-26_11-33-01] $ ",,terminal_output
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-2f756e36-5a16-427a-9b47-60002a9d13571761321558034-2025_10_24-17.59.44.602/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-32bb221d-d676-4384-9ca4-18916339b62b1764092860950-2025_11_25-18.48.03.180/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-34b6dec0-9dde-4479-9af4-cd0656225b821761477933551-2025_10_26-12.25.57.396/source.csv
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,2312,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"12:25:57 PM [info] Activating crowd-code\n12:25:57 PM [info] Recording started\n12:25:57 PM [info] Initializing git provider using file system watchers...\n12:25:58 PM [info] Git repository found\n12:25:58 PM [info] Git provider initialized successfully\n12:25:58 PM [info] Initial git state: [object Object]\n",Log,tab
|
| 3 |
+
3,8187,"TERMINAL",0,0,"",,terminal_command
|
| 4 |
+
4,15085,"TERMINAL",0,0,"",,terminal_command
|
| 5 |
+
5,103178,"TERMINAL",0,0,"",,terminal_command
|
| 6 |
+
6,144952,"TERMINAL",0,0,"bash",,terminal_focus
|
| 7 |
+
7,147182,"TERMINAL",0,0,"git branch",,terminal_command
|
| 8 |
+
8,147205,"TERMINAL",0,0,"]633;C[?1h=\r feat/diffusion-backend[m[m\r\n feat/diffusion-forcing[m[m\r\n feat/diffusion-jittable-sampling[m[m\r\n feat/diffusion-ramp-weight[m[m\r\n* [32mfeat/diffusion-train-from-mae[m[m\r\n feat/diffusion-x-loss[m[m\r\n feat/masked-autoencoder[m[m\r\n feat/notebook-detect-arrayrecord-duplicates[m[m\r\n feat/shortcut-forcing[m[m\r\n main[m[m\r\n prepend-action-maskgit[m[m\r\n vizdoom-dataset[m[m\r\n\r[K[?1l>]0;mihir.mahajan@hai-login2:~/Projects/jasmine",,terminal_output
|
| 9 |
+
9,149188,"TERMINAL",0,0,"^C",,terminal_command
|
| 10 |
+
10,153934,"TERMINAL",0,0,"git status",,terminal_command
|
| 11 |
+
11,153979,"TERMINAL",0,0,"]633;COn branch feat/diffusion-train-from-mae\r\nYour branch is up to date with 'origin/feat/diffusion-train-from-mae'.\r\n\r\nUntracked files:\r\n (use ""git add <file>..."" to include in what will be committed)\r\n\t[31mdata/_vizdoom.ini[m\r\n\t[31mdata/jasmine_data/ViZDoomPPO/[m\r\n\t[31mdata/uv.lock[m\r\n\t[31mjasmine/genie_bak.py[m\r\n\t[31mjasmine/genie_sliding.py[m\r\n\t[31mjasmine/sample_sliding.py[m\r\n\t[31mjasmine/test_dit.py[m\r\n\t[31mslurm/[m\r\n\t[31muv.lock[m\r\n\r\nnothing added to commit but untracked files present (use ""git add"" to track)\r\n]0;mihir.mahajan@hai-login2:~/Projects/jasmine",,terminal_output
|
| 12 |
+
12,270937,"TERMINAL",0,0,"queue",,terminal_command
|
| 13 |
+
13,271006,"TERMINAL",0,0,"]633;C[?1049h[22;0;0t[1;41r(B[m[4l[?7h[H[2JEvery 1.0s: squeue --me[1;124Hhai-login2.haicore.berlin: Sun Oct 26 12:30:28 2025[3;3HJOBID[15GUSER PARTITION NODES CPUS ST[3;53HSUBMIT_TIME[3;74HSTART_TIME TIME TIME_LIMIT NODELIST(REASON)[41;174H",,terminal_output
|
| 14 |
+
14,271756,"TERMINAL",0,0,"[41;1H[?1049l[23;0;0t\r[?1l>]0;mihir.mahajan@hai-login2:~/Projects/jasmine",,terminal_output
|
| 15 |
+
15,954756,"TERMINAL",0,0,"git checkout feat/diffusion-backend",,terminal_command
|
| 16 |
+
16,954806,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 17 |
+
17,954913,"TERMINAL",0,0,"Switched to branch 'feat/diffusion-backend'\r\nYour branch is up to date with 'origin/feat/diffusion-backend'.\r\n]0;mihir.mahajan@hai-login2:~/Projects/jasmine",,terminal_output
|
| 18 |
+
18,955782,"",0,0,"Switched from branch 'feat/diffusion-train-from-mae' to 'feat/diffusion-backend'",,git_branch_checkout
|
| 19 |
+
19,956407,"TERMINAL",0,0,"git pull",,terminal_command
|
| 20 |
+
20,956456,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 21 |
+
21,958218,"TERMINAL",0,0,"remote: Enumerating objects: 15, done.[K\r\nremote: Counting objects: 6% (1/15)[K\rremote: Counting objects: 13% (2/15)[K\rremote: Counting objects: 20% (3/15)[K\rremote: Counting objects: 26% (4/15)[K\rremote: Counting objects: 33% (5/15)[K\rremote: Counting objects: 40% (6/15)[K\rremote: Counting objects: 46% (7/15)[K\rremote: Counting objects: 53% (8/15)[K\rremote: Counting objects: 60% (9/15)[K\rremote: Counting objects: 66% (10/15)[K\rremote: Counting objects: 73% (11/15)[K\rremote: Counting objects: 80% (12/15)[K\rremote: Counting objects: 86% (13/15)[K\rremote: Counting objects: 93% (14/15)[K\rremote: Counting objects: 100% (15/15)[K\rremote: Counting objects: 100% (15/15), done.[K\r\nremote: Compressing objects: 20% (1/5)[K\rremote: Compressing objects: 40% (2/5)[K\rremote: Compressing objects: 60% (3/5)[K\rremote: Compressing objects: 80% (4/5)[K\rremote: Compressing objects: 100% (5/5)[K\rremote: Compressing objects: 100% (5/5), done.[K\r\nremote: Total 15 (delta 10), reused 14 (delta 10), pack-reused 0 (from 0)[K\r\nUnpacking objects: 6% (1/15)\rUnpacking objects: 13% (2/15)\rUnpacking objects: 20% (3/15)\rUnpacking objects: 26% (4/15)\rUnpacking objects: 33% (5/15)\rUnpacking objects: 40% (6/15)\rUnpacking objects: 46% (7/15)\rUnpacking objects: 53% (8/15)\rUnpacking objects: 60% (9/15)\rUnpacking objects: 66% (10/15)\rUnpacking objects: 73% (11/15)\rUnpacking objects: 80% (12/15)\r",,terminal_output
|
| 22 |
+
22,958460,"TERMINAL",0,0,"Unpacking objects: 86% (13/15)\rUnpacking objects: 93% (14/15)\rUnpacking objects: 100% (15/15)\rUnpacking objects: 100% (15/15), 2.64 KiB | 8.00 KiB/s, done.\r\n",,terminal_output
|
| 23 |
+
23,958683,"TERMINAL",0,0,"From github.com:p-doom/jasmine\r\n 64e02f4..93b4c0a feat/diffusion-backend -> origin/feat/diffusion-backend\r\n 58d8c91..41ec2f5 vizdoom-dataset -> origin/vizdoom-dataset\r\n",,terminal_output
|
| 24 |
+
24,958743,"TERMINAL",0,0,"Updating 64e02f4..93b4c0a\r\n",,terminal_output
|
| 25 |
+
25,958801,"TERMINAL",0,0,"Fast-forward\r\n",,terminal_output
|
| 26 |
+
26,958826,"TERMINAL",0,0," jasmine/genie.py | 244 [32m++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++[m[31m----------------------------------------------------------[m\r\n jasmine/models/dynamics.py | 22 [32m+++++++++[m[31m----[m\r\n jasmine/models/tokenizer.py | 1 [32m+[m\r\n jasmine/sample.py | 8 [32m+++[m[31m--[m\r\n jasmine/train_dynamics.py | 6 [32m++[m[31m--[m\r\n jasmine/train_tokenizer.py | 4 [32m++[m[31m-[m\r\n jasmine/utils/nn.py | 28 [32m++++++++[m[31m--------[m\r\n 7 files changed, 182 insertions(+), 131 deletions(-)\r\n]0;mihir.mahajan@hai-login2:~/Projects/jasmine",,terminal_output
|
| 27 |
+
27,1860457,"TERMINAL",0,0,"git status",,terminal_command
|
| 28 |
+
28,1860517,"TERMINAL",0,0,"]633;COn branch feat/diffusion-backend\r\nYour branch is up to date with 'origin/feat/diffusion-backend'.\r\n\r\nUntracked files:\r\n (use ""git add <file>..."" to include in what will be committed)\r\n\t[31mdata/_vizdoom.ini[m\r\n\t[31mdata/jasmine_data/ViZDoomPPO/[m\r\n\t[31mdata/uv.lock[m\r\n\t[31mjasmine/genie_bak.py[m\r\n\t[31mjasmine/genie_sliding.py[m\r\n\t[31mjasmine/sample_sliding.py[m\r\n\t[31mjasmine/test_dit.py[m\r\n\t[31mslurm/[m\r\n\t[31muv.lock[m\r\n\r\nnothing added to commit but untracked files present (use ""git add"" to track)\r\n]0;mihir.mahajan@hai-login2:~/Projects/jasmine",,terminal_output
|
| 29 |
+
29,1868164,"TERMINAL",0,0,"git branch",,terminal_command
|
| 30 |
+
30,1868204,"TERMINAL",0,0,"]633;C[?1h=\r* [32mfeat/diffusion-backend[m[m\r\n feat/diffusion-forcing[m[m\r\n feat/diffusion-jittable-sampling[m[m\r\n feat/diffusion-ramp-weight[m[m\r\n feat/diffusion-train-from-mae[m[m\r\n feat/diffusion-x-loss[m[m\r\n feat/masked-autoencoder[m[m\r\n feat/notebook-detect-arrayrecord-duplicates[m[m\r\n feat/shortcut-forcing[m[m\r\n main[m[m\r\n prepend-action-maskgit[m[m\r\n vizdoom-dataset[m[m\r\n\r[K[?1l>]0;mihir.mahajan@hai-login2:~/Projects/jasmine",,terminal_output
|
| 31 |
+
31,1891700,"TERMINAL",0,0,"git checkout -b ""chore/cleanup-diffusion-branch""",,terminal_command
|
| 32 |
+
32,1891772,"TERMINAL",0,0,"]633;CSwitched to a new branch 'chore/cleanup-diffusion-branch'\r\n]0;mihir.mahajan@hai-login2:~/Projects/jasmine",,terminal_output
|
| 33 |
+
33,1895871,"",0,0,"Switched from branch 'feat/diffusion-backend' to 'chore/cleanup-diffusion-branch'",,git_branch_checkout
|
| 34 |
+
34,1902533,"TERMINAL",0,0,"git checkout feat/diffusion-backend",,terminal_command
|
| 35 |
+
35,1902567,"TERMINAL",0,0,"]633;CSwitched to branch 'feat/diffusion-backend'\r\nYour branch is up to date with 'origin/feat/diffusion-backend'.\r\n]0;mihir.mahajan@hai-login2:~/Projects/jasmine",,terminal_output
|
| 36 |
+
36,1905869,"",0,0,"Switched from branch 'chore/cleanup-diffusion-branch' to 'feat/diffusion-backend'",,git_branch_checkout
|
| 37 |
+
37,1918003,"TERMINAL",0,0,"git checkout -b ""feat/sliding-window-sampling""",,terminal_command
|
| 38 |
+
38,1918066,"TERMINAL",0,0,"]633;CSwitched to a new branch 'feat/sliding-window-sampling'\r\n]0;mihir.mahajan@hai-login2:~/Projects/jasmine",,terminal_output
|
| 39 |
+
39,1920871,"",0,0,"Switched from branch 'feat/diffusion-backend' to 'feat/sliding-window-sampling'",,git_branch_checkout
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-35174d2b-ac4a-4076-abc0-3de665efb30d1764851634497-2025_12_04-13.34.18.756/source.csv
ADDED
|
@@ -0,0 +1,78 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,3781,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"1:34:18 PM [info] Activating crowd-code\n1:34:18 PM [info] Recording started\n1:34:18 PM [info] Initializing git provider using file system watchers...\n1:34:20 PM [info] Retrying git provider initialization...\n1:34:20 PM [info] Git repository found\n1:34:20 PM [info] Git provider initialized successfully\n1:34:21 PM [info] Initial git state: [object Object]\n1:34:21 PM [info] Git repository found\n1:34:21 PM [info] Git provider initialized successfully\n1:34:21 PM [info] Initial git state: [object Object]\n",Log,tab
|
| 3 |
+
3,9132,"TERMINAL",0,0,"",,terminal_command
|
| 4 |
+
4,16099,"TERMINAL",0,0,"",,terminal_command
|
| 5 |
+
5,16109,"TERMINAL",0,0,"]633;C]0;mihir.mahajan@hai-login1:~/Projects/tab-model-eval",,terminal_output
|
| 6 |
+
6,23505,"TERMINAL",0,0,"bash",,terminal_focus
|
| 7 |
+
7,25812,"TERMINAL",0,0,"queue",,terminal_command
|
| 8 |
+
8,25875,"TERMINAL",0,0,"]633;C[?1049h[22;0;0t[1;36r(B[m[4l[?7h[H[2JEvery 1.0s: sque... hai-login1.haicore.berlin: Thu Dec 4 13:34:44 2025[3;3HJOBID[15GUSER PARTITION NODES CPUS ST[3;53HSUBMIT_TIME\r[4d START_TIME TIME TIME_LIMIT NODELIST(REASON)[36;72H",,terminal_output
|
| 9 |
+
9,26838,"TERMINAL",0,0,"[1;67H5[36;72H",,terminal_output
|
| 10 |
+
10,26876,"TERMINAL",0,0,"[36;1H[?1049l[23;0;0t\r[?1l>]0;mihir.mahajan@hai-login1:~/Projects/tab-model-eval",,terminal_output
|
| 11 |
+
11,30522,"TERMINAL",0,0,"clear",,terminal_command
|
| 12 |
+
12,30532,"TERMINAL",0,0,"]633;C[H[2J[3J]0;mihir.mahajan@hai-login1:~/Projects/tab-model-eval",,terminal_output
|
| 13 |
+
13,37004,"data/eval/unit_tests/eval_delete.md",0,0,"cat -n jasmine/baselines/diffusion/sample_diffusion.py\n\n\n\n 1\tfrom dataclasses import dataclass\n 2\timport time\n 3\timport os\n 4\timport optax\n 5\t\n 6\timport dm_pix as pix\n 7\timport einops\n 8\timport jax\n 9\timport jax.numpy as jnp\n 10\timport numpy as np\n 11\timport orbax.checkpoint as ocp\n 12\tfrom PIL import Image, ImageDraw\n 13\timport tyro\n 14\tfrom flax import nnx\n 15\t\n 16\tfrom jasmine.models.genie import GenieDiffusion\n 17\tfrom jasmine.utils.dataloader import get_dataloader\n 18\t\n 19\t\n 20\t@dataclass\n 21\tclass Args:\n 22\t # Experiment\n 23\t seed: int = 0\n 24\t seq_len: int = 16\n 25\t image_channels: int = 3\n 26\t image_height: int = 90\n 27\t image_width: int = 160\n 28\t data_dir: str = ""data/coinrun_episodes""\n 29\t checkpoint: str = """"\n 30\t print_action_indices: bool = True\n 31\t output_dir: str = ""gifs/""\n 32\t # Sampling\n 33\t batch_size: int = 1\n 34\t start_frame: int = 1\n 35\t diffusion_denoise_steps: int = 64\n 36\t diffusion_corrupt_context_factor: float = 0.1\n 37\t # Tokenizer checkpoint\n 38\t tokenizer_dim: int = 512\n 39\t tokenizer_ffn_dim: int = 2048\n 40\t latent_patch_dim: int = 32\n 41\t num_patch_latents: int = 1024\n 42\t patch_size: int = 16\n 43\t tokenizer_num_blocks: int = 4\n 44\t tokenizer_num_heads: int = 8\n 45\t # LAM checkpoint\n 46\t lam_dim: int = 512\n 47\t lam_ffn_dim: int = 2048\n 48\t latent_action_dim: int = 32\n 49\t num_actions: int = 6\n 50\t lam_patch_size: int = 16\n 51\t lam_num_blocks: int = 4\n 52\t lam_num_heads: int = 8\n 53\t use_gt_actions: bool = False\n 54\t # Dynamics checkpoint\n 55\t dyna_dim: int = 512\n 56\t dyna_ffn_dim: int = 2048\n 57\t dyna_num_blocks: int = 6\n 58\t dyna_num_heads: int = 8\n 59\t param_dtype = jnp.float32\n 60\t dtype = jnp.bfloat16\n 61\t use_flash_attention: bool = True\n 62\t\n 63\t\n 64\targs = tyro.cli(Args)\n 65\t\n 66\tif __name__ == ""__main__"":\n 67\t """"""\n 68\t Dimension keys:\n 69\t B: batch size\n 70\t T: number of input (conditioning) frames\n 71\t N: number of patches per frame\n 72\t S: sequence length\n 73\t H: height\n 74\t W: width\n 75\t E: B * (S - 1)\n 76\t """"""\n 77\t jax.distributed.initialize()\n 78\t\n 79\t rng = jax.random.key(args.seed)\n 80\t\n 81\t # --- Load Genie checkpoint ---\n 82\t rngs = nnx.Rngs(rng)\n 83\t genie = GenieDiffusion(\n 84\t # Tokenizer\n 85\t in_dim=args.image_channels,\n 86\t tokenizer_dim=args.tokenizer_dim,\n 87\t tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n 88\t latent_patch_dim=args.latent_patch_dim,\n 89\t num_patch_latents=args.num_patch_latents,\n 90\t patch_size=args.patch_size,\n 91\t tokenizer_num_blocks=args.tokenizer_num_blocks,\n 92\t tokenizer_num_heads=args.tokenizer_num_heads,\n 93\t # LAM\n 94\t lam_dim=args.lam_dim,\n 95\t lam_ffn_dim=args.lam_ffn_dim,\n 96\t latent_action_dim=args.latent_action_dim,\n 97\t num_actions=args.num_actions,\n 98\t lam_patch_size=args.lam_patch_size,\n 99\t lam_num_blocks=args.lam_num_blocks,\n 100\t lam_num_heads=args.lam_num_heads,\n 101\t lam_co_train=False,\n 102\t use_gt_actions=args.use_gt_actions,\n 103\t # Dynamics\n 104\t dyna_dim=args.dyna_dim,\n 105\t dyna_ffn_dim=args.dyna_ffn_dim,\n 106\t dyna_num_blocks=args.dyna_num_blocks,\n 107\t dyna_num_heads=args.dyna_num_heads,\n 108\t param_dtype=args.param_dtype,\n 109\t dtype=args.dtype,\n 110\t use_flash_attention=args.use_flash_attention,\n 111\t diffusion_denoise_steps=args.diffusion_denoise_steps,\n 112\t # FIXME (f.srambical): implement spatiotemporal KV caching and set decode=True\n 113\t decode=False,\n 114\t rngs=rngs,\n 115\t )\n 116\t\n 117\t # Need to delete lam decoder for checkpoint loading\n 118\t if not args.use_gt_actions:\n 119\t assert genie.lam is not None\n 120\t del genie.lam.decoder\n 121\t\n 122\t handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n 123\t handler_registry.add(\n 124\t ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n 125\t )\n 126\t handler_registry.add(\n 127\t ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n 128\t )\n 129\t checkpoint_options = ocp.CheckpointManagerOptions(\n 130\t step_format_fixed_length=6,\n 131\t )\n 132\t checkpoint_manager = ocp.CheckpointManager(\n 133\t args.checkpoint,\n 134\t options=checkpoint_options,\n 135\t handler_registry=handler_registry,\n 136\t )\n 137\t\n 138\t dummy_tx = optax.adamw(\n 139\t learning_rate=optax.linear_schedule(0.0001, 0.0001, 10000),\n 140\t b1=0.9,\n 141\t b2=0.9,\n 142\t weight_decay=1e-4,\n 143\t mu_dtype=args.dtype,\n 144\t )\n 145\t dummy_optimizer = nnx.ModelAndOptimizer(genie, dummy_tx)\n 146\t\n 147\t abstract_optimizer = nnx.eval_shape(lambda: dummy_optimizer)\n 148\t abstract_optimizer_state = nnx.state(abstract_optimizer)\n 149\t restored = checkpoint_manager.restore(\n 150\t checkpoint_manager.latest_step(),\n 151\t args=ocp.args.Composite(\n 152\t model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n 153\t ),\n 154\t )\n 155\t restored_optimizer_state = restored[""model_state""]\n 156\t nnx.update(dummy_optimizer, restored_optimizer_state)\n 157\t\n 158\t # --- Define sampling function ---\n 159\t def _sampling_fn(model: GenieDiffusion, batch: dict) -> jax.Array:\n 160\t """"""Runs Genie.sample with pre-defined generation hyper-parameters.""""""\n 161\t frames = model.sample(\n 162\t batch,\n 163\t args.seq_len,\n 164\t args.diffusion_denoise_steps,\n 165\t args.diffusion_corrupt_context_factor,\n 166\t )\n 167\t return frames\n 168\t\n 169\t # --- Define autoregressive sampling loop ---\n 170\t def _autoreg_sample(\n 171\t genie: GenieDiffusion, rng: jax.Array, batch: dict\n 172\t ) -> jax.Array:\n 173\t batch[""videos""] = batch[""videos""][:, : args.start_frame]\n 174\t batch[""rng""] = rng\n 175\t generated_vid_BSHWC = _sampling_fn(genie, batch)\n 176\t return generated_vid_BSHWC\n 177\t\n 178\t # --- Get video + latent actions ---\n 179\t array_record_files = [\n 180\t os.path.join(args.data_dir, x)\n 181\t for x in os.listdir(args.data_dir)\n 182\t if x.endswith("".array_record"")\n 183\t ]\n 184\t dataloader = get_dataloader(\n 185\t array_record_files,\n 186\t args.seq_len,\n 187\t args.batch_size,\n 188\t args.image_height,\n 189\t args.image_width,\n 190\t args.image_channels,\n 191\t # We don't use workers in order to avoid grain shutdown issues (https://github.com/google/grain/issues/398)\n 192\t num_workers=0,\n 193\t prefetch_buffer_size=1,\n 194\t seed=args.seed,\n 195\t )\n 196\t dataloader = iter(dataloader)\n 197\t batch = next(dataloader)\n 198\t gt_video = jnp.asarray(batch[""videos""], dtype=jnp.float32) / 255.0\n 199\t print(f""DEBUG: gt_video shape: {gt_video.shape}"")\n 200\t batch[""videos""] = gt_video.astype(args.dtype)\n 201\t # Get latent actions for all videos in the batch\n 202\t action_batch_E = None\n 203\t if not args.use_gt_actions:\n 204\t action_batch_E = genie.vq_encode(batch, training=False)\n 205\t batch[""latent_actions""] = action_batch_E\n 206\t print(f""DEBUG: action_batch_E shape: {action_batch_E.shape}"")\n 207\t\n 208\t # --- Sample + evaluate video ---\n 209\t recon_video_BSHWC = _autoreg_sample(genie, rng, batch)\n 210\t print(f""DEBUG: recon_video_BSHWC shape: {recon_video_BSHWC.shape}"")\n 211\t recon_video_BSHWC = recon_video_BSHWC.astype(jnp.float32)\n 212\t print(f""DEBUG: recon_video_BSHWC shape: {recon_video_BSHWC.shape}"")\n 213\t\n 214\t gt = gt_video.clip(0, 1)[:, args.start_frame :]\n 215\t print(f""DEBUG: gt shape: {gt.shape}"")\n 216\t recon = recon_video_BSHWC.clip(0, 1)[:, args.start_frame :]\n 217\t print(f""DEBUG: recon shape: {recon.shape}"")\n 218\t\n 219\t ssim_vmap = jax.vmap(pix.ssim, in_axes=(0, 0))\n 220\t psnr_vmap = jax.vmap(pix.psnr, in_axes=(0, 0))\n 221\t ssim = jnp.asarray(ssim_vmap(gt, recon))\n 222\t psnr = jnp.asarray(psnr_vmap(gt, recon))\n 223\t per_frame_ssim = ssim.mean(0)\n 224\t per_frame_psnr = psnr.mean(0)\n 225\t avg_ssim = ssim.mean()\n 226\t avg_psnr = psnr.mean()\n 227\t\n 228\t print(""Per-frame SSIM:\n"", per_frame_ssim)\n 229\t print(""Per-frame PSNR:\n"", per_frame_psnr)\n 230\t\n 231\t print(f""SSIM: {avg_ssim}"")\n 232\t print(f""PSNR: {avg_psnr}"")\n 233\t\n 234\t # --- Construct video ---\n 235\t true_videos = (gt_video * 255).astype(np.uint8)\n 236\t pred_videos = (recon_video_BSHWC * 255).astype(np.uint8)\n 237\t video_comparison = np.zeros((2, *recon_video_BSHWC.shape), dtype=np.uint8)\n 238\t video_comparison[0] = true_videos[:, : args.seq_len]\n 239\t video_comparison[1] = pred_videos\n 240\t frames = einops.rearrange(video_comparison, ""n b t h w c -> t (b h) (n w) c"")\n 241\t\n 242\t # --- Save video ---\n 243\t imgs = [Image.fromarray(img) for img in frames]\n 244\t # Write actions on each frame, on each row (i.e., for each video in the batch, on the GT row)\n 245\t B = batch[""videos""].shape[0]\n 246\t if action_batch_E is not None:\n 247\t action_batch_BSm11 = jnp.reshape(action_batch_E, (B, args.seq_len - 1, 1))\n 248\t else:\n 249\t action_batch_BSm11 = jnp.reshape(\n 250\t batch[""actions""][:, :-1], (B, args.seq_len - 1, 1)\n 251\t )\n 252\t for t, img in enumerate(imgs[1:]):\n 253\t d = ImageDraw.Draw(img)\n 254\t for row in range(B):\n 255\t if args.print_action_indices:\n 256\t action = action_batch_BSm11[row, t, 0]\n 257\t y_offset = row * batch[""videos""].shape[2] + 2\n 258\t d.text((2, y_offset), f""{action}"", fill=255)\n 259\t\n 260\t os.makedirs(args.output_dir, exist_ok=True)\n 261\t imgs[0].save(\n 262\t os.path.join(args.output_dir, f""generation_{time.time()}.gif""),\n 263\t save_all=True,\n 264\t append_images=imgs[1:],\n 265\t duration=250,\n 266\t loop=0,\n 267\t )\n 268\t # Save predicted videos as PNG image with all frames (skipping first 4) for each item in batch\n 269\t skip = 4\n 270\t # pred_videos: (B, T, H, W, C)\n 271\t B, T, H, W, C = pred_videos.shape\n 272\t\n 273\t for i in range(B):\n 274\t # Predicted\n 275\t pred_strip = np.concatenate(\n 276\t [pred_videos[i, t] for t in range(skip, args.seq_len)], axis=1\n 277\t ) # resulting shape: (H, (T-skip)*W, C)\n 278\t pred_img = Image.fromarray(pred_strip)\n 279\t\n 280\t # Write actions on each frame for the predicted strip\n 281\t if args.print_action_indices:\n 282\t d = ImageDraw.Draw(pred_img)\n 283\t for t in range(skip, args.seq_len):\n 284\t if action_batch_E is not None:\n 285\t action = int(action_batch_BSm11[i, t - 1, 0])\n 286\t else:\n 287\t action = int(batch[""actions""][i, t - 1])\n 288\t x_offset = (t - skip) * pred_videos.shape[3] + 2\n 289\t y_offset = 2\n 290\t d.text((x_offset, y_offset), f""{action}"", fill=255)\n 291\t\n 292\t pred_img.save(\n 293\t os.path.join(\n 294\t args.output_dir, f""generation-{args.seed}-sample{i}-diffusion.png""\n 295\t )\n 296\t )\n 297\t # GT\n 298\t gt_strip = np.concatenate(\n 299\t [true_videos[i, t] for t in range(skip, args.seq_len)], axis=1\n 300\t ) # resulting shape: (H, (T-skip)*W, C)\n 301\t gt_img = Image.fromarray(gt_strip)\n 302\t gt_img.save(\n 303\t os.path.join(args.output_dir, f""generation-{args.seed}-sample{i}-gt.png"")\n 304\t )\n\n\n\n\n# Assistant\n```bash\ncat -n jasmine/baselines/diffusion/sample_diffusion.py | sed -n '195,206p'\n```\n\n 195 )\n 196 dataloader = iter(dataloader)\n 197 batch = next(dataloader)\n 198 gt_video = jnp.asarray(batch[""videos""], dtype=jnp.float32) / 255.0\n 199 print(f""DEBUG: gt_video shape: {gt_video.shape}"")\n 200 batch[""videos""] = gt_video.astype(args.dtype)\n 201 # Get latent actions for all videos in the batch\n 202 action_batch_E = None\n 203 if not args.use_gt_actions:\n 204 action_batch_E = genie.vq_encode(batch, training=False)\n 205 batch[""latent_actions""] = action_batch_E\n 206 print(f""DEBUG: action_batch_E shape: {action_batch_E.shape}"")\n\n\nsed -i '199,199d' jasmine/baselines/diffusion/sample_diffusion.py\n\n\nsed -i '205,205d' jasmine/baselines/diffusion/sample_diffusion.py\n\n\n\ncat -n jasmine/baselines/diffusion/sample_diffusion.py | sed -n '205,216p'\n\n\n\n 205\n 206 # --- Sample + evaluate video ---\n 207 recon_video_BSHWC = _autoreg_sample(genie, rng, batch)\n 208 print(f""DEBUG: recon_video_BSHWC shape: {recon_video_BSHWC.shape}"")\n 209 recon_video_BSHWC = recon_video_BSHWC.astype(jnp.float32)\n 210 print(f""DEBUG: recon_video_BSHWC shape: {recon_video_BSHWC.shape}"")\n 211\n 212 gt = gt_video.clip(0, 1)[:, args.start_frame :]\n 213 print(f""DEBUG: gt shape: {gt.shape}"")\n 214 recon = recon_video_BSHWC.clip(0, 1)[:, args.start_frame :]\n 215 print(f""DEBUG: recon shape: {recon.shape}"")\n 216\n\n\nsed -i '208,208d' jasmine/baselines/diffusion/sample_diffusion.py\n\nsed -i '209,209d' jasmine/baselines/diffusion/sample_diffusion.py\n\nsed -i '211,211d' jasmine/baselines/diffusion/sample_diffusion.py\n\nsed -i '212,212d' jasmine/baselines/diffusion/sample_diffusion.py",markdown,tab
|
| 14 |
+
14,39521,"data/eval/unit_tests/eval_delete.md",14389,0,"",markdown,selection_mouse
|
| 15 |
+
15,39522,"data/eval/unit_tests/eval_delete.md",14388,0,"",markdown,selection_command
|
| 16 |
+
16,40601,"data/eval/unit_tests/eval_delete.md",14324,65,"sed -i '212,212d' jasmine/baselines/diffusion/sample_diffusion.py",markdown,selection_command
|
| 17 |
+
17,42435,"data/eval/unit_tests/eval_delete.md",14388,0,"",markdown,selection_command
|
| 18 |
+
18,44071,"data/eval/unit_tests/eval_delete.md",14324,65,"sed -i '212,212d' jasmine/baselines/diffusion/sample_diffusion.py",markdown,selection_command
|
| 19 |
+
19,45309,"data/eval/unit_tests/eval_delete.md",14324,65,"sed -i '212,212d' jasmine/baselines/diffusion/sample_diffusion.py",markdown,selection_command
|
| 20 |
+
20,46093,"data/eval/unit_tests/eval_delete.md",0,14389,"cat -n jasmine/baselines/diffusion/sample_diffusion.py\n\n\n\n 1\tfrom dataclasses import dataclass\n 2\timport time\n 3\timport os\n 4\timport optax\n 5\t\n 6\timport dm_pix as pix\n 7\timport einops\n 8\timport jax\n 9\timport jax.numpy as jnp\n 10\timport numpy as np\n 11\timport orbax.checkpoint as ocp\n 12\tfrom PIL import Image, ImageDraw\n 13\timport tyro\n 14\tfrom flax import nnx\n 15\t\n 16\tfrom jasmine.models.genie import GenieDiffusion\n 17\tfrom jasmine.utils.dataloader import get_dataloader\n 18\t\n 19\t\n 20\t@dataclass\n 21\tclass Args:\n 22\t # Experiment\n 23\t seed: int = 0\n 24\t seq_len: int = 16\n 25\t image_channels: int = 3\n 26\t image_height: int = 90\n 27\t image_width: int = 160\n 28\t data_dir: str = ""data/coinrun_episodes""\n 29\t checkpoint: str = """"\n 30\t print_action_indices: bool = True\n 31\t output_dir: str = ""gifs/""\n 32\t # Sampling\n 33\t batch_size: int = 1\n 34\t start_frame: int = 1\n 35\t diffusion_denoise_steps: int = 64\n 36\t diffusion_corrupt_context_factor: float = 0.1\n 37\t # Tokenizer checkpoint\n 38\t tokenizer_dim: int = 512\n 39\t tokenizer_ffn_dim: int = 2048\n 40\t latent_patch_dim: int = 32\n 41\t num_patch_latents: int = 1024\n 42\t patch_size: int = 16\n 43\t tokenizer_num_blocks: int = 4\n 44\t tokenizer_num_heads: int = 8\n 45\t # LAM checkpoint\n 46\t lam_dim: int = 512\n 47\t lam_ffn_dim: int = 2048\n 48\t latent_action_dim: int = 32\n 49\t num_actions: int = 6\n 50\t lam_patch_size: int = 16\n 51\t lam_num_blocks: int = 4\n 52\t lam_num_heads: int = 8\n 53\t use_gt_actions: bool = False\n 54\t # Dynamics checkpoint\n 55\t dyna_dim: int = 512\n 56\t dyna_ffn_dim: int = 2048\n 57\t dyna_num_blocks: int = 6\n 58\t dyna_num_heads: int = 8\n 59\t param_dtype = jnp.float32\n 60\t dtype = jnp.bfloat16\n 61\t use_flash_attention: bool = True\n 62\t\n 63\t\n 64\targs = tyro.cli(Args)\n 65\t\n 66\tif __name__ == ""__main__"":\n 67\t """"""\n 68\t Dimension keys:\n 69\t B: batch size\n 70\t T: number of input (conditioning) frames\n 71\t N: number of patches per frame\n 72\t S: sequence length\n 73\t H: height\n 74\t W: width\n 75\t E: B * (S - 1)\n 76\t """"""\n 77\t jax.distributed.initialize()\n 78\t\n 79\t rng = jax.random.key(args.seed)\n 80\t\n 81\t # --- Load Genie checkpoint ---\n 82\t rngs = nnx.Rngs(rng)\n 83\t genie = GenieDiffusion(\n 84\t # Tokenizer\n 85\t in_dim=args.image_channels,\n 86\t tokenizer_dim=args.tokenizer_dim,\n 87\t tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n 88\t latent_patch_dim=args.latent_patch_dim,\n 89\t num_patch_latents=args.num_patch_latents,\n 90\t patch_size=args.patch_size,\n 91\t tokenizer_num_blocks=args.tokenizer_num_blocks,\n 92\t tokenizer_num_heads=args.tokenizer_num_heads,\n 93\t # LAM\n 94\t lam_dim=args.lam_dim,\n 95\t lam_ffn_dim=args.lam_ffn_dim,\n 96\t latent_action_dim=args.latent_action_dim,\n 97\t num_actions=args.num_actions,\n 98\t lam_patch_size=args.lam_patch_size,\n 99\t lam_num_blocks=args.lam_num_blocks,\n 100\t lam_num_heads=args.lam_num_heads,\n 101\t lam_co_train=False,\n 102\t use_gt_actions=args.use_gt_actions,\n 103\t # Dynamics\n 104\t dyna_dim=args.dyna_dim,\n 105\t dyna_ffn_dim=args.dyna_ffn_dim,\n 106\t dyna_num_blocks=args.dyna_num_blocks,\n 107\t dyna_num_heads=args.dyna_num_heads,\n 108\t param_dtype=args.param_dtype,\n 109\t dtype=args.dtype,\n 110\t use_flash_attention=args.use_flash_attention,\n 111\t diffusion_denoise_steps=args.diffusion_denoise_steps,\n 112\t # FIXME (f.srambical): implement spatiotemporal KV caching and set decode=True\n 113\t decode=False,\n 114\t rngs=rngs,\n 115\t )\n 116\t\n 117\t # Need to delete lam decoder for checkpoint loading\n 118\t if not args.use_gt_actions:\n 119\t assert genie.lam is not None\n 120\t del genie.lam.decoder\n 121\t\n 122\t handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n 123\t handler_registry.add(\n 124\t ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n 125\t )\n 126\t handler_registry.add(\n 127\t ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n 128\t )\n 129\t checkpoint_options = ocp.CheckpointManagerOptions(\n 130\t step_format_fixed_length=6,\n 131\t )\n 132\t checkpoint_manager = ocp.CheckpointManager(\n 133\t args.checkpoint,\n 134\t options=checkpoint_options,\n 135\t handler_registry=handler_registry,\n 136\t )\n 137\t\n 138\t dummy_tx = optax.adamw(\n 139\t learning_rate=optax.linear_schedule(0.0001, 0.0001, 10000),\n 140\t b1=0.9,\n 141\t b2=0.9,\n 142\t weight_decay=1e-4,\n 143\t mu_dtype=args.dtype,\n 144\t )\n 145\t dummy_optimizer = nnx.ModelAndOptimizer(genie, dummy_tx)\n 146\t\n 147\t abstract_optimizer = nnx.eval_shape(lambda: dummy_optimizer)\n 148\t abstract_optimizer_state = nnx.state(abstract_optimizer)\n 149\t restored = checkpoint_manager.restore(\n 150\t checkpoint_manager.latest_step(),\n 151\t args=ocp.args.Composite(\n 152\t model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n 153\t ),\n 154\t )\n 155\t restored_optimizer_state = restored[""model_state""]\n 156\t nnx.update(dummy_optimizer, restored_optimizer_state)\n 157\t\n 158\t # --- Define sampling function ---\n 159\t def _sampling_fn(model: GenieDiffusion, batch: dict) -> jax.Array:\n 160\t """"""Runs Genie.sample with pre-defined generation hyper-parameters.""""""\n 161\t frames = model.sample(\n 162\t batch,\n 163\t args.seq_len,\n 164\t args.diffusion_denoise_steps,\n 165\t args.diffusion_corrupt_context_factor,\n 166\t )\n 167\t return frames\n 168\t\n 169\t # --- Define autoregressive sampling loop ---\n 170\t def _autoreg_sample(\n 171\t genie: GenieDiffusion, rng: jax.Array, batch: dict\n 172\t ) -> jax.Array:\n 173\t batch[""videos""] = batch[""videos""][:, : args.start_frame]\n 174\t batch[""rng""] = rng\n 175\t generated_vid_BSHWC = _sampling_fn(genie, batch)\n 176\t return generated_vid_BSHWC\n 177\t\n 178\t # --- Get video + latent actions ---\n 179\t array_record_files = [\n 180\t os.path.join(args.data_dir, x)\n 181\t for x in os.listdir(args.data_dir)\n 182\t if x.endswith("".array_record"")\n 183\t ]\n 184\t dataloader = get_dataloader(\n 185\t array_record_files,\n 186\t args.seq_len,\n 187\t args.batch_size,\n 188\t args.image_height,\n 189\t args.image_width,\n 190\t args.image_channels,\n 191\t # We don't use workers in order to avoid grain shutdown issues (https://github.com/google/grain/issues/398)\n 192\t num_workers=0,\n 193\t prefetch_buffer_size=1,\n 194\t seed=args.seed,\n 195\t )\n 196\t dataloader = iter(dataloader)\n 197\t batch = next(dataloader)\n 198\t gt_video = jnp.asarray(batch[""videos""], dtype=jnp.float32) / 255.0\n 199\t print(f""DEBUG: gt_video shape: {gt_video.shape}"")\n 200\t batch[""videos""] = gt_video.astype(args.dtype)\n 201\t # Get latent actions for all videos in the batch\n 202\t action_batch_E = None\n 203\t if not args.use_gt_actions:\n 204\t action_batch_E = genie.vq_encode(batch, training=False)\n 205\t batch[""latent_actions""] = action_batch_E\n 206\t print(f""DEBUG: action_batch_E shape: {action_batch_E.shape}"")\n 207\t\n 208\t # --- Sample + evaluate video ---\n 209\t recon_video_BSHWC = _autoreg_sample(genie, rng, batch)\n 210\t print(f""DEBUG: recon_video_BSHWC shape: {recon_video_BSHWC.shape}"")\n 211\t recon_video_BSHWC = recon_video_BSHWC.astype(jnp.float32)\n 212\t print(f""DEBUG: recon_video_BSHWC shape: {recon_video_BSHWC.shape}"")\n 213\t\n 214\t gt = gt_video.clip(0, 1)[:, args.start_frame :]\n 215\t print(f""DEBUG: gt shape: {gt.shape}"")\n 216\t recon = recon_video_BSHWC.clip(0, 1)[:, args.start_frame :]\n 217\t print(f""DEBUG: recon shape: {recon.shape}"")\n 218\t\n 219\t ssim_vmap = jax.vmap(pix.ssim, in_axes=(0, 0))\n 220\t psnr_vmap = jax.vmap(pix.psnr, in_axes=(0, 0))\n 221\t ssim = jnp.asarray(ssim_vmap(gt, recon))\n 222\t psnr = jnp.asarray(psnr_vmap(gt, recon))\n 223\t per_frame_ssim = ssim.mean(0)\n 224\t per_frame_psnr = psnr.mean(0)\n 225\t avg_ssim = ssim.mean()\n 226\t avg_psnr = psnr.mean()\n 227\t\n 228\t print(""Per-frame SSIM:\n"", per_frame_ssim)\n 229\t print(""Per-frame PSNR:\n"", per_frame_psnr)\n 230\t\n 231\t print(f""SSIM: {avg_ssim}"")\n 232\t print(f""PSNR: {avg_psnr}"")\n 233\t\n 234\t # --- Construct video ---\n 235\t true_videos = (gt_video * 255).astype(np.uint8)\n 236\t pred_videos = (recon_video_BSHWC * 255).astype(np.uint8)\n 237\t video_comparison = np.zeros((2, *recon_video_BSHWC.shape), dtype=np.uint8)\n 238\t video_comparison[0] = true_videos[:, : args.seq_len]\n 239\t video_comparison[1] = pred_videos\n 240\t frames = einops.rearrange(video_comparison, ""n b t h w c -> t (b h) (n w) c"")\n 241\t\n 242\t # --- Save video ---\n 243\t imgs = [Image.fromarray(img) for img in frames]\n 244\t # Write actions on each frame, on each row (i.e., for each video in the batch, on the GT row)\n 245\t B = batch[""videos""].shape[0]\n 246\t if action_batch_E is not None:\n 247\t action_batch_BSm11 = jnp.reshape(action_batch_E, (B, args.seq_len - 1, 1))\n 248\t else:\n 249\t action_batch_BSm11 = jnp.reshape(\n 250\t batch[""actions""][:, :-1], (B, args.seq_len - 1, 1)\n 251\t )\n 252\t for t, img in enumerate(imgs[1:]):\n 253\t d = ImageDraw.Draw(img)\n 254\t for row in range(B):\n 255\t if args.print_action_indices:\n 256\t action = action_batch_BSm11[row, t, 0]\n 257\t y_offset = row * batch[""videos""].shape[2] + 2\n 258\t d.text((2, y_offset), f""{action}"", fill=255)\n 259\t\n 260\t os.makedirs(args.output_dir, exist_ok=True)\n 261\t imgs[0].save(\n 262\t os.path.join(args.output_dir, f""generation_{time.time()}.gif""),\n 263\t save_all=True,\n 264\t append_images=imgs[1:],\n 265\t duration=250,\n 266\t loop=0,\n 267\t )\n 268\t # Save predicted videos as PNG image with all frames (skipping first 4) for each item in batch\n 269\t skip = 4\n 270\t # pred_videos: (B, T, H, W, C)\n 271\t B, T, H, W, C = pred_videos.shape\n 272\t\n 273\t for i in range(B):\n 274\t # Predicted\n 275\t pred_strip = np.concatenate(\n 276\t [pred_videos[i, t] for t in range(skip, args.seq_len)], axis=1\n 277\t ) # resulting shape: (H, (T-skip)*W, C)\n 278\t pred_img = Image.fromarray(pred_strip)\n 279\t\n 280\t # Write actions on each frame for the predicted strip\n 281\t if args.print_action_indices:\n 282\t d = ImageDraw.Draw(pred_img)\n 283\t for t in range(skip, args.seq_len):\n 284\t if action_batch_E is not None:\n 285\t action = int(action_batch_BSm11[i, t - 1, 0])\n 286\t else:\n 287\t action = int(batch[""actions""][i, t - 1])\n 288\t x_offset = (t - skip) * pred_videos.shape[3] + 2\n 289\t y_offset = 2\n 290\t d.text((x_offset, y_offset), f""{action}"", fill=255)\n 291\t\n 292\t pred_img.save(\n 293\t os.path.join(\n 294\t args.output_dir, f""generation-{args.seed}-sample{i}-diffusion.png""\n 295\t )\n 296\t )\n 297\t # GT\n 298\t gt_strip = np.concatenate(\n 299\t [true_videos[i, t] for t in range(skip, args.seq_len)], axis=1\n 300\t ) # resulting shape: (H, (T-skip)*W, C)\n 301\t gt_img = Image.fromarray(gt_strip)\n 302\t gt_img.save(\n 303\t os.path.join(args.output_dir, f""generation-{args.seed}-sample{i}-gt.png"")\n 304\t )\n\n\n\n\n# Assistant\n```bash\ncat -n jasmine/baselines/diffusion/sample_diffusion.py | sed -n '195,206p'\n```\n\n 195 )\n 196 dataloader = iter(dataloader)\n 197 batch = next(dataloader)\n 198 gt_video = jnp.asarray(batch[""videos""], dtype=jnp.float32) / 255.0\n 199 print(f""DEBUG: gt_video shape: {gt_video.shape}"")\n 200 batch[""videos""] = gt_video.astype(args.dtype)\n 201 # Get latent actions for all videos in the batch\n 202 action_batch_E = None\n 203 if not args.use_gt_actions:\n 204 action_batch_E = genie.vq_encode(batch, training=False)\n 205 batch[""latent_actions""] = action_batch_E\n 206 print(f""DEBUG: action_batch_E shape: {action_batch_E.shape}"")\n\n\nsed -i '199,199d' jasmine/baselines/diffusion/sample_diffusion.py\n\n\nsed -i '205,205d' jasmine/baselines/diffusion/sample_diffusion.py\n\n\n\ncat -n jasmine/baselines/diffusion/sample_diffusion.py | sed -n '205,216p'\n\n\n\n 205\n 206 # --- Sample + evaluate video ---\n 207 recon_video_BSHWC = _autoreg_sample(genie, rng, batch)\n 208 print(f""DEBUG: recon_video_BSHWC shape: {recon_video_BSHWC.shape}"")\n 209 recon_video_BSHWC = recon_video_BSHWC.astype(jnp.float32)\n 210 print(f""DEBUG: recon_video_BSHWC shape: {recon_video_BSHWC.shape}"")\n 211\n 212 gt = gt_video.clip(0, 1)[:, args.start_frame :]\n 213 print(f""DEBUG: gt shape: {gt.shape}"")\n 214 recon = recon_video_BSHWC.clip(0, 1)[:, args.start_frame :]\n 215 print(f""DEBUG: recon shape: {recon.shape}"")\n 216\n\n\nsed -i '208,208d' jasmine/baselines/diffusion/sample_diffusion.py\n\nsed -i '209,209d' jasmine/baselines/diffusion/sample_diffusion.py\n\nsed -i '211,211d' jasmine/baselines/diffusion/sample_diffusion.py\n\nsed -i '212,212d' jasmine/baselines/diffusion/sample_diffusion.py",markdown,selection_command
|
| 21 |
+
21,102756,"TERMINAL",0,0,"",,terminal_command
|
| 22 |
+
22,216425,"data/eval/unit_tests/eval_delete.md",14389,0,"```\n\n# User\n<stdout>\n</stdout>\n\n",markdown,content
|
| 23 |
+
23,216425,"data/eval/unit_tests/eval_delete.md",14324,0,"# Assistant\n```bash\n",markdown,content
|
| 24 |
+
24,216425,"data/eval/unit_tests/eval_delete.md",14323,0,"```\n\n# User\n<stdout>\n</stdout>\n\n",markdown,content
|
| 25 |
+
25,216425,"data/eval/unit_tests/eval_delete.md",14257,0,"# Assistant\n```bash\n",markdown,content
|
| 26 |
+
26,216425,"data/eval/unit_tests/eval_delete.md",14256,0,"```\n\n# User\n<stdout>\n</stdout>\n\n",markdown,content
|
| 27 |
+
27,216425,"data/eval/unit_tests/eval_delete.md",14190,0,"# Assistant\n```bash\n",markdown,content
|
| 28 |
+
28,216425,"data/eval/unit_tests/eval_delete.md",14189,0,"```\n\n# User\n<stdout>\n</stdout>\n\n",markdown,content
|
| 29 |
+
29,216425,"data/eval/unit_tests/eval_delete.md",14123,0,"# Assistant\n```bash\n",markdown,content
|
| 30 |
+
30,216425,"data/eval/unit_tests/eval_delete.md",14121,0,"</stdout>\n",markdown,content
|
| 31 |
+
31,216425,"data/eval/unit_tests/eval_delete.md",13517,1,"# User\n<stdout>",markdown,content
|
| 32 |
+
32,216425,"data/eval/unit_tests/eval_delete.md",13516,0,"```\n",markdown,content
|
| 33 |
+
33,216425,"data/eval/unit_tests/eval_delete.md",13441,0,"# Assistant\n```bash\n",markdown,content
|
| 34 |
+
34,216425,"data/eval/unit_tests/eval_delete.md",13439,0,"# User\n<stdout>\n</stdout>\n",markdown,content
|
| 35 |
+
35,216425,"data/eval/unit_tests/eval_delete.md",13438,0,"```\n",markdown,content
|
| 36 |
+
36,216425,"data/eval/unit_tests/eval_delete.md",13372,0,"\n# Assistant\n```bash\n",markdown,content
|
| 37 |
+
37,216425,"data/eval/unit_tests/eval_delete.md",13371,0,"# User\n<stdout>\n</stdout>\n",markdown,content
|
| 38 |
+
38,216425,"data/eval/unit_tests/eval_delete.md",13370,0,"```\n",markdown,content
|
| 39 |
+
39,216425,"data/eval/unit_tests/eval_delete.md",13304,0,"# Assistant\n```bash\n",markdown,content
|
| 40 |
+
40,216425,"data/eval/unit_tests/eval_delete.md",13302,0,"</stdout>\n",markdown,content
|
| 41 |
+
41,216425,"data/eval/unit_tests/eval_delete.md",12668,0,"# User\n<stdout>\n",markdown,content
|
| 42 |
+
42,216425,"data/eval/unit_tests/eval_delete.md",12564,1,"</stdout>",markdown,content
|
| 43 |
+
43,216425,"data/eval/unit_tests/eval_delete.md",56,1,"# User\n<stdout>",markdown,content
|
| 44 |
+
44,216425,"data/eval/unit_tests/eval_delete.md",55,0,"```\n",markdown,content
|
| 45 |
+
45,216425,"data/eval/unit_tests/eval_delete.md",0,0,"# Assistant\n```bash\n",markdown,content
|
| 46 |
+
46,216429,"data/eval/unit_tests/eval_delete.md",14817,1,"",markdown,content
|
| 47 |
+
47,216429,"data/eval/unit_tests/eval_delete.md",14721,68,"sed -i '212,212d' jasmine/baselines/diffusion/sample_diffusion.py\n```",markdown,content
|
| 48 |
+
48,379094,"data/eval/unit_tests/eval_delete.md",14700,0,"",markdown,selection_mouse
|
| 49 |
+
49,401814,"TERMINAL",0,0,"clear",,terminal_command
|
| 50 |
+
50,401814,"TERMINAL",0,0,"]633;C[H[2J[3J]0;mihir.mahajan@hai-login1:~/Projects/tab-model-eval",,terminal_output
|
| 51 |
+
51,404526,"TERMINAL",0,0,"clear",,terminal_command
|
| 52 |
+
52,464515,"data/eval/unit_tests/eval_delete.md",0,0,"",markdown,tab
|
| 53 |
+
53,464516,"data/eval/unit_tests/eval_delete.md",14593,0,"",markdown,selection_mouse
|
| 54 |
+
54,464570,"data/eval/unit_tests/eval_delete.md",14592,0,"",markdown,selection_command
|
| 55 |
+
55,793294,"data/eval/handcrafted/add_import_hard.md",0,0,"# Assistant\n```bash\npython src/input_pipeline/validation_to_testcases.py --input_file data/temp/validation.jsonl --output_file data/temp/validation_testcases.jsonl\n```\n\n# User\n<stdout>\nTraceback (most recent call last):\n File ""/fast/home/mihir.mahajan/Projects/tab-model-eval/src/input_pipeline/validation_to_testcases.py"", line 47, in <module>\n args = tyro.cli(Args)\n ^^^^\nNameError: name 'tyro' is not defined\n</stdout>\n\n# Assistant\n```bash\ncat -n src/input_pipeline/validation_to_testcases.py\n```\n\n# User\n<stdout>\n 1 import json\n 2 import os\n 3 from dataclasses import dataclass\n 4\n 5\n 6 @dataclass\n 7 class Args:\n 8 input_file: str = ""data/eval/val/validation.jsonl""\n 9 output_file: str = ""data/eval/val/validation_testcases.jsonl""\n 10 task_name: str = ""validation_set""\n 11\n 12\n 13 def convert_to_incremental_jsonl(line, output_file, task_name):\n 14 data = json.loads(line)\n 15\n 16 system_prompt = data.get(""system_prompt"", """")\n 17 conversations = data.get(""conversations"", [])\n 18\n 19 output_lines = []\n 20 context = []\n 21\n 22 for i, conv in enumerate(conversations):\n 23 # Convert ""Assistant""/""User"" to ""assistant""/""user""\n 24 role = conv[""from""].lower()\n 25 content = conv[""value""]\n 26\n 27 # Add current message to context\n 28 context.append({""role"": role, ""content"": content})\n 29\n 30 if i > 0 and role == ""assistant"":\n 31 task_entry = {\n 32 ""system_prompt"": system_prompt,\n 33 ""task_id"": f""{task_name}/{i-1}"",\n 34 ""context"": context[:-1].copy(),\n 35 ""expected_final_response"": content,\n 36 }\n 37 output_lines.append(task_entry)\n 38\n 39 with open(output_file, ""a"") as f:\n 40 for line in output_lines:\n 41 f.write(json.dumps(line) + ""\n"")\n 42\n 43 print(f""Converted {len(output_lines)} entries to {output_file}"")\n 44\n 45\n 46 if __name__ == ""__main__"":\n 47 args = tyro.cli(Args)\n 48\n 49 if os.path.exists(args.output_file):\n 50 os.remove(args.output_file)\n 51\n 52 # iterate over jsonl file and convert to incremental jsonl\n 53 conversation_id = 0\n 54 with open(args.input_file, ""r"") as f:\n 55 for line in f:\n 56 task_name = f""conversation_{conversation_id}/{args.task_name}""\n 57 convert_to_incremental_jsonl(line, args.output_file, task_name)\n 58 conversation_id += 1\n</stdout>\n\n# Assistant\n```bash\nsed -i '1i\import tyro' src/input_pipeline/validation_to_testcases.py\n```\n\n# User\n<stdout>\n</stdout>\n\n\n# Assistant\n```bash\npython src/input_pipeline/validation_to_testcases.py --input_file data/temp/validation.jsonl --output_file data/temp/validation_testcases.jsonl\n```\n\n# User\n<stdout>\nConverted 7 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 4 entries to data/temp/validation_testcases.jsonl\nConverted 7 entries to data/temp/validation_testcases.jsonl\nConverted 4 entries to data/temp/validation_testcases.jsonl\nConverted 8 entries to data/temp/validation_testcases.jsonl\nConverted 8 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 5 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 2 entries to data/temp/validation_testcases.jsonl\nConverted 4 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 5 entries to data/temp/validation_testcases.jsonl\nConverted 6 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\n</stdout>\n",markdown,tab
|
| 56 |
+
56,821305,"data/eval/handcrafted/add_import_easy.md",0,0,"# Assistant\n```bash\npython src/input_pipeline/validation_to_testcases.py --input_file data/temp/validation.jsonl --output_file data/temp/validation_testcases.jsonl\n```\n\n# User\n<stdout>\nTraceback (most recent call last):\n File ""/fast/home/mihir.mahajan/Projects/tab-model-eval/src/input_pipeline/validation_to_testcases.py"", line 57, in <module>\n convert_to_incremental_jsonl(line, args.output_file, task_name)\n ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File ""/fast/home/mihir.mahajan/Projects/tab-model-eval/src/input_pipeline/validation_to_testcases.py"", line 14, in convert_to_incremental_jsonl\n data = json.loads(line)\n ^^^^\nNameError: name 'json' is not defined. Did you forget to import 'json'?\n</stdout>\n\n\n# Assistant\n```bash\ncat -n src/input_pipeline/validation_to_testcases.py\n```\n\n# User\n<stdout>\n 1 import tyro\n 2 import os\n 3 from dataclasses import dataclass\n 4\n 5\n 6 @dataclass\n 7 class Args:\n 8 input_file: str = ""data/eval/val/validation.jsonl""\n 9 output_file: str = ""data/eval/val/validation_testcases.jsonl""\n 10 task_name: str = ""validation_set""\n 11\n 12\n 13 def convert_to_incremental_jsonl(line, output_file, task_name):\n 14 data = json.loads(line)\n 15\n 16 system_prompt = data.get(""system_prompt"", """")\n 17 conversations = data.get(""conversations"", [])\n 18\n 19 output_lines = []\n 20 context = []\n 21\n 22 for i, conv in enumerate(conversations):\n 23 # Convert ""Assistant""/""User"" to ""assistant""/""user""\n 24 role = conv[""from""].lower()\n 25 content = conv[""value""]\n 26\n 27 # Add current message to context\n 28 context.append({""role"": role, ""content"": content})\n 29\n 30 if i > 0 and role == ""assistant"":\n 31 task_entry = {\n 32 ""system_prompt"": system_prompt,\n 33 ""task_id"": f""{task_name}/{i-1}"",\n 34 ""context"": context[:-1].copy(),\n 35 ""expected_final_response"": content,\n 36 }\n 37 output_lines.append(task_entry)\n 38\n 39 with open(output_file, ""a"") as f:\n 40 for line in output_lines:\n 41 f.write(json.dumps(line) + ""\n"")\n 42\n 43 print(f""Converted {len(output_lines)} entries to {output_file}"")\n 44\n 45\n 46 if __name__ == ""__main__"":\n 47 args = tyro.cli(Args)\n 48\n 49 if os.path.exists(args.output_file):\n 50 os.remove(args.output_file)\n 51\n 52 # iterate over jsonl file and convert to incremental jsonl\n 53 conversation_id = 0\n 54 with open(args.input_file, ""r"") as f:\n 55 for line in f:\n 56 task_name = f""conversation_{conversation_id}/{args.task_name}""\n 57 convert_to_incremental_jsonl(line, args.output_file, task_name)\n 58 conversation_id += 1\n</stdout>\n\n# Assistant\n```bash\nsed -i '1i\import json' src/input_pipeline/validation_to_testcases.py\n```\n\n# User\n<stdout>\n</stdout>\n\n# Assistant\n```bash\npython src/input_pipeline/validation_to_testcases.py --input_file data/temp/validation.jsonl --output_file data/temp/validation_testcases.jsonl\n```\n\n# User\n<stdout>\nConverted 7 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 4 entries to data/temp/validation_testcases.jsonl\nConverted 7 entries to data/temp/validation_testcases.jsonl\nConverted 4 entries to data/temp/validation_testcases.jsonl\nConverted 8 entries to data/temp/validation_testcases.jsonl\nConverted 8 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 5 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 2 entries to data/temp/validation_testcases.jsonl\nConverted 4 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 5 entries to data/temp/validation_testcases.jsonl\nConverted 6 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\n</stdout>\n",markdown,tab
|
| 57 |
+
57,829092,"data/eval/unit_tests/add_import_easy.md",0,0,"# Assistant\n```bash\npython src/input_pipeline/validation_to_testcases.py --input_file data/temp/validation.jsonl --output_file data/temp/validation_testcases.jsonl\n```\n\n# User\n<stdout>\nTraceback (most recent call last):\n File ""/fast/home/mihir.mahajan/Projects/tab-model-eval/src/input_pipeline/validation_to_testcases.py"", line 57, in <module>\n convert_to_incremental_jsonl(line, args.output_file, task_name)\n ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File ""/fast/home/mihir.mahajan/Projects/tab-model-eval/src/input_pipeline/validation_to_testcases.py"", line 14, in convert_to_incremental_jsonl\n data = json.loads(line)\n ^^^^\nNameError: name 'json' is not defined. Did you forget to import 'json'?\n</stdout>\n\n\n# Assistant\n```bash\ncat -n src/input_pipeline/validation_to_testcases.py\n```\n\n# User\n<stdout>\n 1 import tyro\n 2 import os\n 3 from dataclasses import dataclass\n 4\n 5\n 6 @dataclass\n 7 class Args:\n 8 input_file: str = ""data/eval/val/validation.jsonl""\n 9 output_file: str = ""data/eval/val/validation_testcases.jsonl""\n 10 task_name: str = ""validation_set""\n 11\n 12\n 13 def convert_to_incremental_jsonl(line, output_file, task_name):\n 14 data = json.loads(line)\n 15\n 16 system_prompt = data.get(""system_prompt"", """")\n 17 conversations = data.get(""conversations"", [])\n 18\n 19 output_lines = []\n 20 context = []\n 21\n 22 for i, conv in enumerate(conversations):\n 23 # Convert ""Assistant""/""User"" to ""assistant""/""user""\n 24 role = conv[""from""].lower()\n 25 content = conv[""value""]\n 26\n 27 # Add current message to context\n 28 context.append({""role"": role, ""content"": content})\n 29\n 30 if i > 0 and role == ""assistant"":\n 31 task_entry = {\n 32 ""system_prompt"": system_prompt,\n 33 ""task_id"": f""{task_name}/{i-1}"",\n 34 ""context"": context[:-1].copy(),\n 35 ""expected_final_response"": content,\n 36 }\n 37 output_lines.append(task_entry)\n 38\n 39 with open(output_file, ""a"") as f:\n 40 for line in output_lines:\n 41 f.write(json.dumps(line) + ""\n"")\n 42\n 43 print(f""Converted {len(output_lines)} entries to {output_file}"")\n 44\n 45\n 46 if __name__ == ""__main__"":\n 47 args = tyro.cli(Args)\n 48\n 49 if os.path.exists(args.output_file):\n 50 os.remove(args.output_file)\n 51\n 52 # iterate over jsonl file and convert to incremental jsonl\n 53 conversation_id = 0\n 54 with open(args.input_file, ""r"") as f:\n 55 for line in f:\n 56 task_name = f""conversation_{conversation_id}/{args.task_name}""\n 57 convert_to_incremental_jsonl(line, args.output_file, task_name)\n 58 conversation_id += 1\n</stdout>\n\n# Assistant\n```bash\nsed -i '1i\import json' src/input_pipeline/validation_to_testcases.py\n```\n\n# User\n<stdout>\n</stdout>\n\n# Assistant\n```bash\npython src/input_pipeline/validation_to_testcases.py --input_file data/temp/validation.jsonl --output_file data/temp/validation_testcases.jsonl\n```\n\n# User\n<stdout>\nConverted 7 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 4 entries to data/temp/validation_testcases.jsonl\nConverted 7 entries to data/temp/validation_testcases.jsonl\nConverted 4 entries to data/temp/validation_testcases.jsonl\nConverted 8 entries to data/temp/validation_testcases.jsonl\nConverted 8 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 5 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 2 entries to data/temp/validation_testcases.jsonl\nConverted 4 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 5 entries to data/temp/validation_testcases.jsonl\nConverted 6 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\n</stdout>\n",markdown,tab
|
| 58 |
+
58,844708,"data/eval/unit_tests/eval_terminal_invocation.md",0,0,"# Assistant\n```bash\npython src/input_pipeline/validation_to_testcases.py --input_file data/temp/validation.jsonl --output_file data/temp/validation_testcases.jsonl\n```\n\n# User\n<stdout>\nTraceback (most recent call last):\n File ""/fast/home/mihir.mahajan/Projects/tab-model-eval/src/input_pipeline/validation_to_testcases.py"", line 57, in <module>\n convert_to_incremental_jsonl(line, args.output_file, task_name)\n ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File ""/fast/home/mihir.mahajan/Projects/tab-model-eval/src/input_pipeline/validation_to_testcases.py"", line 14, in convert_to_incremental_jsonl\n data = json.loads(line)\n ^^^^\nNameError: name 'json' is not defined. Did you forget to import 'json'?\n</stdout>\n\n\n# Assistant\n```bash\ncat -n src/input_pipeline/validation_to_testcases.py\n```\n\n# User\n<stdout>\n 1 import tyro\n 2 import os\n 3 from dataclasses import dataclass\n 4\n 5\n 6 @dataclass\n 7 class Args:\n 8 input_file: str = ""data/eval/val/validation.jsonl""\n 9 output_file: str = ""data/eval/val/validation_testcases.jsonl""\n 10 task_name: str = ""validation_set""\n 11\n 12\n 13 def convert_to_incremental_jsonl(line, output_file, task_name):\n 14 data = json.loads(line)\n 15\n 16 system_prompt = data.get(""system_prompt"", """")\n 17 conversations = data.get(""conversations"", [])\n 18\n 19 output_lines = []\n 20 context = []\n 21\n 22 for i, conv in enumerate(conversations):\n 23 # Convert ""Assistant""/""User"" to ""assistant""/""user""\n 24 role = conv[""from""].lower()\n 25 content = conv[""value""]\n 26\n 27 # Add current message to context\n 28 context.append({""role"": role, ""content"": content})\n 29\n 30 if i > 0 and role == ""assistant"":\n 31 task_entry = {\n 32 ""system_prompt"": system_prompt,\n 33 ""task_id"": f""{task_name}/{i-1}"",\n 34 ""context"": context[:-1].copy(),\n 35 ""expected_final_response"": content,\n 36 }\n 37 output_lines.append(task_entry)\n 38\n 39 with open(output_file, ""a"") as f:\n 40 for line in output_lines:\n 41 f.write(json.dumps(line) + ""\n"")\n 42\n 43 print(f""Converted {len(output_lines)} entries to {output_file}"")\n 44\n 45\n 46 if __name__ == ""__main__"":\n 47 args = tyro.cli(Args)\n 48\n 49 if os.path.exists(args.output_file):\n 50 os.remove(args.output_file)\n 51\n 52 # iterate over jsonl file and convert to incremental jsonl\n 53 conversation_id = 0\n 54 with open(args.input_file, ""r"") as f:\n 55 for line in f:\n 56 task_name = f""conversation_{conversation_id}/{args.task_name}""\n 57 convert_to_incremental_jsonl(line, args.output_file, task_name)\n 58 conversation_id += 1\n</stdout>\n\n# Assistant\n```bash\nsed -i '1i\import json' src/input_pipeline/validation_to_testcases.py\n```\n\n# User\n<stdout>\n</stdout>\n\n# Assistant\n```bash\npython src/input_pipeline/validation_to_testcases.py --input_file data/temp/validation.jsonl --output_file data/temp/validation_testcases.jsonl\n```\n\n# User\n<stdout>\nConverted 7 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 4 entries to data/temp/validation_testcases.jsonl\nConverted 7 entries to data/temp/validation_testcases.jsonl\nConverted 4 entries to data/temp/validation_testcases.jsonl\nConverted 8 entries to data/temp/validation_testcases.jsonl\nConverted 8 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 5 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 2 entries to data/temp/validation_testcases.jsonl\nConverted 4 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 5 entries to data/temp/validation_testcases.jsonl\nConverted 6 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\nConverted 3 entries to data/temp/validation_testcases.jsonl\n</stdout>\n",markdown,tab
|
| 59 |
+
59,855759,"data/eval/unit_tests/eval_terminal_invocation.md",3197,0,"",markdown,selection_mouse
|
| 60 |
+
60,860607,"data/eval/unit_tests/eval_terminal_invocation.md",3066,0,"",markdown,selection_mouse
|
| 61 |
+
61,861097,"data/eval/unit_tests/eval_terminal_invocation.md",3108,0,"",markdown,selection_mouse
|
| 62 |
+
62,861876,"data/eval/unit_tests/eval_terminal_invocation.md",3120,0,"",markdown,selection_mouse
|
| 63 |
+
63,861879,"data/eval/unit_tests/eval_terminal_invocation.md",3119,0,"",markdown,selection_command
|
| 64 |
+
64,861899,"data/eval/unit_tests/eval_terminal_invocation.md",3119,1,"`",markdown,selection_mouse
|
| 65 |
+
65,861905,"data/eval/unit_tests/eval_terminal_invocation.md",3120,0,"",markdown,selection_command
|
| 66 |
+
66,862508,"data/eval/unit_tests/eval_terminal_invocation.md",3128,0,"",markdown,selection_mouse
|
| 67 |
+
67,862519,"data/eval/unit_tests/eval_terminal_invocation.md",3127,0,"",markdown,selection_command
|
| 68 |
+
68,863436,"data/eval/unit_tests/eval_terminal_invocation.md",3120,0,"",markdown,selection_mouse
|
| 69 |
+
69,863437,"data/eval/unit_tests/eval_terminal_invocation.md",3119,0,"",markdown,selection_command
|
| 70 |
+
70,863920,"data/eval/unit_tests/eval_terminal_invocation.md",3128,0,"",markdown,selection_mouse
|
| 71 |
+
71,863921,"data/eval/unit_tests/eval_terminal_invocation.md",3127,0,"",markdown,selection_command
|
| 72 |
+
72,864352,"data/eval/unit_tests/eval_terminal_invocation.md",3137,0,"",markdown,selection_mouse
|
| 73 |
+
73,864352,"data/eval/unit_tests/eval_terminal_invocation.md",3136,0,"",markdown,selection_command
|
| 74 |
+
74,864767,"data/eval/unit_tests/eval_terminal_invocation.md",3147,0,"",markdown,selection_mouse
|
| 75 |
+
75,864767,"data/eval/unit_tests/eval_terminal_invocation.md",3146,0,"",markdown,selection_command
|
| 76 |
+
76,1068619,"data/eval/unit_tests/eval_terminal_invocation.md",0,0,"",markdown,tab
|
| 77 |
+
77,1068620,"data/eval/unit_tests/eval_terminal_invocation.md",3324,0,"",markdown,selection_mouse
|
| 78 |
+
78,1068629,"data/eval/unit_tests/eval_terminal_invocation.md",3323,0,"",markdown,selection_command
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-386afd79-6aa8-4e11-939b-d2205ae947e71760884674230-2025_10_19-16.38.53.787/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-3b2f337f-0bf2-4eb1-81d0-1152f026aa5c1764086742729-2025_11_25-17.06.01.787/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-3e34f5c4-13f2-481c-a23c-176fc713ea5e1761657198697-2025_10_28-14.13.44.256/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-3ee1c8d9-397d-4186-890e-abbebcb677cd1760442276619-2025_10_14-13.49.47.394/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-41e637ef-086e-4e89-977d-363e70d7d3151764500068578-2025_11_30-11.55.34.47/source.csv
ADDED
|
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,2,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=32\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=05:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_32_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python jasmine/train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1536 \\n --min_lr=0 \\n --max_lr=3e-4 \\n --log_image_interval=250 \\n --log \\n --name=tokenizer-batch-size-scaling-32-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 32-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,tab
|
| 3 |
+
2,4925,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"11:55:34 AM [info] Activating crowd-code\n11:55:34 AM [info] Recording started\n11:55:34 AM [info] Initializing git provider using file system watchers...\n11:55:35 AM [info] Git repository found\n11:55:35 AM [info] Git provider initialized successfully\n11:55:35 AM [info] Initial git state: [object Object]\n",Log,tab
|
| 4 |
+
3,46263,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",0,0,"",shellscript,tab
|
| 5 |
+
4,46267,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",1179,0,"",shellscript,selection_mouse
|
| 6 |
+
5,49573,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=05:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_1_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python jasmine/train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=48 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-1-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 1-node sqrt-lr \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,tab
|
| 7 |
+
6,51445,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",0,0,"",shellscript,tab
|
| 8 |
+
7,52258,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",943,0,"",shellscript,selection_mouse
|
| 9 |
+
8,52986,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",944,0,"",shellscript,selection_command
|
| 10 |
+
9,53443,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",943,1,"",shellscript,content
|
| 11 |
+
10,53514,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",942,1,"",shellscript,content
|
| 12 |
+
11,53710,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",941,1,"",shellscript,content
|
| 13 |
+
12,54083,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",941,0,"1",shellscript,content
|
| 14 |
+
13,54084,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",942,0,"",shellscript,selection_keyboard
|
| 15 |
+
14,54132,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",942,0,"0",shellscript,content
|
| 16 |
+
15,54132,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",943,0,"",shellscript,selection_keyboard
|
| 17 |
+
16,54264,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",943,0,"0",shellscript,content
|
| 18 |
+
17,54264,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",944,0,"",shellscript,selection_keyboard
|
| 19 |
+
18,54481,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",944,0,"0",shellscript,content
|
| 20 |
+
19,54481,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",945,0,"",shellscript,selection_keyboard
|
| 21 |
+
20,56715,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",947,0,"\n --log_checkpoint_interval=1000 \",shellscript,content
|
| 22 |
+
21,58155,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",944,0,"",shellscript,selection_command
|
| 23 |
+
22,61500,"TERMINAL",0,0,"",,terminal_command
|
| 24 |
+
23,86690,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch",0,0,"",shellscript,tab
|
| 25 |
+
24,91839,"TERMINAL",0,0,"",,terminal_focus
|
| 26 |
+
25,92728,"TERMINAL",0,0,"source /home/mihir.mahajan/Projects/tab-model-eval/.venv/bin/activate",,terminal_command
|
| 27 |
+
26,92885,"TERMINAL",0,0,"]633;C]0;mihir.mahajan@hai-login2:~/Projects/tab-model-eval",,terminal_output
|
| 28 |
+
27,94149,"TERMINAL",0,0,"bash",,terminal_focus
|
| 29 |
+
28,121987,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch",0,0,"",shellscript,tab
|
| 30 |
+
29,123074,"TERMINAL",0,0,"bash",,terminal_focus
|
| 31 |
+
30,140621,"TERMINAL",0,0,"diff /home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch /home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",,terminal_command
|
| 32 |
+
31,140656,"TERMINAL",0,0,"]633;C3c3\r\n< #SBATCH --nodes=1\r\n---\r\n> #SBATCH --nodes=32\r\n9,11c9,11\r\n< #SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\r\n< #SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\r\n< #SBATCH --job-name=train_tokenizer_batch_size_scaling_1_node\r\n---\r\n> #SBATCH --output=logs/logs_training/%x_%j.log\r\n> #SBATCH --error=logs/logs_training/%x_%j.log\r\n> #SBATCH --job-name=train_tokenizer_batch_size_scaling_32_node\r\n32c32\r\n< --batch_size=48 \\r\n---\r\n> --batch_size=1536 \\r\n34c34\r\n< --max_lr=1e-4 \\r\n---\r\n> --max_lr=3e-4 \\r\n38,39c38,39\r\n< --name=tokenizer-batch-size-scaling-1-node-sqrt-lr-$slurm_job_id \\r\n< --tags tokenizer batch-size-scaling 1-node sqrt-lr \\r\n---\r\n> --name=tokenizer-batch-size-scaling-32-node-sqrt-lr-$slurm_job_id \\r\n> --tags tokenizer batch-size-scaling 32-node sqrt-lr-scaling \\r\n]0;mihir.mahajan@hai-login2:~/Projects/tab-model-eval",,terminal_output
|
| 33 |
+
32,154994,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch",0,0,"",shellscript,tab
|
| 34 |
+
33,160937,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",0,0,"",shellscript,tab
|
| 35 |
+
34,163667,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch",0,0,"",shellscript,tab
|
| 36 |
+
35,164483,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch",187,0,"",shellscript,selection_mouse
|
| 37 |
+
36,167106,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",0,0,"",shellscript,tab
|
| 38 |
+
37,167763,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",249,0,"",shellscript,selection_mouse
|
| 39 |
+
38,168320,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",261,0,"\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log",shellscript,content
|
| 40 |
+
39,168337,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",262,0,"",shellscript,selection_command
|
| 41 |
+
40,168635,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",217,0,"",shellscript,selection_command
|
| 42 |
+
41,168958,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",171,91,"",shellscript,content
|
| 43 |
+
42,171940,"TERMINAL",0,0,"diff /home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch /home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch",,terminal_command
|
| 44 |
+
43,171941,"TERMINAL",0,0,"]633;C3c3\r\n< #SBATCH --nodes=1\r\n---\r\n> #SBATCH --nodes=32\r\n11c11\r\n< #SBATCH --job-name=train_tokenizer_batch_size_scaling_1_node\r\n---\r\n> #SBATCH --job-name=train_tokenizer_batch_size_scaling_32_node\r\n32c32\r\n< --batch_size=48 \\r\n---\r\n> --batch_size=1536 \\r\n34c34\r\n< --max_lr=1e-4 \\r\n---\r\n> --max_lr=3e-4 \\r\n38,39c38,39\r\n< --name=tokenizer-batch-size-scaling-1-node-sqrt-lr-$slurm_job_id \\r\n< --tags tokenizer batch-size-scaling 1-node sqrt-lr \\r\n---\r\n> --name=tokenizer-batch-size-scaling-32-node-sqrt-lr-$slurm_job_id \\r\n> --tags tokenizer batch-size-scaling 32-node sqrt-lr-scaling \\r\n]0;mihir.mahajan@hai-login2:~/Projects/tab-model-eval",,terminal_output
|
| 45 |
+
44,198176,"TERMINAL",0,0,"diff /home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch /home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_16_nodes.sbatch",,terminal_command
|
| 46 |
+
45,198187,"TERMINAL",0,0,"]633;C3c3\r\n< #SBATCH --nodes=1\r\n---\r\n> #SBATCH --nodes=16\r\n11c11\r\n< #SBATCH --job-name=train_tokenizer_batch_size_scaling_1_node\r\n---\r\n> #SBATCH --job-name=train_tokenizer_batch_size_scaling_16_node\r\n32c32\r\n< --batch_size=48 \\r\n---\r\n> --batch_size=768 \\r\n34c34\r\n< --max_lr=1e-4 \\r\n---\r\n> --max_lr=4e-4 \\r\n38,39c38,39\r\n< --name=tokenizer-batch-size-scaling-1-node-sqrt-lr-$slurm_job_id \\r\n< --tags tokenizer batch-size-scaling 1-node sqrt-lr \\r\n---\r\n> --name=tokenizer-batch-size-scaling-16-node-sqrt-lr-$slurm_job_id \\r\n> --tags tokenizer batch-size-scaling 16-node sqrt-lr-scaling \\r\n]0;mihir.mahajan@hai-login2:~/Projects/tab-model-eval",,terminal_output
|
| 47 |
+
46,208288,"TERMINAL",0,0,"diff /home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch /home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_8_nodes.sbatch",,terminal_command
|
| 48 |
+
47,208304,"TERMINAL",0,0,"]633;C3c3\r\n< #SBATCH --nodes=1\r\n---\r\n> #SBATCH --nodes=8\r\n11c11\r\n< #SBATCH --job-name=train_tokenizer_batch_size_scaling_1_node\r\n---\r\n> #SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\r\n32c32\r\n< --batch_size=48 \\r\n---\r\n> --batch_size=384 \\r\n34c34\r\n< --max_lr=1e-4 \\r\n---\r\n> --max_lr=3e-4 \\r\n38,39c38,39\r\n< --name=tokenizer-batch-size-scaling-1-node-sqrt-lr-$slurm_job_id \\r\n< --tags tokenizer batch-size-scaling 1-node sqrt-lr \\r\n---\r\n> --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\r\n> --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\r\n]0;mihir.mahajan@hai-login2:~/Projects/tab-model-eval",,terminal_output
|
| 49 |
+
48,215509,"TERMINAL",0,0,"diff /home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch /home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_4_nodes.sbatch",,terminal_command
|
| 50 |
+
49,215540,"TERMINAL",0,0,"]633;C3c3\r\n< #SBATCH --nodes=1\r\n---\r\n> #SBATCH --nodes=4\r\n11c11\r\n< #SBATCH --job-name=train_tokenizer_batch_size_scaling_1_node\r\n---\r\n> #SBATCH --job-name=train_tokenizer_batch_size_scaling_4_node\r\n32c32\r\n< --batch_size=48 \\r\n---\r\n> --batch_size=192 \\r\n34c34\r\n< --max_lr=1e-4 \\r\n---\r\n> --max_lr=2.00e-4 \\r\n38,39c38,39\r\n< --name=tokenizer-batch-size-scaling-1-node-sqrt-lr-$slurm_job_id \\r\n< --tags tokenizer batch-size-scaling 1-node sqrt-lr \\r\n---\r\n> --name=tokenizer-batch-size-scaling-4-node-sqrt-lr-$slurm_job_id \\r\n> --tags tokenizer batch-size-scaling 4-node sqrt-lr \\r\n]0;mihir.mahajan@hai-login2:~/Projects/tab-model-eval",,terminal_output
|
| 51 |
+
50,227354,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_4_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=4\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=05:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_4_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python jasmine/train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=192 \\n --min_lr=0 \\n --max_lr=2.00e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-4-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 4-node sqrt-lr \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,tab
|
| 52 |
+
51,229213,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_4_nodes.sbatch",1016,0,"",shellscript,selection_mouse
|
| 53 |
+
52,230095,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_4_nodes.sbatch",1015,1,"",shellscript,content
|
| 54 |
+
53,230188,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_4_nodes.sbatch",1014,1,"",shellscript,content
|
| 55 |
+
54,230484,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_4_nodes.sbatch",1013,1,"",shellscript,content
|
| 56 |
+
55,230834,"/home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_4_nodes.sbatch",1012,0,"",shellscript,selection_command
|
| 57 |
+
56,234469,"TERMINAL",0,0,"diff /home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch /home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_4_nodes.sbatch",,terminal_command
|
| 58 |
+
57,234477,"TERMINAL",0,0,"]633;C3c3\r\n< #SBATCH --nodes=1\r\n---\r\n> #SBATCH --nodes=4\r\n11c11\r\n< #SBATCH --job-name=train_tokenizer_batch_size_scaling_1_node\r\n---\r\n> #SBATCH --job-name=train_tokenizer_batch_size_scaling_4_node\r\n32c32\r\n< --batch_size=48 \\r\n---\r\n> --batch_size=192 \\r\n34c34\r\n< --max_lr=1e-4 \\r\n---\r\n> --max_lr=2e-4 \\r\n38,39c38,39\r\n< --name=tokenizer-batch-size-scaling-1-node-sqrt-lr-$slurm_job_id \\r\n< --tags tokenizer batch-size-scaling 1-node sqrt-lr \\r\n---\r\n> --name=tokenizer-batch-size-scaling-4-node-sqrt-lr-$slurm_job_id \\r\n> --tags tokenizer batch-size-scaling 4-node sqrt-lr \\r\n]0;mihir.mahajan@hai-login2:~/Projects/tab-model-eval",,terminal_output
|
| 59 |
+
58,239342,"TERMINAL",0,0,"diff /home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch /home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_2_nodes.sbatch",,terminal_command
|
| 60 |
+
59,239359,"TERMINAL",0,0,"]633;C3c3\r\n< #SBATCH --nodes=1\r\n---\r\n> #SBATCH --nodes=2\r\n11c11\r\n< #SBATCH --job-name=train_tokenizer_batch_size_scaling_1_node\r\n---\r\n> #SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\r\n32c32\r\n< --batch_size=48 \\r\n---\r\n> --batch_size=96 \\r\n34c34\r\n< --max_lr=1e-4 \\r\n---\r\n> --max_lr=1.4e-4 \\r\n38,39c38,39\r\n< --name=tokenizer-batch-size-scaling-1-node-sqrt-lr-$slurm_job_id \\r\n< --tags tokenizer batch-size-scaling 1-node sqrt-lr \\r\n---\r\n> --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\r\n> --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\r\n]0;mihir.mahajan@hai-login2:~/Projects/tab-model-eval",,terminal_output
|
| 61 |
+
60,254499,"TERMINAL",0,0,"cd /home/mihir.mahajan/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr",,terminal_command
|
| 62 |
+
61,255158,"TERMINAL",0,0,"ls",,terminal_command
|
| 63 |
+
62,255158,"TERMINAL",0,0,"]633;Cstart_runs.sh train_tokenizer_1_nodes.sbatch train_tokenizer_4_nodes.sbatch\r\ntester.sh train_tokenizer_2_nodes.sbatch train_tokenizer_8_nodes.sbatch\r\ntrain_tokenizer_16_nodes.sbatch train_tokenizer_32_nodes.sbatch\r\n]0;mihir.mahajan@hai-login2:~/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr",,terminal_output
|
| 64 |
+
63,256089,"TERMINAL",0,0,"cd ..",,terminal_command
|
| 65 |
+
64,265480,"TERMINAL",0,0,"cp -r sqrt_lr sqrt_lr_bak/",,terminal_command
|
| 66 |
+
65,265514,"TERMINAL",0,0,"]633;C]0;mihir.mahajan@hai-login2:~/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer",,terminal_output
|
| 67 |
+
66,267772,"TERMINAL",0,0,"ls",,terminal_command
|
| 68 |
+
67,267772,"TERMINAL",0,0,"]633;C[0m[01;34mconst_lr[0m [01;34mlinear_lr[0m [01;34msqrt_lr[0m [01;34msqrt_lr_bak[0m\r\n]0;mihir.mahajan@hai-login2:~/Projects/jasmine/slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer",,terminal_output
|
| 69 |
+
68,314957,"TERMINAL",0,0,"cd ..",,terminal_command
|
| 70 |
+
69,316867,"TERMINAL",0,0,"cd ..",,terminal_command
|
| 71 |
+
70,317484,"TERMINAL",0,0,"cd ..",,terminal_command
|
| 72 |
+
71,318175,"TERMINAL",0,0,"ls",,terminal_command
|
| 73 |
+
72,320036,"TERMINAL",0,0,"cd ..",,terminal_command
|
| 74 |
+
73,320555,"TERMINAL",0,0,"ls",,terminal_command
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-427a4d3b-1eee-4e1b-ad62-0a837e5c21da1764859633492-2025_12_04-15.47.37.05/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-452e8a0c-f3ce-4668-9b2f-7561bcdbf12a1760538538230-2025_10_15-16.29.25.680/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-4e503427-8b75-4417-944d-0f8b64a7ff841761599689551-2025_10_27-22.15.16.579/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-4e6bed5e-6b9e-447e-859a-ff87eb2aa7061761311558837-2025_10_24-15.13.03.416/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-4ffb36c2-9d46-4459-9b1b-f93f144ad2f61765193679229-2025_12_08-12.35.03.82/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-529e4bba-7d8a-4661-afaa-55c4c5c035291764770291321-2025_12_03-14.58.42.143/source.csv
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,3,"data/eval/handcrafted/eval_file_switch.md",0,0,"# Assistant\n```bash\nls -l slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/\n```\n\n# User\n<stdout>\ntotal 144\n-rw-r--r-- 1 mihir.mahajan mihir.mahajan 1322 Dec 3 14:23 train_tokenizer_16_nodes.sbatch\n-rw-r--r-- 1 mihir.mahajan mihir.mahajan 1309 Dec 3 14:23 train_tokenizer_1_nodes.sbatch\n-rw-r--r-- 1 mihir.mahajan mihir.mahajan 1319 Dec 3 14:23 train_tokenizer_2_nodes.sbatch\n-rw-r--r-- 1 mihir.mahajan mihir.mahajan 1323 Dec 3 14:23 train_tokenizer_32_nodes.sbatch\n-rw-r--r-- 1 mihir.mahajan mihir.mahajan 1310 Dec 3 14:23 train_tokenizer_4_nodes.sbatch\n-rw-r--r-- 1 mihir.mahajan mihir.mahajan 1318 Dec 3 14:23 train_tokenizer_8_nodes.sbatch\n</stdout>\n\n# Assistant\n```bash\ncat -n slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch\n```\n\n# User\n<stdout>\n 1 #!/usr/bin/env bash\n 2\n 3 #SBATCH --nodes=1\n 4 #SBATCH --ntasks-per-node=4\n 5 #SBATCH --time=05:00:00\n 6 #SBATCH --partition=accelerated\n 7 #SBATCH --cpus-per-task=5\n 8 #SBATCH --gres=gpu:4\n 9 #SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n 10 #SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n 11 #SBATCH --job-name=train_tokenizer_batch_size_scaling_1_node\n 12\n 13 # Log the sbatch script\n 14 cat $0\n 15\n 16 module unload mpi/openmpi/5.0\n 17 module unload devel/cuda/12.4\n 18 source .venv/bin/activate\n 19\n 20 array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n 21\n 22 job_name=$SLURM_JOB_NAME\n 23 slurm_job_id=$SLURM_JOB_ID\n 24\n 25 CHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\n 26 mkdir -p $CHECKPOINT_DIR\n 27\n 28 env | grep SLURM\n 29\n 30 srun python jasmine/train_tokenizer.py \\n 31 --ckpt_dir $CHECKPOINT_DIR \\n 32 --batch_size=48 \\n 33 --min_lr=0 \\n 34 --max_lr=1e-4 \\n 35 --log_image_interval=1000 \\n 36 --log_checkpoint_interval=1000 \\n 37 --log \\n 38 --name=tokenizer-batch-size-scaling-1-node-sqrt-lr-$slurm_job_id \\n 39 --tags tokenizer batch-size-scaling 1-node sqrt-lr \\n 40 --entity instant-uv \\n 41 --project jafar \\n 42 --data_dir $array_records_dir\n</stdout>\n\n# Assistant\n```bash\nsed -i '5,5c#SBATCH --time=48:00:00' slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch\n```\n\n\n# Assistant\n```bash\ncat -n slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch\n```\n\n# User\n<stdout>\n</stdout>",markdown,tab
|
| 3 |
+
2,3417,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"2:58:42 PM [info] Activating crowd-code\n2:58:42 PM [info] Recording started\n2:58:42 PM [info] Initializing git provider using file system watchers...\n2:58:43 PM [info] Git repository found\n2:58:43 PM [info] Git provider initialized successfully\n2:58:43 PM [info] Initial git state: [object Object]\n",Log,tab
|
| 4 |
+
3,9859,"TERMINAL",0,0,"",,terminal_command
|
| 5 |
+
4,16890,"TERMINAL",0,0,"",,terminal_command
|
| 6 |
+
5,96256,"TERMINAL",0,0,"",,terminal_command
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-59804d2c-df27-4b41-9fe6-ffdfef8825021760703869419-2025_10_17-14.25.51.972/source.csv
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,3471,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
|
| 3 |
+
3,3682,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"2:25:50 PM [info] Activating crowd-code\n2:25:51 PM [info] Recording started\n2:25:52 PM [info] Initializing git provider using file system watchers...\n2:25:52 PM [info] No workspace folder found\n2:25:55 PM [info] Retrying git provider initialization...\n2:25:55 PM [info] No workspace folder found\n",Log,content
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-5a5585eb-19d8-428e-a477-32e8773045fc1761049920108-2025_10_21-14.32.24.316/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-5f1cd0c2-8bc8-41c4-bed3-625c4b4adb111761657862721-2025_10_28-14.24.45.589/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-5f8ae991-2efe-46c5-b76d-9518b238b91d1761295762673-2025_10_24-10.49.49.33/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-6048c8e1-2439-424f-8372-d4ccd17c2d5f1760522924995-2025_10_15-12.09.21.860/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-6116c825-ee17-41e6-b50b-e2d4d748cd831761555737545-2025_10_27-10.02.40.737/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-66805e21-c18a-477d-a6fd-7ee8129b9a211760898484058-2025_10_19-20.28.33.558/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-6877ad19-9807-438f-8b51-95c34f7ec7121764157755585-2025_11_26-12.49.36.76/source.csv
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,1418,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"12:49:36 PM [info] Activating crowd-code\n12:49:36 PM [info] Recording started\n12:49:36 PM [info] Initializing git provider using file system watchers...\n12:49:36 PM [info] Git repository found\n12:49:36 PM [info] Git provider initialized successfully\n12:49:36 PM [info] Initial git state: [object Object]\n",Log,tab
|
| 3 |
+
3,4025,"TERMINAL",0,0,"",,terminal_command
|
| 4 |
+
4,10769,"TERMINAL",0,0,"",,terminal_command
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-698eb099-f494-4ec6-a868-8174b5764b0d1765283687620-2025_12_09-13.35.26.529/source.csv
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,8859,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"1:35:26 PM [info] Activating crowd-code\n1:35:26 PM [info] Recording started\n1:35:26 PM [info] Initializing git provider using file system watchers...\n1:35:28 PM [info] Git repository found\n1:35:28 PM [info] Git provider initialized successfully\n",Log,tab
|
| 3 |
+
3,14121,"TERMINAL",0,0,"",,terminal_focus
|
| 4 |
+
4,16585,"TERMINAL",0,0,"",,terminal_command
|
| 5 |
+
5,23298,"TERMINAL",0,0,"",,terminal_command
|
| 6 |
+
6,28538,"pyproject.toml",0,0,"# Copyright (c) 2023, NVIDIA CORPORATION. All rights reserved.\n\n[build-system]\nrequires = [""setuptools<80.0.0"", ""pybind11""]\nbuild-backend = ""setuptools.build_meta""\n\n[tool.setuptools]\ninclude-package-data = true\n\n[tool.setuptools.packages.find]\ninclude = [""megatron.core"", ""megatron.core.*""]\n\n[tool.setuptools.dynamic]\nversion = { attr = ""megatron.core.package_info.__version__"" }\nreadme = { file = ""README.md"", content-type = ""text/markdown"" }\n\n[project]\nname = ""megatron-core""\ndynamic = [""version"", ""readme""]\ndescription = ""Megatron Core - a library for efficient and scalable training of transformer based models""\nrequires-python = "">=3.10""\nlicense = { text = ""Apache 2.0"" }\ndependencies = [""torch"", ""numpy<2.0.0"", ""packaging""]\nauthors = [{ name = ""NVIDIA"", email = ""nemo-toolkit@nvidia.com"" }]\nmaintainers = [{ name = ""NVIDIA"", email = ""nemo-toolkit@nvidia.com"" }]\nkeywords = [\n ""NLP"",\n ""NLU"",\n ""deep"",\n ""gpu"",\n ""language"",\n ""learning"",\n ""learning"",\n ""machine"",\n ""nvidia"",\n ""pytorch"",\n ""torch"",\n ""transformer"",\n]\nclassifiers = [\n ""Development Status :: 5 - Production/Stable"",\n ""Environment :: Console"",\n ""Intended Audience :: Developers"",\n ""Intended Audience :: Information Technology"",\n ""Intended Audience :: Science/Research"",\n ""License :: OSI Approved :: BSD License"",\n ""Natural Language :: English"",\n ""Operating System :: OS Independent"",\n ""Programming Language :: Python :: 3"",\n ""Programming Language :: Python :: 3.8"",\n ""Programming Language :: Python :: 3.9"",\n ""Topic :: Scientific/Engineering :: Artificial Intelligence"",\n ""Topic :: Scientific/Engineering :: Image Recognition"",\n ""Topic :: Scientific/Engineering :: Mathematics"",\n ""Topic :: Scientific/Engineering"",\n ""Topic :: Software Development :: Libraries :: Python Modules"",\n ""Topic :: Software Development :: Libraries"",\n ""Topic :: Utilities"",\n]\n\n[project.urls]\nDownload = ""https://github.com/NVIDIA/Megatron-LM/releases""\nHomepage = ""https://github.com/NVIDIA/Megatron-LM/megatron/core""\n\n[project.optional-dependencies]\nmlm = [""flask-restful"", ""sentencepiece"", ""tiktoken"", ""wandb""]\n\ndev = [\n ""tqdm"",\n ""einops~=0.8"",\n ""tensorstore~=0.1,!=0.1.46,!=0.1.72"",\n ""nvtx~=0.2"",\n ""transformers~=4.53"",\n ""multi-storage-client~=0.20"",\n ""opentelemetry-api~=1.33.1"",\n ""setuptools<80.0.0"",\n ""mamba-ssm~=2.2"",\n ""causal-conv1d~=1.5"",\n ""nv-grouped-gemm~=1.1"",\n ""transformer-engine[pytorch]>=2.6.0a0,<2.8.0"",\n ""nvidia-resiliency-ext>=0.4.0a0,<0.5.0"",\n ""nvidia-modelopt[torch]>=0.33.0a0,<0.34.0; sys_platform != 'darwin'"",\n ""megatron-energon[av_decode]~=6.0"",\n ""flashinfer-python"",\n ""onnxscript"",\n]\n\nlts = [\n ""tqdm"",\n ""einops"",\n ""tensorstore!=0.1.46,!=0.1.72"",\n ""nvtx"",\n ""transformers"",\n ""zarr"",\n ""setuptools<80.0.0"",\n]\n\n[dependency-groups]\ntest = [\n ""coverage"",\n ""nltk"",\n ""wrapt"",\n ""pytest==8.3.5"",\n ""pytest-mock"",\n ""pytest-cov"",\n ""pytest-random-order"",\n ""pytest-asyncio"",\n ""pygithub"",\n ""pydantic"",\n ""tensorboard"",\n ""pyyaml"",\n]\n\nbuild = [\n ""setuptools<80.0.0"",\n ""packaging"",\n ""hatchling"",\n ""pybind11"",\n ""Cython>=3.0.0"",\n ""torch"",\n]\nlinting = [\n ""ruff~=0.9.0"",\n ""black==24.4.2"",\n ""isort==5.13.2"",\n ""flake8==7.1.0"",\n ""pylint==3.2.6"",\n]\nci = [""python-gitlab"", ""slack-sdk"", ""pandas""]\nflash_mla = [""flash_mla""]\n\n[tool.uv]\ndefault-groups = [""linting"", ""build"", ""test""]\nno-build-isolation-package = [\n ""transformer-engine"",\n ""transformer-engine-torch"",\n ""mamba-ssm"",\n ""causal-conv1d"",\n ""nv-grouped-gemm"",\n ""flash_mla"",\n]\nlink-mode = ""copy""\nconflicts = [[{ extra = ""lts"" }, { extra = ""dev"" }]]\n# We don't want to install torch, torchvision, and triton\n# because they are already installed in the base image.\noverride-dependencies = [\n ""torch; sys_platform == 'never'"",\n ""torchvision; sys_platform == 'never'"",\n ""triton; sys_platform == 'never'"",\n]\n\n[tool.uv.sources]\nflash_mla = [\n { git = ""https://github.com/deepseek-ai/FlashMLA"", rev = ""9edee0c022cd0938148a18e334203b0aab43aa19"" },\n]\n\n[tool.isort]\nprofile = ""black"" # black-compatible\nline_length = 100 # should match black parameters\npy_version = 310 # python 3.8 as a target version\nknown_first_party = [""megatron""] # FIRSTPARTY section\nknown_third_party = [""transformer_engine""] # THIRDPARTY section\nsections = [""FUTURE"", ""STDLIB"", ""THIRDPARTY"", ""FIRSTPARTY"", ""LOCALFOLDER""]\ndefault_section = ""THIRDPARTY""\nextend_skip = [""setup.py""]\n\n[tool.black]\nline_length = 100\nskip_string_normalization = true\n# recongized by future versions, disallows to reformat code with incompatible versions\n# Matches NeMO version so people working on both codebases don't need two different version of black installed\nrequired_version = ""24""\nskip_magic_trailing_comma = true\ninclude = '\.pyi?$'\nexclude = '''\n/(\n \.git\n | \.venv\n | build\n)/\n'''\n\n[tool.pytest.ini_options]\naddopts = ""--durations=15 -s -rA -x""\ntestpaths = [""tests""]\npython_files = ""test_*.py""\nmarkers = [\n ""internal: mark a test as a test to private/internal functions."",\n ""flaky: mark flaky tests for LTS environment"",\n ""flaky_in_dev: mark flaky tests for DEV environment"",\n]\n\n[tool.coverage.run]\ndata_file = "".coverage_$LOCAL_RANK""\nconcurrency = [""thread"", ""multiprocessing""]\nomit = [\n ""/tmp/*"",\n ""/workspace/tests/*"",\n ""/usr/local/lib/python3.12/dist-packages/*"",\n]\nrelative_files = true\n\n[tool.coverage.paths]\nsource = [""/opt/megatron-lm/""]\n\n[tool.ruff.lint]\n# Enable all `pydocstyle` rules, limiting to those that adhere to the\n# Google convention via `convention = ""google""`, below.\n# select = [""D"", ""F""]\nselect = [""S506""]\n\n# - On top of the Google convention, disable `D417`, which requires\n# documentation for every function parameter.\n# - F841: local variable assigned but never used (exluced to favor readability)\n# TODO: Remove D10 once we are about to release to get all the docstrings written\nignore = [""D417"", ""D10"", ""F841""]\n\n[tool.ruff.lint.pydocstyle]\nconvention = ""google""\n\n# Section to exclude errors for different file types\n[tool.ruff.per-file-ignores]\n# Ignore all directories named `tests`.\n""tests/**"" = [""D""]\n# Ignore all files that end in `_test.py`.\n""*_test.py"" = [""D""]\n# Ignore F401 (import but unused) in __init__.py\n""__init__.py"" = [""F401""]\n",plaintext,tab
|
| 7 |
+
7,89041,"TERMINAL",0,0,"",,terminal_command
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-73f8939e-63ce-4a1f-ae80-e8b595f3c9321764500864397-2025_11_30-12.08.15.208/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-75fcc8f0-691c-4551-9c8c-7a9bcc824a7e1760553685394-2025_10_15-20.41.46.66/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-76116d19-106d-4bbc-8d56-8e736ef694361765124265000-2025_12_07-17.18.05.133/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-7652521e-e335-40e6-8416-914c425804a11765271690051-2025_12_09-10.15.15.804/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
b1ccc08ffefaf10a6c1b392f91212b8c7d87421316664745744fabeefb9fdb16/crowd-code-772b9e7f-7206-4184-bf66-5ac053c11d621761648011932-2025_10_28-11.40.35.184/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|