quickfix: missing newlineke
Browse files
1de052c516cab686515c107385aaf7c3a7e3e5c23c9bc3c0be0cff3df28cd64d/crowd-code-2f4ff312-abac-4732-805a-ec6c245e532e1752745206273-2025_07_17-11.40.45.491/source.csv
CHANGED
|
@@ -1,4 +1,5 @@
|
|
| 1 |
-
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,
|
|
|
|
| 2 |
99,1928731,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 3 |
100,1928792,"TERMINAL",0,0,"[?1049h[22;0;0t[?1h=\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=2\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=24:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:4\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/%x_%j.log\r\n#SBATCH --job-name=train_dynamics_lr_1e-4_larger_ffn\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrec ords_chunked\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraini ng-modelsize-scaling/$job_name\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr- scaling/train_tokenizer_lr_sweep_1e-4\r\n\r\nenv | grep SLURM\r\n\r\nsrun python train_dynamics.py \\r\n --save_ckpt \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=96 \\r\n --tokenizer_ffn_dim=2048 \\r\n --tokenizer_num_blocks=4 \\r\n --lam_ffn_dim=2048 \\r\n --lam_num_blocks=4 \\r\n --dyna_ffn_dim=2048 \\r\n --dyna_num_blocks=6 \\r\n --init_lr=0 \\r\n --max_lr=1.5e-4 \\r\n --log_image_interval=1000 \\r\n --log \\r\n --log_checkpoint_interval=1000 \\r\n --name=dynamics-lr-1e-4-larger-ffn \\r\n --tags dynamics lr-1e-4 larger-ffn mixed-precision flash-attention \\r\n --entity instant-uv \\r\n --project jafar \\r\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\r\n --data_dir $array_records_dir \\r\nSLURM_JOB_USER=tum_dbd0378\r\nSLURM_TASKS_PER_NODE=4(x2)\r\nSLURM_JOB_UID=996262\r\nSLURM_TASK_PID=2571288\r\nSLURM_JOB_GPUS=0,1,2,3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar\r\nSLURMD_NODENAME=hkn0519\r\nSLURM_JOB_START_TIME=1752692957\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1752779357\r\nSLURM_CPUS_ON_NODE=24\r\nSLURM_JOB_CPUS_PER_NODE=24(x2)\r\nSLURM_GPUS_ON_NODE=4\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=5\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=2\r\nSLURM_JOBID=3352437\r\nSLURM_JOB_QOS=normal\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=5\r\nSLURM_NTASKS=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e10.hkn0519\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_NODELIST=hkn[0519-0520]\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=8\r\nSLURM_NNODES=2\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3352437\r\nSLURM_NODEID=0\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=train_dynamics_lr_1e-4_larger_ffn\r\n[7m/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/train_dynamics_lr_1e-4_larger_ffn_3352437.log[27m[K",,terminal_output
|
| 4 |
101,1929130,"TERMINAL",0,0,"\r[K",,terminal_output
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
98,1928682,"TERMINAL",0,0,"less /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/train_dynamics_lr_1e-4_larger_ffn_3352437.log ",,terminal_command
|
| 3 |
99,1928731,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 4 |
100,1928792,"TERMINAL",0,0,"[?1049h[22;0;0t[?1h=\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=2\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=24:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:4\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/%x_%j.log\r\n#SBATCH --job-name=train_dynamics_lr_1e-4_larger_ffn\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrec ords_chunked\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraini ng-modelsize-scaling/$job_name\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr- scaling/train_tokenizer_lr_sweep_1e-4\r\n\r\nenv | grep SLURM\r\n\r\nsrun python train_dynamics.py \\r\n --save_ckpt \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=96 \\r\n --tokenizer_ffn_dim=2048 \\r\n --tokenizer_num_blocks=4 \\r\n --lam_ffn_dim=2048 \\r\n --lam_num_blocks=4 \\r\n --dyna_ffn_dim=2048 \\r\n --dyna_num_blocks=6 \\r\n --init_lr=0 \\r\n --max_lr=1.5e-4 \\r\n --log_image_interval=1000 \\r\n --log \\r\n --log_checkpoint_interval=1000 \\r\n --name=dynamics-lr-1e-4-larger-ffn \\r\n --tags dynamics lr-1e-4 larger-ffn mixed-precision flash-attention \\r\n --entity instant-uv \\r\n --project jafar \\r\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\r\n --data_dir $array_records_dir \\r\nSLURM_JOB_USER=tum_dbd0378\r\nSLURM_TASKS_PER_NODE=4(x2)\r\nSLURM_JOB_UID=996262\r\nSLURM_TASK_PID=2571288\r\nSLURM_JOB_GPUS=0,1,2,3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar\r\nSLURMD_NODENAME=hkn0519\r\nSLURM_JOB_START_TIME=1752692957\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1752779357\r\nSLURM_CPUS_ON_NODE=24\r\nSLURM_JOB_CPUS_PER_NODE=24(x2)\r\nSLURM_GPUS_ON_NODE=4\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=5\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=2\r\nSLURM_JOBID=3352437\r\nSLURM_JOB_QOS=normal\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=5\r\nSLURM_NTASKS=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e10.hkn0519\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_NODELIST=hkn[0519-0520]\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=8\r\nSLURM_NNODES=2\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3352437\r\nSLURM_NODEID=0\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=train_dynamics_lr_1e-4_larger_ffn\r\n[7m/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/train_dynamics_lr_1e-4_larger_ffn_3352437.log[27m[K",,terminal_output
|
| 5 |
101,1929130,"TERMINAL",0,0,"\r[K",,terminal_output
|