Sequence
int64 1
25.2k
| Time
int64 1
858M
| File
stringclasses 830
values | RangeOffset
int64 0
2.21M
| RangeLength
int64 0
168k
| Text
stringlengths 1
4.7M
⌀ | Language
stringclasses 20
values | Type
stringclasses 9
values |
|---|---|---|---|---|---|---|---|
224
| 276,124
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 752
| 0
|
\n
|
shellscript
|
content
|
225
| 276,281
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 753
| 4
| null |
shellscript
|
content
|
226
| 276,451
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 753
| 0
|
256
|
shellscript
|
content
|
227
| 276,456
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 755
| 0
| null |
shellscript
|
selection_command
|
228
| 276,690
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 734
| 0
| null |
shellscript
|
selection_command
|
229
| 277,584
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 715
| 0
| null |
shellscript
|
selection_command
|
230
| 277,815
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 734
| 0
| null |
shellscript
|
selection_command
|
231
| 278,133
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 736
| 0
| null |
shellscript
|
selection_command
|
232
| 278,332
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 738
| 0
| null |
shellscript
|
selection_command
|
233
| 278,579
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 748
| 0
| null |
shellscript
|
selection_command
|
234
| 279,621
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 748
| 4
| null |
shellscript
|
content
|
235
| 279,625
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 747
| 0
| null |
shellscript
|
selection_command
|
236
| 280,007
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 732
| 20
|
--model_dim 256
|
shellscript
|
content
|
237
| 280,009
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 748
| 0
| null |
shellscript
|
selection_command
|
238
| 281,175
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 751
| 0
|
\
|
shellscript
|
content
|
239
| 281,176
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 753
| 0
| null |
shellscript
|
selection_command
|
240
| 283,261
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 0
| 0
| null |
shellscript
|
tab
|
241
| 284,617
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 0
| 0
| null |
shellscript
|
tab
|
242
| 286,708
|
TERMINAL
| 0
| 0
|
so
| null |
terminal_output
|
243
| 286,786
|
TERMINAL
| 0
| 0
|
u
| null |
terminal_output
|
244
| 286,840
|
TERMINAL
| 0
| 0
|
r
| null |
terminal_output
|
245
| 287,023
|
TERMINAL
| 0
| 0
|
c
| null |
terminal_output
|
246
| 287,153
|
TERMINAL
| 0
| 0
|
e
| null |
terminal_output
|
247
| 287,294
|
TERMINAL
| 0
| 0
|
.
| null |
terminal_output
|
248
| 287,384
|
TERMINAL
| 0
| 0
|
v
| null |
terminal_output
|
249
| 287,457
|
TERMINAL
| 0
| 0
|
e
| null |
terminal_output
|
250
| 287,643
|
TERMINAL
| 0
| 0
|
nv_jafar/
| null |
terminal_output
|
251
| 287,949
|
TERMINAL
| 0
| 0
|
b
| null |
terminal_output
|
252
| 288,162
|
TERMINAL
| 0
| 0
|
in/
| null |
terminal_output
|
253
| 288,398
|
TERMINAL
| 0
| 0
|
a
| null |
terminal_output
|
254
| 288,470
|
TERMINAL
| 0
| 0
|
c
| null |
terminal_output
|
255
| 288,660
|
TERMINAL
| 0
| 0
|
tivate
| null |
terminal_output
|
256
| 289,041
|
TERMINAL
| 0
| 0
|
\r\n[?2004l\r]0;tum_ind3695@hkn0505:~/projects/jafar_run_overfit[?2004h(.venv_jafar) [tum_ind3695@hkn0505 jafar_run_overfit]$
| null |
terminal_output
|
257
| 314,000
|
TERMINAL
| 0
| 0
|
[7mslurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh[27m
| null |
terminal_output
|
258
| 315,118
|
TERMINAL
| 0
| 0
|
\r[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[Cslurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh\r[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C
| null |
terminal_output
|
259
| 316,126
|
TERMINAL
| 0
| 0
|
.slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh\r[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C
| null |
terminal_output
|
260
| 316,281
|
TERMINAL
| 0
| 0
|
/slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh\r[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C[C
| null |
terminal_output
|
261
| 316,652
|
TERMINAL
| 0
| 0
|
\r\n[?2004l\r# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv_jafar/bin/activate\r\n\r\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\ntf_records_dir=$ws_dir/data/knoms_tfrecords_200_shards\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\ntags=("overfit_sample", "tokenizer", "debug", "alfred")\r\n\r\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\nenv | grep SLURM\r\n\r\npython train_tokenizer_single_sample.py \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=1 \\r\n --min_lr=4.3e-5 \\r\n --max_lr=4.3e-4 \\r\n --log_image_interval=10 \\r\n --log \\r\n --entity instant-uv \\r\n --project jafar \\r\n --name $job_name \\r\n --tags $tags \\r\n --model_dim 256 \\r\n --data_dir $tf_records_dir\r\n
| null |
terminal_output
|
262
| 316,850
|
TERMINAL
| 0
| 0
|
SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_ind3695\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=991285\r\nSLURM_TASK_PID=707800\r\nSLURM_JOB_GPUS=3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit\r\nSLURMD_NODENAME=hkn0505\r\nSLURM_JOB_START_TIME=1751037996\r\nSLURM_STEP_NODELIST=hkn0505\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1751039796\r\nSLURM_PMI2_SRUN_PORT=39147\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3299579\r\nSLURM_PTY_PORT=34619\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=68\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_NTASKS=1\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e10.hkn0505\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=175\r\nSLURM_NODELIST=hkn0505\r\nSLURM_SRUN_COMM_PORT=41999\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=1\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3299579\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0505\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_NTASKS_PER_NODE=1\r\nSLURM_STEP_LAUNCHER_PORT=41999\r\nSLURM_JOB_GID=502289\r\nSLURM_JOB_NODELIST=hkn0505\r\n
| null |
terminal_output
|
263
| 333,795
|
TERMINAL
| 0
| 0
|
2025-06-27 17:28:41.480315: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\r\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\r\nE0000 00:00:1751038121.746466 708273 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\r\nE0000 00:00:1751038121.796988 708273 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\r\n
| null |
terminal_output
|
264
| 334,796
|
TERMINAL
| 0
| 0
|
W0000 00:00:1751038122.217951 708273 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\r\nW0000 00:00:1751038122.217997 708273 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\r\nW0000 00:00:1751038122.218000 708273 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\r\nW0000 00:00:1751038122.218002 708273 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\r\n
| null |
terminal_output
|
265
| 353,795
|
TERMINAL
| 0
| 0
|
W0000 00:00:1751038141.160274 708273 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\r\nSkipping registering GPU devices...\r\n
| null |
terminal_output
|
266
| 354,794
|
TERMINAL
| 0
| 0
|
Running on 1 devices.\r\n
| null |
terminal_output
|
267
| 356,793
|
TERMINAL
| 0
| 0
|
[34m[1mwandb[0m: Currently logged in as: [33mavocadoali[0m ([33minstant-uv[0m) to [32mhttps://api.wandb.ai[0m. Use [1m`wandb login --relogin`[0m to force relogin\r\n
| null |
terminal_output
|
268
| 356,795
|
TERMINAL
| 0
| 0
|
[34m[1mwandb[0m: Tracking run with wandb version 0.19.11\r\n[34m[1mwandb[0m: Run data is saved locally in [35m[1m/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/wandb/run-20250627_172903-cy8nd25c[0m\r\n[34m[1mwandb[0m: Run [1m`wandb offline`[0m to turn off syncing.\r\n[34m[1mwandb[0m: Syncing run [33minteractive[0m\r\n[34m[1mwandb[0m: ⭐️ View project at [34m[4mhttps://wandb.ai/instant-uv/jafar[0m\r\n[34m[1mwandb[0m: 🚀 View run at [34m[4mhttps://wandb.ai/instant-uv/jafar/runs/cy8nd25c[0m\r\n
| null |
terminal_output
|
269
| 358,794
|
TERMINAL
| 0
| 0
|
2025-06-27 17:29:06.757779: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n
| null |
terminal_output
|
270
| 370,794
|
TERMINAL
| 0
| 0
|
2025-06-27 17:29:18.225656: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n
| null |
terminal_output
|
271
| 372,795
|
TERMINAL
| 0
| 0
|
2025-06-27 17:29:20.377296: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n
| null |
terminal_output
|
272
| 375,792
|
TERMINAL
| 0
| 0
|
Counting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 4770688, 'vq': 32768, 'decoder': 4770672, 'total': 9574128}\r\n
| null |
terminal_output
|
273
| 376,794
|
TERMINAL
| 0
| 0
|
Starting training from step 0...\r\nbatch shape: (1, 16, 90, 160, 3)\r\n
| null |
terminal_output
|
274
| 388,793
|
TERMINAL
| 0
| 0
|
2025-06-27 17:29:35.989084: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-06-27 17:29:35.989504: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-06-27 17:29:35.989525: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-06-27 17:29:35.989616: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-06-27 17:29:35.991060: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n
| null |
terminal_output
|
275
| 426,791
|
TERMINAL
| 0
| 0
|
Step 0, loss: 0.2878085672855377, step time: 50217.286825180054ms\r\nStep 1, loss: 0.24775156378746033, step time: 139.16516304016113ms\r\nStep 2, loss: 0.22453980147838593, step time: 138.41557502746582ms\r\nStep 3, loss: 0.20915493369102478, step time: 137.27498054504395ms\r\n
| null |
terminal_output
|
276
| 427,792
|
TERMINAL
| 0
| 0
|
Step 4, loss: 0.19706685841083527, step time: 136.38997077941895ms\r\nStep 5, loss: 0.18604601919651031, step time: 136.915922164917ms\r\nStep 6, loss: 0.17379429936408997, step time: 136.98315620422363ms\r\nStep 7, loss: 0.16810941696166992, step time: 139.41645622253418ms\r\nStep 8, loss: 0.16610655188560486, step time: 138.90910148620605ms\r\nStep 9, loss: 0.16548341512680054, step time: 137.22848892211914ms\r\n
| null |
terminal_output
|
277
| 428,798
|
TERMINAL
| 0
| 0
|
Step 10, loss: 0.16451306641101837, step time: 138.28635215759277ms\r\nStep 11, loss: 0.16362540423870087, step time: 137.18509674072266ms\r\n
| null |
terminal_output
|
278
| 429,792
|
TERMINAL
| 0
| 0
|
Step 12, loss: 0.16270293295383453, step time: 137.2530460357666ms\r\nStep 13, loss: 0.16208229959011078, step time: 137.0091438293457ms\r\nStep 14, loss: 0.16157127916812897, step time: 137.3128890991211ms\r\nStep 15, loss: 0.16023418307304382, step time: 136.6255283355713ms\r\nStep 16, loss: 0.1580863744020462, step time: 136.76190376281738ms\r\nStep 17, loss: 0.1555618792772293, step time: 136.60311698913574ms\r\nStep 18, loss: 0.15301485359668732, step time: 136.63434982299805ms\r\n
| null |
terminal_output
|
279
| 431,796
|
TERMINAL
| 0
| 0
|
Step 19, loss: 0.15070228278636932, step time: 136.61909103393555ms\r\nStep 20, loss: 0.14873021841049194, step time: 138.07296752929688ms\r\nStep 21, loss: 0.14709101617336273, step time: 137.30311393737793ms\r\nStep 22, loss: 0.14542073011398315, step time: 136.62004470825195ms\r\nStep 23, loss: 0.14299985766410828, step time: 138.46993446350098ms\r\nStep 24, loss: 0.14050206542015076, step time: 138.63539695739746ms\r\nStep 25, loss: 0.13743990659713745, step time: 138.20934295654297ms\r\nStep 26, loss: 0.13456767797470093, step time: 136.69514656066895ms\r\nStep 27, loss: 0.13202443718910217, step time: 137.4068260192871ms\r\nStep 28, loss: 0.13071371614933014, step time: 136.59048080444336ms\r\nStep 29, loss: 0.12923787534236908, step time: 136.30366325378418ms\r\nStep 30, loss: 0.12621724605560303, step time: 137.88890838623047ms\r\nStep 31, loss: 0.12275317311286926, step time: 137.44020462036133ms\r\n
| null |
terminal_output
|
280
| 432,795
|
TERMINAL
| 0
| 0
|
Step 32, loss: 0.12008993327617645, step time: 136.98863983154297ms\r\nStep 33, loss: 0.11796204745769501, step time: 136.55734062194824ms\r\nStep 34, loss: 0.116253562271595, step time: 136.1711025238037ms\r\nStep 35, loss: 0.11444009840488434, step time: 136.2471580505371ms\r\nStep 36, loss: 0.11214287579059601, step time: 136.74283027648926ms\r\nStep 37, loss: 0.10969927161931992, step time: 136.57283782958984ms\r\nStep 38, loss: 0.10776249319314957, step time: 136.54303550720215ms\r\n
| null |
terminal_output
|
281
| 433,797
|
TERMINAL
| 0
| 0
|
Step 39, loss: 0.10700612515211105, step time: 136.8429660797119ms\r\nStep 40, loss: 0.1061176136136055, step time: 143.63932609558105ms\r\nStep 41, loss: 0.10447747260332108, step time: 137.1481418609619ms\r\nStep 42, loss: 0.10248897969722748, step time: 136.8887424468994ms\r\nStep 43, loss: 0.10054730623960495, step time: 136.8556022644043ms\r\nStep 44, loss: 0.09930893033742905, step time: 136.60407066345215ms\r\n
| null |
terminal_output
|
282
| 435,793
|
TERMINAL
| 0
| 0
|
Step 45, loss: 0.09862879663705826, step time: 136.32702827453613ms\r\nStep 46, loss: 0.09765765070915222, step time: 136.71493530273438ms\r\nStep 47, loss: 0.09670083969831467, step time: 136.9166374206543ms\r\nStep 48, loss: 0.09570780396461487, step time: 136.46697998046875ms\r\nStep 49, loss: 0.0949011966586113, step time: 136.81602478027344ms\r\nStep 50, loss: 0.09431950002908707, step time: 137.70270347595215ms\r\nStep 51, loss: 0.09318584948778152, step time: 138.29421997070312ms\r\nStep 52, loss: 0.09204382449388504, step time: 137.2377872467041ms\r\nStep 53, loss: 0.09164949506521225, step time: 137.05110549926758ms\r\nStep 54, loss: 0.09246745705604553, step time: 136.80434226989746ms\r\nStep 55, loss: 0.09276295453310013, step time: 137.25781440734863ms\r\nStep 56, loss: 0.09318467229604721, step time: 136.97123527526855ms\r\nStep 57, loss: 0.09463216364383698, step time: 136.82889938354492ms\r\nStep 58, loss: 0.09674461930990219, step time: 136.49773597717285ms\r\n
| null |
terminal_output
|
283
| 436,593
|
TERMINAL
| 0
| 0
|
Step 59, loss: 0.09831493347883224, step time: 137.01462745666504ms\r\nStep 60, loss: 0.09971342235803604, step time: 137.88175582885742ms\r\nStep 61, loss: 0.10120789706707001, step time: 137.3124122619629ms\r\nStep 62, loss: 0.10248210281133652, step time: 136.9936466217041ms\r\nStep 63, loss: 0.1036282479763031, step time: 136.7807388305664ms\r\n
| null |
terminal_output
|
284
| 436,734
|
TERMINAL
| 0
| 0
|
Step 64, loss: 0.10482157766819, step time: 136.6262435913086ms\r\n
| null |
terminal_output
|
285
| 436,875
|
TERMINAL
| 0
| 0
|
Step 65, loss: 0.10560813546180725, step time: 137.1769905090332ms\r\n
| null |
terminal_output
|
286
| 437,015
|
TERMINAL
| 0
| 0
|
Step 66, loss: 0.10659754276275635, step time: 137.18032836914062ms\r\n
| null |
terminal_output
|
287
| 437,153
|
TERMINAL
| 0
| 0
|
Step 67, loss: 0.1076059639453888, step time: 136.8722915649414ms\r\n
| null |
terminal_output
|
288
| 437,287
|
TERMINAL
| 0
| 0
|
Step 68, loss: 0.10910508036613464, step time: 136.9483470916748ms\r\n
| null |
terminal_output
|
289
| 437,426
|
TERMINAL
| 0
| 0
|
Step 69, loss: 0.10965511202812195, step time: 136.91949844360352ms\r\n
| null |
terminal_output
|
290
| 437,696
|
TERMINAL
| 0
| 0
|
Step 70, loss: 0.10877430438995361, step time: 142.31514930725098ms\r\n
| null |
terminal_output
|
291
| 437,835
|
TERMINAL
| 0
| 0
|
Step 71, loss: 0.10760578513145447, step time: 137.253999710083ms\r\n
| null |
terminal_output
|
292
| 437,973
|
TERMINAL
| 0
| 0
|
Step 72, loss: 0.1065947413444519, step time: 136.67654991149902ms\r\n
| null |
terminal_output
|
293
| 438,110
|
TERMINAL
| 0
| 0
|
Step 73, loss: 0.10574036836624146, step time: 136.82818412780762ms\r\n
| null |
terminal_output
|
294
| 438,248
|
TERMINAL
| 0
| 0
|
Step 74, loss: 0.10474038869142532, step time: 136.4905834197998ms\r\n
| null |
terminal_output
|
295
| 438,389
|
TERMINAL
| 0
| 0
|
Step 75, loss: 0.10362108051776886, step time: 136.49415969848633ms\r\n
| null |
terminal_output
|
296
| 438,528
|
TERMINAL
| 0
| 0
|
Step 76, loss: 0.10265430808067322, step time: 136.89208030700684ms\r\n
| null |
terminal_output
|
297
| 438,666
|
TERMINAL
| 0
| 0
|
Step 77, loss: 0.10095227509737015, step time: 136.96026802062988ms\r\n
| null |
terminal_output
|
298
| 438,856
|
TERMINAL
| 0
| 0
|
Step 78, loss: 0.0991610661149025, step time: 136.56949996948242ms\r\n
| null |
terminal_output
|
299
| 438,959
|
TERMINAL
| 0
| 0
|
^CTraceback (most recent call last):\r\n File "/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/train_tokenizer_single_sample.py", line 240, in <module>\r\n jax.block_until_ready(loss)\r\n File "/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/.venv_jafar/lib/python3.10/site-packages/jax/_src/api.py", line 3117, in block_until_ready\r\n try_to_block(arrays[0])\r\n File "/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/.venv_jafar/lib/python3.10/site-packages/jax/_src/api.py", line 3100, in try_to_block\r\n return x.block_until_ready()\r\nKeyboardInterrupt\r\nTraceback (most recent call last):\r\n File "/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/train_tokenizer_single_sample.py", line 240, in <module>\r\n jax.block_until_ready(loss)\r\n File "/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/.venv_jafar/lib/python3.10/site-packages/jax/_src/api.py", line 3117, in block_until_ready\r\n try_to_block(arrays[0])\r\n File "/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/.venv_jafar/lib/python3.10/site-packages/jax/_src/api.py", line 3100, in try_to_block\r\n return x.block_until_ready()\r\nKeyboardInterrupt\r\n
| null |
terminal_output
|
300
| 439,479
|
TERMINAL
| 0
| 0
|
^C
| null |
terminal_output
|
301
| 439,864
|
TERMINAL
| 0
| 0
|
Exception ignored in atexit callback: <function _start_and_connect_service.<locals>.teardown_atexit at 0x15456c09dfc0>\r\nTraceback (most recent call last):\r\n File "/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/.venv_jafar/lib/python3.10/site-packages/wandb/sdk/lib/service_connection.py", line 94, in teardown_atexit\r\n conn.teardown(hooks.exit_code)\r\n File "/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/.venv_jafar/lib/python3.10/site-packages/wandb/sdk/lib/service_connection.py", line 236, in teardown\r\n return self._proc.join()\r\n File "/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/.venv_jafar/lib/python3.10/site-packages/wandb/sdk/service/service.py", line 251, in join\r\n ret = self._internal_proc.wait()\r\n File "/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/subprocess.py", line 1209, in wait\r\n return self._wait(timeout=timeout)\r\n File "/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/subprocess.py", line 1959, in _wait\r\n (pid, sts) = self._try_wait(0)\r\n File "/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/subprocess.py", line 1917, in _try_wait\r\n (pid, sts) = os.waitpid(self.pid, wait_flags)\r\nKeyboardInterrupt: \r\n
| null |
terminal_output
|
302
| 440,416
|
TERMINAL
| 0
| 0
|
[1;34mwandb[0m: \r\n[1;34mwandb[0m: 🚀 View run [33minteractive[0m at: [34mhttps://wandb.ai/instant-uv/jafar/runs/cy8nd25c[0m\r\n[1;34mwandb[0m: Find logs at: [1;35m../../../../../hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/wandb/run-20250627_172903-cy8nd25c/logs[0m\r\n
| null |
terminal_output
|
303
| 442,363
|
TERMINAL
| 0
| 0
|
]0;tum_ind3695@hkn0505:~/projects/jafar_run_overfit[?2004h(.venv_jafar) [tum_ind3695@hkn0505 jafar_run_overfit]$
| null |
terminal_output
|
304
| 446,578
|
notes.md
| 0
| 0
| null |
markdown
|
tab
|
305
| 449,157
|
notes.md
| 9,144
| 0
| null |
markdown
|
selection_mouse
|
306
| 449,164
|
notes.md
| 9,143
| 0
| null |
markdown
|
selection_command
|
307
| 529,854
|
notes.md
| 9,209
| 0
| null |
markdown
|
selection_mouse
|
308
| 529,858
|
notes.md
| 9,208
| 0
| null |
markdown
|
selection_command
|
309
| 529,869
|
notes.md
| 9,208
| 1
|
8
|
markdown
|
selection_mouse
|
310
| 529,871
|
notes.md
| 9,209
| 0
| null |
markdown
|
selection_command
|
311
| 530,219
|
notes.md
| 9,223
| 0
| null |
markdown
|
selection_mouse
|
312
| 530,227
|
notes.md
| 9,222
| 0
| null |
markdown
|
selection_command
|
313
| 533,938
|
notes.md
| 9,017
| 0
| null |
markdown
|
selection_mouse
|
314
| 533,941
|
notes.md
| 9,016
| 0
| null |
markdown
|
selection_command
|
315
| 535,409
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 0
| 0
| null |
shellscript
|
tab
|
316
| 537,666
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 753
| 0
|
\n
|
shellscript
|
content
|
317
| 538,098
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 758
| 0
|
-
|
shellscript
|
content
|
318
| 538,099
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 759
| 0
| null |
shellscript
|
selection_keyboard
|
319
| 538,260
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 759
| 0
|
-
|
shellscript
|
content
|
320
| 538,261
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 760
| 0
| null |
shellscript
|
selection_keyboard
|
321
| 538,360
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 760
| 0
|
shellscript
|
content
|
|
322
| 538,361
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 761
| 0
| null |
shellscript
|
selection_keyboard
|
323
| 538,659
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 760
| 1
| null |
shellscript
|
content
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.