Sequence
int64 1
25.2k
| Time
int64 1
858M
| File
stringclasses 830
values | RangeOffset
int64 0
2.21M
| RangeLength
int64 0
168k
| Text
stringlengths 1
4.7M
⌀ | Language
stringclasses 20
values | Type
stringclasses 9
values |
|---|---|---|---|---|---|---|---|
624
| 1,076,254
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 749
| 0
|
u
|
shellscript
|
content
|
625
| 1,076,255
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 750
| 0
| null |
shellscript
|
selection_keyboard
|
626
| 1,076,424
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 750
| 0
|
m
|
shellscript
|
content
|
627
| 1,076,425
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 751
| 0
| null |
shellscript
|
selection_keyboard
|
628
| 1,076,792
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 751
| 0
|
_
|
shellscript
|
content
|
629
| 1,076,793
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 752
| 0
| null |
shellscript
|
selection_keyboard
|
630
| 1,077,069
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 752
| 0
|
b
|
shellscript
|
content
|
631
| 1,077,070
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 753
| 0
| null |
shellscript
|
selection_keyboard
|
632
| 1,077,221
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 753
| 0
|
l
|
shellscript
|
content
|
633
| 1,077,222
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 754
| 0
| null |
shellscript
|
selection_keyboard
|
634
| 1,077,371
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 754
| 0
|
o
|
shellscript
|
content
|
635
| 1,077,372
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 755
| 0
| null |
shellscript
|
selection_keyboard
|
636
| 1,078,007
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 755
| 0
|
cks 4 \
|
shellscript
|
content
|
637
| 1,078,359
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 761
| 0
| null |
shellscript
|
selection_command
|
638
| 1,078,830
|
notes.md
| 0
| 0
| null |
markdown
|
tab
|
639
| 1,080,498
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 0
| 0
| null |
shellscript
|
tab
|
640
| 1,083,932
|
notes.md
| 0
| 0
| null |
markdown
|
tab
|
641
| 1,085,196
|
TERMINAL
| 0
| 0
| null |
terminal_output
|
|
642
| 1,085,657
|
TERMINAL
| 0
| 0
|
./slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| null |
terminal_output
|
643
| 1,086,163
|
TERMINAL
| 0
| 0
|
\r\n[?2004l\r# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv_jafar/bin/activate\r\n\r\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\ntf_records_dir=$ws_dir/data/knoms_tfrecords_200_shards\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\ntags="overfit_sample tokenizer debug alfred"\r\n\r\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\nenv | grep SLURM\r\n\r\npython train_tokenizer_single_sample.py \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=1 \\r\n --min_lr=4.3e-5 \\r\n --max_lr=4.3e-4 \\r\n --log_image_interval=10 \\r\n --log \\r\n --entity instant-uv \\r\n --project jafar \\r\n --name $job_name \\r\n --tags $tags \\r\n --model_dim 64 \\r\n --num_blocks 4 \\r\n --data_dir $tf_records_dir\r\n
| null |
terminal_output
|
644
| 1,086,310
|
TERMINAL
| 0
| 0
|
SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_ind3695\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=991285\r\nSLURM_TASK_PID=707800\r\nSLURM_JOB_GPUS=3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit\r\nSLURMD_NODENAME=hkn0505\r\nSLURM_JOB_START_TIME=1751037996\r\nSLURM_STEP_NODELIST=hkn0505\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1751039796\r\nSLURM_PMI2_SRUN_PORT=39147\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3299579\r\nSLURM_PTY_PORT=34619\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=68\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_NTASKS=1\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e10.hkn0505\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=175\r\nSLURM_NODELIST=hkn0505\r\nSLURM_SRUN_COMM_PORT=41999\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=1\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3299579\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0505\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_NTASKS_PER_NODE=1\r\nSLURM_STEP_LAUNCHER_PORT=41999\r\nSLURM_JOB_GID=502289\r\nSLURM_JOB_NODELIST=hkn0505\r\n
| null |
terminal_output
|
645
| 1,088,423
|
TERMINAL
| 0
| 0
|
2025-06-27 17:41:16.474711: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\r\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\r\nE0000 00:00:1751038876.487429 713983 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\r\nE0000 00:00:1751038876.491509 713983 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\r\nW0000 00:00:1751038876.503669 713983 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\r\nW0000 00:00:1751038876.503687 713983 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\r\nW0000 00:00:1751038876.503690 713983 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\r\nW0000 00:00:1751038876.503693 713983 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\r\n
| null |
terminal_output
|
646
| 1,091,423
|
TERMINAL
| 0
| 0
|
W0000 00:00:1751038879.498193 713983 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\r\nSkipping registering GPU devices...\r\n
| null |
terminal_output
|
647
| 1,091,903
|
TERMINAL
| 0
| 0
|
Running on 1 devices.\r\n
| null |
terminal_output
|
648
| 1,092,699
|
TERMINAL
| 0
| 0
|
[34m[1mwandb[0m: Currently logged in as: [33mavocadoali[0m ([33minstant-uv[0m) to [32mhttps://api.wandb.ai[0m. Use [1m`wandb login --relogin`[0m to force relogin\r\n
| null |
terminal_output
|
649
| 1,093,292
|
TERMINAL
| 0
| 0
|
[34m[1mwandb[0m: Tracking run with wandb version 0.19.11\r\n[34m[1mwandb[0m: Run data is saved locally in [35m[1m/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/wandb/run-20250627_174120-dtxtnkas[0m\r\n[34m[1mwandb[0m: Run [1m`wandb offline`[0m to turn off syncing.\r\n[34m[1mwandb[0m: Syncing run [33minteractive[0m\r\n[34m[1mwandb[0m: ⭐️ View project at [34m[4mhttps://wandb.ai/instant-uv/jafar[0m\r\n[34m[1mwandb[0m: 🚀 View run at [34m[4mhttps://wandb.ai/instant-uv/jafar/runs/dtxtnkas[0m\r\n
| null |
terminal_output
|
650
| 1,094,711
|
TERMINAL
| 0
| 0
|
2025-06-27 17:41:22.771251: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n
| null |
terminal_output
|
651
| 1,104,248
|
TERMINAL
| 0
| 0
|
2025-06-27 17:41:32.328489: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n
| null |
terminal_output
|
652
| 1,106,285
|
TERMINAL
| 0
| 0
|
2025-06-27 17:41:34.361462: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n
| null |
terminal_output
|
653
| 1,108,425
|
TERMINAL
| 0
| 0
|
Counting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 156736, 'vq': 32768, 'decoder': 156720, 'total': 346224}\r\n
| null |
terminal_output
|
654
| 1,109,020
|
TERMINAL
| 0
| 0
|
Starting training from step 0...\r\nbatch shape: (1, 16, 90, 160, 3)\r\n
| null |
terminal_output
|
655
| 1,115,397
|
TERMINAL
| 0
| 0
|
2025-06-27 17:41:43.474422: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-06-27 17:41:43.474719: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-06-27 17:41:43.474742: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-06-27 17:41:43.474799: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-06-27 17:41:43.475581: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n
| null |
terminal_output
|
656
| 1,139,430
|
TERMINAL
| 0
| 0
|
Step 0, loss: 0.22973881661891937, step time: 30171.963691711426ms\r\n
| null |
terminal_output
|
657
| 1,139,483
|
TERMINAL
| 0
| 0
|
Step 1, loss: 0.22567498683929443, step time: 51.11837387084961ms\r\n
| null |
terminal_output
|
658
| 1,139,535
|
TERMINAL
| 0
| 0
|
Step 2, loss: 0.22210054099559784, step time: 50.873756408691406ms\r\n
| null |
terminal_output
|
659
| 1,139,634
|
TERMINAL
| 0
| 0
|
Step 3, loss: 0.21824738383293152, step time: 50.57096481323242ms\r\n
| null |
terminal_output
|
660
| 1,139,790
|
TERMINAL
| 0
| 0
|
Step 4, loss: 0.2147761434316635, step time: 50.531625747680664ms\r\nStep 5, loss: 0.2116047590970993, step time: 49.07846450805664ms\r\nStep 6, loss: 0.20854488015174866, step time: 48.74777793884277ms\r\nStep 7, loss: 0.20581576228141785, step time: 48.25401306152344ms\r\n
| null |
terminal_output
|
661
| 1,139,891
|
TERMINAL
| 0
| 0
|
Step 8, loss: 0.2033093422651291, step time: 48.099517822265625ms\r\nStep 9, loss: 0.2007577270269394, step time: 48.25139045715332ms\r\n
| null |
terminal_output
|
662
| 1,140,695
|
TERMINAL
| 0
| 0
|
Step 10, loss: 0.19851168990135193, step time: 49.56221580505371ms\r\nStep 11, loss: 0.19617383182048798, step time: 48.66170883178711ms\r\n
| null |
terminal_output
|
663
| 1,140,795
|
TERMINAL
| 0
| 0
|
Step 12, loss: 0.19404467940330505, step time: 48.393964767456055ms\r\nStep 13, loss: 0.19187328219413757, step time: 48.616647720336914ms\r\n
| null |
terminal_output
|
664
| 1,140,847
|
TERMINAL
| 0
| 0
|
Step 14, loss: 0.18987049162387848, step time: 48.28047752380371ms\r\n
| null |
terminal_output
|
665
| 1,140,899
|
TERMINAL
| 0
| 0
|
Step 15, loss: 0.18780189752578735, step time: 48.294782638549805ms\r\n
| null |
terminal_output
|
666
| 1,141,100
|
TERMINAL
| 0
| 0
|
Step 16, loss: 0.1859811544418335, step time: 48.09260368347168ms\r\nStep 17, loss: 0.1842631846666336, step time: 48.24066162109375ms\r\nStep 18, loss: 0.1826377660036087, step time: 48.32100868225098ms\r\nStep 19, loss: 0.18085257709026337, step time: 48.25425148010254ms\r\n
| null |
terminal_output
|
667
| 1,141,374
|
TERMINAL
| 0
| 0
|
Step 20, loss: 0.17928507924079895, step time: 49.29542541503906ms\r\nStep 21, loss: 0.17767782509326935, step time: 48.53558540344238ms\r\nStep 22, loss: 0.17619723081588745, step time: 48.42019081115723ms\r\n
| null |
terminal_output
|
668
| 1,141,475
|
TERMINAL
| 0
| 0
|
Step 23, loss: 0.17467747628688812, step time: 48.22421073913574ms\r\nStep 24, loss: 0.17312972247600555, step time: 48.16746711730957ms\r\n
| null |
terminal_output
|
669
| 1,141,724
|
TERMINAL
| 0
| 0
|
Step 25, loss: 0.17165374755859375, step time: 48.19846153259277ms\r\nStep 26, loss: 0.17034600675106049, step time: 48.27237129211426ms\r\nStep 27, loss: 0.16892999410629272, step time: 48.62570762634277ms\r\nStep 28, loss: 0.16756761074066162, step time: 48.035621643066406ms\r\nStep 29, loss: 0.1661963015794754, step time: 48.21372032165527ms\r\n
| null |
terminal_output
|
670
| 1,141,896
|
TERMINAL
| 0
| 0
|
Step 30, loss: 0.16494277119636536, step time: 49.306631088256836ms\r\n
| null |
terminal_output
|
671
| 1,141,949
|
TERMINAL
| 0
| 0
|
Step 31, loss: 0.16365596652030945, step time: 48.6299991607666ms\r\n
| null |
terminal_output
|
672
| 1,142,046
|
TERMINAL
| 0
| 0
|
Step 32, loss: 0.16249655187129974, step time: 48.61855506896973ms\r\nStep 33, loss: 0.16120009124279022, step time: 48.310279846191406ms\r\n
| null |
terminal_output
|
673
| 1,142,147
|
TERMINAL
| 0
| 0
|
Step 34, loss: 0.16007155179977417, step time: 48.25901985168457ms\r\nStep 35, loss: 0.15878090262413025, step time: 48.39801788330078ms\r\n
| null |
terminal_output
|
674
| 1,142,346
|
TERMINAL
| 0
| 0
|
Step 36, loss: 0.15763361752033234, step time: 48.25305938720703ms\r\nStep 37, loss: 0.1564590185880661, step time: 48.352718353271484ms\r\nStep 38, loss: 0.15543945133686066, step time: 48.21133613586426ms\r\nStep 39, loss: 0.15439952909946442, step time: 48.277854919433594ms\r\n
| null |
terminal_output
|
675
| 1,142,575
|
TERMINAL
| 0
| 0
|
Step 40, loss: 0.15315622091293335, step time: 49.272775650024414ms\r\nStep 41, loss: 0.15220825374126434, step time: 48.54297637939453ms\r\n
| null |
terminal_output
|
676
| 1,142,722
|
TERMINAL
| 0
| 0
|
Step 42, loss: 0.15134300291538239, step time: 48.703670501708984ms\r\nStep 43, loss: 0.15032324194908142, step time: 48.316001892089844ms\r\nStep 44, loss: 0.149506613612175, step time: 48.36773872375488ms\r\n
| null |
terminal_output
|
677
| 1,142,821
|
TERMINAL
| 0
| 0
|
Step 45, loss: 0.14860518276691437, step time: 48.094749450683594ms\r\nStep 46, loss: 0.14767403900623322, step time: 48.261165618896484ms\r\n
| null |
terminal_output
|
678
| 1,142,973
|
TERMINAL
| 0
| 0
|
Step 47, loss: 0.1468399316072464, step time: 48.30765724182129ms\r\nStep 48, loss: 0.14600539207458496, step time: 48.22206497192383ms\r\nStep 49, loss: 0.14503760635852814, step time: 48.39611053466797ms\r\n
| null |
terminal_output
|
679
| 1,143,246
|
TERMINAL
| 0
| 0
|
Step 50, loss: 0.1442810297012329, step time: 49.155235290527344ms\r\nStep 51, loss: 0.14335674047470093, step time: 48.76852035522461ms\r\nStep 52, loss: 0.14240196347236633, step time: 48.488616943359375ms\r\n
| null |
terminal_output
|
680
| 1,143,349
|
TERMINAL
| 0
| 0
|
Step 53, loss: 0.14156025648117065, step time: 48.29525947570801ms\r\nStep 54, loss: 0.14068448543548584, step time: 48.43616485595703ms\r\n
| null |
terminal_output
|
681
| 1,143,596
|
TERMINAL
| 0
| 0
|
Step 55, loss: 0.13976573944091797, step time: 48.24948310852051ms\r\nStep 56, loss: 0.13888418674468994, step time: 48.241376876831055ms\r\nStep 57, loss: 0.13811372220516205, step time: 48.236846923828125ms\r\nStep 58, loss: 0.13715946674346924, step time: 48.2172966003418ms\r\nStep 59, loss: 0.13630743324756622, step time: 48.35391044616699ms\r\n
| null |
terminal_output
|
682
| 1,143,775
|
TERMINAL
| 0
| 0
|
Step 60, loss: 0.13544805347919464, step time: 49.73959922790527ms\r\n
| null |
terminal_output
|
683
| 1,143,924
|
TERMINAL
| 0
| 0
|
Step 61, loss: 0.13468116521835327, step time: 52.21271514892578ms\r\nStep 62, loss: 0.1337975114583969, step time: 48.66600036621094ms\r\nStep 63, loss: 0.13309070467948914, step time: 48.11453819274902ms\r\n
| null |
terminal_output
|
684
| 1,144,127
|
TERMINAL
| 0
| 0
|
Step 64, loss: 0.13232079148292542, step time: 48.374176025390625ms\r\nStep 65, loss: 0.13143107295036316, step time: 48.17509651184082ms\r\nStep 66, loss: 0.13057038187980652, step time: 48.310041427612305ms\r\nStep 67, loss: 0.12974940240383148, step time: 48.227787017822266ms\r\n
| null |
terminal_output
|
685
| 1,144,228
|
TERMINAL
| 0
| 0
|
Step 68, loss: 0.12883210182189941, step time: 48.37393760681152ms\r\nStep 69, loss: 0.12799714505672455, step time: 48.35200309753418ms\r\n
| null |
terminal_output
|
686
| 1,144,452
|
TERMINAL
| 0
| 0
|
Step 70, loss: 0.12717531621456146, step time: 49.127817153930664ms\r\nStep 71, loss: 0.12642090022563934, step time: 48.70462417602539ms\r\n
| null |
terminal_output
|
687
| 1,144,601
|
TERMINAL
| 0
| 0
|
Step 72, loss: 0.12560580670833588, step time: 48.22230339050293ms\r\nStep 73, loss: 0.12476280331611633, step time: 48.45380783081055ms\r\nStep 74, loss: 0.12407097965478897, step time: 48.215627670288086ms\r\n
| null |
terminal_output
|
688
| 1,144,851
|
TERMINAL
| 0
| 0
|
Step 75, loss: 0.1233859434723854, step time: 48.12121391296387ms\r\nStep 76, loss: 0.12275543063879013, step time: 48.40588569641113ms\r\nStep 77, loss: 0.12196243554353714, step time: 48.16579818725586ms\r\nStep 78, loss: 0.12130146473646164, step time: 49.008846282958984ms\r\nStep 79, loss: 0.12060396373271942, step time: 48.2635498046875ms\r\n
| null |
terminal_output
|
689
| 1,145,025
|
TERMINAL
| 0
| 0
|
Step 80, loss: 0.11992035806179047, step time: 49.05271530151367ms\r\n
| null |
terminal_output
|
690
| 1,145,172
|
TERMINAL
| 0
| 0
|
Step 81, loss: 0.11925981938838959, step time: 49.17168617248535ms\r\nStep 82, loss: 0.11857131123542786, step time: 48.26545715332031ms\r\nStep 83, loss: 0.11796504259109497, step time: 48.502206802368164ms\r\n
| null |
terminal_output
|
691
| 1,145,379
|
TERMINAL
| 0
| 0
|
Step 84, loss: 0.11743718385696411, step time: 48.51245880126953ms\r\nStep 85, loss: 0.11669611185789108, step time: 48.42638969421387ms\r\nStep 86, loss: 0.11613167822360992, step time: 48.51245880126953ms\r\nStep 87, loss: 0.11553380638360977, step time: 48.34175109863281ms\r\n
| null |
terminal_output
|
692
| 1,145,475
|
TERMINAL
| 0
| 0
|
Step 88, loss: 0.11495697498321533, step time: 48.36869239807129ms\r\nStep 89, loss: 0.11442873626947403, step time: 48.2487678527832ms\r\n
| null |
terminal_output
|
693
| 1,145,850
|
TERMINAL
| 0
| 0
|
Step 90, loss: 0.11386150121688843, step time: 49.219608306884766ms\r\nStep 91, loss: 0.113355353474617, step time: 48.57182502746582ms\r\nStep 92, loss: 0.11278309673070908, step time: 48.246145248413086ms\r\nStep 93, loss: 0.11220993846654892, step time: 48.467159271240234ms\r\nStep 94, loss: 0.1116456538438797, step time: 48.17533493041992ms\r\n
| null |
terminal_output
|
694
| 1,145,901
|
TERMINAL
| 0
| 0
|
Step 95, loss: 0.11118030548095703, step time: 48.27284812927246ms\r\n
| null |
terminal_output
|
695
| 1,146,101
|
TERMINAL
| 0
| 0
|
Step 96, loss: 0.11055561900138855, step time: 48.27713966369629ms\r\nStep 97, loss: 0.10996095836162567, step time: 48.30574989318848ms\r\nStep 98, loss: 0.10936656594276428, step time: 48.41756820678711ms\r\nStep 99, loss: 0.10877495259046555, step time: 48.235416412353516ms\r\n
| null |
terminal_output
|
696
| 1,146,327
|
TERMINAL
| 0
| 0
|
Step 100, loss: 0.10819685459136963, step time: 49.37624931335449ms\r\nStep 101, loss: 0.10764206200838089, step time: 48.6602783203125ms\r\n
| null |
terminal_output
|
697
| 1,146,477
|
TERMINAL
| 0
| 0
|
Step 102, loss: 0.10705095529556274, step time: 48.17914962768555ms\r\nStep 103, loss: 0.10665144771337509, step time: 48.38085174560547ms\r\nStep 104, loss: 0.10608810186386108, step time: 48.31194877624512ms\r\n
| null |
terminal_output
|
698
| 1,146,727
|
TERMINAL
| 0
| 0
|
Step 105, loss: 0.10568448901176453, step time: 48.33173751831055ms\r\nStep 106, loss: 0.10507740080356598, step time: 48.180341720581055ms\r\nStep 107, loss: 0.10465841740369797, step time: 48.32196235656738ms\r\nStep 108, loss: 0.10413363575935364, step time: 48.16460609436035ms\r\nStep 109, loss: 0.10361927002668381, step time: 48.267364501953125ms\r\n
| null |
terminal_output
|
699
| 1,146,902
|
TERMINAL
| 0
| 0
|
Step 110, loss: 0.10319382697343826, step time: 49.67546463012695ms\r\n
| null |
terminal_output
|
700
| 1,147,050
|
TERMINAL
| 0
| 0
|
Step 111, loss: 0.10276873409748077, step time: 48.65455627441406ms\r\nStep 112, loss: 0.10226455330848694, step time: 48.413991928100586ms\r\nStep 113, loss: 0.10178731381893158, step time: 48.39205741882324ms\r\n
| null |
terminal_output
|
701
| 1,147,101
|
TERMINAL
| 0
| 0
|
Step 114, loss: 0.10129542648792267, step time: 48.32267761230469ms\r\n
| null |
terminal_output
|
702
| 1,147,153
|
TERMINAL
| 0
| 0
|
Step 115, loss: 0.10072394460439682, step time: 48.65121841430664ms\r\n
| null |
terminal_output
|
703
| 1,147,253
|
TERMINAL
| 0
| 0
|
Step 116, loss: 0.1003187894821167, step time: 48.288583755493164ms\r\nStep 117, loss: 0.09988361597061157, step time: 48.55227470397949ms\r\n
| null |
terminal_output
|
704
| 1,147,355
|
TERMINAL
| 0
| 0
|
Step 118, loss: 0.0993257388472557, step time: 48.32887649536133ms\r\nStep 119, loss: 0.09882426261901855, step time: 48.33626747131348ms\r\n
| null |
terminal_output
|
705
| 1,147,731
|
TERMINAL
| 0
| 0
|
Step 120, loss: 0.09837417304515839, step time: 49.399614334106445ms\r\nStep 121, loss: 0.09783116728067398, step time: 48.6602783203125ms\r\nStep 122, loss: 0.09730947017669678, step time: 48.55084419250488ms\r\nStep 123, loss: 0.09676717221736908, step time: 48.3705997467041ms\r\nStep 124, loss: 0.09622417390346527, step time: 48.25854301452637ms\r\n
| null |
terminal_output
|
706
| 1,147,981
|
TERMINAL
| 0
| 0
|
Step 125, loss: 0.09569689631462097, step time: 48.227548599243164ms\r\nStep 126, loss: 0.0952029824256897, step time: 48.26545715332031ms\r\nStep 127, loss: 0.09465636312961578, step time: 48.416852951049805ms\r\nStep 128, loss: 0.09412790834903717, step time: 48.12932014465332ms\r\nStep 129, loss: 0.0936170443892479, step time: 48.375844955444336ms\r\n
| null |
terminal_output
|
707
| 1,148,153
|
TERMINAL
| 0
| 0
|
Step 130, loss: 0.093131884932518, step time: 49.25131797790527ms\r\n
| null |
terminal_output
|
708
| 1,148,354
|
TERMINAL
| 0
| 0
|
Step 131, loss: 0.09257060289382935, step time: 48.7973690032959ms\r\nStep 132, loss: 0.0920131504535675, step time: 48.58708381652832ms\r\nStep 133, loss: 0.09145741909742355, step time: 48.545122146606445ms\r\nStep 134, loss: 0.0909249410033226, step time: 48.39038848876953ms\r\n
| null |
terminal_output
|
709
| 1,148,407
|
TERMINAL
| 0
| 0
|
Step 135, loss: 0.09030286222696304, step time: 48.3551025390625ms\r\n
| null |
terminal_output
|
710
| 1,148,459
|
TERMINAL
| 0
| 0
|
Step 136, loss: 0.08976317942142487, step time: 48.25997352600098ms\r\n
| null |
terminal_output
|
711
| 1,148,610
|
TERMINAL
| 0
| 0
|
Step 137, loss: 0.0892477035522461, step time: 48.37822914123535ms\r\nStep 138, loss: 0.08866812288761139, step time: 48.30431938171387ms\r\nStep 139, loss: 0.0880734845995903, step time: 48.38824272155762ms\r\n
| null |
terminal_output
|
712
| 1,148,930
|
TERMINAL
| 0
| 0
|
Step 140, loss: 0.0875922441482544, step time: 49.527645111083984ms\r\nStep 141, loss: 0.08698724210262299, step time: 48.86651039123535ms\r\nStep 142, loss: 0.0864284560084343, step time: 48.56228828430176ms\r\nStep 143, loss: 0.08598388731479645, step time: 48.454999923706055ms\r\n
| null |
terminal_output
|
713
| 1,148,983
|
TERMINAL
| 0
| 0
|
Step 144, loss: 0.08545909821987152, step time: 49.10683631896973ms\r\n
| null |
terminal_output
|
714
| 1,149,038
|
TERMINAL
| 0
| 0
|
Step 145, loss: 0.0847507044672966, step time: 49.45039749145508ms\r\n
| null |
terminal_output
|
715
| 1,149,089
|
TERMINAL
| 0
| 0
|
Step 146, loss: 0.08411858975887299, step time: 49.73936080932617ms\r\n
| null |
terminal_output
|
716
| 1,149,139
|
TERMINAL
| 0
| 0
|
Step 147, loss: 0.0835011824965477, step time: 49.61681365966797ms\r\n
| null |
terminal_output
|
717
| 1,149,241
|
TERMINAL
| 0
| 0
|
Step 148, loss: 0.08299552649259567, step time: 49.53289031982422ms\r\nStep 149, loss: 0.08232547342777252, step time: 48.88153076171875ms\r\n
| null |
terminal_output
|
718
| 1,149,618
|
TERMINAL
| 0
| 0
|
Step 150, loss: 0.08171946555376053, step time: 48.98548126220703ms\r\nStep 151, loss: 0.08112937211990356, step time: 48.65670204162598ms\r\nStep 152, loss: 0.0805525928735733, step time: 48.24471473693848ms\r\nStep 153, loss: 0.08003890514373779, step time: 48.270225524902344ms\r\nStep 154, loss: 0.07941573113203049, step time: 48.09427261352539ms\r\n
| null |
terminal_output
|
719
| 1,149,864
|
TERMINAL
| 0
| 0
|
Step 155, loss: 0.07884304970502853, step time: 48.2175350189209ms\r\nStep 156, loss: 0.07828739285469055, step time: 48.26045036315918ms\r\nStep 157, loss: 0.07778231799602509, step time: 48.24638366699219ms\r\nStep 158, loss: 0.07719840854406357, step time: 48.14004898071289ms\r\nStep 159, loss: 0.07671979069709778, step time: 48.47073554992676ms\r\n
| null |
terminal_output
|
720
| 1,150,237
|
TERMINAL
| 0
| 0
|
Step 160, loss: 0.07620523869991302, step time: 49.1943359375ms\r\nStep 161, loss: 0.0755499079823494, step time: 48.63595962524414ms\r\nStep 162, loss: 0.07505421340465546, step time: 48.3548641204834ms\r\nStep 163, loss: 0.07471846044063568, step time: 48.364877700805664ms\r\nStep 164, loss: 0.07420185953378677, step time: 48.378705978393555ms\r\n
| null |
terminal_output
|
721
| 1,150,386
|
TERMINAL
| 0
| 0
|
Step 165, loss: 0.07378396391868591, step time: 48.250675201416016ms\r\nStep 166, loss: 0.07334044575691223, step time: 48.074960708618164ms\r\nStep 167, loss: 0.07295762747526169, step time: 48.271894454956055ms\r\n
| null |
terminal_output
|
722
| 1,150,487
|
TERMINAL
| 0
| 0
|
Step 168, loss: 0.07263373583555222, step time: 48.19941520690918ms\r\nStep 169, loss: 0.07218320667743683, step time: 48.32053184509277ms\r\n
| null |
terminal_output
|
723
| 1,150,665
|
TERMINAL
| 0
| 0
|
Step 170, loss: 0.0718797817826271, step time: 49.59750175476074ms\r\n
| null |
terminal_output
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.