Sequence
int64
1
25.2k
Time
int64
1
858M
File
stringclasses
830 values
RangeOffset
int64
0
2.21M
RangeLength
int64
0
168k
Text
stringlengths
1
4.7M
Language
stringclasses
20 values
Type
stringclasses
9 values
624
1,076,254
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
749
0
u
shellscript
content
625
1,076,255
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
750
0
null
shellscript
selection_keyboard
626
1,076,424
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
750
0
m
shellscript
content
627
1,076,425
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
751
0
null
shellscript
selection_keyboard
628
1,076,792
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
751
0
_
shellscript
content
629
1,076,793
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
752
0
null
shellscript
selection_keyboard
630
1,077,069
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
752
0
b
shellscript
content
631
1,077,070
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
753
0
null
shellscript
selection_keyboard
632
1,077,221
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
753
0
l
shellscript
content
633
1,077,222
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
754
0
null
shellscript
selection_keyboard
634
1,077,371
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
754
0
o
shellscript
content
635
1,077,372
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
755
0
null
shellscript
selection_keyboard
636
1,078,007
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
755
0
cks 4 \
shellscript
content
637
1,078,359
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
761
0
null
shellscript
selection_command
638
1,078,830
notes.md
0
0
null
markdown
tab
639
1,080,498
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
0
0
null
shellscript
tab
640
1,083,932
notes.md
0
0
null
markdown
tab
641
1,085,196
TERMINAL
0
0

null
terminal_output
642
1,085,657
TERMINAL
0
0
./slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
null
terminal_output
643
1,086,163
TERMINAL
0
0
\r\n[?2004l\r# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv_jafar/bin/activate\r\n\r\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\ntf_records_dir=$ws_dir/data/knoms_tfrecords_200_shards\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\ntags="overfit_sample tokenizer debug alfred"\r\n\r\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\nenv | grep SLURM\r\n\r\npython train_tokenizer_single_sample.py \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=1 \\r\n --min_lr=4.3e-5 \\r\n --max_lr=4.3e-4 \\r\n --log_image_interval=10 \\r\n --log \\r\n --entity instant-uv \\r\n --project jafar \\r\n --name $job_name \\r\n --tags $tags \\r\n --model_dim 64 \\r\n --num_blocks 4 \\r\n --data_dir $tf_records_dir\r\n
null
terminal_output
644
1,086,310
TERMINAL
0
0
SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_ind3695\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=991285\r\nSLURM_TASK_PID=707800\r\nSLURM_JOB_GPUS=3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit\r\nSLURMD_NODENAME=hkn0505\r\nSLURM_JOB_START_TIME=1751037996\r\nSLURM_STEP_NODELIST=hkn0505\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1751039796\r\nSLURM_PMI2_SRUN_PORT=39147\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3299579\r\nSLURM_PTY_PORT=34619\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=68\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_NTASKS=1\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e10.hkn0505\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=175\r\nSLURM_NODELIST=hkn0505\r\nSLURM_SRUN_COMM_PORT=41999\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=1\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3299579\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0505\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_NTASKS_PER_NODE=1\r\nSLURM_STEP_LAUNCHER_PORT=41999\r\nSLURM_JOB_GID=502289\r\nSLURM_JOB_NODELIST=hkn0505\r\n
null
terminal_output
645
1,088,423
TERMINAL
0
0
2025-06-27 17:41:16.474711: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\r\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\r\nE0000 00:00:1751038876.487429 713983 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\r\nE0000 00:00:1751038876.491509 713983 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\r\nW0000 00:00:1751038876.503669 713983 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\r\nW0000 00:00:1751038876.503687 713983 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\r\nW0000 00:00:1751038876.503690 713983 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\r\nW0000 00:00:1751038876.503693 713983 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\r\n
null
terminal_output
646
1,091,423
TERMINAL
0
0
W0000 00:00:1751038879.498193 713983 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\r\nSkipping registering GPU devices...\r\n
null
terminal_output
647
1,091,903
TERMINAL
0
0
Running on 1 devices.\r\n
null
terminal_output
648
1,092,699
TERMINAL
0
0
wandb: Currently logged in as: avocadoali (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n
null
terminal_output
649
1,093,292
TERMINAL
0
0
wandb: Tracking run with wandb version 0.19.11\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/wandb/run-20250627_174120-dtxtnkas\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run interactive\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/dtxtnkas\r\n
null
terminal_output
650
1,094,711
TERMINAL
0
0
2025-06-27 17:41:22.771251: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n
null
terminal_output
651
1,104,248
TERMINAL
0
0
2025-06-27 17:41:32.328489: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n
null
terminal_output
652
1,106,285
TERMINAL
0
0
2025-06-27 17:41:34.361462: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n
null
terminal_output
653
1,108,425
TERMINAL
0
0
Counting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 156736, 'vq': 32768, 'decoder': 156720, 'total': 346224}\r\n
null
terminal_output
654
1,109,020
TERMINAL
0
0
Starting training from step 0...\r\nbatch shape: (1, 16, 90, 160, 3)\r\n
null
terminal_output
655
1,115,397
TERMINAL
0
0
2025-06-27 17:41:43.474422: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-06-27 17:41:43.474719: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-06-27 17:41:43.474742: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-06-27 17:41:43.474799: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-06-27 17:41:43.475581: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n
null
terminal_output
656
1,139,430
TERMINAL
0
0
Step 0, loss: 0.22973881661891937, step time: 30171.963691711426ms\r\n
null
terminal_output
657
1,139,483
TERMINAL
0
0
Step 1, loss: 0.22567498683929443, step time: 51.11837387084961ms\r\n
null
terminal_output
658
1,139,535
TERMINAL
0
0
Step 2, loss: 0.22210054099559784, step time: 50.873756408691406ms\r\n
null
terminal_output
659
1,139,634
TERMINAL
0
0
Step 3, loss: 0.21824738383293152, step time: 50.57096481323242ms\r\n
null
terminal_output
660
1,139,790
TERMINAL
0
0
Step 4, loss: 0.2147761434316635, step time: 50.531625747680664ms\r\nStep 5, loss: 0.2116047590970993, step time: 49.07846450805664ms\r\nStep 6, loss: 0.20854488015174866, step time: 48.74777793884277ms\r\nStep 7, loss: 0.20581576228141785, step time: 48.25401306152344ms\r\n
null
terminal_output
661
1,139,891
TERMINAL
0
0
Step 8, loss: 0.2033093422651291, step time: 48.099517822265625ms\r\nStep 9, loss: 0.2007577270269394, step time: 48.25139045715332ms\r\n
null
terminal_output
662
1,140,695
TERMINAL
0
0
Step 10, loss: 0.19851168990135193, step time: 49.56221580505371ms\r\nStep 11, loss: 0.19617383182048798, step time: 48.66170883178711ms\r\n
null
terminal_output
663
1,140,795
TERMINAL
0
0
Step 12, loss: 0.19404467940330505, step time: 48.393964767456055ms\r\nStep 13, loss: 0.19187328219413757, step time: 48.616647720336914ms\r\n
null
terminal_output
664
1,140,847
TERMINAL
0
0
Step 14, loss: 0.18987049162387848, step time: 48.28047752380371ms\r\n
null
terminal_output
665
1,140,899
TERMINAL
0
0
Step 15, loss: 0.18780189752578735, step time: 48.294782638549805ms\r\n
null
terminal_output
666
1,141,100
TERMINAL
0
0
Step 16, loss: 0.1859811544418335, step time: 48.09260368347168ms\r\nStep 17, loss: 0.1842631846666336, step time: 48.24066162109375ms\r\nStep 18, loss: 0.1826377660036087, step time: 48.32100868225098ms\r\nStep 19, loss: 0.18085257709026337, step time: 48.25425148010254ms\r\n
null
terminal_output
667
1,141,374
TERMINAL
0
0
Step 20, loss: 0.17928507924079895, step time: 49.29542541503906ms\r\nStep 21, loss: 0.17767782509326935, step time: 48.53558540344238ms\r\nStep 22, loss: 0.17619723081588745, step time: 48.42019081115723ms\r\n
null
terminal_output
668
1,141,475
TERMINAL
0
0
Step 23, loss: 0.17467747628688812, step time: 48.22421073913574ms\r\nStep 24, loss: 0.17312972247600555, step time: 48.16746711730957ms\r\n
null
terminal_output
669
1,141,724
TERMINAL
0
0
Step 25, loss: 0.17165374755859375, step time: 48.19846153259277ms\r\nStep 26, loss: 0.17034600675106049, step time: 48.27237129211426ms\r\nStep 27, loss: 0.16892999410629272, step time: 48.62570762634277ms\r\nStep 28, loss: 0.16756761074066162, step time: 48.035621643066406ms\r\nStep 29, loss: 0.1661963015794754, step time: 48.21372032165527ms\r\n
null
terminal_output
670
1,141,896
TERMINAL
0
0
Step 30, loss: 0.16494277119636536, step time: 49.306631088256836ms\r\n
null
terminal_output
671
1,141,949
TERMINAL
0
0
Step 31, loss: 0.16365596652030945, step time: 48.6299991607666ms\r\n
null
terminal_output
672
1,142,046
TERMINAL
0
0
Step 32, loss: 0.16249655187129974, step time: 48.61855506896973ms\r\nStep 33, loss: 0.16120009124279022, step time: 48.310279846191406ms\r\n
null
terminal_output
673
1,142,147
TERMINAL
0
0
Step 34, loss: 0.16007155179977417, step time: 48.25901985168457ms\r\nStep 35, loss: 0.15878090262413025, step time: 48.39801788330078ms\r\n
null
terminal_output
674
1,142,346
TERMINAL
0
0
Step 36, loss: 0.15763361752033234, step time: 48.25305938720703ms\r\nStep 37, loss: 0.1564590185880661, step time: 48.352718353271484ms\r\nStep 38, loss: 0.15543945133686066, step time: 48.21133613586426ms\r\nStep 39, loss: 0.15439952909946442, step time: 48.277854919433594ms\r\n
null
terminal_output
675
1,142,575
TERMINAL
0
0
Step 40, loss: 0.15315622091293335, step time: 49.272775650024414ms\r\nStep 41, loss: 0.15220825374126434, step time: 48.54297637939453ms\r\n
null
terminal_output
676
1,142,722
TERMINAL
0
0
Step 42, loss: 0.15134300291538239, step time: 48.703670501708984ms\r\nStep 43, loss: 0.15032324194908142, step time: 48.316001892089844ms\r\nStep 44, loss: 0.149506613612175, step time: 48.36773872375488ms\r\n
null
terminal_output
677
1,142,821
TERMINAL
0
0
Step 45, loss: 0.14860518276691437, step time: 48.094749450683594ms\r\nStep 46, loss: 0.14767403900623322, step time: 48.261165618896484ms\r\n
null
terminal_output
678
1,142,973
TERMINAL
0
0
Step 47, loss: 0.1468399316072464, step time: 48.30765724182129ms\r\nStep 48, loss: 0.14600539207458496, step time: 48.22206497192383ms\r\nStep 49, loss: 0.14503760635852814, step time: 48.39611053466797ms\r\n
null
terminal_output
679
1,143,246
TERMINAL
0
0
Step 50, loss: 0.1442810297012329, step time: 49.155235290527344ms\r\nStep 51, loss: 0.14335674047470093, step time: 48.76852035522461ms\r\nStep 52, loss: 0.14240196347236633, step time: 48.488616943359375ms\r\n
null
terminal_output
680
1,143,349
TERMINAL
0
0
Step 53, loss: 0.14156025648117065, step time: 48.29525947570801ms\r\nStep 54, loss: 0.14068448543548584, step time: 48.43616485595703ms\r\n
null
terminal_output
681
1,143,596
TERMINAL
0
0
Step 55, loss: 0.13976573944091797, step time: 48.24948310852051ms\r\nStep 56, loss: 0.13888418674468994, step time: 48.241376876831055ms\r\nStep 57, loss: 0.13811372220516205, step time: 48.236846923828125ms\r\nStep 58, loss: 0.13715946674346924, step time: 48.2172966003418ms\r\nStep 59, loss: 0.13630743324756622, step time: 48.35391044616699ms\r\n
null
terminal_output
682
1,143,775
TERMINAL
0
0
Step 60, loss: 0.13544805347919464, step time: 49.73959922790527ms\r\n
null
terminal_output
683
1,143,924
TERMINAL
0
0
Step 61, loss: 0.13468116521835327, step time: 52.21271514892578ms\r\nStep 62, loss: 0.1337975114583969, step time: 48.66600036621094ms\r\nStep 63, loss: 0.13309070467948914, step time: 48.11453819274902ms\r\n
null
terminal_output
684
1,144,127
TERMINAL
0
0
Step 64, loss: 0.13232079148292542, step time: 48.374176025390625ms\r\nStep 65, loss: 0.13143107295036316, step time: 48.17509651184082ms\r\nStep 66, loss: 0.13057038187980652, step time: 48.310041427612305ms\r\nStep 67, loss: 0.12974940240383148, step time: 48.227787017822266ms\r\n
null
terminal_output
685
1,144,228
TERMINAL
0
0
Step 68, loss: 0.12883210182189941, step time: 48.37393760681152ms\r\nStep 69, loss: 0.12799714505672455, step time: 48.35200309753418ms\r\n
null
terminal_output
686
1,144,452
TERMINAL
0
0
Step 70, loss: 0.12717531621456146, step time: 49.127817153930664ms\r\nStep 71, loss: 0.12642090022563934, step time: 48.70462417602539ms\r\n
null
terminal_output
687
1,144,601
TERMINAL
0
0
Step 72, loss: 0.12560580670833588, step time: 48.22230339050293ms\r\nStep 73, loss: 0.12476280331611633, step time: 48.45380783081055ms\r\nStep 74, loss: 0.12407097965478897, step time: 48.215627670288086ms\r\n
null
terminal_output
688
1,144,851
TERMINAL
0
0
Step 75, loss: 0.1233859434723854, step time: 48.12121391296387ms\r\nStep 76, loss: 0.12275543063879013, step time: 48.40588569641113ms\r\nStep 77, loss: 0.12196243554353714, step time: 48.16579818725586ms\r\nStep 78, loss: 0.12130146473646164, step time: 49.008846282958984ms\r\nStep 79, loss: 0.12060396373271942, step time: 48.2635498046875ms\r\n
null
terminal_output
689
1,145,025
TERMINAL
0
0
Step 80, loss: 0.11992035806179047, step time: 49.05271530151367ms\r\n
null
terminal_output
690
1,145,172
TERMINAL
0
0
Step 81, loss: 0.11925981938838959, step time: 49.17168617248535ms\r\nStep 82, loss: 0.11857131123542786, step time: 48.26545715332031ms\r\nStep 83, loss: 0.11796504259109497, step time: 48.502206802368164ms\r\n
null
terminal_output
691
1,145,379
TERMINAL
0
0
Step 84, loss: 0.11743718385696411, step time: 48.51245880126953ms\r\nStep 85, loss: 0.11669611185789108, step time: 48.42638969421387ms\r\nStep 86, loss: 0.11613167822360992, step time: 48.51245880126953ms\r\nStep 87, loss: 0.11553380638360977, step time: 48.34175109863281ms\r\n
null
terminal_output
692
1,145,475
TERMINAL
0
0
Step 88, loss: 0.11495697498321533, step time: 48.36869239807129ms\r\nStep 89, loss: 0.11442873626947403, step time: 48.2487678527832ms\r\n
null
terminal_output
693
1,145,850
TERMINAL
0
0
Step 90, loss: 0.11386150121688843, step time: 49.219608306884766ms\r\nStep 91, loss: 0.113355353474617, step time: 48.57182502746582ms\r\nStep 92, loss: 0.11278309673070908, step time: 48.246145248413086ms\r\nStep 93, loss: 0.11220993846654892, step time: 48.467159271240234ms\r\nStep 94, loss: 0.1116456538438797, step time: 48.17533493041992ms\r\n
null
terminal_output
694
1,145,901
TERMINAL
0
0
Step 95, loss: 0.11118030548095703, step time: 48.27284812927246ms\r\n
null
terminal_output
695
1,146,101
TERMINAL
0
0
Step 96, loss: 0.11055561900138855, step time: 48.27713966369629ms\r\nStep 97, loss: 0.10996095836162567, step time: 48.30574989318848ms\r\nStep 98, loss: 0.10936656594276428, step time: 48.41756820678711ms\r\nStep 99, loss: 0.10877495259046555, step time: 48.235416412353516ms\r\n
null
terminal_output
696
1,146,327
TERMINAL
0
0
Step 100, loss: 0.10819685459136963, step time: 49.37624931335449ms\r\nStep 101, loss: 0.10764206200838089, step time: 48.6602783203125ms\r\n
null
terminal_output
697
1,146,477
TERMINAL
0
0
Step 102, loss: 0.10705095529556274, step time: 48.17914962768555ms\r\nStep 103, loss: 0.10665144771337509, step time: 48.38085174560547ms\r\nStep 104, loss: 0.10608810186386108, step time: 48.31194877624512ms\r\n
null
terminal_output
698
1,146,727
TERMINAL
0
0
Step 105, loss: 0.10568448901176453, step time: 48.33173751831055ms\r\nStep 106, loss: 0.10507740080356598, step time: 48.180341720581055ms\r\nStep 107, loss: 0.10465841740369797, step time: 48.32196235656738ms\r\nStep 108, loss: 0.10413363575935364, step time: 48.16460609436035ms\r\nStep 109, loss: 0.10361927002668381, step time: 48.267364501953125ms\r\n
null
terminal_output
699
1,146,902
TERMINAL
0
0
Step 110, loss: 0.10319382697343826, step time: 49.67546463012695ms\r\n
null
terminal_output
700
1,147,050
TERMINAL
0
0
Step 111, loss: 0.10276873409748077, step time: 48.65455627441406ms\r\nStep 112, loss: 0.10226455330848694, step time: 48.413991928100586ms\r\nStep 113, loss: 0.10178731381893158, step time: 48.39205741882324ms\r\n
null
terminal_output
701
1,147,101
TERMINAL
0
0
Step 114, loss: 0.10129542648792267, step time: 48.32267761230469ms\r\n
null
terminal_output
702
1,147,153
TERMINAL
0
0
Step 115, loss: 0.10072394460439682, step time: 48.65121841430664ms\r\n
null
terminal_output
703
1,147,253
TERMINAL
0
0
Step 116, loss: 0.1003187894821167, step time: 48.288583755493164ms\r\nStep 117, loss: 0.09988361597061157, step time: 48.55227470397949ms\r\n
null
terminal_output
704
1,147,355
TERMINAL
0
0
Step 118, loss: 0.0993257388472557, step time: 48.32887649536133ms\r\nStep 119, loss: 0.09882426261901855, step time: 48.33626747131348ms\r\n
null
terminal_output
705
1,147,731
TERMINAL
0
0
Step 120, loss: 0.09837417304515839, step time: 49.399614334106445ms\r\nStep 121, loss: 0.09783116728067398, step time: 48.6602783203125ms\r\nStep 122, loss: 0.09730947017669678, step time: 48.55084419250488ms\r\nStep 123, loss: 0.09676717221736908, step time: 48.3705997467041ms\r\nStep 124, loss: 0.09622417390346527, step time: 48.25854301452637ms\r\n
null
terminal_output
706
1,147,981
TERMINAL
0
0
Step 125, loss: 0.09569689631462097, step time: 48.227548599243164ms\r\nStep 126, loss: 0.0952029824256897, step time: 48.26545715332031ms\r\nStep 127, loss: 0.09465636312961578, step time: 48.416852951049805ms\r\nStep 128, loss: 0.09412790834903717, step time: 48.12932014465332ms\r\nStep 129, loss: 0.0936170443892479, step time: 48.375844955444336ms\r\n
null
terminal_output
707
1,148,153
TERMINAL
0
0
Step 130, loss: 0.093131884932518, step time: 49.25131797790527ms\r\n
null
terminal_output
708
1,148,354
TERMINAL
0
0
Step 131, loss: 0.09257060289382935, step time: 48.7973690032959ms\r\nStep 132, loss: 0.0920131504535675, step time: 48.58708381652832ms\r\nStep 133, loss: 0.09145741909742355, step time: 48.545122146606445ms\r\nStep 134, loss: 0.0909249410033226, step time: 48.39038848876953ms\r\n
null
terminal_output
709
1,148,407
TERMINAL
0
0
Step 135, loss: 0.09030286222696304, step time: 48.3551025390625ms\r\n
null
terminal_output
710
1,148,459
TERMINAL
0
0
Step 136, loss: 0.08976317942142487, step time: 48.25997352600098ms\r\n
null
terminal_output
711
1,148,610
TERMINAL
0
0
Step 137, loss: 0.0892477035522461, step time: 48.37822914123535ms\r\nStep 138, loss: 0.08866812288761139, step time: 48.30431938171387ms\r\nStep 139, loss: 0.0880734845995903, step time: 48.38824272155762ms\r\n
null
terminal_output
712
1,148,930
TERMINAL
0
0
Step 140, loss: 0.0875922441482544, step time: 49.527645111083984ms\r\nStep 141, loss: 0.08698724210262299, step time: 48.86651039123535ms\r\nStep 142, loss: 0.0864284560084343, step time: 48.56228828430176ms\r\nStep 143, loss: 0.08598388731479645, step time: 48.454999923706055ms\r\n
null
terminal_output
713
1,148,983
TERMINAL
0
0
Step 144, loss: 0.08545909821987152, step time: 49.10683631896973ms\r\n
null
terminal_output
714
1,149,038
TERMINAL
0
0
Step 145, loss: 0.0847507044672966, step time: 49.45039749145508ms\r\n
null
terminal_output
715
1,149,089
TERMINAL
0
0
Step 146, loss: 0.08411858975887299, step time: 49.73936080932617ms\r\n
null
terminal_output
716
1,149,139
TERMINAL
0
0
Step 147, loss: 0.0835011824965477, step time: 49.61681365966797ms\r\n
null
terminal_output
717
1,149,241
TERMINAL
0
0
Step 148, loss: 0.08299552649259567, step time: 49.53289031982422ms\r\nStep 149, loss: 0.08232547342777252, step time: 48.88153076171875ms\r\n
null
terminal_output
718
1,149,618
TERMINAL
0
0
Step 150, loss: 0.08171946555376053, step time: 48.98548126220703ms\r\nStep 151, loss: 0.08112937211990356, step time: 48.65670204162598ms\r\nStep 152, loss: 0.0805525928735733, step time: 48.24471473693848ms\r\nStep 153, loss: 0.08003890514373779, step time: 48.270225524902344ms\r\nStep 154, loss: 0.07941573113203049, step time: 48.09427261352539ms\r\n
null
terminal_output
719
1,149,864
TERMINAL
0
0
Step 155, loss: 0.07884304970502853, step time: 48.2175350189209ms\r\nStep 156, loss: 0.07828739285469055, step time: 48.26045036315918ms\r\nStep 157, loss: 0.07778231799602509, step time: 48.24638366699219ms\r\nStep 158, loss: 0.07719840854406357, step time: 48.14004898071289ms\r\nStep 159, loss: 0.07671979069709778, step time: 48.47073554992676ms\r\n
null
terminal_output
720
1,150,237
TERMINAL
0
0
Step 160, loss: 0.07620523869991302, step time: 49.1943359375ms\r\nStep 161, loss: 0.0755499079823494, step time: 48.63595962524414ms\r\nStep 162, loss: 0.07505421340465546, step time: 48.3548641204834ms\r\nStep 163, loss: 0.07471846044063568, step time: 48.364877700805664ms\r\nStep 164, loss: 0.07420185953378677, step time: 48.378705978393555ms\r\n
null
terminal_output
721
1,150,386
TERMINAL
0
0
Step 165, loss: 0.07378396391868591, step time: 48.250675201416016ms\r\nStep 166, loss: 0.07334044575691223, step time: 48.074960708618164ms\r\nStep 167, loss: 0.07295762747526169, step time: 48.271894454956055ms\r\n
null
terminal_output
722
1,150,487
TERMINAL
0
0
Step 168, loss: 0.07263373583555222, step time: 48.19941520690918ms\r\nStep 169, loss: 0.07218320667743683, step time: 48.32053184509277ms\r\n
null
terminal_output
723
1,150,665
TERMINAL
0
0
Step 170, loss: 0.0718797817826271, step time: 49.59750175476074ms\r\n
null
terminal_output