Sequence
int64 1
25.2k
| Time
int64 1
858M
| File
stringclasses 830
values | RangeOffset
int64 0
2.21M
| RangeLength
int64 0
168k
| Text
stringlengths 1
4.7M
⌀ | Language
stringclasses 20
values | Type
stringclasses 9
values |
|---|---|---|---|---|---|---|---|
324
| 538,810
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 760
| 0
|
n
|
shellscript
|
content
|
325
| 538,812
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 761
| 0
| null |
shellscript
|
selection_keyboard
|
326
| 539,022
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 761
| 0
|
u
|
shellscript
|
content
|
327
| 539,022
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 762
| 0
| null |
shellscript
|
selection_keyboard
|
328
| 539,192
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 762
| 0
|
m
|
shellscript
|
content
|
329
| 539,193
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 763
| 0
| null |
shellscript
|
selection_keyboard
|
330
| 539,387
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 763
| 0
|
_
|
shellscript
|
content
|
331
| 539,389
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 764
| 0
| null |
shellscript
|
selection_keyboard
|
332
| 539,614
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 764
| 0
|
h
|
shellscript
|
content
|
333
| 539,615
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 765
| 0
| null |
shellscript
|
selection_keyboard
|
334
| 539,713
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 765
| 0
|
e
|
shellscript
|
content
|
335
| 539,714
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 766
| 0
| null |
shellscript
|
selection_keyboard
|
336
| 539,824
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 766
| 0
|
a
|
shellscript
|
content
|
337
| 539,825
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 767
| 0
| null |
shellscript
|
selection_keyboard
|
338
| 539,922
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 767
| 0
|
d
|
shellscript
|
content
|
339
| 539,923
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 768
| 0
| null |
shellscript
|
selection_keyboard
|
340
| 540,609
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 768
| 0
|
s 4 \
|
shellscript
|
content
|
341
| 541,258
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 772
| 0
| null |
shellscript
|
selection_command
|
342
| 541,939
|
notes.md
| 0
| 0
| null |
markdown
|
tab
|
343
| 543,257
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 0
| 0
| null |
shellscript
|
tab
|
344
| 550,658
|
train_tokenizer_single_sample.py
| 0
| 0
| null |
python
|
tab
|
345
| 552,030
|
train_tokenizer_single_sample.py
| 0
| 0
| null |
python
|
selection_command
|
346
| 555,241
|
train_tokenizer_single_sample.py
| 1,193
| 0
| null |
python
|
selection_command
|
347
| 556,218
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 0
| 0
| null |
shellscript
|
tab
|
348
| 558,866
|
notes.md
| 0
| 0
| null |
markdown
|
tab
|
349
| 559,617
|
TERMINAL
| 0
| 0
|
./slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| null |
terminal_output
|
350
| 561,432
|
TERMINAL
| 0
| 0
|
\r\n[?2004l\r# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv_jafar/bin/activate\r\n\r\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\ntf_records_dir=$ws_dir/data/knoms_tfrecords_200_shards\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\ntags=("overfit_sample", "tokenizer", "debug", "alfred")\r\n\r\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\nenv | grep SLURM\r\n\r\npython train_tokenizer_single_sample.py \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=1 \\r\n --min_lr=4.3e-5 \\r\n --max_lr=4.3e-4 \\r\n --log_image_interval=10 \\r\n --log \\r\n --entity instant-uv \\r\n --project jafar \\r\n --name $job_name \\r\n --tags $tags \\r\n --model_dim 256 \\r\n --num_heads 4 \\r\n --data_dir $tf_records_dir\r\n
| null |
terminal_output
|
351
| 561,574
|
TERMINAL
| 0
| 0
|
SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_ind3695\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=991285\r\nSLURM_TASK_PID=707800\r\nSLURM_JOB_GPUS=3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit\r\nSLURMD_NODENAME=hkn0505\r\nSLURM_JOB_START_TIME=1751037996\r\nSLURM_STEP_NODELIST=hkn0505\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1751039796\r\nSLURM_PMI2_SRUN_PORT=39147\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3299579\r\nSLURM_PTY_PORT=34619\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=68\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_NTASKS=1\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e10.hkn0505\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=175\r\nSLURM_NODELIST=hkn0505\r\nSLURM_SRUN_COMM_PORT=41999\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=1\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3299579\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0505\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_NTASKS_PER_NODE=1\r\nSLURM_STEP_LAUNCHER_PORT=41999\r\nSLURM_JOB_GID=502289\r\nSLURM_JOB_NODELIST=hkn0505\r\n
| null |
terminal_output
|
352
| 564,360
|
TERMINAL
| 0
| 0
|
2025-06-27 17:32:32.410771: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\r\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\r\nE0000 00:00:1751038352.423499 710438 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\r\nE0000 00:00:1751038352.427576 710438 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\r\nW0000 00:00:1751038352.439609 710438 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\r\nW0000 00:00:1751038352.439631 710438 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\r\nW0000 00:00:1751038352.439633 710438 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\r\nW0000 00:00:1751038352.439635 710438 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\r\n
| null |
terminal_output
|
353
| 568,084
|
TERMINAL
| 0
| 0
|
W0000 00:00:1751038356.161548 710438 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\r\nSkipping registering GPU devices...\r\n
| null |
terminal_output
|
354
| 568,407
|
TERMINAL
| 0
| 0
|
Running on 1 devices.\r\n
| null |
terminal_output
|
355
| 569,170
|
TERMINAL
| 0
| 0
|
[34m[1mwandb[0m: Currently logged in as: [33mavocadoali[0m ([33minstant-uv[0m) to [32mhttps://api.wandb.ai[0m. Use [1m`wandb login --relogin`[0m to force relogin\r\n
| null |
terminal_output
|
356
| 569,734
|
TERMINAL
| 0
| 0
|
[34m[1mwandb[0m: Tracking run with wandb version 0.19.11\r\n[34m[1mwandb[0m: Run data is saved locally in [35m[1m/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/wandb/run-20250627_173237-7qr0fzyr[0m\r\n[34m[1mwandb[0m: Run [1m`wandb offline`[0m to turn off syncing.\r\n[34m[1mwandb[0m: Syncing run [33minteractive[0m\r\n[34m[1mwandb[0m: ⭐️ View project at [34m[4mhttps://wandb.ai/instant-uv/jafar[0m\r\n[34m[1mwandb[0m: 🚀 View run at [34m[4mhttps://wandb.ai/instant-uv/jafar/runs/7qr0fzyr[0m\r\n
| null |
terminal_output
|
357
| 571,046
|
TERMINAL
| 0
| 0
|
2025-06-27 17:32:39.124222: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n
| null |
terminal_output
|
358
| 582,803
|
TERMINAL
| 0
| 0
|
2025-06-27 17:32:50.748109: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n
| null |
terminal_output
|
359
| 584,801
|
TERMINAL
| 0
| 0
|
2025-06-27 17:32:52.829743: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n
| null |
terminal_output
|
360
| 587,585
|
TERMINAL
| 0
| 0
|
Counting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 4770688, 'vq': 32768, 'decoder': 4770672, 'total': 9574128}\r\n
| null |
terminal_output
|
361
| 588,360
|
TERMINAL
| 0
| 0
|
Starting training from step 0...\r\nbatch shape: (1, 16, 90, 160, 3)\r\n
| null |
terminal_output
|
362
| 599,986
|
TERMINAL
| 0
| 0
|
2025-06-27 17:33:08.064139: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-06-27 17:33:08.064567: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-06-27 17:33:08.064588: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-06-27 17:33:08.064681: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-06-27 17:33:08.066182: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n
| null |
terminal_output
|
363
| 604,251
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 0
| 0
| null |
shellscript
|
tab
|
364
| 608,318
|
TERMINAL
| 0
| 0
|
^C2025-06-27 17:33:16.399639: F external/xla/xla/service/gpu/autotuning/gemm_fusion_autotuner.cc:1136] Non-OK-status: executable.status()\r\nStatus: INTERNAL: ptxas exited with non-zero error code 2, output: - Failure occured when compiling fusion gemm_fusion_dot.228 with config '{block_m:64,block_n:128,block_k:16,split_k:1,num_stages:4,num_warps:2,num_ctas:1}'\r\nFused HLO computation:\r\n%gemm_fusion_dot.228_computation (parameter_0.159: f32[1,16,920,256], parameter_1.159: f32[256,256]) -> f32[1,16,920,256] {\r\n %parameter_0.159 = f32[1,16,920,256]{3,2,1,0} parameter(0)\r\n %bitcast.16502 = f32[14720,256]{1,0} bitcast(%parameter_0.159), metadata={op_name="jit(train_step)/jit(main)/transpose(jvp(TokenizerVQVAE))/decoder/checkpoint/STBlock_7/add_any" source_file="/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/utils/nn.py" source_line=64}\r\n %parameter_1.159 = f32[256,256]{1,0} parameter(1)\r\n %dot.1723 = f32[14720,256]{1,0} dot(%bitcast.16502, %parameter_1.159), lhs_contracting_dims={1}, rhs_contracting_dims={1}, metadata={op_name="jit(train_step)/jit(main)/transpose(jvp(TokenizerVQVAE))/decoder/checkpoint/STBlock_7/Dense_0/dot_general" source_file="/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/.venv_jafar/lib/python3.10/site-packages/flax/linen/linear.py" source_line=287}\r\n ROOT %bitcast.16503 = f32[1,16,920,256]{3,2,1,0} bitcast(%dot.1723), metadata={op_name="jit(train_step)/jit(main)/transpose(jvp(TokenizerVQVAE))/decoder/checkpoint/STBlock_7/Dense_0/dot_general" source_file="/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_run_overfit/.venv_jafar/lib/python3.10/site-packages/flax/linen/linear.py" source_line=287}\r\n}\r\n
| null |
terminal_output
|
365
| 608,812
|
TERMINAL
| 0
| 0
|
^C./slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh: line 34: 710438 Aborted (core dumped) python train_tokenizer_single_sample.py --ckpt_dir $CHECKPOINT_DIR --batch_size=1 --min_lr=4.3e-5 --max_lr=4.3e-4 --log_image_interval=10 --log --entity instant-uv --project jafar --name $job_name --tags $tags --model_dim 256 --num_heads 4 --data_dir $tf_records_dir\r\n]0;tum_ind3695@hkn0505:~/projects/jafar_run_overfit[?2004h(.venv_jafar) [tum_ind3695@hkn0505 jafar_run_overfit]$
| null |
terminal_output
|
366
| 612,406
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 0
| 0
| null |
shellscript
|
tab
|
367
| 615,274
|
notes.md
| 0
| 0
| null |
markdown
|
tab
|
368
| 617,242
|
notes.md
| 9,299
| 0
| null |
markdown
|
selection_mouse
|
369
| 619,710
|
notes.md
| 9,116
| 0
| null |
markdown
|
selection_mouse
|
370
| 620,062
|
notes.md
| 9,116
| 0
|
\n
|
markdown
|
content
|
371
| 620,323
|
notes.md
| 9,117
| 0
|
\n
|
markdown
|
content
|
372
| 620,441
|
notes.md
| 9,118
| 0
|
\n
|
markdown
|
content
|
373
| 620,701
|
notes.md
| 9,118
| 0
| null |
markdown
|
selection_command
|
374
| 620,851
|
notes.md
| 9,117
| 0
| null |
markdown
|
selection_command
|
375
| 621,079
|
notes.md
| 9,117
| 0
|
{'encoder': 4770688, 'vq': 32768, 'decoder': 4770672, 'total': 9574128}
|
markdown
|
content
|
376
| 621,085
|
notes.md
| 9,187
| 0
| null |
markdown
|
selection_command
|
377
| 622,530
|
slurm/dev/alfred/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh
| 0
| 0
| null |
shellscript
|
tab
|
378
| 623,494
|
notes.md
| 0
| 0
| null |
markdown
|
tab
|
379
| 624,123
|
notes.md
| 9,116
| 0
| null |
markdown
|
selection_command
|
380
| 624,322
|
notes.md
| 9,115
| 0
| null |
markdown
|
selection_command
|
381
| 624,490
|
notes.md
| 9,114
| 0
| null |
markdown
|
selection_command
|
382
| 624,656
|
notes.md
| 9,102
| 0
| null |
markdown
|
selection_command
|
383
| 625,293
|
notes.md
| 9,113
| 0
|
\n
|
markdown
|
content
|
384
| 625,693
|
notes.md
| 9,114
| 1
|
\n
|
markdown
|
selection_command
|
385
| 626,545
|
notes.md
| 9,114
| 0
| null |
markdown
|
selection_command
|
386
| 629,711
|
notes.md
| 9,114
| 1
|
\n
|
markdown
|
selection_command
|
387
| 629,907
|
notes.md
| 9,032
| 82
|
{'encoder': 18.978.432, 'vq': 32.768, 'decoder': 18.978.416, 'total': 37.989.616}\n
|
markdown
|
selection_command
|
388
| 630,259
|
notes.md
| 9,018
| 96
|
num_heads = 8\n{'encoder': 18.978.432, 'vq': 32.768, 'decoder': 18.978.416, 'total': 37.989.616}\n
|
markdown
|
selection_command
|
389
| 630,454
|
notes.md
| 9,003
| 111
|
num_blocks = 8\nnum_heads = 8\n{'encoder': 18.978.432, 'vq': 32.768, 'decoder': 18.978.416, 'total': 37.989.616}\n
|
markdown
|
selection_command
|
390
| 630,627
|
notes.md
| 8,988
| 126
|
patch_size = 4\nnum_blocks = 8\nnum_heads = 8\n{'encoder': 18.978.432, 'vq': 32.768, 'decoder': 18.978.416, 'total': 37.989.616}\n
|
markdown
|
selection_command
|
391
| 630,779
|
notes.md
| 8,969
| 145
|
num_latents = 1024\npatch_size = 4\nnum_blocks = 8\nnum_heads = 8\n{'encoder': 18.978.432, 'vq': 32.768, 'decoder': 18.978.416, 'total': 37.989.616}\n
|
markdown
|
selection_command
|
392
| 630,940
|
notes.md
| 8,953
| 161
|
latent_dim = 32\nnum_latents = 1024\npatch_size = 4\nnum_blocks = 8\nnum_heads = 8\n{'encoder': 18.978.432, 'vq': 32.768, 'decoder': 18.978.416, 'total': 37.989.616}\n
|
markdown
|
selection_command
|
393
| 631,112
|
notes.md
| 8,937
| 177
|
model_dim = 256\nlatent_dim = 32\nnum_latents = 1024\npatch_size = 4\nnum_blocks = 8\nnum_heads = 8\n{'encoder': 18.978.432, 'vq': 32.768, 'decoder': 18.978.416, 'total': 37.989.616}\n
|
markdown
|
selection_command
|
394
| 632,145
|
notes.md
| 8,926
| 188
|
**9.6mio**\nmodel_dim = 256\nlatent_dim = 32\nnum_latents = 1024\npatch_size = 4\nnum_blocks = 8\nnum_heads = 8\n{'encoder': 18.978.432, 'vq': 32.768, 'decoder': 18.978.416, 'total': 37.989.616}\n
|
markdown
|
selection_command
|
395
| 632,962
|
notes.md
| 8,926
| 0
| null |
markdown
|
selection_command
|
396
| 633,303
|
notes.md
| 8,937
| 0
| null |
markdown
|
selection_command
|
397
| 633,553
|
notes.md
| 8,953
| 0
| null |
markdown
|
selection_command
|
398
| 633,582
|
notes.md
| 8,969
| 0
| null |
markdown
|
selection_command
|
399
| 633,610
|
notes.md
| 8,988
| 0
| null |
markdown
|
selection_command
|
400
| 633,642
|
notes.md
| 9,003
| 0
| null |
markdown
|
selection_command
|
401
| 633,680
|
notes.md
| 9,018
| 0
| null |
markdown
|
selection_command
|
402
| 633,710
|
notes.md
| 9,032
| 0
| null |
markdown
|
selection_command
|
403
| 633,746
|
notes.md
| 9,114
| 0
| null |
markdown
|
selection_command
|
404
| 633,781
|
notes.md
| 9,115
| 0
| null |
markdown
|
selection_command
|
405
| 633,970
|
notes.md
| 9,116
| 0
| null |
markdown
|
selection_command
|
406
| 634,237
|
notes.md
| 9,116
| 0
|
**9.6mio**\nmodel_dim = 256\nlatent_dim = 32\nnum_latents = 1024\npatch_size = 4\nnum_blocks = 8\nnum_heads = 8\n{'encoder': 18.978.432, 'vq': 32.768, 'decoder': 18.978.416, 'total': 37.989.616}\n\n
|
markdown
|
content
|
407
| 634,244
|
notes.md
| 9,116
| 0
| null |
markdown
|
selection_command
|
408
| 634,705
|
notes.md
| 9,127
| 0
| null |
markdown
|
selection_command
|
409
| 634,883
|
notes.md
| 9,143
| 0
| null |
markdown
|
selection_command
|
410
| 635,288
|
notes.md
| 9,127
| 0
| null |
markdown
|
selection_command
|
411
| 635,707
|
notes.md
| 9,143
| 0
| null |
markdown
|
selection_command
|
412
| 635,954
|
notes.md
| 9,159
| 0
| null |
markdown
|
selection_command
|
413
| 635,971
|
notes.md
| 9,178
| 0
| null |
markdown
|
selection_command
|
414
| 636,121
|
notes.md
| 9,193
| 0
| null |
markdown
|
selection_command
|
415
| 636,317
|
notes.md
| 9,208
| 0
| null |
markdown
|
selection_command
|
416
| 636,965
|
notes.md
| 9,221
| 0
| null |
markdown
|
selection_command
|
417
| 637,252
|
notes.md
| 9,220
| 1
| null |
markdown
|
content
|
418
| 637,991
|
notes.md
| 9,220
| 0
|
4
|
markdown
|
content
|
419
| 637,992
|
notes.md
| 9,221
| 0
| null |
markdown
|
selection_keyboard
|
420
| 638,180
|
notes.md
| 9,220
| 0
| null |
markdown
|
selection_command
|
421
| 638,290
|
notes.md
| 9,205
| 0
| null |
markdown
|
selection_command
|
422
| 638,722
|
notes.md
| 9,220
| 0
| null |
markdown
|
selection_command
|
423
| 638,854
|
notes.md
| 9,234
| 0
| null |
markdown
|
selection_command
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.