mihirma commited on
Commit
7b1f77e
·
verified ·
1 Parent(s): 67113ed

Add files using upload-large-folder tool

Browse files
Files changed (50) hide show
  1. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-03ba0d47-b31c-4c6c-ac79-ac489ea685411761933076688-2025_10_31-18.51.24.953/source.csv +0 -0
  2. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-0aba6691-eb18-4884-b411-e702ed3c10ea1762435058937-2025_11_06-14.17.43.16/source.csv +17 -0
  3. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-0ca86193-5df2-4b75-a3df-f99e6edf1a161761842294336-2025_10_30-17.38.18.780/source.csv +0 -0
  4. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-13abb9ec-f400-4b07-a3bb-8f8000a533f01766266331804-2025_12_20-22.32.20.901/source.csv +44 -0
  5. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-14a0bec6-a4a9-4b59-996b-ff826d7cb6461762440546881-2025_11_06-15.49.10.583/source.csv +22 -0
  6. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-1e1272f3-7d2e-4b26-8662-ef59b0a62e821764410118523-2025_11_29-10.55.25.324/source.csv +0 -0
  7. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-222fe98e-29ac-4b20-9a65-fe2e31f8eb701751128122769-2025_06_28-09.28.47.536/source.csv +29 -0
  8. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-25373d4e-68c9-48b1-a295-2d5d418610141764451376541-2025_11_29-22.22.59.470/source.csv +0 -0
  9. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-266a35fd-8e9a-45ca-bab6-73f8e208e8e31764440451347-2025_11_29-19.20.59.166/source.csv +10 -0
  10. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-2bb200ce-4bc8-4bc3-9354-29e24db5d38e1752063967983-2025_07_09-14.26.42.463/source.csv +0 -0
  11. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-3126c08d-d6fe-466c-828c-ddc956b35e091762367702624-2025_11_05-19.35.09.972/source.csv +12 -0
  12. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-3334d6a5-5e72-43ad-a8da-1813285fd7a51758271588339-2025_09_19-10.46.34.637/source.csv +0 -0
  13. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-395f734f-3624-4a17-97d3-4e353b215a5c1755368459655-2025_08_16-20.21.02.372/source.csv +0 -0
  14. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-3bffc30a-6fb0-48dd-ab21-3fe225eb22c51757148592262-2025_09_06-10.49.57.58/source.csv +0 -0
  15. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-3fe74aa6-92c3-405a-9c6c-49c34aec593b1762364457821-2025_11_05-18.41.02.400/source.csv +0 -0
  16. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-46c4a757-02fe-4d2b-80c9-10277f3bc6421757061314730-2025_09_05-10.35.23.869/source.csv +4 -0
  17. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-48fbb415-6db9-4d35-b548-561e828791bf1751383187013-2025_07_01-17.19.57.60/source.csv +4 -0
  18. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-506a7e8b-c022-493f-90ec-d205964468e01762418679057-2025_11_06-09.44.42.572/source.csv +0 -0
  19. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-5f817451-b135-49e8-83cd-ad3d90c69e831764422803521-2025_11_29-14.26.46.563/source.csv +106 -0
  20. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-63ef9b25-3351-4174-890b-f49574ab1a3c1758993863516-2025_09_27-19.24.25.912/source.csv +512 -0
  21. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-6685e05b-001a-49a4-91eb-69d4235b607a1764445464368-2025_11_29-20.44.27.618/source.csv +54 -0
  22. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-67ee49a3-cbd3-48c6-b0d1-b2c0dde2809c1764447402473-2025_11_29-21.16.46.562/source.csv +110 -0
  23. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-76df12ce-63e6-4f0f-97c2-bd8d132564c31762368713747-2025_11_05-19.51.58.593/source.csv +17 -0
  24. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-7bb39648-a065-4847-8b07-5024f699fc131758998987531-2025_09_27-20.49.55.850/source.csv +0 -0
  25. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-7f860396-c5c8-4f1f-8ce7-04e005748e611754402256906-2025_08_05-15.57.44.850/source.csv +0 -0
  26. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-847fd734-88e4-48ed-bb91-f08cf4e2ebd91762367961128-2025_11_05-19.39.26.989/source.csv +559 -0
  27. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-84bc9952-c4b0-4456-bdc2-984faf53684f1751163593750-2025_06_28-19.19.55.196/source.csv +9 -0
  28. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-971339fb-c463-429e-a956-f7bb98fdea341755623101217-2025_08_19-19.05.03.822/source.csv +0 -0
  29. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-978ed4eb-d9c9-4380-b981-e501087459181750623968304-2025_06_22-13.26.11.394/source.csv +7 -0
  30. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-9deb94c8-8483-468e-9e97-468dc0f8f81a1766266039649-2025_12_20-22.27.31.492/source.csv +6 -0
  31. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-a046d89c-9bee-4791-a53b-dd74376ff2861755421590581-2025_08_17-11.06.44.343/source.csv +0 -0
  32. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-af9d387a-db29-4e3d-9a10-63e0995d4e191758702840888-2025_09_24-10.34.06.824/source.csv +0 -0
  33. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-b00cd52f-686b-4cad-89ec-cf5dcdc287a11753702370531-2025_07_28-13.32.59.505/source.csv +0 -0
  34. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-b656ac92-3ae1-4979-b90f-67b013ce79bf1762368361483-2025_11_05-19.46.06.776/source.csv +0 -0
  35. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-b8ef8518-bf52-489e-b2a9-5e402fd02c471760857710761-2025_10_19-09.08.39.428/source.csv +7 -0
  36. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-c00a9926-ed80-4aa8-b582-d6f932b7ac281754468444238-2025_08_06-10.20.46.993/source.csv +0 -0
  37. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-c3986074-5718-42fa-9de0-e87adef2d7e21764449343379-2025_11_29-21.49.07.420/source.csv +204 -0
  38. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-c8c1182e-9d15-45c9-9466-b9340fc3403b1754557906412-2025_08_07-11.11.49.271/source.csv +27 -0
  39. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-cb2b7623-a503-48be-abf1-496fd7fc45d81755596347163-2025_08_19-11.39.13.701/source.csv +0 -0
  40. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-ccf7a99f-954c-4a69-b08a-cef496899ce51762421470970-2025_11_06-10.31.14.52/source.csv +22 -0
  41. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-ce6b4bab-9c2d-4454-8220-eaa00cd55d321764450093947-2025_11_29-22.01.38.36/source.csv +83 -0
  42. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-d5c0f30b-efff-4cba-9421-06a8c78914011759343086476-2025_10_01-20.24.54.587/source.csv +0 -0
  43. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-dbb18611-0352-4268-ab16-fb8c7e2b82741763481981825-2025_11_18-17.06.23.992/source.csv +98 -0
  44. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-e1878deb-64ca-4d0d-aaf1-5699f0eabd291756538623447-2025_08_30-08.23.51.174/source.csv +0 -0
  45. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-ea3b8a60-f42f-4cc5-977e-bdd5170b12601763540725942-2025_11_19-09.25.37.711/source.csv +0 -0
  46. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-f0382786-979c-4a6d-8e9b-f5977f18eb4f1753726151187-2025_07_28-20.09.13.67/source.csv +0 -0
  47. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-f1861737-541e-498a-8209-7a659cb20b861757609758410-2025_09_11-18.56.00.140/source.csv +2 -0
  48. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-f6f489a9-a5c9-44d3-8bfa-e950be10e08d1763478651049-2025_11_18-16.10.57.25/source.csv +0 -0
  49. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-faba6583-b2c9-4b94-9ba6-9f240428520a1750722089894-2025_06_23-23.49.28.299/source.csv +151 -0
  50. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-fc31d59a-d5b8-4dc1-8760-333f6ac9e55c1763046484160-2025_11_13-16.08.05.840/source.csv +0 -0
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-03ba0d47-b31c-4c6c-ac79-ac489ea685411761933076688-2025_10_31-18.51.24.953/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-0aba6691-eb18-4884-b411-e702ed3c10ea1762435058937-2025_11_06-14.17.43.16/source.csv ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,13,"Untitled-1",0,0,"",plaintext,tab
3
+ 2,134,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"2:17:43 PM [info] Activating crowd-code\n2:17:43 PM [info] Recording started\n2:17:43 PM [info] Initializing git provider using file system watchers...\n2:17:43 PM [info] No workspace folder found\n",Log,tab
4
+ 3,2022,"extension-output-pdoom-org.crowd-code-#1-crowd-code",194,0,"2:17:45 PM [info] Retrying git provider initialization...\n2:17:45 PM [info] No workspace folder found\n",Log,content
5
+ 4,47988,"Untitled-1",0,0,"",plaintext,tab
6
+ 5,53515,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
7
+ 6,54656,"Untitled-1",0,0,"",plaintext,tab
8
+ 7,57479,"Untitled-2",0,0,"",plaintext,tab
9
+ 8,58471,"Untitled-2",0,0,"hello world\n",plaintext,content
10
+ 9,58512,"TERMINAL",0,0,"undefinedfranzsrambical@MBF6N9WFVKFV ~ % echo VSCode test",,terminal_command
11
+ 10,58512,"TERMINAL",0,0,"]633;CVSCode test\r\n% \r \r",,terminal_output
12
+ 11,58923,"Untitled-2",0,0,"",plaintext,selection_command
13
+ 12,58925,"Untitled-2",0,0,"hello world\n",plaintext,content
14
+ 13,58985,"TERMINAL",0,0,"echo VSCode test",,terminal_command
15
+ 14,58986,"TERMINAL",0,0,"]633;CVSCode test\r\n% \r \r",,terminal_output
16
+ 15,63927,"Untitled-2",0,24,"",plaintext,content
17
+ 16,63948,"Untitled-1",0,0,"",plaintext,tab
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-0ca86193-5df2-4b75-a3df-f99e6edf1a161761842294336-2025_10_30-17.38.18.780/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-13abb9ec-f400-4b07-a3bb-8f8000a533f01766266331804-2025_12_20-22.32.20.901/source.csv ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,2,"train.py",0,0,"import ray\nfrom sglang.srt.constants import GPU_MEMORY_TYPE_KV_CACHE, GPU_MEMORY_TYPE_WEIGHTS\n\ntry:\n from sglang.srt.constants import GPU_MEMORY_TYPE_CUDA_GRAPH\nexcept ImportError:\n GPU_MEMORY_TYPE_CUDA_GRAPH = None\n\nfrom miles.ray.placement_group import create_placement_groups, create_rollout_manager, create_training_models\nfrom miles.utils.arguments import parse_args\nfrom miles.utils.logging_utils import configure_logger\nfrom miles.utils.misc import should_run_periodic_action\nfrom miles.utils.tracking_utils import init_tracking\n\n\ndef train(args):\n configure_logger()\n # allocate the GPUs\n pgs = create_placement_groups(args)\n init_tracking(args)\n\n # create the rollout manager, with sglang engines inside.\n # need to initialize rollout manager first to calculate num_rollout\n rollout_manager, num_rollout_per_epoch = create_rollout_manager(args, pgs[""rollout""])\n\n # create the actor and critic models\n actor_model, critic_model = create_training_models(args, pgs, rollout_manager)\n\n if args.offload_rollout:\n ray.get(rollout_manager.onload.remote(tags=[GPU_MEMORY_TYPE_WEIGHTS]))\n\n # always update weight first so that sglang has the loaded weights from training.\n actor_model.update_weights()\n\n if args.check_weight_update_equal:\n ray.get(rollout_manager.check_weights.remote(action=""compare""))\n\n if args.offload_rollout:\n if GPU_MEMORY_TYPE_CUDA_GRAPH is not None:\n ray.get(rollout_manager.onload.remote(tags=[GPU_MEMORY_TYPE_CUDA_GRAPH]))\n ray.get(rollout_manager.onload.remote(tags=[GPU_MEMORY_TYPE_KV_CACHE]))\n\n # special case for eval-only\n if args.num_rollout == 0 and args.eval_interval is not None:\n ray.get(rollout_manager.eval.remote(rollout_id=0))\n\n def offload_train():\n if args.offload_train:\n if args.use_critic:\n critic_model.offload()\n if rollout_id >= args.num_critic_only_steps:\n actor_model.offload()\n else:\n actor_model.offload()\n else:\n actor_model.clear_memory()\n\n def onload_rollout():\n if args.offload_rollout:\n ray.get(rollout_manager.onload.remote(tags=[GPU_MEMORY_TYPE_WEIGHTS]))\n\n # train loop.\n # note that for async training, one can change the position of the sync operation(ray.get).\n for rollout_id in range(args.start_rollout_id, args.num_rollout):\n if args.eval_interval is not None and rollout_id == 0:\n ray.get(rollout_manager.eval.remote(rollout_id))\n\n rollout_data_ref = ray.get(rollout_manager.generate.remote(rollout_id))\n\n if args.offload_rollout:\n ray.get(rollout_manager.offload.remote())\n\n if args.use_critic:\n critic_train_handle = critic_model.async_train(rollout_id, rollout_data_ref)\n if rollout_id >= args.num_critic_only_steps:\n ray.get(actor_model.async_train(rollout_id, rollout_data_ref))\n ray.get(critic_train_handle)\n else:\n ray.get(actor_model.async_train(rollout_id, rollout_data_ref))\n\n if should_run_periodic_action(rollout_id, args.save_interval, num_rollout_per_epoch):\n if (not args.use_critic) or (rollout_id >= args.num_critic_only_steps):\n actor_model.save_model(rollout_id)\n if args.use_critic:\n critic_model.save_model(rollout_id)\n if args.rollout_global_dataset:\n ray.get(rollout_manager.save.remote(rollout_id))\n\n offload_train()\n onload_rollout()\n actor_model.update_weights()\n\n if args.offload_rollout:\n if GPU_MEMORY_TYPE_CUDA_GRAPH is not None:\n ray.get(rollout_manager.onload.remote(tags=[GPU_MEMORY_TYPE_CUDA_GRAPH]))\n ray.get(rollout_manager.onload.remote(tags=[GPU_MEMORY_TYPE_KV_CACHE]))\n\n if should_run_periodic_action(rollout_id, args.eval_interval, num_rollout_per_epoch):\n ray.get(rollout_manager.eval.remote(rollout_id))\n\n ray.get(rollout_manager.dispose.remote())\n\n\nif __name__ == ""__main__"":\n args = parse_args()\n train(args)\n",python,tab
3
+ 2,154,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"10:32:20 PM [info] Activating crowd-code\n10:32:20 PM [info] Recording started\n10:32:20 PM [info] Initializing git provider using file system watchers...\n",Log,tab
4
+ 3,181,"extension-output-pdoom-org.crowd-code-#1-crowd-code",153,0,"10:32:21 PM [info] Git repository found\n10:32:21 PM [info] Git provider initialized successfully\n10:32:21 PM [info] Initial git state: [object Object]\n",Log,content
5
+ 4,800734,"train.py",0,0,"",python,tab
6
+ 5,841318,"train.py",1218,0,"",python,selection_command
7
+ 6,883121,"train.py",1091,0,"",python,selection_command
8
+ 7,2043278,"scripts/run-qwen3-0.6B-torch-sft-crowd-code.sh",0,0,"#!/bin/bash\n\n# for rerun the task\npkill -9 sglang\nsleep 3\nray stop --force\npkill -9 ray\npkill -9 python\nsleep 3\npkill -9 ray\npkill -9 python\n\n# set -ex\n\n# will prevent ray from buffering stdout/stderr\nexport PYTHONUNBUFFERED=1\nexport CUDA_VISIBLE_DEVICES=0,1\n\nNVLINK_COUNT=$(nvidia-smi | grep -o ""NVLink"" | wc -l)\nif [ ""$NVLINK_COUNT"" -gt 0 ]; then\n HAS_NVLINK=1\nelse\n HAS_NVLINK=0\nfi\necho ""HAS_NVLINK: $HAS_NVLINK (detected $NVLINK_COUNT NVLink references)""\n\n\n# --- 1. DYNAMIC HOST IP DETECTION (CRITICAL FOR SLURM) ---\n# Don't hardcode IP. Get the actual IP of the current node.\nexport HEAD_NODE_IP=$(hostname -I | awk '{print $1}')\necho ""Detected Head Node IP: ${HEAD_NODE_IP}""\n\n# --- 2. PROXY CONFIGURATION ---\n# Ensure local traffic doesn't go through a corporate proxy\nexport no_proxy=""${HEAD_NODE_IP},localhost,127.0.0.1,0.0.0.0""\nexport NO_PROXY=""${HEAD_NODE_IP},localhost,127.0.0.1,0.0.0.0""\n\n# --- 3. DEBUGGING & STABILITY ENV VARS ---\n# Force NCCL/Distributed into a robust mode to prevent initialization hangs\n# export NCCL_P2P_DISABLE=1\n# export NCCL_IB_DISABLE=1\nexport NCCL_DEBUG=INFO\nexport TORCH_DISTRIBUTED_DEBUG=INFO\n\nSCRIPT_DIR=""$(cd -- ""$(dirname -- ""${BASH_SOURCE[0]}"")"" &>/dev/null && pwd)""\n\nRUN_ID=${RUN_ID:-""run_$(date +%Y%m%d_%H%M%S)""}\nLOAD_SAVE_PATH=""/fast/project/HFMI_SynergyUnit/tab_model/huggingface/shared_data/${RUN_ID}/checkpoints""\n\nCKPT_ARGS=(\n --hf-checkpoint /fast/project/HFMI_SynergyUnit/tab_model/huggingface/Qwen3-0.6B\n --load /fast/project/HFMI_SynergyUnit/tab_model/huggingface/Qwen3-0.6B\n --ref-load /fast/project/HFMI_SynergyUnit/tab_model/huggingface/Qwen3-0.6B\n)\n\nSFT_ARGS=(\n --rollout-function-path miles.rollout.sft_rollout.generate_rollout\n --prompt-data /fast/project/HFMI_SynergyUnit/tab_model/huggingface/nemo_hf_part_jsonl_4k_tokens.parquet\n --input-key messages\n --apply-chat-template\n --rollout-shuffle\n --num-epoch 3\n --rollout-batch-size 16\n --global-batch-size 16\n\n --loss-type sft_loss\n --calculate-per-token-loss\n --disable-compute-advantages-and-returns\n --num-rollout 2000\n --debug-train-only\n)\n\nOPTIMIZER_ARGS=(\n --optimizer adam\n --lr 1e-5\n --lr-decay-style WSD\n --lr-wsd-decay-style linear\n --lr-warmup-iters 100\n --lr-decay-iters 2000\n --lr-wsd-decay-iters 500\n --weight-decay 0.1\n --adam-beta1 0.9\n --adam-beta2 0.98\n)\n\nWANDB_ARGS=(\n --use-wandb\n --wandb-project crowd-pilot-miles\n --wandb-team instant-uv\n --wandb-group qwen3-0.6b-sft-torch\n)\n\nTRAIN_BACKEND_ARGS=(\n --train-backend fsdp\n --update-weight-buffer-size 536870912\n --gradient-checkpointing\n --attn-implementation flash_attention_3\n --train-env-vars '{""PYTORCH_CUDA_ALLOC_CONF"":""expandable_segments:True""}'\n --actor-num-gpus-per-node 2\n)\n\nPERF_ARGS=(\n --use-dynamic-batch-size\n --max-tokens-per-gpu 9216\n)\n\nMISC_ARGS=(\n --actor-num-nodes 1\n --actor-num-gpus-per-node 2\n --colocate\n --rollout-max-context-len 8192\n --rollout-max-prompt-len 8000\n --rollout-max-response-len 8192\n --use-fault-tolerance\n --dump-details /fast/project/HFMI_SynergyUnit/tab_model/huggingface/shared_data/qwen3-600M-fsdp-1116-noref/dump_details\n)\n\n# launch the master node of ray in container - 2 GPUs for training\nexport MASTER_ADDR=${MASTER_ADDR:-""127.0.0.1""}\npython3 -m ray.scripts.scripts start --head \\n --node-ip-address=${HEAD_NODE_IP} \\n --num-gpus 2 \\n --num-cpus 4 \\n --memory=214748364800 \\n --disable-usage-stats \\n --dashboard-host=0.0.0.0 \\n --dashboard-port=8265 \\n --port=6379\n\necho ""Ray started. Waiting for Dashboard to be ready...""\n\n# --- 4. WAIT FOR DASHBOARD (FIX FOR 504 ERROR) ---\n# Loop until the dashboard port accepts connections\nfor i in {1..30}; do\n if curl -s ""http://${HEAD_NODE_IP}:8265"" > /dev/null; then\n echo ""Dashboard is up!""\n break\n fi\n echo ""Waiting for Ray Dashboard...""\n sleep 2\ndone\n# Add a small safety buffer\nsleep 5\n\n# Build runtime env\nRUNTIME_ENV_JSON=""{\n \""env_vars\"": {\n \""PYTHONPATH\"": \""/fast/project/HFMI_SynergyUnit/mihir/Megatron-LM/\"",\n \""CUDA_DEVICE_MAX_CONNECTIONS\"": \""1\"",\n \""NCCL_NVLS_ENABLE\"": \""${HAS_NVLINK}\"",\n \""PYTORCH_CUDA_ALLOC_CONF\"": \""expandable_segments:True\""\n }\n}""\n\npython3 -m ray.scripts.scripts job submit --address=""http://${HEAD_NODE_IP}:8265"" \\n --runtime-env-json=""${RUNTIME_ENV_JSON}"" \\n -- python3 train.py \\n ${CKPT_ARGS[@]} \\n ${SFT_ARGS[@]} \\n ${OPTIMIZER_ARGS[@]} \\n ${WANDB_ARGS[@]} \\n ${SGLANG_ARGS[@]} \\n ${TRAIN_BACKEND_ARGS[@]} \\n ${PERF_ARGS[@]} \\n ${MISC_ARGS[@]}\n\n\n\n",shellscript,tab
9
+ 8,2047087,"train.py",0,0,"",python,tab
10
+ 9,2047968,"scripts/run-qwen3-0.6B-torch-sft-crowd-code.sh",0,0,"",shellscript,tab
11
+ 10,2048525,"train.py",0,0,"",python,tab
12
+ 11,2475294,"train_sft.py",0,0,"#!/usr/bin/env python3\n""""""\nRay-free SFT Training Script for Miles.\n\nThis script provides a simplified training path for Supervised Fine-Tuning (SFT)\nthat bypasses Ray entirely and uses torchrun for distributed training.\n\nUsage:\n torchrun --nproc_per_node=2 train_sft.py \\n --hf-checkpoint /path/to/model \\n --prompt-data /path/to/data.parquet \\n --input-key messages \\n --apply-chat-template \\n ...\n\nThis is equivalent to the Ray-based SFT with --debug-train-only, but without\nthe Ray overhead.\n""""""\n\nimport copy\nimport logging\nimport os\nfrom argparse import Namespace\nfrom datetime import timedelta\nfrom itertools import accumulate\nfrom pathlib import Path\n\nimport torch\nimport torch.distributed as dist\nfrom torch.distributed.device_mesh import init_device_mesh\nfrom tqdm import tqdm\nfrom transformers import AutoConfig\n\nfrom miles.backends.fsdp_utils import checkpoint\nfrom miles.backends.fsdp_utils.actor import (\n apply_fsdp2,\n get_logprob_and_entropy_with_cp,\n sum_of_sample_mean,\n)\nfrom miles.backends.fsdp_utils.data_packing import pack_sequences, unpack_sequences\nfrom miles.backends.fsdp_utils.lr_scheduler import get_lr_scheduler\nfrom miles.rollout.data_source import RolloutDataSource\nfrom miles.utils import tracking_utils\nfrom miles.utils.arguments import parse_args\nfrom miles.utils.data import Dataset, get_minimum_num_micro_batch_size\nfrom miles.utils.distributed_utils import get_gloo_group, init_gloo_group\nfrom miles.utils.logging_utils import configure_logger\nfrom miles.utils.mask_utils import MultiTurnLossMaskGenerator\nfrom miles.utils.metric_utils import compute_rollout_step\nfrom miles.utils.misc import load_function, should_run_periodic_action\nfrom miles.utils.processing_utils import load_processor, load_tokenizer\nfrom miles.utils.profile_utils import TrainProfiler\nfrom miles.utils.timer import timer\nfrom miles.utils.tracking_utils import init_tracking\nfrom miles.utils.types import Sample\n\nlogger = logging.getLogger(__name__)\n\n\nclass SFTTrainer:\n """"""\n A simplified trainer for SFT that runs without Ray.\n\n This class combines the functionality of:\n - FSDPTrainRayActor (model initialization, FSDP wrapping, training)\n - RolloutManager (data loading via generate_rollout)\n - The main training loop from train.py\n """"""\n\n def __init__(self, args: Namespace):\n self.args = args\n self.device = torch.device(""cuda"")\n\n # Setup distributed\n self._init_distributed()\n\n # Setup device mesh for FSDP\n self._setup_device_mesh()\n\n torch.manual_seed(args.seed)\n\n # Initialize tracking on rank 0\n if dist.get_rank() == 0:\n init_tracking(args, primary=True)\n\n # Load tokenizer and config\n self._load_tokenizer_and_config()\n\n # Initialize data source\n self._init_data_source()\n\n # Initialize model\n self._init_model()\n\n # Initialize optimizer and scheduler\n self._init_optimizer()\n\n # Load checkpoint if available\n self._load_checkpoint()\n\n # Initialize profiler\n self.prof = TrainProfiler(args)\n self.prof.on_init_end()\n\n logger.info(f""[Rank {dist.get_rank()}] SFTTrainer initialized successfully"")\n\n def _init_distributed(self):\n """"""Initialize distributed training.""""""\n # torchrun sets these environment variables\n local_rank = int(os.environ.get(""LOCAL_RANK"", 0))\n torch.cuda.set_device(f""cuda:{local_rank}"")\n\n backend = self.args.distributed_backend\n dist.init_process_group(\n backend=backend,\n timeout=timedelta(minutes=self.args.distributed_timeout_minutes),\n )\n init_gloo_group()\n\n self.args.rank = dist.get_rank()\n self.args.world_size = dist.get_world_size()\n\n logger.info(\n f""[Rank {self.args.rank}] Distributed initialized: ""\n f""world_size={self.args.world_size}, local_rank={local_rank}""\n )\n\n def _setup_device_mesh(self):\n """"""Setup device mesh for FSDP (pure DP mode for SFT).""""""\n world_size = dist.get_world_size()\n rank = dist.get_rank()\n\n # For SFT, we use pure DP (cp_size=1)\n self.cp_size = getattr(self.args, ""context_parallel_size"", 1)\n self.dp_size = world_size // self.cp_size\n\n self.mesh = init_device_mesh(\n ""cuda"",\n mesh_shape=(self.dp_size, self.cp_size),\n mesh_dim_names=(""dp"", ""cp""),\n )\n\n self.dp_group = self.mesh.get_group(""dp"")\n self.cp_group = self.mesh.get_group(""cp"")\n self.dp_mesh = self.mesh[""dp""]\n\n self.dp_rank = rank // self.cp_size\n self.cp_rank = rank % self.cp_size\n\n logger.info(\n f""[Rank {rank}] Device mesh: dp_size={self.dp_size}, cp_size={self.cp_size}, ""\n f""dp_rank={self.dp_rank}, cp_rank={self.cp_rank}""\n )\n\n def _load_tokenizer_and_config(self):\n """"""Load tokenizer and model config sequentially to avoid race conditions.""""""\n for i in range(dist.get_world_size()):\n if i == dist.get_rank():\n self.hf_config = AutoConfig.from_pretrained(\n self.args.hf_checkpoint, trust_remote_code=True\n )\n self.tokenizer = load_tokenizer(\n self.args.hf_checkpoint, trust_remote_code=True\n )\n self.processor = None\n if self.args.multimodal_keys:\n self.processor = load_processor(\n self.args.hf_checkpoint, trust_remote_code=True\n )\n dist.barrier(group=get_gloo_group())\n\n # Initialize loss mask generator for SFT\n self.mask_generator = MultiTurnLossMaskGenerator(\n self.tokenizer,\n tokenizer_type=getattr(self.args, ""loss_mask_type"", None),\n )\n\n def _init_data_source(self):\n """"""Initialize the data source for SFT training.""""""\n self.data_source = RolloutDataSource(self.args)\n\n # Calculate num_rollout from dataset size\n if self.args.num_rollout is None:\n num_rollout_per_epoch = len(self.data_source.dataset) // self.args.rollout_batch_size\n self.args.num_rollout = num_rollout_per_epoch * self.args.num_epoch\n self.num_rollout_per_epoch = num_rollout_per_epoch\n else:\n self.num_rollout_per_epoch = None\n\n if getattr(self.args, ""start_rollout_id"", None) is None:\n self.args.start_rollout_id = 0\n\n logger.info(\n f""[Rank {dist.get_rank()}] Data source initialized: ""\n f""dataset_size={len(self.data_source.dataset)}, ""\n f""num_rollout={self.args.num_rollout}""\n )\n\n def _get_init_weight_context_manager(self):\n """"""Get context manager for model initialization.""""""\n from accelerate import init_empty_weights\n\n use_meta_tensor = not self.hf_config.tie_word_embeddings\n\n def cpu_init_weights():\n return torch.device(""cpu"")\n\n if use_meta_tensor:\n return init_empty_weights if dist.get_rank() != 0 else cpu_init_weights\n else:\n return cpu_init_weights\n\n def _fsdp2_load_full_state_dict(self, model, full_state, device_mesh, cpu_offload):\n """"""Load full state dict into FSDP2 model with broadcast from rank 0.""""""\n from torch.distributed.checkpoint.state_dict import (\n StateDictOptions,\n set_model_state_dict,\n )\n\n if dist.get_rank() == 0:\n model = model.to(device=torch.cuda.current_device(), non_blocking=True)\n else:\n model = model.to_empty(device=torch.cuda.current_device())\n\n is_cpu_offload = cpu_offload is not None\n options = StateDictOptions(\n full_state_dict=True, cpu_offload=is_cpu_offload, broadcast_from_rank0=True\n )\n\n set_model_state_dict(model, full_state, options=options)\n\n for _name, buf in model.named_buffers():\n dist.broadcast(buf, src=0)\n\n if is_cpu_offload:\n model.to(""cpu"", non_blocking=True)\n for buf in model.buffers():\n buf.data = buf.data.to(torch.cuda.current_device())\n\n return model\n\n def _get_model_cls(self):\n """"""Get the appropriate model class based on config.""""""\n if hasattr(self.hf_config, ""vision_config""):\n from transformers import AutoModelForImageTextToText\n\n return AutoModelForImageTextToText\n else:\n from transformers import AutoModelForCausalLM\n\n return AutoModelForCausalLM\n\n def _init_model(self):\n """"""Initialize and wrap model with FSDP.""""""\n self.fsdp_cpu_offload = getattr(self.args, ""fsdp_cpu_offload"", False)\n\n init_context = self._get_init_weight_context_manager()\n\n with init_context():\n model = self._get_model_cls().from_pretrained(\n self.args.hf_checkpoint,\n trust_remote_code=True,\n attn_implementation=self.args.attn_implementation,\n )\n\n model.train()\n full_state = model.state_dict()\n\n model = apply_fsdp2(\n model, mesh=self.dp_mesh, cpu_offload=self.fsdp_cpu_offload, args=self.args\n )\n\n model = self._fsdp2_load_full_state_dict(\n model,\n full_state,\n self.dp_mesh,\n cpu_offload=True if self.fsdp_cpu_offload else None,\n )\n\n self.model = model\n\n if self.args.gradient_checkpointing:\n self.model.gradient_checkpointing_enable()\n\n logger.info(f""[Rank {dist.get_rank()}] Model initialized with FSDP"")\n\n def _init_optimizer(self):\n """"""Initialize optimizer and learning rate scheduler.""""""\n if self.args.optimizer == ""adam"":\n self.optimizer = torch.optim.AdamW(\n self.model.parameters(),\n lr=self.args.lr,\n betas=(self.args.adam_beta1, self.args.adam_beta2),\n eps=self.args.adam_eps,\n weight_decay=self.args.weight_decay,\n )\n else:\n raise ValueError(f""Unsupported optimizer: {self.args.optimizer}"")\n\n self.lr_scheduler = get_lr_scheduler(self.args, self.optimizer)\n self.global_step = 0\n self.micro_step = 0\n\n def _load_checkpoint(self):\n """"""Load checkpoint if available.""""""\n checkpoint_payload = checkpoint.load(self)\n checkpoint.finalize_load(self, checkpoint_payload)\n\n def generate_sft_rollout(self, rollout_id: int) -> list[Sample]:\n """"""Generate SFT rollout data (tokenize and create loss masks).""""""\n samples = self.data_source.get_samples(self.args.rollout_batch_size)\n\n result = []\n for i, (sample,) in enumerate(samples):\n messages = sample.prompt\n token_ids, loss_mask = self.mask_generator.get_loss_mask(messages)\n response_length = self.mask_generator.get_response_lengths([loss_mask])[0]\n\n sample.tokens = token_ids\n sample.response_length = response_length\n sample.reward = 0\n sample.loss_mask = loss_mask[-response_length:]\n result.append(sample)\n\n if i == 0 and rollout_id == 0 and dist.get_rank() == 0:\n logger.info(\n f""SFT rollout sample: tokens_len={len(token_ids)}, ""\n f""response_length={response_length}""\n )\n\n return result\n\n def _convert_samples_to_train_data(self, samples: list[Sample]) -> dict:\n """"""Convert samples to training data format.""""""\n train_data = {\n ""tokens"": [sample.tokens for sample in samples],\n ""response_lengths"": [sample.response_length for sample in samples],\n ""rewards"": [0.0 for _ in samples],\n ""raw_reward"": [0.0 for _ in samples],\n ""truncated"": [0 for _ in samples],\n ""sample_indices"": [sample.index for sample in samples],\n }\n\n loss_masks = []\n for sample in samples:\n if sample.loss_mask is None:\n sample.loss_mask = [1] * sample.response_length\n loss_masks.append(sample.loss_mask)\n train_data[""loss_masks""] = loss_masks\n\n return train_data\n\n def _split_train_data_by_dp(self, data: dict) -> dict:\n """"""Split training data for current DP rank.""""""\n total_lengths = [len(t) for t in data[""tokens""]]\n data[""total_lengths""] = total_lengths\n\n # Simple round-robin partitioning\n partition = list(range(self.dp_rank, len(total_lengths), self.dp_size))\n\n rollout_data = {""partition"": partition, ""total_lengths"": total_lengths}\n\n for key in [\n ""tokens"",\n ""response_lengths"",\n ""rewards"",\n ""raw_reward"",\n ""truncated"",\n ""loss_masks"",\n ""sample_indices"",\n ]:\n if key in data:\n rollout_data[key] = [data[key][j] for j in partition]\n\n return rollout_data\n\n def _packed_data(self, rollout_data: dict) -> tuple[list[dict], list[int]]:\n """"""Pack variable-length sequences for efficient processing.""""""\n tokens = rollout_data[""tokens""]\n\n packed_batches = []\n mbs_size_list = []\n local_batch_size = self.args.global_batch_size // self.dp_size\n\n if self.args.use_dynamic_batch_size:\n max_tokens = self.args.max_tokens_per_gpu\n if self.cp_size > 1:\n max_tokens = max_tokens * self.cp_size\n\n for i in range(0, len(tokens), local_batch_size):\n mbs_size_list.append(\n get_minimum_num_micro_batch_size(\n [len(t) for t in rollout_data[""tokens""][i : i + local_batch_size]],\n max_tokens,\n )\n )\n num_microbatches = torch.tensor(\n mbs_size_list, dtype=torch.int, device=torch.cuda.current_device()\n )\n dist.all_reduce(num_microbatches, op=dist.ReduceOp.MAX, group=self.dp_group)\n num_microbatches = num_microbatches.tolist()\n else:\n num_microbatches = [\n self.args.global_batch_size // (self.args.micro_batch_size * self.dp_size)\n ] * (len(tokens) // local_batch_size)\n\n start = 0\n for mbs_size in num_microbatches:\n end = start + local_batch_size\n # Create dummy advantages/returns for SFT (not used but required by pack_sequences)\n dummy_advantages = [\n torch.zeros(rollout_data[""response_lengths""][i])\n for i in range(start, end)\n ]\n packed_batches.extend(\n pack_sequences(\n rollout_data[""tokens""][start:end],\n rollout_data[""loss_masks""][start:end],\n rollout_data[""rewards""][start:end],\n rollout_data[""raw_reward""][start:end],\n rollout_data[""response_lengths""][start:end],\n dummy_advantages, # advantages\n dummy_advantages, # returns\n num_packs=mbs_size,\n )\n )\n start = end\n\n grad_accum = list(accumulate(num_microbatches))\n return packed_batches, grad_accum\n\n def _get_model_inputs_args(self, packed_sequence: dict) -> dict:\n """"""Prepare model input arguments from packed sequence.""""""\n input_ids = packed_sequence[""tokens""].unsqueeze(0)\n position_ids = packed_sequence[""position_ids""].unsqueeze(0)\n\n return {\n ""input_ids"": input_ids,\n ""position_ids"": position_ids,\n ""attention_mask"": None,\n }\n\n def _compute_log_prob(self, packed_batches: list[dict]) -> None:\n """"""Compute log probabilities for SFT loss.""""""\n with timer(""log_probs""), torch.no_grad():\n for batch in tqdm(\n packed_batches,\n desc=""log_probs"",\n disable=dist.get_rank() != 0,\n ):\n model_args = self._get_model_inputs_args(batch)\n logits = self.model(**model_args).logits.squeeze(0).float()\n log_probs_result, entropy_result = get_logprob_and_entropy_with_cp(\n logits=logits,\n target_tokens=batch[""tokens""],\n cp_rank=self.cp_rank,\n cp_size=self.cp_size,\n cp_group=self.cp_group,\n model_input_ids=model_args[""input_ids""],\n allow_compile=True,\n temperature=self.args.rollout_temperature,\n )\n batch[""log_probs""] = log_probs_result\n batch[""entropy""] = entropy_result\n\n def _compute_sft_loss(self, unpacked_batches: list[dict], logits: torch.Tensor):\n """"""Compute SFT loss (negative log likelihood).""""""\n loss_masks = [\n batch[""loss_masks""].to(device=logits.device) for batch in unpacked_batches\n ]\n response_lengths = [batch[""response_lengths""] for batch in unpacked_batches]\n log_probs = torch.cat(\n [batch[""cur_log_probs""] for batch in unpacked_batches], dim=0\n )\n loss = -sum_of_sample_mean(log_probs, response_lengths, loss_masks)\n\n if log_probs.numel() == 0:\n loss += 0 * logits.sum()\n\n return loss, {""loss"": loss.detach()}\n\n def _train_step(\n self,\n packed_batch: dict,\n reported_accum: dict,\n mbs_id: int,\n grad_accum: list[int],\n ):\n """"""Execute one training step.""""""\n model_args = self._get_model_inputs_args(packed_batch)\n logits = self.model(**model_args).logits.squeeze(0).float()\n\n log_probs, entropy_result = get_logprob_and_entropy_with_cp(\n logits=logits,\n target_tokens=packed_batch[""tokens""],\n cp_rank=self.cp_rank,\n cp_size=self.cp_size,\n cp_group=self.cp_group,\n model_input_ids=model_args[""input_ids""],\n allow_compile=True,\n temperature=self.args.rollout_temperature,\n )\n packed_batch[""cur_log_probs""] = log_probs\n packed_batch[""entropy""] = entropy_result\n\n unpacked_batches = unpack_sequences(packed_batch)\n loss, reported = self._compute_sft_loss(unpacked_batches, logits)\n\n # Scale loss for gradient accumulation\n loss = loss * self.dp_size / self.args.global_batch_size\n loss.backward()\n\n for k, v in reported.items():\n reported_accum.setdefault(k, []).append(v)\n\n if (mbs_id + 1) in grad_accum:\n grad_norm = torch.nn.utils.clip_grad_norm_(\n self.model.parameters(), self.args.clip_grad\n )\n grad_norm = float(grad_norm)\n\n self.optimizer.step()\n self.lr_scheduler.step()\n self.optimizer.zero_grad(set_to_none=True)\n\n # Aggregate and log metrics\n aggregated = {k: torch.stack(v).sum().item() for k, v in reported_accum.items()}\n reduced_aggregated = [None] * self.dp_size\n dist.all_gather_object(reduced_aggregated, aggregated, group=self.dp_group)\n aggregated = {}\n for k in reported_accum.keys():\n aggregated[k] = sum([r[k] for r in reduced_aggregated]) / self.args.global_batch_size\n reported_accum.clear()\n\n if dist.get_rank() == 0:\n log_dict = {\n f""train/{k}"": (val.item() if torch.is_tensor(val) else val)\n for k, val in aggregated.items()\n }\n log_dict[""train/grad_norm""] = grad_norm\n\n lr_values = self.lr_scheduler.get_last_lr()\n for gid, _group in enumerate(self.optimizer.param_groups):\n log_dict[f""train/lr_{gid}""] = lr_values[gid]\n\n logger.info(f""step {self.global_step}: {log_dict}"")\n log_dict[""train/step""] = self.global_step\n tracking_utils.log(self.args, log_dict, step_key=""train/step"")\n\n self.global_step += 1\n\n def train_one_rollout(self, rollout_id: int):\n """"""Execute one rollout's worth of training.""""""\n # Generate SFT data\n samples = self.generate_sft_rollout(rollout_id)\n\n # Convert to training format\n train_data = self._convert_samples_to_train_data(samples)\n\n # Split by DP rank\n rollout_data = self._split_train_data_by_dp(train_data)\n\n # Create advantages/returns (dummy for SFT)\n rollout_data[""advantages""] = rollout_data[""returns""] = [\n torch.tensor([0.0] * rollout_data[""response_lengths""][i])\n for i in range(len(rollout_data[""rewards""]))\n ]\n\n # Pack sequences\n packed_batches, grad_accum = self._packed_data(rollout_data)\n\n if len(grad_accum) == 0:\n logger.warning(f""[Rank {dist.get_rank()}] No batches to train on rollout {rollout_id}"")\n return\n\n # Compute log probs for logging\n self._compute_log_prob(packed_batches)\n\n # Training loop\n with timer(""actor_train""):\n reported_accum = {}\n self.optimizer.zero_grad(set_to_none=True)\n\n for mbs_id, packed_batch in enumerate(\n tqdm(packed_batches, desc=""actor_train"", disable=dist.get_rank() != 0)\n ):\n self._train_step(packed_batch, reported_accum, mbs_id, grad_accum)\n\n self.prof.step(rollout_id=rollout_id)\n\n def save_model(self, iteration: int):\n """"""Save model checkpoint.""""""\n if self.args.save is None:\n return\n checkpoint.save(self, iteration)\n\n def train(self):\n """"""Main training loop.""""""\n logger.info(\n f""[Rank {dist.get_rank()}] Starting training: ""\n f""rollout_id {self.args.start_rollout_id} -> {self.args.num_rollout}""\n )\n\n for rollout_id in range(self.args.start_rollout_id, self.args.num_rollout):\n self.train_one_rollout(rollout_id)\n\n # Save checkpoint periodically\n if should_run_periodic_action(\n rollout_id, self.args.save_interval, self.num_rollout_per_epoch\n ):\n self.save_model(rollout_id)\n\n logger.info(f""[Rank {dist.get_rank()}] Training completed!"")\n\n\ndef set_sft_defaults(args: Namespace) -> Namespace:\n """"""Set default values appropriate for SFT training.""""""\n # These flags are normally set by debug_train_only in Ray mode\n if not hasattr(args, ""debug_train_only""):\n args.debug_train_only = True\n\n # SFT-specific defaults\n if not hasattr(args, ""loss_type"") or args.loss_type is None:\n args.loss_type = ""sft_loss""\n\n if not hasattr(args, ""advantage_estimator"") or args.advantage_estimator is None:\n args.advantage_estimator = ""grpo""\n\n if not hasattr(args, ""n_samples_per_prompt"") or args.n_samples_per_prompt is None:\n args.n_samples_per_prompt = 1\n\n if not hasattr(args, ""rollout_global_dataset""):\n args.rollout_global_dataset = True\n\n if not hasattr(args, ""context_parallel_size"") or args.context_parallel_size is None:\n args.context_parallel_size = 1\n\n if not hasattr(args, ""attn_implementation"") or args.attn_implementation is None:\n args.attn_implementation = ""flash_attention_2""\n\n if not hasattr(args, ""rollout_temperature"") or args.rollout_temperature is None:\n args.rollout_temperature = 1.0\n\n # Ensure we don't try to use rollout engines\n args.offload_train = False\n args.offload_rollout = False\n args.colocate = False\n\n return args\n\n\ndef main():\n configure_logger()\n\n # Parse arguments\n args = parse_args()\n\n # Apply SFT-specific defaults\n args = set_sft_defaults(args)\n\n # Create trainer and run\n trainer = SFTTrainer(args)\n trainer.train()\n\n\nif __name__ == ""__main__"":\n main()\n\n",python,tab
13
+ 12,2476526,"train_sft.py",1243,0,"",python,selection_keyboard
14
+ 13,2476626,"train_sft.py",2719,0,"",python,selection_keyboard
15
+ 14,2476877,"train_sft.py",3992,0,"",python,selection_keyboard
16
+ 15,2476909,"train_sft.py",5612,0,"",python,selection_keyboard
17
+ 16,2476942,"train_sft.py",7187,0,"",python,selection_keyboard
18
+ 17,2476975,"train_sft.py",8649,0,"",python,selection_keyboard
19
+ 18,2477009,"train_sft.py",10081,0,"",python,selection_keyboard
20
+ 19,2477042,"train_sft.py",11752,0,"",python,selection_keyboard
21
+ 20,2477075,"train_sft.py",13254,0,"",python,selection_keyboard
22
+ 21,2477108,"train_sft.py",15059,0,"",python,selection_keyboard
23
+ 22,2477141,"train_sft.py",16777,0,"",python,selection_keyboard
24
+ 23,2477175,"train_sft.py",18365,0,"",python,selection_keyboard
25
+ 24,2477208,"train_sft.py",20098,0,"",python,selection_keyboard
26
+ 25,2477241,"train_sft.py",21632,0,"",python,selection_keyboard
27
+ 26,2477275,"train_sft.py",23157,0,"",python,selection_keyboard
28
+ 27,2477307,"train_sft.py",24080,0,"",python,selection_keyboard
29
+ 28,2477660,"train_sft.py",22989,0,"",python,selection_keyboard
30
+ 29,2477911,"train_sft.py",21446,0,"",python,selection_keyboard
31
+ 30,2477942,"train_sft.py",19889,0,"",python,selection_keyboard
32
+ 31,2477975,"train_sft.py",18201,0,"",python,selection_keyboard
33
+ 32,2478009,"train_sft.py",16595,0,"",python,selection_keyboard
34
+ 33,2478042,"train_sft.py",14830,0,"",python,selection_keyboard
35
+ 34,2478077,"train_sft.py",13074,0,"",python,selection_keyboard
36
+ 35,2478108,"train_sft.py",11536,0,"",python,selection_keyboard
37
+ 36,2478141,"train_sft.py",9899,0,"",python,selection_keyboard
38
+ 37,2478174,"train_sft.py",8549,0,"",python,selection_keyboard
39
+ 38,2478208,"train_sft.py",7060,0,"",python,selection_keyboard
40
+ 39,2478241,"train_sft.py",5403,0,"",python,selection_keyboard
41
+ 40,2478275,"train_sft.py",3822,0,"",python,selection_keyboard
42
+ 41,2478309,"train_sft.py",2599,0,"",python,selection_keyboard
43
+ 42,2478335,"train_sft.py",1033,0,"",python,selection_keyboard
44
+ 43,2478374,"train_sft.py",0,0,"",python,selection_keyboard
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-14a0bec6-a4a9-4b59-996b-ff826d7cb6461762440546881-2025_11_06-15.49.10.583/source.csv ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 2,55,"TERMINAL",0,0,"Crowd Pilot Prime",,terminal_focus
3
+ 3,151,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"3:49:10 PM [info] Activating crowd-code\n3:49:10 PM [info] Recording started\n3:49:10 PM [info] Initializing git provider using file system watchers...\n3:49:10 PM [info] No workspace folder found\n",Log,tab
4
+ 4,2029,"extension-output-pdoom-org.crowd-code-#1-crowd-code",194,0,"3:49:12 PM [info] Retrying git provider initialization...\n3:49:12 PM [info] No workspace folder found\n",Log,content
5
+ 5,103052,"Untitled-1",0,0,"",plaintext,tab
6
+ 6,103929,"TERMINAL",0,0,"Test",,terminal_focus
7
+ 7,105383,"Untitled-1",0,0,"hello world\n",plaintext,content
8
+ 8,105612,"TERMINAL",0,0,"echo VSCode test",,terminal_command
9
+ 9,105613,"TERMINAL",0,0,"]633;CVSCode test\r\n% \r \r",,terminal_output
10
+ 10,109511,"TERMINAL",0,0,"Test",,terminal_focus
11
+ 11,110379,"Untitled-1",0,0,"",plaintext,selection_command
12
+ 12,110383,"Untitled-1",0,0,"hello world\n",plaintext,content
13
+ 13,110510,"TERMINAL",0,0,"echo VSCode test",,terminal_command
14
+ 14,110511,"TERMINAL",0,0,"]633;CVSCode test\r\n% \r \r",,terminal_output
15
+ 15,111139,"Untitled-1",0,0,"",plaintext,selection_command
16
+ 16,111141,"Untitled-1",0,0,"hello world\n",plaintext,content
17
+ 17,111169,"TERMINAL",0,0,"echo VSCode test",,terminal_command
18
+ 18,111170,"TERMINAL",0,0,"]633;CVSCode test\r\n% \r \r",,terminal_output
19
+ 19,114703,"Untitled-1",0,0,"",plaintext,selection_command
20
+ 20,114705,"Untitled-1",0,0,"hello world\n",plaintext,content
21
+ 21,114736,"TERMINAL",0,0,"echo VSCode test",,terminal_command
22
+ 22,114737,"TERMINAL",0,0,"]633;CVSCode test\r\n% \r \r",,terminal_output
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-1e1272f3-7d2e-4b26-8662-ef59b0a62e821764410118523-2025_11_29-10.55.25.324/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-222fe98e-29ac-4b20-9a65-fe2e31f8eb701751128122769-2025_06_28-09.28.47.536/source.csv ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,2,"train_tokenizer.py",0,0,"from dataclasses import dataclass, field\nimport os\nimport time\n\nimport einops\nfrom flax.training import orbax_utils\nfrom flax.training.train_state import TrainState\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax\nfrom orbax.checkpoint import PyTreeCheckpointer\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\n\nfrom models.tokenizer import TokenizerVQVAE\nfrom utils.dataloader import get_dataloader\n\nts = int(time.time())\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 300_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = ""data_tfrecords/coinrun""\n checkpoint: str = """"\n # Optimization\n vq_beta: float = 0.25\n batch_size: int = 48\n min_lr: float = 3e-4\n max_lr: float = 3e-4\n warmup_steps: int = 10000\n # Tokenizer\n model_dim: int = 512\n latent_dim: int = 32\n num_latents: int = 1024\n patch_size: int = 4\n num_blocks: int = 8\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.01\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_tokenizer""\n tags: list[str] = field(default_factory=lambda: [""tokenizer""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_gradients: bool = False\n\n\nargs = tyro.cli(Args)\n\n\ndef tokenizer_loss_fn(params, state, inputs):\n # --- Compute loss ---\n outputs = state.apply_fn(\n params,\n inputs,\n training=True,\n rngs={""params"": inputs[""rng""], ""dropout"": inputs[""dropout_rng""]},\n )\n mse = jnp.square(inputs[""videos""] - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = inputs[""videos""].clip(0, 1).reshape(-1, *inputs[""videos""].shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = pix.psnr(gt, recon).mean()\n ssim = pix.ssim(gt, recon).mean()\n _, index_counts = jnp.unique_counts(\n jnp.ravel(outputs[""indices""]), size=args.num_latents, fill_value=0\n )\n codebook_usage = (index_counts != 0).mean()\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=codebook_usage,\n )\n return loss, (outputs[""recon""], metrics)\n\n\n@jax.jit\ndef train_step(state, inputs):\n grad_fn = jax.value_and_grad(tokenizer_loss_fn, has_aux=True, allow_int=True)\n (loss, (recon, metrics)), grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n if args.log_gradients:\n metrics[""encoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""encoder""]\n )\n metrics[""vq_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""vq""]\n )\n metrics[""decoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""decoder""]\n )\n return state, loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.PRNGKey(args.seed)\n if args.log and jax.process_index() == 0:\n wandb.init(\n entity=args.entity,\n project=args.project,\n name=args.name,\n tags=args.tags,\n group=""debug"",\n config=args\n )\n\n # --- Initialize model ---\n tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n )\n rng, _rng = jax.random.split(rng)\n image_shape = (args.image_height, args.image_width, args.image_channels)\n inputs = dict(\n videos=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len, *image_shape),\n dtype=jnp.float32,\n ),\n )\n init_params = tokenizer.init(_rng, inputs)\n\n # --- Initialize optimizer ---\n lr_schedule = optax.warmup_cosine_decay_schedule(\n args.min_lr, args.max_lr, args.warmup_steps, args.num_steps\n )\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4)\n train_state = TrainState.create(apply_fn=tokenizer.apply, params=init_params, tx=tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n train_state = jax.device_put(train_state, replicated_sharding)\n\n # --- Load checkpoint ---\n step = 0\n if args.checkpoint:\n restore_target = {""model"": train_state}\n restore_args = orbax_utils.restore_args_from_target(restore_target)\n train_state.params[""params""].update(\n PyTreeCheckpointer()\n .restore(args.checkpoint, item=restore_target, restore_args=restore_args)[\n ""model""\n ]\n .params[""params""]\n )\n # Assume checkpoint is of the form tokenizer_<timestamp>_<step>\n step += int(args.checkpoint.split(""_"")[-1])\n\n # --- TRAIN LOOP ---\n tfrecord_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".tfrecord"")\n ]\n dataloader = get_dataloader(\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n tfrecord_files,\n args.seq_len,\n args.batch_size,\n *image_shape,\n )\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng, _rng_dropout = jax.random.split(rng, 3)\n\n videos_sharding = NamedSharding(\n mesh, PartitionSpec(""data"", None, None, None, None)\n )\n videos = jax.make_array_from_process_local_data(videos_sharding, videos)\n\n inputs = dict(videos=videos, rng=_rng, dropout_rng=_rng_dropout)\n start_time = time.time()\n train_state, loss, recon, metrics = train_step(train_state, inputs)\n elapsed_time = (time.time() - start_time) * 1000\n print(f""Step {step}, loss: {loss}, step time: {elapsed_time}ms"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n ""step_time_ms"": elapsed_time,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0]\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n if step % args.log_checkpoint_interval == 0:\n ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,\n )\n if step >= args.num_steps:\n break\n",python,tab
3
+ 2,62,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
4
+ 3,97,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"9:28:47 AM [info] Activating crowd-code\n9:28:47 AM [info] Recording started\n9:28:47 AM [info] Initializing git provider using file system watchers...\n9:28:47 AM [info] Git repository found\n9:28:47 AM [info] Git provider initialized successfully\n",Log,content
5
+ 4,195,"extension-output-pdoom-org.crowd-code-#1-crowd-code",245,0,"9:28:47 AM [info] Initial git state: [object Object]\n",Log,content
6
+ 5,109366,"train_tokenizer.py",0,0,"",python,tab
7
+ 6,120114,"train_tokenizer.py",0,0,"Switched from branch 'dataloader-reproducibility-test' to 'main'",python,git_branch_checkout
8
+ 7,135115,"train_tokenizer.py",6657,0,"",python,selection_mouse
9
+ 8,135131,"train_tokenizer.py",6656,0,"",python,selection_command
10
+ 9,584544,"utils/dataloader.py",0,0,"import functools\nimport jax\n\nimport tensorflow as tf\n\n# reserve GPU memory for JAX only if tensorflow is built with GPU support\ntf.config.experimental.set_visible_devices([], ""GPU"")\n\n\n# --- TensorFlow function for processing: slicing, normalization ---\ndef _tf_process_episode(episode_tensor, seq_len, image_h, image_w, image_c):\n """"""\n Processes a raw episode tensor in TensorFlow.\n Takes a full episode, extracts a random sequence, and normalizes it.\n Args:\n episode_tensor: A TensorFlow tensor representing a full video episode.\n Expected shape: (dynamic_length, image_h, image_w, image_c)\n Expected dtype: e.g., tf.uint8 (raw pixel values)\n seq_len: The desired length of the sub-sequence to extract.\n image_h: The height of each frame.\n image_w: The width of each frame.\n image_c: The number of channels in each frame.\n Returns:\n A TensorFlow tensor representing the processed video sequence.\n Shape: (seq_len, image_h, image_w, image_c)\n Dtype: tf.float32 (normalized pixel values)\n """"""\n current_episode_len = tf.shape(episode_tensor)[0]\n\n max_start_idx = current_episode_len - seq_len\n\n start_idx = tf.random.uniform(\n shape=(), minval=0, maxval=max_start_idx + 1, dtype=tf.int32\n )\n\n seq = episode_tensor[start_idx : start_idx + seq_len]\n\n seq = tf.cast(seq, tf.float32) / 255.0\n\n # Ensure the final shape is statically known for batching.\n # tf.reshape is robust, but tf.ensure_shape or set_shape can also be used if confident.\n processed_sequence = tf.reshape(seq, [seq_len, image_h, image_w, image_c])\n\n return processed_sequence\n\n\ndef _parse_tfrecord_fn(example_proto, image_h, image_w, image_c):\n feature_description = {\n ""height"": tf.io.FixedLenFeature([], tf.int64),\n ""width"": tf.io.FixedLenFeature([], tf.int64),\n ""channels"": tf.io.FixedLenFeature([], tf.int64),\n ""sequence_length"": tf.io.FixedLenFeature([], tf.int64),\n ""raw_video"": tf.io.FixedLenFeature([], tf.string),\n }\n example = tf.io.parse_single_example(example_proto, feature_description)\n\n video_shape = (example[""sequence_length""], image_h, image_w, image_c)\n\n episode_tensor = tf.io.decode_raw(example[""raw_video""], out_type=tf.uint8)\n episode_tensor = tf.reshape(episode_tensor, video_shape)\n\n episode_tensor = tf.ensure_shape(episode_tensor, [None, image_h, image_w, image_c])\n return episode_tensor\n\n\ndef get_dataloader(\n tfrecord_paths: list[str], # List of TFRecord file paths\n seq_len: int,\n global_batch_size: int,\n image_h: int,\n image_w: int,\n image_c: int,\n shuffle_buffer_size: int = 1000,\n num_parallel_calls: int = tf.data.AUTOTUNE,\n seed: int = 42,\n):\n """"""\n Creates a tf.data.Dataset pipeline from TFRecord files.\n """"""\n if not tfrecord_paths:\n raise ValueError(""tfrecord_paths list cannot be empty."")\n\n process_id = jax.process_index()\n num_processes = jax.process_count()\n\n assert (\n global_batch_size % num_processes == 0\n ), ""Global batch size {global_batch_size} \\n must be divisible by the number of JAX processes {num_processes} for proper sharding.""\n per_process_batch_size = global_batch_size // num_processes\n\n dataset = tf.data.TFRecordDataset(\n tfrecord_paths, num_parallel_reads=tf.data.AUTOTUNE\n )\n\n dataset = dataset.shard(num_shards=num_processes, index=process_id)\n\n # (f.srambical) NOTE: For TFRecords, it's often good to have a large shuffle buffer.\n if shuffle_buffer_size > 0:\n dataset = dataset.shuffle(\n buffer_size=shuffle_buffer_size, seed=seed, reshuffle_each_iteration=True\n )\n parse_fn = functools.partial(\n _parse_tfrecord_fn, image_h=image_h, image_w=image_w, image_c=image_c\n )\n dataset = dataset.map(parse_fn, num_parallel_calls=num_parallel_calls)\n\n tf_process_fn = functools.partial(\n _tf_process_episode,\n seq_len=seq_len,\n image_h=image_h,\n image_w=image_w,\n image_c=image_c,\n )\n dataset = dataset.map(tf_process_fn, num_parallel_calls=num_parallel_calls)\n\n dataset = dataset.repeat(None)\n dataset = dataset.batch(per_process_batch_size, drop_remainder=True)\n dataset = dataset.prefetch(tf.data.AUTOTUNE)\n\n return dataset.as_numpy_iterator()\n",python,tab
11
+ 10,587078,"utils/dataloader.py",4313,0,"",python,selection_command
12
+ 11,651295,"utils/dataloader.py",0,0,"",python,selection_command
13
+ 12,652011,"utils/dataloader.py",17,0,"",python,selection_command
14
+ 13,652136,"utils/dataloader.py",28,0,"",python,selection_command
15
+ 14,652214,"utils/dataloader.py",29,0,"",python,selection_command
16
+ 15,652563,"utils/dataloader.py",36,0,"",python,selection_command
17
+ 16,655020,"utils/dataloader.py",4380,0,"",python,selection_command
18
+ 17,655495,"utils/dataloader.py",4341,0,"",python,selection_command
19
+ 18,655578,"utils/dataloader.py",4345,0,"",python,selection_command
20
+ 19,655774,"utils/dataloader.py",4352,0,"",python,selection_command
21
+ 20,655989,"utils/dataloader.py",4359,0,"",python,selection_command
22
+ 21,656093,"utils/dataloader.py",4340,0,"",python,selection_command
23
+ 22,656226,"utils/dataloader.py",4345,0,"",python,selection_command
24
+ 23,656410,"utils/dataloader.py",4340,0,"",python,selection_command
25
+ 24,656615,"utils/dataloader.py",4295,0,"",python,selection_command
26
+ 25,656753,"utils/dataloader.py",4303,0,"",python,selection_command
27
+ 26,656931,"utils/dataloader.py",4305,0,"",python,selection_command
28
+ 27,657095,"utils/dataloader.py",4312,0,"",python,selection_command
29
+ 28,657232,"utils/dataloader.py",4313,0,"",python,selection_command
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-25373d4e-68c9-48b1-a295-2d5d418610141764451376541-2025_11_29-22.22.59.470/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-266a35fd-8e9a-45ca-bab6-73f8e208e8e31764440451347-2025_11_29-19.20.59.166/source.csv ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,11,"Untitled-1",0,0,"",plaintext,tab
3
+ 2,115,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"7:20:59 PM [info] Activating crowd-code\n7:20:59 PM [info] Recording started\n7:20:59 PM [info] Initializing git provider using file system watchers...\n7:20:59 PM [info] No workspace folder found\n",Log,tab
4
+ 3,1023,"Untitled-1",0,0,"",plaintext,tab
5
+ 4,2180,"TERMINAL",0,0,"Test",,terminal_focus
6
+ 5,2186,"Untitled-1",0,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
7
+ 6,2904,"Untitled-1",46,0,"",plaintext,selection_command
8
+ 7,2997,"Untitled-1",39,0,"",plaintext,selection_command
9
+ 8,3156,"Untitled-1",32,0,"",plaintext,selection_command
10
+ 9,3286,"Untitled-1",0,0,"",plaintext,selection_command
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-2bb200ce-4bc8-4bc3-9354-29e24db5d38e1752063967983-2025_07_09-14.26.42.463/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-3126c08d-d6fe-466c-828c-ddc956b35e091762367702624-2025_11_05-19.35.09.972/source.csv ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,2,"package.json",0,0,"{\n ""name"": ""crowd-pilot"",\n ""displayName"": ""crowd-pilot"",\n ""description"": ""Teaching language models to code like humans."",\n ""version"": ""0.0.1"",\n ""engines"": {\n ""vscode"": ""^1.99.3""\n },\n ""categories"": [\n ""Other""\n ],\n ""activationEvents"": [],\n ""main"": ""./out/extension.js"",\n ""contributes"": {\n ""commands"": [\n {\n ""command"": ""crowd-pilot.helloWorld"",\n ""title"": ""Hello World""\n }\n ]\n },\n ""scripts"": {\n ""vscode:prepublish"": ""npm run compile"",\n ""compile"": ""tsc -p ./"",\n ""watch"": ""tsc -watch -p ./"",\n ""pretest"": ""npm run compile && npm run lint"",\n ""lint"": ""eslint src"",\n ""test"": ""vscode-test""\n },\n ""devDependencies"": {\n ""@types/vscode"": ""^1.105.0"",\n ""@types/mocha"": ""^10.0.10"",\n ""@types/node"": ""22.x"",\n ""@typescript-eslint/eslint-plugin"": ""^8.45.0"",\n ""@typescript-eslint/parser"": ""^8.45.0"",\n ""eslint"": ""^9.36.0"",\n ""typescript"": ""^5.9.3"",\n ""@vscode/test-cli"": ""^0.0.11"",\n ""@vscode/test-electron"": ""^2.5.2""\n }\n}\n",json,tab
3
+ 2,329,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"7:35:09 PM [info] Activating crowd-code\n7:35:09 PM [info] Recording started\n7:35:09 PM [info] Initializing git provider using file system watchers...\n7:35:10 PM [error] Not a git repository: EntryNotFound (FileSystemError): Error: ENOENT: no such file or directory, stat '/home/franz.srambical/crowd-pilot/crowd-pilot-extension/.git'\n",Log,tab
4
+ 3,1458,"package.json",0,0,"",json,tab
5
+ 4,141551,"TERMINAL",0,0,"",,terminal_focus
6
+ 5,182907,"TERMINAL",0,0,"",,terminal_focus
7
+ 6,186311,"TERMINAL",0,0,"ls",,terminal_command
8
+ 7,186317,"TERMINAL",0,0,"]633;CCHANGELOG.md eslint.config.mjs node_modules package.json package-lock.json README.md src tsconfig.json vsc-extension-quickstart.md\r\n]0;franz.srambical@hai-login1:~/crowd-pilot/crowd-pilot-extension",,terminal_output
9
+ 8,190535,"TERMINAL",0,0,"cd ..",,terminal_command
10
+ 9,195620,"TERMINAL",0,0,"rm -r crowd-pilot-extension/",,terminal_command
11
+ 10,195670,"TERMINAL",0,0,"]633;C",,terminal_output
12
+ 11,201188,"TERMINAL",0,0,"]0;franz.srambical@hai-login1:~/crowd-pilot",,terminal_output
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-3334d6a5-5e72-43ad-a8da-1813285fd7a51758271588339-2025_09_19-10.46.34.637/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-395f734f-3624-4a17-97d3-4e353b215a5c1755368459655-2025_08_16-20.21.02.372/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-3bffc30a-6fb0-48dd-ab21-3fe225eb22c51757148592262-2025_09_06-10.49.57.58/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-3fe74aa6-92c3-405a-9c6c-49c34aec593b1762364457821-2025_11_05-18.41.02.400/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-46c4a757-02fe-4d2b-80c9-10277f3bc6421757061314730-2025_09_05-10.35.23.869/source.csv ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,2,"models/tokenizer.py",0,0,"from typing import Dict, Tuple\n\nimport flax.nnx as nnx\nimport jax.numpy as jnp\nimport jax\n\nfrom utils.preprocess import patchify, unpatchify\nfrom utils.nn import STTransformer, VectorQuantizer\n\n\nclass TokenizerVQVAE(nnx.Module):\n """"""\n ST-ViVit VQ-VAE\n\n Dimension keys:\n B: batch size\n T: sequence length\n N: number of patches per frame\n L: latent dimension\n D: B * T * N\n H: height\n W: width\n C: number of channels\n P: patch token dimension (patch_size^2 * C)\n """"""\n\n def __init__(\n self,\n in_dim: int,\n model_dim: int,\n ffn_dim: int,\n latent_dim: int,\n num_latents: int,\n patch_size: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n codebook_dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n rngs: nnx.Rngs,\n ):\n self.in_dim = in_dim\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.latent_dim = latent_dim\n self.num_latents = num_latents\n self.patch_size = patch_size\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.codebook_dropout = codebook_dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n\n self.encoder = STTransformer(\n self.in_dim * self.patch_size**2,\n self.model_dim,\n self.ffn_dim,\n self.latent_dim,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.vq = VectorQuantizer(\n self.latent_dim,\n self.num_latents,\n self.codebook_dropout,\n self.dtype,\n rngs=rngs,\n )\n self.out_dim = self.in_dim * self.patch_size**2\n self.decoder = STTransformer(\n self.latent_dim,\n self.model_dim,\n self.ffn_dim,\n self.out_dim,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n\n def __call__(\n self, batch: Dict[str, jax.Array], training: bool = True\n ) -> Dict[str, jax.Array]:\n H, W = batch[""videos""].shape[2:4]\n videos_BTHWC = batch[""videos""]\n outputs = self.vq_encode(videos_BTHWC, training)\n z_q_BTNL = outputs[""z_q""]\n recon_BTHWC = self.decoder(z_q_BTNL)\n recon_BTHWC = recon_BTHWC.astype(jnp.float32)\n recon_BTHWC = nnx.sigmoid(recon_BTHWC)\n recon_BTHWC = recon_BTHWC.astype(self.dtype)\n recon_BTHWC = unpatchify(recon_BTHWC, self.patch_size, H, W)\n outputs[""recon""] = recon_BTHWC\n return outputs\n\n def vq_encode(\n self, videos: jax.Array, training: bool = True\n ) -> Dict[str, jax.Array]:\n # --- Preprocess + encode ---\n B, T = videos.shape[:2]\n patch_BTNP = patchify(videos, self.patch_size)\n N = patch_BTNP.shape[2]\n x_BTNL = self.encoder(patch_BTNP)\n\n # --- Vector quantize ---\n x_DL = x_BTNL.reshape(B * T * N, self.latent_dim)\n z_q_DL, z_DL, emb_DL, indices_D = self.vq(x_DL, training)\n z_q_BTNL = z_q_DL.reshape(B, T, N, self.latent_dim)\n indices_BTN = indices_D.reshape(B, T, N)\n return dict(z_q=z_q_BTNL, z=z_DL, emb=emb_DL, indices=indices_BTN)\n\n def decode(self, indices_BTN: jax.Array, video_hw: Tuple[int, int]) -> jax.Array:\n z_BTNL = self.vq.codebook[indices_BTN]\n recon_BTNP = self.decoder(z_BTNL)\n recon_BTNP = recon_BTNP.astype(jnp.float32)\n recon_BTNP = nnx.sigmoid(recon_BTNP)\n recon_BTNP = recon_BTNP.astype(self.dtype)\n return unpatchify(recon_BTNP, self.patch_size, *video_hw)\n",python,tab
3
+ 2,89,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"10:35:23 AM [info] Activating crowd-code\n10:35:23 AM [info] Recording started\n10:35:23 AM [info] Initializing git provider using file system watchers...\n10:35:23 AM [info] Git repository found\n10:35:23 AM [info] Git provider initialized successfully\n",Log,tab
4
+ 3,125,"extension-output-pdoom-org.crowd-code-#1-crowd-code",250,0,"10:35:23 AM [info] Initial git state: [object Object]\n",Log,content
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-48fbb415-6db9-4d35-b548-561e828791bf1751383187013-2025_07_01-17.19.57.60/source.csv ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,2,"utils/dataloader.py",0,0,"import functools\nimport jax\n\nimport tensorflow as tf\n\n# reserve GPU memory for JAX only if tensorflow is built with GPU support\ntf.config.experimental.set_visible_devices([], ""GPU"")\n\n\n# --- TensorFlow function for processing: slicing, normalization ---\ndef _tf_process_episode(episode_tensor, seq_len, image_h, image_w, image_c):\n """"""\n Processes a raw episode tensor in TensorFlow.\n Takes a full episode, extracts a random sequence, and normalizes it.\n Args:\n episode_tensor: A TensorFlow tensor representing a full video episode.\n Expected shape: (dynamic_length, image_h, image_w, image_c)\n Expected dtype: e.g., tf.uint8 (raw pixel values)\n seq_len: The desired length of the sub-sequence to extract.\n image_h: The height of each frame.\n image_w: The width of each frame.\n image_c: The number of channels in each frame.\n Returns:\n A TensorFlow tensor representing the processed video sequence.\n Shape: (seq_len, image_h, image_w, image_c)\n Dtype: tf.float32 (normalized pixel values)\n """"""\n current_episode_len = tf.shape(episode_tensor)[0]\n\n max_start_idx = current_episode_len - seq_len\n\n start_idx = tf.random.uniform(\n shape=(), minval=0, maxval=max_start_idx + 1, dtype=tf.int32\n )\n\n seq = episode_tensor[start_idx : start_idx + seq_len]\n\n seq = tf.cast(seq, tf.float32) / 255.0\n\n # Ensure the final shape is statically known for batching.\n # tf.reshape is robust, but tf.ensure_shape or set_shape can also be used if confident.\n processed_sequence = tf.reshape(seq, [seq_len, image_h, image_w, image_c])\n\n return processed_sequence\n\n\ndef _parse_tfrecord_fn(example_proto, image_h, image_w, image_c):\n feature_description = {\n ""height"": tf.io.FixedLenFeature([], tf.int64),\n ""width"": tf.io.FixedLenFeature([], tf.int64),\n ""channels"": tf.io.FixedLenFeature([], tf.int64),\n ""sequence_length"": tf.io.FixedLenFeature([], tf.int64),\n ""raw_video"": tf.io.FixedLenFeature([], tf.string),\n }\n example = tf.io.parse_single_example(example_proto, feature_description)\n\n video_shape = (example[""sequence_length""], image_h, image_w, image_c)\n\n episode_tensor = tf.io.decode_raw(example[""raw_video""], out_type=tf.uint8)\n episode_tensor = tf.reshape(episode_tensor, video_shape)\n\n episode_tensor = tf.ensure_shape(episode_tensor, [None, image_h, image_w, image_c])\n return episode_tensor\n\n\ndef get_dataloader(\n tfrecord_paths: list[str], # List of TFRecord file paths\n seq_len: int,\n global_batch_size: int,\n image_h: int,\n image_w: int,\n image_c: int,\n shuffle_buffer_size: int = 1000,\n num_parallel_calls: int = tf.data.AUTOTUNE,\n seed: int = 42,\n):\n """"""\n Creates a tf.data.Dataset pipeline from TFRecord files.\n """"""\n if not tfrecord_paths:\n raise ValueError(""tfrecord_paths list cannot be empty."")\n\n process_id = jax.process_index()\n num_processes = jax.process_count()\n\n assert (\n global_batch_size % num_processes == 0\n ), ""Global batch size {global_batch_size} \\n must be divisible by the number of JAX processes {num_processes} for proper sharding.""\n per_process_batch_size = global_batch_size // num_processes\n\n dataset = tf.data.TFRecordDataset(\n tfrecord_paths, num_parallel_reads=tf.data.AUTOTUNE\n )\n\n dataset = dataset.shard(num_shards=num_processes, index=process_id)\n\n # (f.srambical) NOTE: For TFRecords, it's often good to have a large shuffle buffer.\n if shuffle_buffer_size > 0:\n dataset = dataset.shuffle(\n buffer_size=shuffle_buffer_size, seed=seed, reshuffle_each_iteration=True\n )\n parse_fn = functools.partial(\n _parse_tfrecord_fn, image_h=image_h, image_w=image_w, image_c=image_c\n )\n dataset = dataset.map(parse_fn, num_parallel_calls=num_parallel_calls)\n\n tf_process_fn = functools.partial(\n _tf_process_episode,\n seq_len=seq_len,\n image_h=image_h,\n image_w=image_w,\n image_c=image_c,\n )\n dataset = dataset.map(tf_process_fn, num_parallel_calls=num_parallel_calls)\n\n dataset = dataset.repeat(None)\n dataset = dataset.batch(per_process_batch_size, drop_remainder=True)\n dataset = dataset.prefetch(tf.data.AUTOTUNE)\n\n return dataset.as_numpy_iterator()\n",python,tab
3
+ 2,121,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
4
+ 3,886,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"5:19:57 PM [info] Activating crowd-code\n5:19:57 PM [info] Recording started\n5:19:57 PM [info] Initializing git provider using file system watchers...\n5:19:57 PM [info] Git repository found\n5:19:57 PM [info] Git provider initialized successfully\n5:19:57 PM [info] Initial git state: [object Object]\n",Log,content
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-506a7e8b-c022-493f-90ec-d205964468e01762418679057-2025_11_06-09.44.42.572/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-5f817451-b135-49e8-83cd-ad3d90c69e831764422803521-2025_11_29-14.26.46.563/source.csv ADDED
@@ -0,0 +1,106 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 2,119,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"2:26:46 PM [info] Activating crowd-code\n2:26:46 PM [info] Recording started\n2:26:46 PM [info] Initializing git provider using file system watchers...\n2:26:46 PM [info] No workspace folder found\n",Log,tab
3
+ 3,1785,"Untitled-1",0,0,"",plaintext,tab
4
+ 4,8014,"TERMINAL",0,0,"Test",,terminal_focus
5
+ 5,8018,"Untitled-1",0,0,"// crowd-pilot mock insert\n",plaintext,content
6
+ 6,14335,"Untitled-1",27,0," ",plaintext,content
7
+ 7,14384,"Untitled-1",28,0,"",plaintext,selection_keyboard
8
+ 8,17027,"Untitled-1",27,1,"",plaintext,content
9
+ 9,17440,"Untitled-1",0,0,"",plaintext,selection_command
10
+ 10,17927,"Untitled-1",27,0,"const crowdPilotMockInsert = async (req, res) => {",plaintext,content
11
+ 11,17930,"Untitled-1",77,0,"",plaintext,selection_command
12
+ 12,18887,"Untitled-1",76,0,"",plaintext,selection_command
13
+ 13,19642,"Untitled-1",26,51,"",plaintext,content
14
+ 14,19651,"Untitled-1",0,0,"",plaintext,selection_command
15
+ 15,20927,"Untitled-1",0,26,"",plaintext,content
16
+ 16,23781,"Untitled-1",0,0," // replaced by crowd-pilot",plaintext,content
17
+ 17,28652,"Untitled-1",0,27,"",plaintext,content
18
+ 18,32471,"Untitled-1",0,0,"// crowd-pilot mock insert\n",plaintext,content
19
+ 19,35166,"Untitled-1",27,0," // replaced by crowd-pilot",plaintext,content
20
+ 20,40747,"Untitled-1",54,0," mock insert",plaintext,content
21
+ 21,41943,"Untitled-1",65,0,"",plaintext,selection_command
22
+ 22,42514,"Untitled-1",26,40,"",plaintext,content
23
+ 23,42524,"Untitled-1",0,0,"",plaintext,selection_command
24
+ 24,42869,"Untitled-1",0,26,"",plaintext,content
25
+ 25,259053,"Untitled-1",0,0,"",plaintext,tab
26
+ 26,320097,"Untitled-1",0,0,"// crowd-pilot mock insert\n",plaintext,content
27
+ 27,322120,"Untitled-1",0,0,"",plaintext,selection_command
28
+ 28,325219,"Untitled-1",27,0,"",plaintext,selection_command
29
+ 29,325368,"Untitled-1",0,0,"",plaintext,selection_command
30
+ 30,326434,"Untitled-1",27,0,"",plaintext,selection_command
31
+ 31,327065,"Untitled-1",0,0,"",plaintext,selection_command
32
+ 32,327494,"Untitled-1",27,0,"",plaintext,selection_command
33
+ 33,327669,"Untitled-1",0,0,"",plaintext,selection_command
34
+ 34,328074,"Untitled-1",27,0," // replaced by crowd-pilot",plaintext,content
35
+ 35,334354,"Untitled-1",27,0,"",plaintext,selection_command
36
+ 36,334754,"Untitled-1",0,0,"",plaintext,selection_command
37
+ 37,334956,"Untitled-1",27,0,"",plaintext,selection_command
38
+ 38,335352,"Untitled-1",0,0,"",plaintext,selection_command
39
+ 39,338682,"Untitled-1",0,26,"",plaintext,content
40
+ 40,339703,"Untitled-1",1,0,"",plaintext,selection_command
41
+ 41,341925,"Untitled-1",1,0,"d",plaintext,content
42
+ 42,341927,"Untitled-1",2,0,"",plaintext,selection_keyboard
43
+ 43,342108,"Untitled-1",2,0,"d",plaintext,content
44
+ 44,342111,"Untitled-1",3,0,"",plaintext,selection_keyboard
45
+ 45,342457,"Untitled-1",2,0,"",plaintext,selection_command
46
+ 46,342898,"Untitled-1",0,30,"",plaintext,content
47
+ 47,344478,"Untitled-1",0,0,"// crowd-pilot mock insert\n",plaintext,content
48
+ 48,345380,"Untitled-1",27,0," // replaced by crowd-pilot",plaintext,content
49
+ 49,345858,"Untitled-1",27,27,"",plaintext,content
50
+ 50,346851,"Untitled-1",27,0,"// crowd-pilot mock insert\n",plaintext,content
51
+ 51,352488,"Untitled-1",27,0,"",plaintext,selection_command
52
+ 52,353793,"Untitled-1",54,0,"",plaintext,selection_command
53
+ 53,353922,"Untitled-1",27,0,"",plaintext,selection_command
54
+ 54,354141,"Untitled-1",0,0,"",plaintext,selection_command
55
+ 55,354542,"Untitled-1",27,0,"",plaintext,selection_command
56
+ 56,354840,"Untitled-1",0,0,"",plaintext,selection_command
57
+ 57,355330,"Untitled-1",54,0," // replaced by crowd-pilot",plaintext,content
58
+ 58,359823,"Untitled-1",0,26,"",plaintext,content
59
+ 59,360305,"Untitled-1",0,0,"// crowd-pilot mock insert\n",plaintext,content
60
+ 60,361249,"Untitled-1",28,0,"",plaintext,selection_command
61
+ 61,361421,"Untitled-1",27,0,"",plaintext,selection_command
62
+ 62,361591,"Untitled-1",0,0,"",plaintext,selection_command
63
+ 63,361946,"Untitled-1",27,0," // replaced by crowd-pilot",plaintext,content
64
+ 64,362415,"Untitled-1",0,26,"",plaintext,content
65
+ 65,538480,"Untitled-1",0,1,"",plaintext,content
66
+ 66,538482,"Untitled-1",1,0,"",plaintext,selection_command
67
+ 67,538810,"Untitled-1",0,28,"",plaintext,content
68
+ 68,539108,"Untitled-1",0,27,"",plaintext,content
69
+ 69,539113,"Untitled-1",1,0,"",plaintext,selection_command
70
+ 70,539444,"Untitled-1",0,27,"",plaintext,content
71
+ 71,553294,"Untitled-1",0,0,"// crowd-pilot mock insert\n",plaintext,content
72
+ 72,555984,"Untitled-1",0,0,"",plaintext,selection_command
73
+ 73,556512,"Untitled-1",27,0," // replaced by crowd-pilot",plaintext,content
74
+ 74,557430,"Untitled-1",0,26,"",plaintext,content
75
+ 75,558025,"Untitled-1",0,0,"// crowd-pilot mock insert\n",plaintext,content
76
+ 76,558583,"Untitled-1",27,0," // replaced by crowd-pilot",plaintext,content
77
+ 77,559571,"Untitled-1",27,27,"",plaintext,content
78
+ 78,559844,"Untitled-1",27,0,"// crowd-pilot mock insert\n",plaintext,content
79
+ 79,560020,"Untitled-1",54,0," // replaced by crowd-pilot",plaintext,content
80
+ 80,560244,"Untitled-1",54,27,"",plaintext,content
81
+ 81,560435,"Untitled-1",54,0,"// crowd-pilot mock insert\n",plaintext,content
82
+ 82,560652,"Untitled-1",81,0," // replaced by crowd-pilot",plaintext,content
83
+ 83,560845,"Untitled-1",81,27,"",plaintext,content
84
+ 84,561028,"Untitled-1",81,0,"// crowd-pilot mock insert\n",plaintext,content
85
+ 85,561222,"Untitled-1",108,0," // replaced by crowd-pilot",plaintext,content
86
+ 86,561999,"Untitled-1",135,0,"j",plaintext,content
87
+ 87,562001,"Untitled-1",136,0,"",plaintext,selection_keyboard
88
+ 88,562154,"Untitled-1",136,0,"j",plaintext,content
89
+ 89,562156,"Untitled-1",137,0,"",plaintext,selection_keyboard
90
+ 90,562566,"Untitled-1",136,0,"",plaintext,selection_command
91
+ 91,562748,"Untitled-1",164,0,"",plaintext,selection_command
92
+ 92,563786,"Untitled-1",138,27," // replaced by crowd-pilot",plaintext,selection_command
93
+ 93,564097,"Untitled-1",108,57," // replaced by crowd-pilotjj\n // replaced by crowd-pilot",plaintext,selection_command
94
+ 94,564260,"Untitled-1",81,84,"// crowd-pilot mock insert\n // replaced by crowd-pilotjj\n // replaced by crowd-pilot",plaintext,selection_command
95
+ 95,564438,"Untitled-1",54,111,"// crowd-pilot mock insert\n// crowd-pilot mock insert\n // replaced by crowd-pilotjj\n // replaced by crowd-pilot",plaintext,selection_command
96
+ 96,564600,"Untitled-1",27,138,"// crowd-pilot mock insert\n// crowd-pilot mock insert\n// crowd-pilot mock insert\n // replaced by crowd-pilotjj\n // replaced by crowd-pilot",plaintext,selection_command
97
+ 97,564789,"Untitled-1",0,165,"// crowd-pilot mock insert\n// crowd-pilot mock insert\n// crowd-pilot mock insert\n// crowd-pilot mock insert\n // replaced by crowd-pilotjj\n // replaced by crowd-pilot",plaintext,selection_command
98
+ 98,565053,"Untitled-1",0,165,"",plaintext,content
99
+ 99,577385,"Untitled-1",0,0,"// crowd-pilot mock insert\n",plaintext,content
100
+ 100,578345,"Untitled-1",0,0,"",plaintext,selection_command
101
+ 101,579764,"Untitled-1",27,0,"",plaintext,selection_command
102
+ 102,580043,"Untitled-1",0,0,"",plaintext,selection_command
103
+ 103,587897,"Untitled-1",27,0," // replaced by crowd-pilot",plaintext,content
104
+ 104,591971,"Untitled-1",0,26,"",plaintext,content
105
+ 105,592879,"Untitled-1",1,0,"",plaintext,selection_command
106
+ 106,594302,"Untitled-1",0,0,"// crowd-pilot mock insert\n",plaintext,content
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-63ef9b25-3351-4174-890b-f49574ab1a3c1758993863516-2025_09_27-19.24.25.912/source.csv ADDED
@@ -0,0 +1,512 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,1,"index.html",0,0,"<html>\n<head>\n <style>\n body {\n margin: 20px;\n margin-left: 12%;\n margin-right:12%;\n a {\n color: green;\n text-decoration-style: dotted;\n }\n }\n </style>\n</head>\n<p>\n Franz Srambical\n <br>\n ============\n <br>\nHi! I'm Franz. I started and scaled <a href=""https://pdoom.org"">p(doom)</a>, a Discord-based research community, from zero >200 members. We work on addressing core blockers towards general intelligence that cannot be solved by scaling up compute. We work on everything from kernel-level optimizations to large-scale distributed systems for pre-training and reinforcement learning. A lot of our current work involves finding, investigating and exploiting novel data troves <a href=""https://pdoom.org/crowd_code.html"">[1]</a> <a href=""https://pdoom.org/jasmine.html"">[2]</a>, and building open infrastructure/codebases.\nI am ex-distributed systems software engineer at <a href=""https://celonis.com"">Celonis</a>, dropped out of uni (Informatics at <a href=""https://tum.de"">TUM</a>), minored in Computational Neuroscience (courses at <a href=""https://www.gsn.uni-muenchen.de"">LMU Graduate School of Systemic Neurosciences</a>), <a href=""https://www.youtube.com/watch?v=N5nVSXV9Hbk&t=21971s"">ex-linux kernel developer</a>, ex-<a href=""public/vwa.pdf"">'alignment researcher'</a> in my high school years, ex-RA at <a href=""https://aidos.group"">Bastian Rieck's lab</a>, ex-'BCI researcher' & founding member at neuroTUM.\n<br>\n<br>\nI want to create AGI (i.e. design architectures that push the pareto-frontier of intelligent systems).\n<br>\n<br>\nIf any of our research work interests you, check out the <a href=""https://pdoom.org/blog.html"">p(doom) blog</a>. I also have <a href=""https://www.linkedin.com/in/franz-srambical-418630178/"" >linkedin</a>, <a href=""https://twitter.com/lemergenz"" >twitter</a> and <a href=""https://scholar.google.com/citations?user=W26dT4EAAAAJ&hl=en&oi=ao"" >google scholar</a>.\n<br>\n<br>\nList of preprints accumulated throughout my education that are not worthy of publication (& thus not on arxiv), but valuable to some nonetheless:\n<ul>\n <li>\n <a href=""https://pdoom.org/jax_assert.html"">\n A blog post on performance-degradation free value assertions in JAX using a very recent and still private JAX API.\n </a>\n </li>\n <li>\n <a href=""https://pdoom.org/gae_rlax.html"">\n A blog post on how a lot of PPO implementations are technically wrong (including Deepmind's reference implementation in RLax).\n </a>\n </li>\n <li>\n <a href=""https://pdoom.org/ppo.html"">\n A blog post on how practically everyone in LLM post-training is currently implicitly using REINFORCE with baseline, clipping and a likelihood ratio, and not PPO (in the classical sense).\n </a>\n </li>\n <li>\n <a href=""https://pdoom.org/thesis.html"">\n A blog post on a simple mental model that permits straightforward contextualization of the current research frontier and extrapolation of what the most important future research directions are going to be.\n </a>\n </li>\n <li>\n <a href=""https://pdoom.org/causal_mask.html"">\n A blog post on why the Transformer's causal mask is its ultimate feature, not a bug\n </a>\n </li>\n <li>\n <a href=""public/vwa.pdf"">\n A 30-page paper on AGI alignment written in 2018/19 during my high school years (unfortunately in German)\n </a>\n </li>\n <li>\n <a href=""public/cfr.pdf"">\n A review paper on superhuman poker bots\n </a>\n </li>\n <li>\n <a href=""public/mup-lr-warmup.pdf"">\n A 2-page writeup on the necessity of learning rate warmup under muParametrization\n </a> (with <a href=""https://github.com/emergenz/mup-lr-warmup"">code</a>)\n </li>\n <li>\n <a href=""public/causal_mask_poster.pdf"">\n A poster on why the Transformer's causal mask is its ultimate feature, not a bug\n </a>\n </li>\n <li>\n <a href=""public/wenn_besitzen_unfair_ist.pdf"">\n A poster on the 99-year leasehold system in Singapore as a means to mitigate generational wealth (unfortunately in German)\n </a>\n </li>\n <li>\n <a href=""public/panoptic-3d-reconstruction.pdf"">\n A crappy 'it's just x but with y'-type computer vision paper on 2% better (and 3% worse) panoptic 3D reconstruction (which included one-shot finetuning a 1B parameter diffusion model hours before the deadline)\n </a>\n </li>\n <li>\n <a href=""https://docs.google.com/document/d/14xx883ywhbJeaPz13S2lu9NY5RBwUzbNnh6K-o-Y06I/edit?usp=sharing"">\n A 2-page doc outlining my thoughts on whether machines can think \n </a>\n </li>\n</ul>\n\nIncomplete list of talks I have given:\n<ul>\n <li>\n <a href=""https://docs.google.com/presentation/d/1fq_JiOP9zZS0w_fi9sZxMb-9TmKnleugl9ik3sJK_sc/edit?usp=sharing"">\n Talk on the necessity of learning rate warmup under muParametrization (and a real-time case study on the absurd speed of modern AI research)\n </a>\n </li>\n <li>\n <a href=""https://docs.google.com/presentation/d/1V-lfRj54czkQiZw9ziEnFPdByIuM6lkuSF5gKmHvbQQ/edit?usp=sharing"">\n Talk on AlphaFold 3, motivating its architectural design through the lens of the Transformer architecture and its modern variants\n </a> \n </li>\n <li>\n <a href=""https://docs.google.com/presentation/d/1jxDhbyrCtgme_ebzchh8qIIlr8vriu4H7FClAphuDpU/edit?usp=sharing"">\n Talk at MunichNLP on p(doom), adaptive compute at inference time and predicting text-based protein function descriptions directly from sequence, bypassing structure\n </a> \n </li>\n <li>\n <a href=""https://docs.google.com/presentation/d/1JVqy-0HdfE7POWw5LWiD02fYMa7VKgi2fFkYaxibJKI/edit?usp=sharing"">\n Talk on the 'translation gap' between core and applied machine learning research, scaling protein function prediction as neural machine translation, ESM-3, AlphaFold 3, the causal mask, muTransfer and ARC-AGI\n </a> \n </li>\n <li>\n <a href=""https://docs.google.com/presentation/d/1KXu-bkNRr_0VLh1KdZazt3u3IFlFPtfkAu4AeYFqnAw/edit?usp=sharing"">\n Talk on beating modern dynamic graph classification baselines using a 1-layer LSTM\n </a> \n </li>\n <li>\n <a href=""public/counting_neurons_and_ultrasonic_communication.pdf"">\n On neurons that count and ultrasonic communication in frogs (I have catastrophically forgotten the contents of this talk)\n </a>\n </li>\n</ul>\n</p>\n<br>\n<br>\nPS: no free lunch is a myth\n</html>",html,tab
3
+ 2,93,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"7:24:25 PM [info] Activating crowd-code\n7:24:25 PM [info] Recording started\n7:24:25 PM [info] Initializing git provider using file system watchers...\n7:24:25 PM [info] Git repository found\n7:24:25 PM [info] Git provider initialized successfully\n",Log,tab
4
+ 3,141,"extension-output-pdoom-org.crowd-code-#1-crowd-code",245,0,"7:24:25 PM [info] Initial git state: [object Object]\n",Log,content
5
+ 4,1594,"index.html",0,0,"",html,tab
6
+ 5,20948,"index.html",251,0,"",html,selection_command
7
+ 6,22220,"index.html",257,0,"\n <br>",html,content
8
+ 7,22243,"index.html",260,0,"",html,selection_command
9
+ 8,22991,"index.html",257,0,"\n ",html,content
10
+ 9,23678,"index.html",258,2,"",html,content
11
+ 10,24194,"index.html",258,0,"Y",html,content
12
+ 11,24196,"index.html",259,0,"",html,selection_keyboard
13
+ 12,24361,"index.html",259,0,"o",html,content
14
+ 13,24364,"index.html",260,0,"",html,selection_keyboard
15
+ 14,24409,"index.html",260,0,"u",html,content
16
+ 15,24412,"index.html",261,0,"",html,selection_keyboard
17
+ 16,24540,"index.html",261,0," ",html,content
18
+ 17,24542,"index.html",262,0,"",html,selection_keyboard
19
+ 18,24678,"index.html",262,0,"c",html,content
20
+ 19,24682,"index.html",263,0,"",html,selection_keyboard
21
+ 20,24760,"index.html",263,0,"a",html,content
22
+ 21,24764,"index.html",264,0,"",html,selection_keyboard
23
+ 22,24876,"index.html",264,0,"n",html,content
24
+ 23,24880,"index.html",265,0,"",html,selection_keyboard
25
+ 24,24974,"index.html",265,0," ",html,content
26
+ 25,24978,"index.html",266,0,"",html,selection_keyboard
27
+ 26,25147,"index.html",266,0,"r",html,content
28
+ 27,25150,"index.html",267,0,"",html,selection_keyboard
29
+ 28,25232,"index.html",267,0,"e",html,content
30
+ 29,25234,"index.html",268,0,"",html,selection_keyboard
31
+ 30,25323,"index.html",268,0,"a",html,content
32
+ 31,25324,"index.html",269,0,"",html,selection_keyboard
33
+ 32,25489,"index.html",269,0,"c",html,content
34
+ 33,25491,"index.html",270,0,"",html,selection_keyboard
35
+ 34,25617,"index.html",270,0,"h",html,content
36
+ 35,25619,"index.html",271,0,"",html,selection_keyboard
37
+ 36,25729,"index.html",271,0," ",html,content
38
+ 37,25733,"index.html",272,0,"",html,selection_keyboard
39
+ 38,25981,"index.html",272,0,"m",html,content
40
+ 39,25985,"index.html",273,0,"",html,selection_keyboard
41
+ 40,26067,"index.html",273,0,"e",html,content
42
+ 41,26072,"index.html",274,0,"",html,selection_keyboard
43
+ 42,26140,"index.html",274,0," ",html,content
44
+ 43,26144,"index.html",275,0,"",html,selection_keyboard
45
+ 44,26315,"index.html",275,0,"a",html,content
46
+ 45,26319,"index.html",276,0,"",html,selection_keyboard
47
+ 46,26413,"index.html",276,0,"t",html,content
48
+ 47,26416,"index.html",277,0,"",html,selection_keyboard
49
+ 48,26509,"index.html",277,0," ",html,content
50
+ 49,26512,"index.html",278,0,"",html,selection_keyboard
51
+ 50,28201,"index.html",278,0,"<",html,content
52
+ 51,28205,"index.html",279,0,"",html,selection_keyboard
53
+ 52,28345,"index.html",279,0,">",html,content
54
+ 53,28347,"index.html",280,0,"",html,selection_keyboard
55
+ 54,29472,"index.html",279,1,"",html,content
56
+ 55,29713,"index.html",278,1,"",html,content
57
+ 56,31961,"index.html",278,0,"\",html,content
58
+ 57,31963,"index.html",279,0,"",html,selection_keyboard
59
+ 58,32519,"index.html",279,0,"<",html,content
60
+ 59,32522,"index.html",280,0,"",html,selection_keyboard
61
+ 60,33713,"index.html",279,1,"",html,content
62
+ 61,33846,"index.html",278,1,"",html,content
63
+ 62,37499,"index.html",278,0,"`",html,content
64
+ 63,37504,"index.html",279,0,"",html,selection_keyboard
65
+ 64,37712,"index.html",279,0,"`",html,content
66
+ 65,37714,"index.html",280,0,"",html,selection_keyboard
67
+ 66,38335,"index.html",279,1,"",html,content
68
+ 67,38477,"index.html",278,1,"",html,content
69
+ 68,53829,"index.html",278,0,"''",html,content
70
+ 69,53834,"index.html",279,0,"",html,selection_keyboard
71
+ 70,54138,"index.html",279,0,">",html,content
72
+ 71,54144,"index.html",280,0,"",html,selection_keyboard
73
+ 72,54489,"index.html",279,1,"",html,content
74
+ 73,54623,"index.html",278,2,"",html,content
75
+ 74,54799,"index.html",278,0,"\",html,content
76
+ 75,54801,"index.html",279,0,"",html,selection_keyboard
77
+ 76,55466,"index.html",279,0,">",html,content
78
+ 77,55470,"index.html",280,0,"",html,selection_keyboard
79
+ 78,56351,"index.html",280,0," ",html,content
80
+ 79,56355,"index.html",281,0,"",html,selection_keyboard
81
+ 80,56576,"index.html",281,0,"\",html,content
82
+ 81,56577,"index.html",282,0,"",html,selection_keyboard
83
+ 82,57061,"index.html",282,0,">",html,content
84
+ 83,57063,"index.html",283,0,"",html,selection_keyboard
85
+ 84,57638,"index.html",282,1,"",html,content
86
+ 85,59055,"index.html",282,0,">",html,content
87
+ 86,59062,"index.html",283,0,"",html,selection_keyboard
88
+ 87,59874,"index.html",282,0,"",html,selection_command
89
+ 88,59992,"index.html",281,0,"",html,selection_command
90
+ 89,60196,"index.html",280,0,"",html,selection_command
91
+ 90,60559,"index.html",279,0,"",html,selection_command
92
+ 91,60874,"index.html",280,0,"",html,selection_command
93
+ 92,61015,"index.html",279,1,"",html,content
94
+ 93,61448,"index.html",279,0,"<",html,content
95
+ 94,61450,"index.html",280,0,"",html,selection_keyboard
96
+ 95,61743,"index.html",279,0,"",html,selection_command
97
+ 96,62222,"index.html",280,0,"",html,selection_command
98
+ 97,64577,"index.html",279,0,"",html,selection_command
99
+ 98,64701,"index.html",280,0,"",html,selection_command
100
+ 99,64831,"index.html",281,0,"",html,selection_command
101
+ 100,65068,"index.html",280,1,"",html,content
102
+ 101,67202,"index.html",280,0,"m",html,content
103
+ 102,67208,"index.html",281,0,"",html,selection_keyboard
104
+ 103,67674,"index.html",281,0,"y",html,content
105
+ 104,67677,"index.html",282,0,"",html,selection_keyboard
106
+ 105,67889,"index.html",282,0,"_",html,content
107
+ 106,67890,"index.html",283,0,"",html,selection_keyboard
108
+ 107,68061,"index.html",283,0,"f",html,content
109
+ 108,68062,"index.html",284,0,"",html,selection_keyboard
110
+ 109,68174,"index.html",284,0,"i",html,content
111
+ 110,68175,"index.html",285,0,"",html,selection_keyboard
112
+ 111,68273,"index.html",285,0,"r",html,content
113
+ 112,68274,"index.html",286,0,"",html,selection_keyboard
114
+ 113,68321,"index.html",286,0,"s",html,content
115
+ 114,68322,"index.html",287,0,"",html,selection_keyboard
116
+ 115,68413,"index.html",287,0,"t",html,content
117
+ 116,68414,"index.html",288,0,"",html,selection_keyboard
118
+ 117,68640,"index.html",288,0,"_",html,content
119
+ 118,68642,"index.html",289,0,"",html,selection_keyboard
120
+ 119,68910,"index.html",289,0,"n",html,content
121
+ 120,68912,"index.html",290,0,"",html,selection_keyboard
122
+ 121,68973,"index.html",290,0,"a",html,content
123
+ 122,68974,"index.html",291,0,"",html,selection_keyboard
124
+ 123,69099,"index.html",291,0,"m",html,content
125
+ 124,69100,"index.html",292,0,"",html,selection_keyboard
126
+ 125,69171,"index.html",292,0,"e",html,content
127
+ 126,69173,"index.html",293,0,"",html,selection_keyboard
128
+ 127,71229,"index.html",295,0,"@<my_last_name\>.com",html,content
129
+ 128,71231,"index.html",315,0,"",html,selection_command
130
+ 129,71510,"index.html",314,0,"",html,selection_command
131
+ 130,71679,"index.html",312,0,"",html,selection_command
132
+ 131,71833,"index.html",309,0,"",html,selection_command
133
+ 132,71982,"index.html",297,0,"",html,selection_command
134
+ 133,72090,"index.html",293,0,"",html,selection_command
135
+ 134,72458,"index.html",294,0,"",html,selection_command
136
+ 135,72600,"index.html",295,0,"",html,selection_command
137
+ 136,72778,"index.html",296,0,"",html,selection_command
138
+ 137,72929,"index.html",297,0,"",html,selection_command
139
+ 138,73648,"index.html",210,6177,"",html,content
140
+ 139,75092,"index.html",209,0,"",html,selection_command
141
+ 140,75386,"index.html",211,0," Franz Srambical\n <br>\n ============\n <br>\nYou can reach me at \<my_first_name\>@<my_last_name\>.com\n <br>\nHi! I'm Franz. I started and scaled <a href=""https://pdoom.org"">p(doom)</a>, a Discord-based research community, from zero >200 members. We work on addressing core blockers towards general intelligence that cannot be solved by scaling up compute. We work on everything from kernel-level optimizations to large-scale distributed systems for pre-training and reinforcement learning. A lot of our current work involves finding, investigating and exploiting novel data troves <a href=""https://pdoom.org/crowd_code.html"">[1]</a> <a href=""https://pdoom.org/jasmine.html"">[2]</a>, and building open infrastructure/codebases.\nI am ex-distributed systems software engineer at <a href=""https://celonis.com"">Celonis</a>, dropped out of uni (Informatics at <a href=""https://tum.de"">TUM</a>), minored in Computational Neuroscience (courses at <a href=""https://www.gsn.uni-muenchen.de"">LMU Graduate School of Systemic Neurosciences</a>), <a href=""https://www.youtube.com/watch?v=N5nVSXV9Hbk&t=21971s"">ex-linux kernel developer</a>, ex-<a href=""public/vwa.pdf"">'alignment researcher'</a> in my high school years, ex-RA at <a href=""https://aidos.group"">Bastian Rieck's lab</a>, ex-'BCI researcher' & founding member at neuroTUM.\n<br>\n<br>\nI want to create AGI (i.e. design architectures that push the pareto-frontier of intelligent systems).\n<br>\n<br>\nIf any of our research work interests you, check out the <a href=""https://pdoom.org/blog.html"">p(doom) blog</a>. I also have <a href=""https://www.linkedin.com/in/franz-srambical-418630178/"" >linkedin</a>, <a href=""https://twitter.com/lemergenz"" >twitter</a> and <a href=""https://scholar.google.com/citations?user=W26dT4EAAAAJ&hl=en&oi=ao"" >google scholar</a>.\n<br>\n<br>\nList of preprints accumulated throughout my education that are not worthy of publication (& thus not on arxiv), but valuable to some nonetheless:\n<ul>\n <li>\n <a href=""https://pdoom.org/jax_assert.html"">\n A blog post on performance-degradation free value assertions in JAX using a very recent and still private JAX API.\n </a>\n </li>\n <li>\n <a href=""https://pdoom.org/gae_rlax.html"">\n A blog post on how a lot of PPO implementations are technically wrong (including Deepmind's reference implementation in RLax).\n </a>\n </li>\n <li>\n <a href=""https://pdoom.org/ppo.html"">\n A blog post on how practically everyone in LLM post-training is currently implicitly using REINFORCE with baseline, clipping and a likelihood ratio, and not PPO (in the classical sense).\n </a>\n </li>\n <li>\n <a href=""https://pdoom.org/thesis.html"">\n A blog post on a simple mental model that permits straightforward contextualization of the current research frontier and extrapolation of what the most important future research directions are going to be.\n </a>\n </li>\n <li>\n <a href=""https://pdoom.org/causal_mask.html"">\n A blog post on why the Transformer's causal mask is its ultimate feature, not a bug\n </a>\n </li>\n <li>\n <a href=""public/vwa.pdf"">\n A 30-page paper on AGI alignment written in 2018/19 during my high school years (unfortunately in German)\n </a>\n </li>\n <li>\n <a href=""public/cfr.pdf"">\n A review paper on superhuman poker bots\n </a>\n </li>\n <li>\n <a href=""public/mup-lr-warmup.pdf"">\n A 2-page writeup on the necessity of learning rate warmup under muParametrization\n </a> (with <a href=""https://github.com/emergenz/mup-lr-warmup"">code</a>)\n </li>\n <li>\n <a href=""public/causal_mask_poster.pdf"">\n A poster on why the Transformer's causal mask is its ultimate feature, not a bug\n </a>\n </li>\n <li>\n <a href=""public/wenn_besitzen_unfair_ist.pdf"">\n A poster on the 99-year leasehold system in Singapore as a means to mitigate generational wealth (unfortunately in German)\n </a>\n </li>\n <li>\n <a href=""public/panoptic-3d-reconstruction.pdf"">\n A crappy 'it's just x but with y'-type computer vision paper on 2% better (and 3% worse) panoptic 3D reconstruction (which included one-shot finetuning a 1B parameter diffusion model hours before the deadline)\n </a>\n </li>\n <li>\n <a href=""https://docs.google.com/document/d/14xx883ywhbJeaPz13S2lu9NY5RBwUzbNnh6K-o-Y06I/edit?usp=sharing"">\n A 2-page doc outlining my thoughts on whether machines can think \n </a>\n </li>\n</ul>\n\nIncomplete list of talks I have given:\n<ul>\n <li>\n <a href=""https://docs.google.com/presentation/d/1fq_JiOP9zZS0w_fi9sZxMb-9TmKnleugl9ik3sJK_sc/edit?usp=sharing"">\n Talk on the necessity of learning rate warmup under muParametrization (and a real-time case study on the absurd speed of modern AI research)\n </a>\n </li>\n <li>\n <a href=""https://docs.google.com/presentation/d/1V-lfRj54czkQiZw9ziEnFPdByIuM6lkuSF5gKmHvbQQ/edit?usp=sharing"">\n Talk on AlphaFold 3, motivating its architectural design through the lens of the Transformer architecture and its modern variants\n </a> \n </li>\n <li>\n <a href=""https://docs.google.com/presentation/d/1jxDhbyrCtgme_ebzchh8qIIlr8vriu4H7FClAphuDpU/edit?usp=sharing"">\n Talk at MunichNLP on p(doom), adaptive compute at inference time and predicting text-based protein function descriptions directly from sequence, bypassing structure\n </a> \n </li>\n <li>\n <a href=""https://docs.google.com/presentation/d/1JVqy-0HdfE7POWw5LWiD02fYMa7VKgi2fFkYaxibJKI/edit?usp=sharing"">\n Talk on the 'translation gap' between core and applied machine learning research, scaling protein function prediction as neural machine translation, ESM-3, AlphaFold 3, the causal mask, muTransfer and ARC-AGI\n </a> \n </li>\n <li>\n <a href=""https://docs.google.com/presentation/d/1KXu-bkNRr_0VLh1KdZazt3u3IFlFPtfkAu4AeYFqnAw/edit?usp=sharing"">\n Talk on beating modern dynamic graph classification baselines using a 1-layer LSTM\n </a> \n </li>\n <li>\n <a href=""public/counting_neurons_and_ultrasonic_communication.pdf"">\n On neurons that count and ultrasonic communication in frogs (I have catastrophically forgotten the contents of this talk)\n </a>\n </li>\n</ul>\n",html,content
142
+ 141,75391,"index.html",297,0,"",html,selection_command
143
+ 142,76267,"index.html",297,1,"m",html,selection_command
144
+ 143,76327,"index.html",297,12,"my_last_name",html,selection_command
145
+ 144,76569,"index.html",297,15,"my_last_name\>.",html,selection_command
146
+ 145,76803,"index.html",297,14,"my_last_name\>",html,selection_command
147
+ 146,76966,"index.html",297,13,"my_last_name\",html,selection_command
148
+ 147,77115,"index.html",297,12,"my_last_name",html,selection_command
149
+ 148,77737,"index.html",297,12,"",html,content
150
+ 149,79172,"index.html",296,0,"",html,selection_command
151
+ 150,79684,"index.html",297,0,"",html,selection_command
152
+ 151,82146,"index.html",296,0,"",html,selection_command
153
+ 152,82452,"index.html",296,0,"'",html,content
154
+ 153,82459,"index.html",297,0,"",html,selection_keyboard
155
+ 154,82642,"index.html",296,0,"",html,selection_command
156
+ 155,83706,"index.html",296,1,"\",html,content
157
+ 156,84461,"index.html",297,0,"",html,selection_command
158
+ 157,89934,"index.html",301,0,"",html,selection_command
159
+ 158,90100,"index.html",307,0,"",html,selection_command
160
+ 159,90560,"index.html",301,0,"",html,selection_command
161
+ 160,91966,"index.html",293,0,"",html,selection_command
162
+ 161,92083,"index.html",280,0,"",html,selection_command
163
+ 162,92378,"index.html",281,0,"",html,selection_command
164
+ 163,92533,"index.html",293,0,"",html,selection_command
165
+ 164,92928,"index.html",294,0,"",html,selection_command
166
+ 165,93147,"index.html",295,0,"",html,selection_command
167
+ 166,93394,"index.html",296,0,"",html,selection_command
168
+ 167,94818,"index.html",296,8,"",html,content
169
+ 168,97748,"index.html",296,0,"\",html,content
170
+ 169,97751,"index.html",297,0,"",html,selection_keyboard
171
+ 170,98436,"index.html",297,0,"<",html,content
172
+ 171,98441,"index.html",298,0,"",html,selection_keyboard
173
+ 172,98982,"index.html",298,0,"\",html,content
174
+ 173,98986,"index.html",299,0,"",html,selection_keyboard
175
+ 174,99296,"index.html",299,0,">",html,content
176
+ 175,99300,"index.html",300,0,"",html,selection_keyboard
177
+ 176,99607,"index.html",299,0,"",html,selection_command
178
+ 177,99750,"index.html",298,0,"",html,selection_command
179
+ 178,100306,"index.html",297,0,"",html,selection_command
180
+ 179,123509,"index.html",300,0,"",html,selection_command
181
+ 180,123840,"index.html",300,0,".",html,content
182
+ 181,123842,"index.html",301,0,"",html,selection_keyboard
183
+ 182,125585,"index.html",301,0,"o",html,content
184
+ 183,125591,"index.html",302,0,"",html,selection_keyboard
185
+ 184,125665,"index.html",302,0,"r",html,content
186
+ 185,125669,"index.html",303,0,"",html,selection_keyboard
187
+ 186,125698,"index.html",303,0,"g",html,content
188
+ 187,125700,"index.html",304,0,"",html,selection_keyboard
189
+ 188,125960,"index.html",303,0,"",html,selection_command
190
+ 189,126109,"index.html",302,0,"",html,selection_command
191
+ 190,126268,"index.html",301,0,"",html,selection_command
192
+ 191,126411,"index.html",300,0,"",html,selection_command
193
+ 192,126581,"index.html",299,0,"",html,selection_command
194
+ 193,126729,"index.html",298,0,"",html,selection_command
195
+ 194,136731,"index.html",298,0,"o",html,content
196
+ 195,136735,"index.html",299,0,"",html,selection_keyboard
197
+ 196,136761,"index.html",299,0,"u",html,content
198
+ 197,136762,"index.html",300,0,"",html,selection_keyboard
199
+ 198,137160,"index.html",299,1,"",html,content
200
+ 199,137410,"index.html",298,1,"",html,content
201
+ 200,137710,"index.html",298,0,"o",html,content
202
+ 201,137711,"index.html",299,0,"",html,selection_keyboard
203
+ 202,137780,"index.html",299,0,"u",html,content
204
+ 203,137781,"index.html",300,0,"",html,selection_keyboard
205
+ 204,137853,"index.html",300,0,"r",html,content
206
+ 205,137855,"index.html",301,0,"",html,selection_keyboard
207
+ 206,138094,"index.html",301,0,"_",html,content
208
+ 207,138097,"index.html",302,0,"",html,selection_keyboard
209
+ 208,138971,"index.html",302,0,"o",html,content
210
+ 209,138974,"index.html",303,0,"",html,selection_keyboard
211
+ 210,139034,"index.html",303,0,"r",html,content
212
+ 211,139036,"index.html",304,0,"",html,selection_keyboard
213
+ 212,139226,"index.html",304,0,"g",html,content
214
+ 213,139227,"index.html",305,0,"",html,selection_keyboard
215
+ 214,141457,"index.html",304,0,"",html,selection_command
216
+ 215,142619,"index.html",305,0,"",html,selection_command
217
+ 216,142829,"index.html",304,1,"",html,content
218
+ 217,142993,"index.html",303,1,"",html,content
219
+ 218,143157,"index.html",302,1,"",html,content
220
+ 219,143298,"index.html",302,0,"r",html,content
221
+ 220,143300,"index.html",303,0,"",html,selection_keyboard
222
+ 221,143404,"index.html",303,0,"e",html,content
223
+ 222,143406,"index.html",304,0,"",html,selection_keyboard
224
+ 223,143471,"index.html",304,0,"s",html,content
225
+ 224,143473,"index.html",305,0,"",html,selection_keyboard
226
+ 225,143607,"index.html",305,0,"e",html,content
227
+ 226,143608,"index.html",306,0,"",html,selection_keyboard
228
+ 227,143687,"index.html",306,0,"a",html,content
229
+ 228,143689,"index.html",307,0,"",html,selection_keyboard
230
+ 229,143771,"index.html",307,0,"r",html,content
231
+ 230,143772,"index.html",308,0,"",html,selection_keyboard
232
+ 231,143951,"index.html",308,0,"c",html,content
233
+ 232,143954,"index.html",309,0,"",html,selection_keyboard
234
+ 233,144061,"index.html",309,0,"h",html,content
235
+ 234,144063,"index.html",310,0,"",html,selection_keyboard
236
+ 235,144281,"index.html",310,0,"_",html,content
237
+ 236,144284,"index.html",311,0,"",html,selection_keyboard
238
+ 237,144494,"index.html",311,0,"o",html,content
239
+ 238,144495,"index.html",312,0,"",html,selection_keyboard
240
+ 239,144575,"index.html",312,0,"r",html,content
241
+ 240,144576,"index.html",313,0,"",html,selection_keyboard
242
+ 241,144749,"index.html",313,0,"g",html,content
243
+ 242,144751,"index.html",314,0,"",html,selection_keyboard
244
+ 243,144927,"index.html",313,0,"",html,selection_command
245
+ 244,146184,"index.html",314,0,"",html,selection_command
246
+ 245,146718,"index.html",313,0,"",html,selection_command
247
+ 246,162862,"index.html",298,0,"",html,selection_command
248
+ 247,163265,"index.html",298,16,"",html,content
249
+ 248,164085,"index.html",297,0,"",html,selection_command
250
+ 249,164433,"index.html",298,0,"",html,selection_command
251
+ 250,164591,"index.html",299,0,"",html,selection_command
252
+ 251,164761,"index.html",300,0,"",html,selection_command
253
+ 252,165148,"index.html",300,4,"",html,content
254
+ 253,165150,"index.html",299,0,"",html,selection_command
255
+ 254,165512,"index.html",298,0,"",html,selection_command
256
+ 255,166262,"index.html",298,0,"t",html,content
257
+ 256,166265,"index.html",299,0,"",html,selection_keyboard
258
+ 257,166421,"index.html",299,0,"h",html,content
259
+ 258,166423,"index.html",300,0,"",html,selection_keyboard
260
+ 259,167227,"index.html",300,0,"e",html,content
261
+ 260,167233,"index.html",301,0,"",html,selection_keyboard
262
+ 261,167822,"index.html",301,0,"_",html,content
263
+ 262,167826,"index.html",302,0,"",html,selection_keyboard
264
+ 263,168179,"index.html",302,0,"p",html,content
265
+ 264,168181,"index.html",303,0,"",html,selection_keyboard
266
+ 265,169027,"index.html",302,1,"",html,content
267
+ 266,170916,"index.html",302,0,"p",html,content
268
+ 267,170921,"index.html",303,0,"",html,selection_keyboard
269
+ 268,171109,"index.html",303,0,"d",html,content
270
+ 269,171112,"index.html",304,0,"",html,selection_keyboard
271
+ 270,171212,"index.html",304,0,"o",html,content
272
+ 271,171214,"index.html",305,0,"",html,selection_keyboard
273
+ 272,171531,"index.html",305,0,"o",html,content
274
+ 273,171533,"index.html",306,0,"",html,selection_keyboard
275
+ 274,171724,"index.html",306,0,"m",html,content
276
+ 275,171727,"index.html",307,0,"",html,selection_keyboard
277
+ 276,172230,"index.html",307,0,"_",html,content
278
+ 277,172232,"index.html",308,0,"",html,selection_keyboard
279
+ 278,173134,"index.html",308,0,"o",html,content
280
+ 279,173138,"index.html",309,0,"",html,selection_keyboard
281
+ 280,173213,"index.html",309,0,"r",html,content
282
+ 281,173217,"index.html",310,0,"",html,selection_keyboard
283
+ 282,173379,"index.html",310,0,"g",html,content
284
+ 283,173382,"index.html",311,0,"",html,selection_keyboard
285
+ 284,173598,"index.html",311,0,"_",html,content
286
+ 285,173600,"index.html",312,0,"",html,selection_keyboard
287
+ 286,173835,"index.html",312,0,"d",html,content
288
+ 287,173841,"index.html",313,0,"",html,selection_keyboard
289
+ 288,173889,"index.html",313,0,"o",html,content
290
+ 289,173891,"index.html",314,0,"",html,selection_keyboard
291
+ 290,173980,"index.html",314,0,"m",html,content
292
+ 291,173982,"index.html",315,0,"",html,selection_keyboard
293
+ 292,174087,"index.html",315,0,"a",html,content
294
+ 293,174090,"index.html",316,0,"",html,selection_keyboard
295
+ 294,174461,"index.html",315,1,"",html,content
296
+ 295,174610,"index.html",314,1,"",html,content
297
+ 296,174780,"index.html",313,1,"",html,content
298
+ 297,174921,"index.html",312,1,"",html,content
299
+ 298,175176,"index.html",312,0,"w",html,content
300
+ 299,175178,"index.html",313,0,"",html,selection_keyboard
301
+ 300,175295,"index.html",313,0,"e",html,content
302
+ 301,175296,"index.html",314,0,"",html,selection_keyboard
303
+ 302,175459,"index.html",314,0,"b",html,content
304
+ 303,175460,"index.html",315,0,"",html,selection_keyboard
305
+ 304,177144,"index.html",314,1,"",html,content
306
+ 305,177305,"index.html",313,1,"",html,content
307
+ 306,177453,"index.html",312,1,"",html,content
308
+ 307,177494,"index.html",312,0,"a",html,content
309
+ 308,177495,"index.html",313,0,"",html,selection_keyboard
310
+ 309,177581,"index.html",313,0,"d",html,content
311
+ 310,177583,"index.html",314,0,"",html,selection_keyboard
312
+ 311,178830,"index.html",313,1,"",html,content
313
+ 312,179427,"index.html",312,1,"",html,content
314
+ 313,179927,"index.html",311,1,"",html,content
315
+ 314,180171,"index.html",310,1,"",html,content
316
+ 315,180205,"index.html",309,1,"",html,content
317
+ 316,180238,"index.html",308,1,"",html,content
318
+ 317,180272,"index.html",307,1,"",html,content
319
+ 318,180306,"index.html",306,1,"",html,content
320
+ 319,180340,"index.html",305,1,"",html,content
321
+ 320,180374,"index.html",304,1,"",html,content
322
+ 321,180407,"index.html",303,1,"",html,content
323
+ 322,180511,"index.html",302,1,"",html,content
324
+ 323,180764,"index.html",301,1,"",html,content
325
+ 324,180796,"index.html",300,1,"",html,content
326
+ 325,180907,"index.html",299,1,"",html,content
327
+ 326,181099,"index.html",298,1,"",html,content
328
+ 327,181467,"index.html",297,0,"",html,selection_command
329
+ 328,181743,"index.html",298,0,"",html,selection_command
330
+ 329,181895,"index.html",299,0,"",html,selection_command
331
+ 330,182309,"index.html",300,0,"",html,selection_command
332
+ 331,182643,"index.html",300,0,".",html,content
333
+ 332,182645,"index.html",301,0,"",html,selection_keyboard
334
+ 333,182894,"index.html",301,0,"o",html,content
335
+ 334,182896,"index.html",302,0,"",html,selection_keyboard
336
+ 335,182954,"index.html",302,0,"r",html,content
337
+ 336,182957,"index.html",303,0,"",html,selection_keyboard
338
+ 337,183026,"index.html",303,0,"g",html,content
339
+ 338,183029,"index.html",304,0,"",html,selection_keyboard
340
+ 339,183231,"index.html",303,0,"",html,selection_command
341
+ 340,183477,"index.html",302,0,"",html,selection_command
342
+ 341,183610,"index.html",301,0,"",html,selection_command
343
+ 342,183765,"index.html",300,0,"",html,selection_command
344
+ 343,183894,"index.html",299,0,"",html,selection_command
345
+ 344,184033,"index.html",298,0,"",html,selection_command
346
+ 345,186117,"index.html",297,0,"",html,selection_command
347
+ 346,187736,"index.html",298,0,"",html,selection_command
348
+ 347,216666,"index.html",298,0,"o",html,content
349
+ 348,216669,"index.html",299,0,"",html,selection_keyboard
350
+ 349,216712,"index.html",299,0,"u",html,content
351
+ 350,216713,"index.html",300,0,"",html,selection_keyboard
352
+ 351,216787,"index.html",300,0,"r",html,content
353
+ 352,216790,"index.html",301,0,"",html,selection_keyboard
354
+ 353,217010,"index.html",301,0,"_",html,content
355
+ 354,217012,"index.html",302,0,"",html,selection_keyboard
356
+ 355,217240,"index.html",302,0,"d",html,content
357
+ 356,217243,"index.html",303,0,"",html,selection_keyboard
358
+ 357,217345,"index.html",303,0,"o",html,content
359
+ 358,217347,"index.html",304,0,"",html,selection_keyboard
360
+ 359,217425,"index.html",304,0,"m",html,content
361
+ 360,217427,"index.html",305,0,"",html,selection_keyboard
362
+ 361,217545,"index.html",305,0,"a",html,content
363
+ 362,217547,"index.html",306,0,"",html,selection_keyboard
364
+ 363,217616,"index.html",306,0,"i",html,content
365
+ 364,217618,"index.html",307,0,"",html,selection_keyboard
366
+ 365,217686,"index.html",307,0,"n",html,content
367
+ 366,217688,"index.html",308,0,"",html,selection_keyboard
368
+ 367,217843,"index.html",307,0,"",html,selection_command
369
+ 368,220954,"index.html",258,56,"You can reach me at \<my_first_name\>@\<our_domain\>.org",html,selection_command
370
+ 369,282815,"index.html",307,0,"",html,selection_command
371
+ 370,428719,"index.html",258,56,"You can reach me at <span style=""unicode-bidi:bidi-override; direction: rtl;"">gro.moodp@znarf</span>",html,content
372
+ 371,430826,"index.html",365,0,"",html,selection_mouse
373
+ 372,430835,"index.html",364,0,"",html,selection_command
374
+ 373,436409,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
375
+ 374,437731,"index.html",0,0,"",html,tab
376
+ 375,442872,"TERMINAL",0,0,"",,terminal_focus
377
+ 376,444165,"TERMINAL",0,0,"open .",,terminal_command
378
+ 377,444218,"TERMINAL",0,0,"]633;C",,terminal_output
379
+ 378,444275,"TERMINAL",0,0,"% \r \r",,terminal_output
380
+ 379,460114,"index.html",258,100,"You can reach me at \<my_first_name\>@\<our_domain\>.org",html,content
381
+ 380,461454,"index.html",263,0,"",html,selection_command
382
+ 381,461844,"index.html",258,0,"",html,selection_command
383
+ 382,471150,"index.html",262,0,"",html,selection_command
384
+ 383,471399,"index.html",266,0,"",html,selection_command
385
+ 384,471426,"index.html",272,0,"",html,selection_command
386
+ 385,471462,"index.html",275,0,"",html,selection_command
387
+ 386,471493,"index.html",278,0,"",html,selection_command
388
+ 387,472167,"index.html",279,0,"",html,selection_command
389
+ 388,478767,"index.html",291,0,"",html,selection_mouse
390
+ 389,479728,"index.html",258,56,"You can reach me at \<my_first_name\>@\<our_domain\>.org",html,selection_command
391
+ 390,492189,"index.html",291,0,"",html,selection_command
392
+ 391,496416,"index.html",320,0,"",html,selection_command
393
+ 392,497371,"index.html",321,0,"\n <br>",html,content
394
+ 393,497390,"index.html",324,0,"",html,selection_command
395
+ 394,525770,"index.html",279,0,"",html,selection_mouse
396
+ 395,527877,"index.html",280,0,"",html,selection_command
397
+ 396,528119,"index.html",279,1,"",html,content
398
+ 397,528344,"index.html",279,0,"@&lt;",html,content
399
+ 398,528346,"index.html",284,0,"",html,selection_keyboard
400
+ 399,528962,"index.html",283,0,"",html,selection_command
401
+ 400,529259,"index.html",282,0,"",html,selection_command
402
+ 401,529508,"index.html",281,0,"",html,selection_command
403
+ 402,529597,"index.html",280,0,"",html,selection_command
404
+ 403,529756,"index.html",279,0,"",html,selection_command
405
+ 404,530251,"index.html",278,1,"",html,content
406
+ 405,530373,"index.html",277,0,"",html,selection_command
407
+ 406,530590,"index.html",278,0,"",html,selection_command
408
+ 407,530848,"index.html",279,0,"",html,selection_command
409
+ 408,530873,"index.html",280,0,"",html,selection_command
410
+ 409,530907,"index.html",281,0,"",html,selection_command
411
+ 410,530939,"index.html",282,0,"",html,selection_command
412
+ 411,530973,"index.html",283,0,"",html,selection_command
413
+ 412,531006,"index.html",284,0,"",html,selection_command
414
+ 413,531040,"index.html",285,0,"",html,selection_command
415
+ 414,531073,"index.html",286,0,"",html,selection_command
416
+ 415,531107,"index.html",287,0,"",html,selection_command
417
+ 416,531139,"index.html",288,0,"",html,selection_command
418
+ 417,531173,"index.html",289,0,"",html,selection_command
419
+ 418,531206,"index.html",290,0,"",html,selection_command
420
+ 419,531245,"index.html",291,0,"",html,selection_command
421
+ 420,531273,"index.html",292,0,"",html,selection_command
422
+ 421,531306,"index.html",293,0,"",html,selection_command
423
+ 422,531338,"index.html",294,0,"",html,selection_command
424
+ 423,531372,"index.html",295,0,"",html,selection_command
425
+ 424,531405,"index.html",296,0,"",html,selection_command
426
+ 425,531439,"index.html",297,0,"",html,selection_command
427
+ 426,531913,"index.html",296,0,"",html,selection_command
428
+ 427,532781,"index.html",297,0,"",html,selection_command
429
+ 428,532943,"index.html",298,0,"",html,selection_command
430
+ 429,533179,"index.html",297,1,"",html,content
431
+ 430,534808,"index.html",296,1,"",html,content
432
+ 431,535324,"index.html",296,0,"@&lt;",html,content
433
+ 432,535328,"index.html",301,0,"",html,selection_keyboard
434
+ 433,540124,"index.html",300,0,"",html,selection_command
435
+ 434,540593,"index.html",299,0,"",html,selection_command
436
+ 435,541730,"index.html",299,1,"t",html,selection_command
437
+ 436,541826,"index.html",298,2,"lt",html,selection_command
438
+ 437,541938,"index.html",298,2,"",html,content
439
+ 438,542391,"index.html",298,0,"g",html,content
440
+ 439,542395,"index.html",299,0,"",html,selection_keyboard
441
+ 440,542612,"index.html",299,0,"t",html,content
442
+ 441,542621,"index.html",300,0,"",html,selection_keyboard
443
+ 442,542868,"index.html",299,0,"",html,selection_command
444
+ 443,543165,"index.html",298,0,"",html,selection_command
445
+ 444,544114,"index.html",299,0,"",html,selection_command
446
+ 445,544322,"index.html",300,0,"",html,selection_command
447
+ 446,544976,"index.html",299,0,"",html,selection_command
448
+ 447,545162,"index.html",298,0,"",html,selection_command
449
+ 448,545331,"index.html",297,0,"",html,selection_command
450
+ 449,545573,"index.html",296,0,"",html,selection_command
451
+ 450,552504,"index.html",296,1,"",html,content
452
+ 451,552628,"index.html",283,0,"",html,selection_command
453
+ 452,552769,"index.html",282,0,"",html,selection_command
454
+ 453,553049,"index.html",280,0,"",html,selection_command
455
+ 454,553236,"index.html",278,0,"",html,selection_command
456
+ 455,553593,"index.html",278,1,"",html,content
457
+ 456,554351,"index.html",279,0,"",html,selection_command
458
+ 457,554596,"index.html",280,0,"",html,selection_command
459
+ 458,554627,"index.html",281,0,"",html,selection_command
460
+ 459,554660,"index.html",282,0,"",html,selection_command
461
+ 460,554695,"index.html",283,0,"",html,selection_command
462
+ 461,554728,"index.html",284,0,"",html,selection_command
463
+ 462,554761,"index.html",285,0,"",html,selection_command
464
+ 463,554797,"index.html",286,0,"",html,selection_command
465
+ 464,554991,"index.html",287,0,"",html,selection_command
466
+ 465,556225,"index.html",314,0,"&gt;",html,content
467
+ 466,556225,"index.html",312,2,"",html,content
468
+ 467,556225,"index.html",302,0,"&lt;",html,content
469
+ 468,556225,"index.html",300,2,"",html,content
470
+ 469,581018,"index.html",282,0,"",html,selection_command
471
+ 470,581160,"index.html",281,0,"",html,selection_command
472
+ 471,581269,"index.html",279,0,"",html,selection_command
473
+ 472,581646,"index.html",278,0,"",html,selection_command
474
+ 473,583688,"index.html",278,0,"<",html,content
475
+ 474,583691,"index.html",279,0,"",html,selection_keyboard
476
+ 475,583759,"index.html",279,0,">",html,content
477
+ 476,583762,"index.html",280,0,"",html,selection_keyboard
478
+ 477,584253,"index.html",279,0,"",html,selection_command
479
+ 478,584486,"index.html",279,0,"c",html,content
480
+ 479,584488,"index.html",280,0,"",html,selection_keyboard
481
+ 480,584626,"index.html",280,0,"o",html,content
482
+ 481,584628,"index.html",281,0,"",html,selection_keyboard
483
+ 482,584701,"index.html",281,0,"d",html,content
484
+ 483,584703,"index.html",282,0,"",html,selection_keyboard
485
+ 484,584753,"index.html",282,0,"e",html,content
486
+ 485,584754,"index.html",283,0,"",html,selection_keyboard
487
+ 486,585006,"index.html",282,0,"",html,selection_command
488
+ 487,585695,"index.html",328,0,"",html,selection_command
489
+ 488,586173,"index.html",328,0,"<",html,content
490
+ 489,586175,"index.html",329,0,"",html,selection_keyboard
491
+ 490,586288,"index.html",329,0,">",html,content
492
+ 491,586290,"index.html",330,0,"",html,selection_keyboard
493
+ 492,586646,"index.html",329,0,"",html,selection_command
494
+ 493,586931,"index.html",329,0,"/",html,content
495
+ 494,586933,"index.html",330,0,"",html,selection_keyboard
496
+ 495,587039,"index.html",330,0,"code",html,content
497
+ 496,587241,"index.html",334,0,"c",html,content
498
+ 497,587243,"index.html",335,0,"",html,selection_keyboard
499
+ 498,587344,"index.html",335,0,"o",html,content
500
+ 499,587345,"index.html",336,0,"",html,selection_keyboard
501
+ 500,587457,"index.html",336,0,"d",html,content
502
+ 501,587458,"index.html",337,0,"",html,selection_keyboard
503
+ 502,587488,"index.html",337,0,"e",html,content
504
+ 503,587489,"index.html",338,0,"",html,selection_keyboard
505
+ 504,587754,"index.html",337,1,"",html,content
506
+ 505,587904,"index.html",336,1,"",html,content
507
+ 506,588044,"index.html",335,1,"",html,content
508
+ 507,588175,"index.html",334,1,"",html,content
509
+ 508,588257,"index.html",333,0,"",html,selection_command
510
+ 509,618072,"index.html",258,0,"",html,selection_command
511
+ 510,618922,"index.html",251,0,"",html,selection_command
512
+ 511,619267,"index.html",236,0,"",html,selection_command
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-6685e05b-001a-49a4-91eb-69d4235b607a1764445464368-2025_11_29-20.44.27.618/source.csv ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 2,117,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"8:44:27 PM [info] Activating crowd-code\n8:44:27 PM [info] Recording started\n8:44:27 PM [info] Initializing git provider using file system watchers...\n8:44:27 PM [info] No workspace folder found\n",Log,tab
3
+ 3,2030,"extension-output-pdoom-org.crowd-code-#1-crowd-code",194,0,"8:44:29 PM [info] Retrying git provider initialization...\n8:44:29 PM [info] No workspace folder found\n",Log,content
4
+ 4,73787,"Untitled-1",0,0,"",plaintext,tab
5
+ 5,74471,"TERMINAL",0,0,"Test",,terminal_focus
6
+ 6,74474,"Untitled-1",0,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
7
+ 7,75274,"Untitled-1",46,0,"",plaintext,selection_command
8
+ 8,75351,"Untitled-1",39,0,"",plaintext,selection_command
9
+ 9,75502,"Untitled-1",32,0,"",plaintext,selection_command
10
+ 10,75646,"Untitled-1",0,0,"",plaintext,selection_command
11
+ 11,76093,"Untitled-1",0,0,"\n",plaintext,content
12
+ 12,76693,"Untitled-1",0,0,"\n",plaintext,content
13
+ 13,77849,"Untitled-1",0,0,"",plaintext,selection_command
14
+ 14,137441,"Untitled-1",0,78,"",plaintext,content
15
+ 15,143149,"Untitled-1",0,0,"/* crowd-pilot: replacement */\nREPLACED LINE 1\nREPLACED LINE 2",plaintext,content
16
+ 16,145392,"Untitled-1",61,1,"",plaintext,content
17
+ 17,145854,"Untitled-1",47,14,"",plaintext,content
18
+ 18,145979,"Untitled-1",46,1,"",plaintext,content
19
+ 19,146740,"Untitled-1",31,15,"",plaintext,content
20
+ 20,147149,"Untitled-1",30,1,"",plaintext,content
21
+ 21,147612,"Untitled-1",0,30,"",plaintext,content
22
+ 22,150661,"Untitled-1",0,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
23
+ 23,151667,"Untitled-1",46,0,"",plaintext,selection_command
24
+ 24,151761,"Untitled-1",39,0,"",plaintext,selection_command
25
+ 25,151910,"Untitled-1",32,0,"",plaintext,selection_command
26
+ 26,152050,"Untitled-1",0,0,"",plaintext,selection_command
27
+ 27,153296,"Untitled-1",0,0,"\n",plaintext,content
28
+ 28,153771,"Untitled-1",0,0,"\n",plaintext,content
29
+ 29,154902,"Untitled-1",0,0,"",plaintext,selection_command
30
+ 30,189329,"Untitled-1",34,13,"/* crowd-pilot: replacement */\nREPLACED LINE 1\nREPLACED LINE 2",plaintext,content
31
+ 31,191206,"Untitled-1",97,30,"",plaintext,content
32
+ 32,192387,"Untitled-1",2,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
33
+ 33,213059,"Untitled-1",1,0,"",plaintext,selection_command
34
+ 34,213141,"Untitled-1",2,0,"",plaintext,selection_command
35
+ 35,213263,"Untitled-1",34,0,"",plaintext,selection_command
36
+ 36,213517,"Untitled-1",41,0,"",plaintext,selection_command
37
+ 37,213547,"Untitled-1",48,0,"",plaintext,selection_command
38
+ 38,213580,"Untitled-1",78,0,"",plaintext,selection_command
39
+ 39,213613,"Untitled-1",110,0,"",plaintext,selection_command
40
+ 40,213647,"Untitled-1",141,0,"",plaintext,selection_command
41
+ 41,213680,"Untitled-1",157,0,"",plaintext,selection_command
42
+ 42,214171,"Untitled-1",141,0,"",plaintext,selection_command
43
+ 43,214607,"Untitled-1",141,16,"",plaintext,content
44
+ 44,214969,"Untitled-1",141,16,"",plaintext,content
45
+ 45,215292,"Untitled-1",110,0,"",plaintext,selection_command
46
+ 46,215612,"Untitled-1",78,0,"",plaintext,selection_command
47
+ 47,217205,"Untitled-1",78,32,"",plaintext,content
48
+ 48,217552,"Untitled-1",78,31,"",plaintext,content
49
+ 49,217766,"Untitled-1",48,0,"",plaintext,selection_command
50
+ 50,217866,"Untitled-1",41,0,"",plaintext,selection_command
51
+ 51,217997,"Untitled-1",34,0,"",plaintext,selection_command
52
+ 52,218243,"Untitled-1",2,0,"",plaintext,selection_command
53
+ 53,218277,"Untitled-1",1,0,"",plaintext,selection_command
54
+ 54,218310,"Untitled-1",0,0,"",plaintext,selection_command
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-67ee49a3-cbd3-48c6-b0d1-b2c0dde2809c1764447402473-2025_11_29-21.16.46.562/source.csv ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,40,"Untitled-1",0,0,"",plaintext,tab
3
+ 2,56,"Untitled-1",0,0,"\n",plaintext,content
4
+ 3,120,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"9:16:46 PM [info] Activating crowd-code\n9:16:46 PM [info] Recording started\n9:16:46 PM [info] Initializing git provider using file system watchers...\n9:16:46 PM [info] No workspace folder found\n",Log,tab
5
+ 4,1065,"Untitled-1",0,0,"",plaintext,tab
6
+ 5,1728,"Untitled-1",0,0,"",plaintext,selection_command
7
+ 6,2181,"Untitled-1",0,0,"\n",plaintext,content
8
+ 7,2591,"Untitled-1",0,0,"",plaintext,selection_command
9
+ 8,4308,"Untitled-1",2,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
10
+ 9,8930,"Untitled-1",1,0,"",plaintext,selection_command
11
+ 10,9293,"Untitled-1",41,36,"/* crowd-pilot: replacement */\nREPLACED LINE 1\nREPLACED LINE 2",plaintext,content
12
+ 11,9694,"TERMINAL",0,0,"undefinedfranzsrambical@MBF6N9WFVKFV ~ % echo ""Hello World""",,terminal_command
13
+ 12,9695,"TERMINAL",0,0,"]633;CHello World\r\n% \r \r",,terminal_output
14
+ 13,10598,"Untitled-1",34,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
15
+ 14,11295,"Untitled-1",66,13,"/* crowd-pilot: replacement */\nREPLACED LINE 1\nREPLACED LINE 2",plaintext,content
16
+ 15,11753,"Untitled-1",129,67,"",plaintext,content
17
+ 16,12980,"Untitled-1",2,0,"",plaintext,selection_command
18
+ 17,13912,"Untitled-1",2,160,"",plaintext,content
19
+ 18,14581,"Untitled-1",1,0,"",plaintext,selection_command
20
+ 19,14713,"Untitled-1",0,0,"",plaintext,selection_command
21
+ 20,15843,"Untitled-1",1,0,"",plaintext,selection_command
22
+ 21,16043,"Untitled-1",0,0,"",plaintext,selection_command
23
+ 22,16824,"Untitled-1",1,0,"",plaintext,selection_command
24
+ 23,16957,"Untitled-1",2,0,"",plaintext,selection_command
25
+ 24,17872,"TERMINAL",0,0,"echo ""Hello World""",,terminal_command
26
+ 25,17873,"TERMINAL",0,0,"]633;CHello World\r\n% \r \r",,terminal_output
27
+ 26,18854,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
28
+ 27,21927,"TERMINAL",0,0,"",,terminal_focus
29
+ 28,21930,"Untitled-1",0,0,"",plaintext,tab
30
+ 29,24480,"TERMINAL",0,0,"zsh",,terminal_focus
31
+ 30,28713,"Untitled-1",1,0,"",plaintext,selection_command
32
+ 31,28814,"Untitled-1",0,0,"",plaintext,selection_command
33
+ 32,31125,"Untitled-1",0,0,"\n",plaintext,content
34
+ 33,32693,"Untitled-1",0,0,"",plaintext,selection_command
35
+ 34,36217,"Untitled-1",1,0,"",plaintext,selection_command
36
+ 35,36378,"Untitled-1",2,0,"",plaintext,selection_command
37
+ 36,37422,"Untitled-1",3,0,"",plaintext,selection_command
38
+ 37,39092,"Untitled-1",2,0,"",plaintext,selection_command
39
+ 38,39217,"Untitled-1",1,0,"",plaintext,selection_command
40
+ 39,39388,"Untitled-1",0,0,"",plaintext,selection_command
41
+ 40,42146,"Untitled-1",1,0,"",plaintext,selection_command
42
+ 41,42987,"Untitled-1",2,0,"",plaintext,selection_command
43
+ 42,43256,"Untitled-1",1,0,"",plaintext,selection_command
44
+ 43,43417,"Untitled-1",0,0,"",plaintext,selection_command
45
+ 44,44473,"Untitled-1",1,0,"",plaintext,selection_command
46
+ 45,44613,"Untitled-1",2,0,"",plaintext,selection_command
47
+ 46,44762,"Untitled-1",3,0,"",plaintext,selection_command
48
+ 47,45072,"Untitled-1",3,0,"\n",plaintext,content
49
+ 48,45336,"Untitled-1",4,0,"\n",plaintext,content
50
+ 49,45489,"Untitled-1",5,0,"\n",plaintext,content
51
+ 50,45613,"Untitled-1",6,0,"\n",plaintext,content
52
+ 51,45751,"Untitled-1",7,0,"\n",plaintext,content
53
+ 52,45882,"Untitled-1",8,0,"\n",plaintext,content
54
+ 53,47013,"Untitled-1",8,0,"",plaintext,selection_command
55
+ 54,47097,"Untitled-1",7,0,"",plaintext,selection_command
56
+ 55,47346,"Untitled-1",6,0,"",plaintext,selection_command
57
+ 56,47379,"Untitled-1",5,0,"",plaintext,selection_command
58
+ 57,47409,"Untitled-1",4,0,"",plaintext,selection_command
59
+ 58,47441,"Untitled-1",3,0,"",plaintext,selection_command
60
+ 59,47475,"Untitled-1",2,0,"",plaintext,selection_command
61
+ 60,47510,"Untitled-1",1,0,"",plaintext,selection_command
62
+ 61,47544,"Untitled-1",0,0,"",plaintext,selection_command
63
+ 62,51467,"Untitled-1",2,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
64
+ 63,54118,"Untitled-1",1,0,"",plaintext,selection_command
65
+ 64,54248,"Untitled-1",0,0,"",plaintext,selection_command
66
+ 65,60485,"Untitled-1",34,13,"/* crowd-pilot: replacement */\nREPLACED LINE 1\nREPLACED LINE 2",plaintext,content
67
+ 66,61161,"Untitled-1",97,31,"",plaintext,content
68
+ 67,67015,"TERMINAL",0,0,"echo ""Hello World""",,terminal_command
69
+ 68,67016,"TERMINAL",0,0,"]633;CHello World\r\n% \r \r",,terminal_output
70
+ 69,68043,"Untitled-1",1,0,"",plaintext,selection_command
71
+ 70,68143,"Untitled-1",0,0,"",plaintext,selection_command
72
+ 71,70887,"Untitled-1",1,0,"",plaintext,selection_command
73
+ 72,71014,"Untitled-1",2,0,"",plaintext,selection_command
74
+ 73,71264,"Untitled-1",34,0,"",plaintext,selection_command
75
+ 74,71297,"Untitled-1",65,0,"",plaintext,selection_command
76
+ 75,71326,"Untitled-1",81,0,"",plaintext,selection_command
77
+ 76,71360,"Untitled-1",97,0,"",plaintext,selection_command
78
+ 77,71394,"Untitled-1",98,0,"",plaintext,selection_command
79
+ 78,71428,"Untitled-1",99,0,"",plaintext,selection_command
80
+ 79,71461,"Untitled-1",100,0,"",plaintext,selection_command
81
+ 80,71494,"Untitled-1",101,0,"",plaintext,selection_command
82
+ 81,71528,"Untitled-1",102,0,"",plaintext,selection_command
83
+ 82,71786,"Untitled-1",101,0,"",plaintext,selection_command
84
+ 83,72032,"Untitled-1",100,0,"",plaintext,selection_command
85
+ 84,72063,"Untitled-1",99,0,"",plaintext,selection_command
86
+ 85,72095,"Untitled-1",98,0,"",plaintext,selection_command
87
+ 86,72126,"Untitled-1",97,0,"",plaintext,selection_command
88
+ 87,72159,"Untitled-1",81,0,"",plaintext,selection_command
89
+ 88,72192,"Untitled-1",65,0,"",plaintext,selection_command
90
+ 89,72225,"Untitled-1",34,0,"",plaintext,selection_command
91
+ 90,72258,"Untitled-1",2,0,"",plaintext,selection_command
92
+ 91,72293,"Untitled-1",1,0,"",plaintext,selection_command
93
+ 92,72327,"Untitled-1",0,0,"",plaintext,selection_command
94
+ 93,76826,"Untitled-1",1,0,"",plaintext,selection_command
95
+ 94,76916,"Untitled-1",2,0,"",plaintext,selection_command
96
+ 95,78658,"Untitled-1",2,101,"",plaintext,content
97
+ 96,80308,"Untitled-1",2,0,"\n",plaintext,content
98
+ 97,80578,"Untitled-1",3,0,"\n",plaintext,content
99
+ 98,80919,"Untitled-1",4,0,"\n",plaintext,content
100
+ 99,81058,"Untitled-1",5,0,"\n",plaintext,content
101
+ 100,81197,"Untitled-1",6,0,"\n",plaintext,content
102
+ 101,81342,"Untitled-1",7,0,"\n",plaintext,content
103
+ 102,81601,"Untitled-1",7,0,"",plaintext,selection_command
104
+ 103,81850,"Untitled-1",6,0,"",plaintext,selection_command
105
+ 104,81880,"Untitled-1",5,0,"",plaintext,selection_command
106
+ 105,81912,"Untitled-1",4,0,"",plaintext,selection_command
107
+ 106,81944,"Untitled-1",3,0,"",plaintext,selection_command
108
+ 107,81974,"Untitled-1",2,0,"",plaintext,selection_command
109
+ 108,82009,"Untitled-1",1,0,"",plaintext,selection_command
110
+ 109,82042,"Untitled-1",0,0,"",plaintext,selection_command
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-76df12ce-63e6-4f0f-97c2-bd8d132564c31762368713747-2025_11_05-19.51.58.593/source.csv ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 2,445,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"7:51:58 PM [info] Activating crowd-code\n7:51:58 PM [info] Recording started\n7:51:58 PM [info] Initializing git provider using file system watchers...\n7:51:58 PM [info] No workspace folder found\n",Log,tab
3
+ 3,8991,"Untitled-1",0,0,"",plaintext,tab
4
+ 4,11195,"TERMINAL",0,0,"Test",,terminal_focus
5
+ 5,11245,"Untitled-1",0,0,"hello world\n",plaintext,content
6
+ 6,11656,"TERMINAL",0,0,"echo VSCode test",,terminal_command
7
+ 7,11659,"TERMINAL",0,0,"]633;CVSCode test\r\n% \r \r",,terminal_output
8
+ 8,27017,"Untitled-1",0,0,"",plaintext,selection_command
9
+ 9,27019,"Untitled-1",0,0,"hello world\n",plaintext,content
10
+ 10,27056,"TERMINAL",0,0,"echo VSCode test",,terminal_command
11
+ 11,27057,"TERMINAL",0,0,"]633;CVSCode test\r\n% \r \r",,terminal_output
12
+ 12,143014,"Untitled-1",12,12,"",plaintext,content
13
+ 13,143312,"Untitled-1",0,0,"",plaintext,selection_command
14
+ 14,143674,"Untitled-1",0,12,"",plaintext,content
15
+ 15,157904,"Untitled-1",0,0,"hello world\n",plaintext,content
16
+ 16,158186,"TERMINAL",0,0,"echo VSCode test",,terminal_command
17
+ 17,158188,"TERMINAL",0,0,"]633;CVSCode test\r\n% \r \r",,terminal_output
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-7bb39648-a065-4847-8b07-5024f699fc131758998987531-2025_09_27-20.49.55.850/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-7f860396-c5c8-4f1f-8ce7-04e005748e611754402256906-2025_08_05-15.57.44.850/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-847fd734-88e4-48ed-bb91-f08cf4e2ebd91762367961128-2025_11_05-19.39.26.989/source.csv ADDED
@@ -0,0 +1,559 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 2,474,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"7:39:26 PM [info] Activating crowd-code\n7:39:26 PM [info] Recording started\n7:39:26 PM [info] Initializing git provider using file system watchers...\n7:39:27 PM [info] Git repository found\n7:39:27 PM [info] Git provider initialized successfully\n",Log,tab
3
+ 3,702,"extension-output-pdoom-org.crowd-code-#1-crowd-code",245,0,"7:39:27 PM [info] Initial git state: [object Object]\n",Log,content
4
+ 4,5342,"TERMINAL",0,0,"",,terminal_focus
5
+ 5,7581,"TERMINAL",0,0,"npx --package yo --package generator-code -- yo code",,terminal_command
6
+ 6,7630,"TERMINAL",0,0,"]633;C",,terminal_output
7
+ 7,8387,"TERMINAL",0,0,"⠙",,terminal_output
8
+ 8,8467,"TERMINAL",0,0,"⠹",,terminal_output
9
+ 9,8547,"TERMINAL",0,0,"⠸",,terminal_output
10
+ 10,8630,"TERMINAL",0,0,"⠼",,terminal_output
11
+ 11,8749,"TERMINAL",0,0,"⠴Need to install the following packages:\r\nyo@5.1.0\r\ngenerator-code@1.11.13\r\nOk to proceed? (y) ",,terminal_output
12
+ 12,9905,"TERMINAL",0,0,"y",,terminal_output
13
+ 13,10186,"TERMINAL",0,0,"\r\r\n\r\n⠙",,terminal_output
14
+ 14,10270,"TERMINAL",0,0,"⠹",,terminal_output
15
+ 15,10349,"TERMINAL",0,0,"⠸",,terminal_output
16
+ 16,10430,"TERMINAL",0,0,"⠼",,terminal_output
17
+ 17,10511,"TERMINAL",0,0,"⠴",,terminal_output
18
+ 18,10630,"TERMINAL",0,0,"⠦",,terminal_output
19
+ 19,10691,"TERMINAL",0,0,"⠧",,terminal_output
20
+ 20,10772,"TERMINAL",0,0,"⠇",,terminal_output
21
+ 21,10853,"TERMINAL",0,0,"⠏",,terminal_output
22
+ 22,10935,"TERMINAL",0,0,"⠋",,terminal_output
23
+ 23,11016,"TERMINAL",0,0,"⠙",,terminal_output
24
+ 24,11095,"TERMINAL",0,0,"⠹",,terminal_output
25
+ 25,11176,"TERMINAL",0,0,"⠸",,terminal_output
26
+ 26,11256,"TERMINAL",0,0,"⠼",,terminal_output
27
+ 27,11336,"TERMINAL",0,0,"⠴",,terminal_output
28
+ 28,11475,"TERMINAL",0,0,"⠦",,terminal_output
29
+ 29,11557,"TERMINAL",0,0,"⠧",,terminal_output
30
+ 30,11636,"TERMINAL",0,0,"⠇",,terminal_output
31
+ 31,11720,"TERMINAL",0,0,"⠏",,terminal_output
32
+ 32,11801,"TERMINAL",0,0,"⠋",,terminal_output
33
+ 33,11879,"TERMINAL",0,0,"⠙",,terminal_output
34
+ 34,11957,"TERMINAL",0,0,"⠹",,terminal_output
35
+ 35,12038,"TERMINAL",0,0,"⠸",,terminal_output
36
+ 36,12117,"TERMINAL",0,0,"⠼",,terminal_output
37
+ 37,12199,"TERMINAL",0,0,"⠴",,terminal_output
38
+ 38,12281,"TERMINAL",0,0,"⠦",,terminal_output
39
+ 39,12362,"TERMINAL",0,0,"⠧",,terminal_output
40
+ 40,12442,"TERMINAL",0,0,"⠇",,terminal_output
41
+ 41,12523,"TERMINAL",0,0,"⠏",,terminal_output
42
+ 42,12605,"TERMINAL",0,0,"⠋",,terminal_output
43
+ 43,12687,"TERMINAL",0,0,"⠙",,terminal_output
44
+ 44,12766,"TERMINAL",0,0,"⠹",,terminal_output
45
+ 45,12846,"TERMINAL",0,0,"⠸",,terminal_output
46
+ 46,12930,"TERMINAL",0,0,"⠼",,terminal_output
47
+ 47,13010,"TERMINAL",0,0,"⠴",,terminal_output
48
+ 48,13091,"TERMINAL",0,0,"⠦",,terminal_output
49
+ 49,13173,"TERMINAL",0,0,"⠧",,terminal_output
50
+ 50,13253,"TERMINAL",0,0,"⠇",,terminal_output
51
+ 51,13332,"TERMINAL",0,0,"⠏",,terminal_output
52
+ 52,13412,"TERMINAL",0,0,"⠋",,terminal_output
53
+ 53,13495,"TERMINAL",0,0,"⠙",,terminal_output
54
+ 54,13575,"TERMINAL",0,0,"⠹",,terminal_output
55
+ 55,13655,"TERMINAL",0,0,"⠸",,terminal_output
56
+ 56,13736,"TERMINAL",0,0,"⠼",,terminal_output
57
+ 57,13815,"TERMINAL",0,0,"⠴",,terminal_output
58
+ 58,13898,"TERMINAL",0,0,"⠦",,terminal_output
59
+ 59,13979,"TERMINAL",0,0,"⠧",,terminal_output
60
+ 60,14058,"TERMINAL",0,0,"⠇",,terminal_output
61
+ 61,14139,"TERMINAL",0,0,"⠏",,terminal_output
62
+ 62,14219,"TERMINAL",0,0,"⠋",,terminal_output
63
+ 63,14302,"TERMINAL",0,0,"⠙",,terminal_output
64
+ 64,14382,"TERMINAL",0,0,"⠹",,terminal_output
65
+ 65,14461,"TERMINAL",0,0,"⠸",,terminal_output
66
+ 66,14543,"TERMINAL",0,0,"⠼",,terminal_output
67
+ 67,14624,"TERMINAL",0,0,"⠴",,terminal_output
68
+ 68,14706,"TERMINAL",0,0,"⠦",,terminal_output
69
+ 69,14784,"TERMINAL",0,0,"⠧",,terminal_output
70
+ 70,14866,"TERMINAL",0,0,"⠇",,terminal_output
71
+ 71,14947,"TERMINAL",0,0,"⠏",,terminal_output
72
+ 72,15026,"TERMINAL",0,0,"⠋",,terminal_output
73
+ 73,15106,"TERMINAL",0,0,"⠙",,terminal_output
74
+ 74,15188,"TERMINAL",0,0,"⠹",,terminal_output
75
+ 75,15267,"TERMINAL",0,0,"⠸",,terminal_output
76
+ 76,15349,"TERMINAL",0,0,"⠼",,terminal_output
77
+ 77,15427,"TERMINAL",0,0,"⠴",,terminal_output
78
+ 78,15508,"TERMINAL",0,0,"⠦",,terminal_output
79
+ 79,15590,"TERMINAL",0,0,"⠧",,terminal_output
80
+ 80,15671,"TERMINAL",0,0,"⠇",,terminal_output
81
+ 81,15750,"TERMINAL",0,0,"⠏",,terminal_output
82
+ 82,15834,"TERMINAL",0,0,"⠋",,terminal_output
83
+ 83,15911,"TERMINAL",0,0,"⠙",,terminal_output
84
+ 84,15994,"TERMINAL",0,0,"⠹",,terminal_output
85
+ 85,16074,"TERMINAL",0,0,"⠸",,terminal_output
86
+ 86,16153,"TERMINAL",0,0,"⠼",,terminal_output
87
+ 87,16236,"TERMINAL",0,0,"⠴",,terminal_output
88
+ 88,16319,"TERMINAL",0,0,"⠦",,terminal_output
89
+ 89,16402,"TERMINAL",0,0,"⠧",,terminal_output
90
+ 90,16493,"TERMINAL",0,0,"⠇",,terminal_output
91
+ 91,16569,"TERMINAL",0,0,"⠏",,terminal_output
92
+ 92,16648,"TERMINAL",0,0,"⠋",,terminal_output
93
+ 93,16731,"TERMINAL",0,0,"⠙",,terminal_output
94
+ 94,16811,"TERMINAL",0,0,"⠹",,terminal_output
95
+ 95,16892,"TERMINAL",0,0,"⠸",,terminal_output
96
+ 96,16974,"TERMINAL",0,0,"⠼",,terminal_output
97
+ 97,17355,"TERMINAL",0,0,"⠴",,terminal_output
98
+ 98,17447,"TERMINAL",0,0,"⠦",,terminal_output
99
+ 99,17528,"TERMINAL",0,0,"⠧",,terminal_output
100
+ 100,17598,"TERMINAL",0,0,"⠇",,terminal_output
101
+ 101,17684,"TERMINAL",0,0,"⠏",,terminal_output
102
+ 102,17805,"TERMINAL",0,0,"⠋",,terminal_output
103
+ 103,17921,"TERMINAL",0,0,"⠙",,terminal_output
104
+ 104,18047,"TERMINAL",0,0,"⠹",,terminal_output
105
+ 105,18124,"TERMINAL",0,0,"⠸",,terminal_output
106
+ 106,18204,"TERMINAL",0,0,"⠼",,terminal_output
107
+ 107,18285,"TERMINAL",0,0,"⠴",,terminal_output
108
+ 108,18365,"TERMINAL",0,0,"⠦",,terminal_output
109
+ 109,18445,"TERMINAL",0,0,"⠧",,terminal_output
110
+ 110,18526,"TERMINAL",0,0,"⠇",,terminal_output
111
+ 111,18606,"TERMINAL",0,0,"⠏",,terminal_output
112
+ 112,18685,"TERMINAL",0,0,"⠋",,terminal_output
113
+ 113,18765,"TERMINAL",0,0,"⠙",,terminal_output
114
+ 114,18849,"TERMINAL",0,0,"⠹",,terminal_output
115
+ 115,18928,"TERMINAL",0,0,"⠸",,terminal_output
116
+ 116,19009,"TERMINAL",0,0,"⠼",,terminal_output
117
+ 117,19091,"TERMINAL",0,0,"⠴",,terminal_output
118
+ 118,19175,"TERMINAL",0,0,"⠦",,terminal_output
119
+ 119,19253,"TERMINAL",0,0,"⠧",,terminal_output
120
+ 120,19338,"TERMINAL",0,0,"⠇",,terminal_output
121
+ 121,19419,"TERMINAL",0,0,"⠏",,terminal_output
122
+ 122,19499,"TERMINAL",0,0,"⠋",,terminal_output
123
+ 123,19644,"TERMINAL",0,0,"⠙",,terminal_output
124
+ 124,19789,"TERMINAL",0,0,"⠹",,terminal_output
125
+ 125,19863,"TERMINAL",0,0,"⠸",,terminal_output
126
+ 126,19948,"TERMINAL",0,0,"⠼",,terminal_output
127
+ 127,20027,"TERMINAL",0,0,"⠴",,terminal_output
128
+ 128,20110,"TERMINAL",0,0,"⠦",,terminal_output
129
+ 129,20191,"TERMINAL",0,0,"⠧",,terminal_output
130
+ 130,20316,"TERMINAL",0,0,"⠇",,terminal_output
131
+ 131,20398,"TERMINAL",0,0,"⠏",,terminal_output
132
+ 132,20478,"TERMINAL",0,0,"⠋",,terminal_output
133
+ 133,20559,"TERMINAL",0,0,"⠙",,terminal_output
134
+ 134,20637,"TERMINAL",0,0,"⠹",,terminal_output
135
+ 135,20718,"TERMINAL",0,0,"⠸",,terminal_output
136
+ 136,20800,"TERMINAL",0,0,"⠼",,terminal_output
137
+ 137,20881,"TERMINAL",0,0,"⠴",,terminal_output
138
+ 138,20959,"TERMINAL",0,0,"⠦",,terminal_output
139
+ 139,21044,"TERMINAL",0,0,"⠧",,terminal_output
140
+ 140,21124,"TERMINAL",0,0,"⠇",,terminal_output
141
+ 141,21203,"TERMINAL",0,0,"⠏",,terminal_output
142
+ 142,21287,"TERMINAL",0,0,"⠋",,terminal_output
143
+ 143,21368,"TERMINAL",0,0,"⠙",,terminal_output
144
+ 144,21449,"TERMINAL",0,0,"⠹",,terminal_output
145
+ 145,21531,"TERMINAL",0,0,"⠸",,terminal_output
146
+ 146,21612,"TERMINAL",0,0,"⠼",,terminal_output
147
+ 147,21696,"TERMINAL",0,0,"⠴",,terminal_output
148
+ 148,21773,"TERMINAL",0,0,"⠦",,terminal_output
149
+ 149,21852,"TERMINAL",0,0,"⠧",,terminal_output
150
+ 150,21935,"TERMINAL",0,0,"⠇",,terminal_output
151
+ 151,22016,"TERMINAL",0,0,"⠏",,terminal_output
152
+ 152,22095,"TERMINAL",0,0,"⠋",,terminal_output
153
+ 153,22176,"TERMINAL",0,0,"⠙",,terminal_output
154
+ 154,22261,"TERMINAL",0,0,"⠹",,terminal_output
155
+ 155,22338,"TERMINAL",0,0,"⠸",,terminal_output
156
+ 156,22419,"TERMINAL",0,0,"⠼",,terminal_output
157
+ 157,22498,"TERMINAL",0,0,"⠴",,terminal_output
158
+ 158,22579,"TERMINAL",0,0,"⠦",,terminal_output
159
+ 159,22663,"TERMINAL",0,0,"⠧",,terminal_output
160
+ 160,22748,"TERMINAL",0,0,"⠇",,terminal_output
161
+ 161,22827,"TERMINAL",0,0,"⠏",,terminal_output
162
+ 162,22909,"TERMINAL",0,0,"⠋",,terminal_output
163
+ 163,22989,"TERMINAL",0,0,"⠙",,terminal_output
164
+ 164,23070,"TERMINAL",0,0,"⠹",,terminal_output
165
+ 165,23152,"TERMINAL",0,0,"⠸",,terminal_output
166
+ 166,23232,"TERMINAL",0,0,"⠼",,terminal_output
167
+ 167,23336,"TERMINAL",0,0,"⠴",,terminal_output
168
+ 168,23417,"TERMINAL",0,0,"⠦",,terminal_output
169
+ 169,23513,"TERMINAL",0,0,"⠧",,terminal_output
170
+ 170,23594,"TERMINAL",0,0,"⠇",,terminal_output
171
+ 171,23677,"TERMINAL",0,0,"⠏",,terminal_output
172
+ 172,23757,"TERMINAL",0,0,"⠋",,terminal_output
173
+ 173,23837,"TERMINAL",0,0,"⠙",,terminal_output
174
+ 174,23917,"TERMINAL",0,0,"⠹",,terminal_output
175
+ 175,23998,"TERMINAL",0,0,"⠸",,terminal_output
176
+ 176,24078,"TERMINAL",0,0,"⠼",,terminal_output
177
+ 177,24193,"TERMINAL",0,0,"⠴",,terminal_output
178
+ 178,24289,"TERMINAL",0,0,"⠦",,terminal_output
179
+ 179,24348,"TERMINAL",0,0,"⠧",,terminal_output
180
+ 180,24428,"TERMINAL",0,0,"⠇",,terminal_output
181
+ 181,24511,"TERMINAL",0,0,"⠏",,terminal_output
182
+ 182,24593,"TERMINAL",0,0,"⠋",,terminal_output
183
+ 183,24675,"TERMINAL",0,0,"⠙",,terminal_output
184
+ 184,24797,"TERMINAL",0,0,"⠹",,terminal_output
185
+ 185,24873,"TERMINAL",0,0,"⠸",,terminal_output
186
+ 186,24949,"TERMINAL",0,0,"⠼",,terminal_output
187
+ 187,25031,"TERMINAL",0,0,"⠴",,terminal_output
188
+ 188,25112,"TERMINAL",0,0,"⠦",,terminal_output
189
+ 189,25194,"TERMINAL",0,0,"⠧",,terminal_output
190
+ 190,25274,"TERMINAL",0,0,"⠇",,terminal_output
191
+ 191,25355,"TERMINAL",0,0,"⠏",,terminal_output
192
+ 192,25434,"TERMINAL",0,0,"⠋",,terminal_output
193
+ 193,25515,"TERMINAL",0,0,"⠙",,terminal_output
194
+ 194,25596,"TERMINAL",0,0,"⠹",,terminal_output
195
+ 195,25677,"TERMINAL",0,0,"⠸",,terminal_output
196
+ 196,25758,"TERMINAL",0,0,"⠼",,terminal_output
197
+ 197,25838,"TERMINAL",0,0,"⠴",,terminal_output
198
+ 198,25919,"TERMINAL",0,0,"⠦",,terminal_output
199
+ 199,25998,"TERMINAL",0,0,"⠧",,terminal_output
200
+ 200,26079,"TERMINAL",0,0,"⠇",,terminal_output
201
+ 201,26161,"TERMINAL",0,0,"⠏",,terminal_output
202
+ 202,26244,"TERMINAL",0,0,"⠋",,terminal_output
203
+ 203,26323,"TERMINAL",0,0,"⠙",,terminal_output
204
+ 204,26404,"TERMINAL",0,0,"⠹",,terminal_output
205
+ 205,26485,"TERMINAL",0,0,"⠸",,terminal_output
206
+ 206,26580,"TERMINAL",0,0,"⠼",,terminal_output
207
+ 207,26648,"TERMINAL",0,0,"⠴",,terminal_output
208
+ 208,26741,"TERMINAL",0,0,"⠦",,terminal_output
209
+ 209,26820,"TERMINAL",0,0,"⠧",,terminal_output
210
+ 210,26909,"TERMINAL",0,0,"⠇",,terminal_output
211
+ 211,26988,"TERMINAL",0,0,"⠏",,terminal_output
212
+ 212,27071,"TERMINAL",0,0,"⠋",,terminal_output
213
+ 213,27150,"TERMINAL",0,0,"⠙",,terminal_output
214
+ 214,27256,"TERMINAL",0,0,"⠹",,terminal_output
215
+ 215,27367,"TERMINAL",0,0,"⠸",,terminal_output
216
+ 216,27440,"TERMINAL",0,0,"⠼",,terminal_output
217
+ 217,27520,"TERMINAL",0,0,"⠴",,terminal_output
218
+ 218,27604,"TERMINAL",0,0,"⠦",,terminal_output
219
+ 219,27686,"TERMINAL",0,0,"⠧",,terminal_output
220
+ 220,27775,"TERMINAL",0,0,"⠇",,terminal_output
221
+ 221,27889,"TERMINAL",0,0,"⠏",,terminal_output
222
+ 222,27946,"TERMINAL",0,0,"⠋",,terminal_output
223
+ 223,28029,"TERMINAL",0,0,"⠙",,terminal_output
224
+ 224,28108,"TERMINAL",0,0,"⠹",,terminal_output
225
+ 225,28188,"TERMINAL",0,0,"⠸",,terminal_output
226
+ 226,28275,"TERMINAL",0,0,"⠼",,terminal_output
227
+ 227,28364,"TERMINAL",0,0,"⠴",,terminal_output
228
+ 228,28455,"TERMINAL",0,0,"⠦",,terminal_output
229
+ 229,28535,"TERMINAL",0,0,"⠧",,terminal_output
230
+ 230,28615,"TERMINAL",0,0,"⠇",,terminal_output
231
+ 231,28696,"TERMINAL",0,0,"⠏",,terminal_output
232
+ 232,28785,"TERMINAL",0,0,"⠋",,terminal_output
233
+ 233,28864,"TERMINAL",0,0,"⠙",,terminal_output
234
+ 234,28947,"TERMINAL",0,0,"⠹",,terminal_output
235
+ 235,29029,"TERMINAL",0,0,"⠸",,terminal_output
236
+ 236,29113,"TERMINAL",0,0,"⠼",,terminal_output
237
+ 237,29194,"TERMINAL",0,0,"⠴",,terminal_output
238
+ 238,29278,"TERMINAL",0,0,"⠦",,terminal_output
239
+ 239,29354,"TERMINAL",0,0,"⠧",,terminal_output
240
+ 240,29435,"TERMINAL",0,0,"⠇",,terminal_output
241
+ 241,29531,"TERMINAL",0,0,"⠏",,terminal_output
242
+ 242,29610,"TERMINAL",0,0,"⠋",,terminal_output
243
+ 243,29703,"TERMINAL",0,0,"⠙",,terminal_output
244
+ 244,29786,"TERMINAL",0,0,"⠹",,terminal_output
245
+ 245,29946,"TERMINAL",0,0,"⠸",,terminal_output
246
+ 246,30035,"TERMINAL",0,0,"⠼",,terminal_output
247
+ 247,30256,"TERMINAL",0,0,"⠴",,terminal_output
248
+ 248,30487,"TERMINAL",0,0,"⠦",,terminal_output
249
+ 249,30589,"TERMINAL",0,0,"⠧",,terminal_output
250
+ 250,30705,"TERMINAL",0,0,"⠇",,terminal_output
251
+ 251,30817,"TERMINAL",0,0,"⠏",,terminal_output
252
+ 252,30943,"TERMINAL",0,0,"⠋",,terminal_output
253
+ 253,31591,"TERMINAL",0,0,"⠙",,terminal_output
254
+ 254,31756,"TERMINAL",0,0,"⠹",,terminal_output
255
+ 255,31867,"TERMINAL",0,0,"⠸",,terminal_output
256
+ 256,31932,"TERMINAL",0,0,"npm warn deprecated boolean@3.2.0: Package no longer supported. Contact Support at https://www.npmjs.com/support for more info.\r\n⠸",,terminal_output
257
+ 257,32016,"TERMINAL",0,0,"⠼",,terminal_output
258
+ 258,32107,"TERMINAL",0,0,"⠴",,terminal_output
259
+ 259,32189,"TERMINAL",0,0,"⠦",,terminal_output
260
+ 260,32276,"TERMINAL",0,0,"⠧",,terminal_output
261
+ 261,32389,"TERMINAL",0,0,"⠇",,terminal_output
262
+ 262,32535,"TERMINAL",0,0,"⠏",,terminal_output
263
+ 263,32620,"TERMINAL",0,0,"⠋",,terminal_output
264
+ 264,32734,"TERMINAL",0,0,"npm warn deprecated npmlog@2.0.4: This package is no longer supported.\r\n⠋npm warn deprecated are-we-there-yet@1.1.7: This package is no longer supported.\r\n⠙npm warn deprecated gauge@1.2.7: This package is no longer supported.\r\n⠙",,terminal_output
265
+ 265,32815,"TERMINAL",0,0,"⠹",,terminal_output
266
+ 266,32896,"TERMINAL",0,0,"⠸",,terminal_output
267
+ 267,32980,"TERMINAL",0,0,"⠼",,terminal_output
268
+ 268,33057,"TERMINAL",0,0,"⠴",,terminal_output
269
+ 269,33138,"TERMINAL",0,0,"⠦",,terminal_output
270
+ 270,33221,"TERMINAL",0,0,"⠧",,terminal_output
271
+ 271,33301,"TERMINAL",0,0,"⠇",,terminal_output
272
+ 272,33382,"TERMINAL",0,0,"⠏",,terminal_output
273
+ 273,33464,"TERMINAL",0,0,"⠋",,terminal_output
274
+ 274,33546,"TERMINAL",0,0,"⠙",,terminal_output
275
+ 275,33628,"TERMINAL",0,0,"⠹",,terminal_output
276
+ 276,33710,"TERMINAL",0,0,"⠸",,terminal_output
277
+ 277,33790,"TERMINAL",0,0,"⠼",,terminal_output
278
+ 278,33877,"TERMINAL",0,0,"⠴",,terminal_output
279
+ 279,33966,"TERMINAL",0,0,"⠦",,terminal_output
280
+ 280,34046,"TERMINAL",0,0,"⠧",,terminal_output
281
+ 281,34129,"TERMINAL",0,0,"⠇",,terminal_output
282
+ 282,34208,"TERMINAL",0,0,"⠏",,terminal_output
283
+ 283,34295,"TERMINAL",0,0,"⠋",,terminal_output
284
+ 284,34381,"TERMINAL",0,0,"⠙",,terminal_output
285
+ 285,34463,"TERMINAL",0,0,"⠹",,terminal_output
286
+ 286,34544,"TERMINAL",0,0,"⠸",,terminal_output
287
+ 287,34623,"TERMINAL",0,0,"⠼",,terminal_output
288
+ 288,34704,"TERMINAL",0,0,"⠴",,terminal_output
289
+ 289,34782,"TERMINAL",0,0,"⠦",,terminal_output
290
+ 290,34861,"TERMINAL",0,0,"⠧",,terminal_output
291
+ 291,34943,"TERMINAL",0,0,"⠇",,terminal_output
292
+ 292,35023,"TERMINAL",0,0,"⠏",,terminal_output
293
+ 293,35104,"TERMINAL",0,0,"⠋",,terminal_output
294
+ 294,35184,"TERMINAL",0,0,"⠙",,terminal_output
295
+ 295,35262,"TERMINAL",0,0,"⠹",,terminal_output
296
+ 296,35344,"TERMINAL",0,0,"⠸",,terminal_output
297
+ 297,35423,"TERMINAL",0,0,"⠼",,terminal_output
298
+ 298,35503,"TERMINAL",0,0,"⠴",,terminal_output
299
+ 299,35583,"TERMINAL",0,0,"⠦",,terminal_output
300
+ 300,35665,"TERMINAL",0,0,"⠧",,terminal_output
301
+ 301,35743,"TERMINAL",0,0,"⠇",,terminal_output
302
+ 302,35823,"TERMINAL",0,0,"⠏",,terminal_output
303
+ 303,35904,"TERMINAL",0,0,"⠋",,terminal_output
304
+ 304,35985,"TERMINAL",0,0,"⠙",,terminal_output
305
+ 305,36065,"TERMINAL",0,0,"⠹",,terminal_output
306
+ 306,36147,"TERMINAL",0,0,"⠸",,terminal_output
307
+ 307,36229,"TERMINAL",0,0,"⠼",,terminal_output
308
+ 308,36310,"TERMINAL",0,0,"⠴",,terminal_output
309
+ 309,36390,"TERMINAL",0,0,"⠦",,terminal_output
310
+ 310,36471,"TERMINAL",0,0,"⠧",,terminal_output
311
+ 311,36551,"TERMINAL",0,0,"⠇",,terminal_output
312
+ 312,36632,"TERMINAL",0,0,"⠏",,terminal_output
313
+ 313,36713,"TERMINAL",0,0,"⠋",,terminal_output
314
+ 314,36828,"TERMINAL",0,0,"⠙",,terminal_output
315
+ 315,36908,"TERMINAL",0,0,"⠹",,terminal_output
316
+ 316,36989,"TERMINAL",0,0,"⠸",,terminal_output
317
+ 317,37071,"TERMINAL",0,0,"⠼",,terminal_output
318
+ 318,37152,"TERMINAL",0,0,"⠴",,terminal_output
319
+ 319,37323,"TERMINAL",0,0,"⠦",,terminal_output
320
+ 320,37402,"TERMINAL",0,0,"",,terminal_output
321
+ 321,39009,"TERMINAL",0,0,"\r\n _-----_ ╭──────────────────────────╮\r\n | | │ Welcome to the Visual │\r\n |--(o)--| │ Studio Code Extension │\r\n `---------´ │ generator! │\r\n ( _´U`_ ) ╰──────────────────────────╯\r\n /___A___\ /\r\n | ~ | \r\n __'.___.'__ \r\n ´ ` |° ´ Y ` \r\n\r\n",,terminal_output
322
+ 322,39520,"TERMINAL",0,0,"? What type of extension do you want to create? (Use arrow keys)\r\n❯ New Extension (TypeScript) \r\n New Extension (JavaScript) \r\n New Color Theme \r\n New Language Support \r\n New Code Snippets \r\n New Keymap \r\n New Extension Pack \r\n New Language Pack (Localization) \r\n New Web Extension (TypeScript) \r\n New Notebook Renderer (TypeScript) [?25l",,terminal_output
323
+ 323,41704,"TERMINAL",0,0,"? What type of extension do you want to create? New Extension (TypeScript)[?25l\r\n[?25h? What's the name of your extension? () ",,terminal_output
324
+ 324,42919,"TERMINAL",0,0,"? What's the name of your extension? c",,terminal_output
325
+ 325,43103,"TERMINAL",0,0,"? What's the name of your extension? cr",,terminal_output
326
+ 326,43219,"TERMINAL",0,0,"? What's the name of your extension? cro",,terminal_output
327
+ 327,43319,"TERMINAL",0,0,"? What's the name of your extension? crow",,terminal_output
328
+ 328,43480,"TERMINAL",0,0,"? What's the name of your extension? crowd",,terminal_output
329
+ 329,44319,"TERMINAL",0,0,"? What's the name of your extension? crowd-",,terminal_output
330
+ 330,44570,"TERMINAL",0,0,"? What's the name of your extension? crowd-p",,terminal_output
331
+ 331,44651,"TERMINAL",0,0,"? What's the name of your extension? crowd-pi",,terminal_output
332
+ 332,44904,"TERMINAL",0,0,"? What's the name of your extension? crowd-pil",,terminal_output
333
+ 333,44996,"TERMINAL",0,0,"? What's the name of your extension? crowd-pilo",,terminal_output
334
+ 334,45086,"TERMINAL",0,0,"? What's the name of your extension? crowd-pilot",,terminal_output
335
+ 335,45370,"TERMINAL",0,0,"? What's the name of your extension? crowd-pilot-",,terminal_output
336
+ 336,45460,"TERMINAL",0,0,"? What's the name of your extension? crowd-pilot-e",,terminal_output
337
+ 337,45747,"TERMINAL",0,0,"? What's the name of your extension? crowd-pilot-ex",,terminal_output
338
+ 338,45850,"TERMINAL",0,0,"? What's the name of your extension? crowd-pilot-ext",,terminal_output
339
+ 339,45904,"TERMINAL",0,0,"? What's the name of your extension? crowd-pilot-exte",,terminal_output
340
+ 340,46014,"TERMINAL",0,0,"? What's the name of your extension? crowd-pilot-exten",,terminal_output
341
+ 341,46316,"TERMINAL",0,0,"? What's the name of your extension? crowd-pilot-extens? What's the name of your extension? crowd-pilot-extensi? What's the name of your extension? crowd-pilot-extensio? What's the name of your extension? crowd-pilot-extension",,terminal_output
342
+ 342,46568,"TERMINAL",0,0,"? What's the name of your extension? crowd-pilot-extension\r\n[?25h? What's the identifier of your extension? (crowd-pilot-extension) ",,terminal_output
343
+ 343,47595,"TERMINAL",0,0,"? What's the identifier of your extension? c",,terminal_output
344
+ 344,47768,"TERMINAL",0,0,"? What's the identifier of your extension? cr",,terminal_output
345
+ 345,47944,"TERMINAL",0,0,"? What's the identifier of your extension? cro",,terminal_output
346
+ 346,48043,"TERMINAL",0,0,"? What's the identifier of your extension? crow",,terminal_output
347
+ 347,48192,"TERMINAL",0,0,"? What's the identifier of your extension? crowd",,terminal_output
348
+ 348,48393,"TERMINAL",0,0,"? What's the identifier of your extension? crowd-",,terminal_output
349
+ 349,48577,"TERMINAL",0,0,"? What's the identifier of your extension? crowd-p",,terminal_output
350
+ 350,48659,"TERMINAL",0,0,"? What's the identifier of your extension? crowd-pi",,terminal_output
351
+ 351,49008,"TERMINAL",0,0,"? What's the identifier of your extension? crowd-pil? What's the identifier of your extension? crowd-pilo",,terminal_output
352
+ 352,49130,"TERMINAL",0,0,"? What's the identifier of your extension? crowd-pilot",,terminal_output
353
+ 353,49294,"TERMINAL",0,0,"? What's the identifier of your extension? crowd-pilot\r\n[?25h? What's the description of your extension? () ",,terminal_output
354
+ 354,55046,"TERMINAL",0,0,"? What's the description of your extension? T",,terminal_output
355
+ 355,55407,"TERMINAL",0,0,"? What's the description of your extension? Te",,terminal_output
356
+ 356,55610,"TERMINAL",0,0,"? What's the description of your extension? Tea",,terminal_output
357
+ 357,55674,"TERMINAL",0,0,"? What's the description of your extension? Teac",,terminal_output
358
+ 358,55787,"TERMINAL",0,0,"? What's the description of your extension? Teach",,terminal_output
359
+ 359,55930,"TERMINAL",0,0,"? What's the description of your extension? Teachi",,terminal_output
360
+ 360,56047,"TERMINAL",0,0,"? What's the description of your extension? Teachin? What's the description of your extension? Teaching",,terminal_output
361
+ 361,56218,"TERMINAL",0,0,"? What's the description of your extension? Teaching ",,terminal_output
362
+ 362,56317,"TERMINAL",0,0,"? What's the description of your extension? Teaching l",,terminal_output
363
+ 363,56401,"TERMINAL",0,0,"? What's the description of your extension? Teaching la? What's the description of your extension? Teaching lan",,terminal_output
364
+ 364,56502,"TERMINAL",0,0,"? What's the description of your extension? Teaching lang",,terminal_output
365
+ 365,56587,"TERMINAL",0,0,"? What's the description of your extension? Teaching langu",,terminal_output
366
+ 366,56787,"TERMINAL",0,0,"? What's the description of your extension? Teaching langua? What's the description of your extension? Teaching languag? What's the description of your extension? Teaching language",,terminal_output
367
+ 367,56997,"TERMINAL",0,0,"? What's the description of your extension? Teaching language ",,terminal_output
368
+ 368,57066,"TERMINAL",0,0,"? What's the description of your extension? Teaching language m",,terminal_output
369
+ 369,57150,"TERMINAL",0,0,"? What's the description of your extension? Teaching language mo",,terminal_output
370
+ 370,57320,"TERMINAL",0,0,"? What's the description of your extension? Teaching language mod",,terminal_output
371
+ 371,57396,"TERMINAL",0,0,"? What's the description of your extension? Teaching language mode",,terminal_output
372
+ 372,57582,"TERMINAL",0,0,"? What's the description of your extension? Teaching language model",,terminal_output
373
+ 373,57734,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models",,terminal_output
374
+ 374,57785,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models ",,terminal_output
375
+ 375,57864,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models t? What's the description of your extension? Teaching language models tr",,terminal_output
376
+ 376,57935,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models tro",,terminal_output
377
+ 377,58119,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models tro ",,terminal_output
378
+ 378,58211,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models tro c",,terminal_output
379
+ 379,58298,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models tro co",,terminal_output
380
+ 380,58837,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models tro ",,terminal_output
381
+ 381,59004,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models ",,terminal_output
382
+ 382,59360,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models t",,terminal_output
383
+ 383,59475,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to",,terminal_output
384
+ 384,59717,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to ? What's the description of your extension? Teaching language models to c",,terminal_output
385
+ 385,59886,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to co",,terminal_output
386
+ 386,59966,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to cod",,terminal_output
387
+ 387,60158,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to codd",,terminal_output
388
+ 388,60245,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to codde",,terminal_output
389
+ 389,60443,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to codde ",,terminal_output
390
+ 390,60601,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to codde l? What's the description of your extension? Teaching language models to codde li",,terminal_output
391
+ 391,60881,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to codde l",,terminal_output
392
+ 392,60994,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to codde ",,terminal_output
393
+ 393,61135,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to codde",,terminal_output
394
+ 394,61299,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to codd",,terminal_output
395
+ 395,61462,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to cod",,terminal_output
396
+ 396,61550,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to cod ",,terminal_output
397
+ 397,61714,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to cod e",,terminal_output
398
+ 398,61987,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to cod ",,terminal_output
399
+ 399,62146,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to cod? What's the description of your extension? Teaching language models to code",,terminal_output
400
+ 400,62301,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to code ",,terminal_output
401
+ 401,62401,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to code l? What's the description of your extension? Teaching language models to code li",,terminal_output
402
+ 402,62601,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to code lik? What's the description of your extension? Teaching language models to code like",,terminal_output
403
+ 403,62763,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to code like ? What's the description of your extension? Teaching language models to code like h",,terminal_output
404
+ 404,62935,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to code like hu",,terminal_output
405
+ 405,63093,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to code like hum",,terminal_output
406
+ 406,63253,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to code like huma? What's the description of your extension? Teaching language models to code like human",,terminal_output
407
+ 407,63453,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to code like humans",,terminal_output
408
+ 408,63535,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to code like humans.",,terminal_output
409
+ 409,64002,"TERMINAL",0,0,"? What's the description of your extension? Teaching language models to code like humans.\r\n[?25h? Initialize a git repository? (Y/n) ",,terminal_output
410
+ 410,65236,"TERMINAL",0,0,"? Initialize a git repository? (Y/n) n",,terminal_output
411
+ 411,66247,"TERMINAL",0,0,"? Initialize a git repository? No\r\n[?25h? Which bundler to use? (Use arrow keys)\r\n❯ unbundled \r\n webpack \r\n esbuild [?25l",,terminal_output
412
+ 412,67286,"TERMINAL",0,0,"? Which bundler to use? unbundled[?25l\r\n[?25h? Which package manager to use? (Use arrow keys)\r\n❯ npm \r\n yarn \r\n pnpm [?25l",,terminal_output
413
+ 413,68168,"TERMINAL",0,0,"? Which package manager to use? npm[?25l\r\n[?25h\r\nWriting in /Users/franzsrambical/Documents/pdoom/crowd-pilot/crowd-pilot...\r\n create crowd-pilot/.vscode/extensions.json\r\n create crowd-pilot/.vscode/launch.json\r\n create crowd-pilot/.vscode/settings.json\r\n create crowd-pilot/.vscode/tasks.json\r\n create crowd-pilot/package.json\r\n create crowd-pilot/tsconfig.json\r\n create crowd-pilot/.vscodeignore\r\n create crowd-pilot/vsc-extension-quickstart.md\r\n create crowd-pilot/README.md\r\n create crowd-pilot/CHANGELOG.md\r\n create crowd-pilot/src/extension.ts\r\n create crowd-pilot/src/test/extension.test.ts\r\n create crowd-pilot/.vscode-test.mjs\r\n create crowd-pilot/eslint.config.mjs\r\n\r\nChanges to package.json were detected.\r\n\r\nRunning npm install for you to install the required dependencies.\r\n",,terminal_output
414
+ 414,68620,"TERMINAL",0,0,"⠙",,terminal_output
415
+ 415,68703,"TERMINAL",0,0,"⠹",,terminal_output
416
+ 416,68785,"TERMINAL",0,0,"⠸",,terminal_output
417
+ 417,68865,"TERMINAL",0,0,"⠼",,terminal_output
418
+ 418,68948,"TERMINAL",0,0,"⠴",,terminal_output
419
+ 419,69099,"TERMINAL",0,0,"⠦",,terminal_output
420
+ 420,69180,"TERMINAL",0,0,"⠧",,terminal_output
421
+ 421,69263,"TERMINAL",0,0,"⠇",,terminal_output
422
+ 422,69353,"TERMINAL",0,0,"⠏",,terminal_output
423
+ 423,69432,"TERMINAL",0,0,"⠋",,terminal_output
424
+ 424,69524,"TERMINAL",0,0,"⠙",,terminal_output
425
+ 425,69589,"TERMINAL",0,0,"⠹",,terminal_output
426
+ 426,69699,"TERMINAL",0,0,"⠸",,terminal_output
427
+ 427,69777,"TERMINAL",0,0,"⠼",,terminal_output
428
+ 428,69856,"TERMINAL",0,0,"⠴",,terminal_output
429
+ 429,69937,"TERMINAL",0,0,"⠦",,terminal_output
430
+ 430,70018,"TERMINAL",0,0,"⠧",,terminal_output
431
+ 431,70195,"TERMINAL",0,0,"⠇",,terminal_output
432
+ 432,70274,"TERMINAL",0,0,"⠏",,terminal_output
433
+ 433,70355,"TERMINAL",0,0,"⠋",,terminal_output
434
+ 434,70437,"TERMINAL",0,0,"⠙",,terminal_output
435
+ 435,70514,"TERMINAL",0,0,"⠹",,terminal_output
436
+ 436,70595,"TERMINAL",0,0,"⠸",,terminal_output
437
+ 437,70677,"TERMINAL",0,0,"⠼",,terminal_output
438
+ 438,70756,"TERMINAL",0,0,"⠴",,terminal_output
439
+ 439,70836,"TERMINAL",0,0,"⠦",,terminal_output
440
+ 440,70970,"TERMINAL",0,0,"⠧",,terminal_output
441
+ 441,71049,"TERMINAL",0,0,"⠇",,terminal_output
442
+ 442,71131,"TERMINAL",0,0,"⠏",,terminal_output
443
+ 443,71212,"TERMINAL",0,0,"⠋",,terminal_output
444
+ 444,71291,"TERMINAL",0,0,"⠙",,terminal_output
445
+ 445,71372,"TERMINAL",0,0,"⠹",,terminal_output
446
+ 446,71452,"TERMINAL",0,0,"⠸",,terminal_output
447
+ 447,71531,"TERMINAL",0,0,"⠼",,terminal_output
448
+ 448,71655,"TERMINAL",0,0,"⠴",,terminal_output
449
+ 449,71737,"TERMINAL",0,0,"⠦",,terminal_output
450
+ 450,71818,"TERMINAL",0,0,"⠧",,terminal_output
451
+ 451,71901,"TERMINAL",0,0,"⠇",,terminal_output
452
+ 452,71986,"TERMINAL",0,0,"⠏",,terminal_output
453
+ 453,72061,"TERMINAL",0,0,"⠋",,terminal_output
454
+ 454,72142,"TERMINAL",0,0,"⠙",,terminal_output
455
+ 455,72222,"TERMINAL",0,0,"⠹",,terminal_output
456
+ 456,72302,"TERMINAL",0,0,"⠸",,terminal_output
457
+ 457,72385,"TERMINAL",0,0,"⠼",,terminal_output
458
+ 458,72468,"TERMINAL",0,0,"⠴",,terminal_output
459
+ 459,72548,"TERMINAL",0,0,"⠦",,terminal_output
460
+ 460,72628,"TERMINAL",0,0,"⠧",,terminal_output
461
+ 461,72710,"TERMINAL",0,0,"⠇",,terminal_output
462
+ 462,72794,"TERMINAL",0,0,"⠏",,terminal_output
463
+ 463,72878,"TERMINAL",0,0,"⠋",,terminal_output
464
+ 464,72960,"TERMINAL",0,0,"⠙",,terminal_output
465
+ 465,73041,"TERMINAL",0,0,"⠹",,terminal_output
466
+ 466,73122,"TERMINAL",0,0,"⠸",,terminal_output
467
+ 467,73205,"TERMINAL",0,0,"⠼",,terminal_output
468
+ 468,73280,"TERMINAL",0,0,"⠴",,terminal_output
469
+ 469,73398,"TERMINAL",0,0,"⠦",,terminal_output
470
+ 470,73467,"TERMINAL",0,0,"⠧",,terminal_output
471
+ 471,73546,"TERMINAL",0,0,"⠇",,terminal_output
472
+ 472,73626,"TERMINAL",0,0,"⠏",,terminal_output
473
+ 473,73711,"TERMINAL",0,0,"⠋",,terminal_output
474
+ 474,73789,"TERMINAL",0,0,"⠙",,terminal_output
475
+ 475,73869,"TERMINAL",0,0,"⠹",,terminal_output
476
+ 476,73950,"TERMINAL",0,0,"⠸",,terminal_output
477
+ 477,74032,"TERMINAL",0,0,"⠼",,terminal_output
478
+ 478,74109,"TERMINAL",0,0,"⠴",,terminal_output
479
+ 479,74190,"TERMINAL",0,0,"⠦",,terminal_output
480
+ 480,74270,"TERMINAL",0,0,"⠧",,terminal_output
481
+ 481,74350,"TERMINAL",0,0,"⠇",,terminal_output
482
+ 482,74431,"TERMINAL",0,0,"⠏",,terminal_output
483
+ 483,74511,"TERMINAL",0,0,"⠋",,terminal_output
484
+ 484,74591,"TERMINAL",0,0,"⠙",,terminal_output
485
+ 485,74673,"TERMINAL",0,0,"⠹",,terminal_output
486
+ 486,74752,"TERMINAL",0,0,"⠸",,terminal_output
487
+ 487,74835,"TERMINAL",0,0,"⠼",,terminal_output
488
+ 488,74966,"TERMINAL",0,0,"⠴",,terminal_output
489
+ 489,75045,"TERMINAL",0,0,"⠦",,terminal_output
490
+ 490,75126,"TERMINAL",0,0,"⠧",,terminal_output
491
+ 491,75205,"TERMINAL",0,0,"⠇",,terminal_output
492
+ 492,75288,"TERMINAL",0,0,"⠏",,terminal_output
493
+ 493,75374,"TERMINAL",0,0,"⠋",,terminal_output
494
+ 494,75454,"TERMINAL",0,0,"⠙",,terminal_output
495
+ 495,75534,"TERMINAL",0,0,"⠹",,terminal_output
496
+ 496,75613,"TERMINAL",0,0,"⠸",,terminal_output
497
+ 497,75694,"TERMINAL",0,0,"⠼",,terminal_output
498
+ 498,75775,"TERMINAL",0,0,"⠴",,terminal_output
499
+ 499,75854,"TERMINAL",0,0,"⠦",,terminal_output
500
+ 500,75935,"TERMINAL",0,0,"⠧",,terminal_output
501
+ 501,76015,"TERMINAL",0,0,"⠇",,terminal_output
502
+ 502,76111,"TERMINAL",0,0,"⠏",,terminal_output
503
+ 503,76193,"TERMINAL",0,0,"⠋",,terminal_output
504
+ 504,76273,"TERMINAL",0,0,"⠙",,terminal_output
505
+ 505,76354,"TERMINAL",0,0,"⠹",,terminal_output
506
+ 506,76437,"TERMINAL",0,0,"⠸",,terminal_output
507
+ 507,76513,"TERMINAL",0,0,"⠼",,terminal_output
508
+ 508,76653,"TERMINAL",0,0,"⠴",,terminal_output
509
+ 509,76720,"TERMINAL",0,0,"⠦",,terminal_output
510
+ 510,76800,"TERMINAL",0,0,"⠧",,terminal_output
511
+ 511,76878,"TERMINAL",0,0,"⠇",,terminal_output
512
+ 512,76959,"TERMINAL",0,0,"⠏",,terminal_output
513
+ 513,77039,"TERMINAL",0,0,"⠋",,terminal_output
514
+ 514,77118,"TERMINAL",0,0,"⠙",,terminal_output
515
+ 515,77199,"TERMINAL",0,0,"⠹",,terminal_output
516
+ 516,77279,"TERMINAL",0,0,"⠸",,terminal_output
517
+ 517,77358,"TERMINAL",0,0,"⠼",,terminal_output
518
+ 518,77442,"TERMINAL",0,0,"⠴",,terminal_output
519
+ 519,77521,"TERMINAL",0,0,"⠦",,terminal_output
520
+ 520,77603,"TERMINAL",0,0,"⠧",,terminal_output
521
+ 521,77683,"TERMINAL",0,0,"⠇",,terminal_output
522
+ 522,77764,"TERMINAL",0,0,"⠏",,terminal_output
523
+ 523,77845,"TERMINAL",0,0,"⠋",,terminal_output
524
+ 524,77926,"TERMINAL",0,0,"⠙",,terminal_output
525
+ 525,78007,"TERMINAL",0,0,"⠹",,terminal_output
526
+ 526,78089,"TERMINAL",0,0,"⠸",,terminal_output
527
+ 527,78169,"TERMINAL",0,0,"⠼",,terminal_output
528
+ 528,78252,"TERMINAL",0,0,"⠴",,terminal_output
529
+ 529,78581,"TERMINAL",0,0,"⠦",,terminal_output
530
+ 530,78684,"TERMINAL",0,0,"⠧",,terminal_output
531
+ 531,78754,"TERMINAL",0,0,"⠇",,terminal_output
532
+ 532,78900,"TERMINAL",0,0,"⠏",,terminal_output
533
+ 533,78940,"TERMINAL",0,0,"npm warn deprecated inflight@1.0.6: This module is not supported, and leaks memory. Do not use it. Check out lru-cache if you want a good and tested way to coalesce async requests by a key value, which is much more comprehensive and powerful.\r\n⠋",,terminal_output
534
+ 534,79019,"TERMINAL",0,0,"⠙",,terminal_output
535
+ 535,79140,"TERMINAL",0,0,"⠹npm warn deprecated glob@7.2.3: Glob versions prior to v9 are no longer supported\r\n⠹",,terminal_output
536
+ 536,79216,"TERMINAL",0,0,"⠸",,terminal_output
537
+ 537,79298,"TERMINAL",0,0,"⠼",,terminal_output
538
+ 538,79420,"TERMINAL",0,0,"⠴",,terminal_output
539
+ 539,79487,"TERMINAL",0,0,"⠦",,terminal_output
540
+ 540,79570,"TERMINAL",0,0,"⠧",,terminal_output
541
+ 541,79761,"TERMINAL",0,0,"⠇⠏",,terminal_output
542
+ 542,79856,"TERMINAL",0,0,"⠋",,terminal_output
543
+ 543,79928,"TERMINAL",0,0,"⠙",,terminal_output
544
+ 544,80025,"TERMINAL",0,0,"⠹",,terminal_output
545
+ 545,80088,"TERMINAL",0,0,"⠸",,terminal_output
546
+ 546,80169,"TERMINAL",0,0,"⠼",,terminal_output
547
+ 547,80249,"TERMINAL",0,0,"⠴",,terminal_output
548
+ 548,80330,"TERMINAL",0,0,"⠦",,terminal_output
549
+ 549,80429,"TERMINAL",0,0,"⠧",,terminal_output
550
+ 550,80495,"TERMINAL",0,0,"⠇",,terminal_output
551
+ 551,80571,"TERMINAL",0,0,"⠏",,terminal_output
552
+ 552,80656,"TERMINAL",0,0,"⠋",,terminal_output
553
+ 553,80735,"TERMINAL",0,0,"⠙",,terminal_output
554
+ 554,80842,"TERMINAL",0,0,"⠹",,terminal_output
555
+ 555,80942,"TERMINAL",0,0,"⠸",,terminal_output
556
+ 556,80999,"TERMINAL",0,0,"⠼",,terminal_output
557
+ 557,81081,"TERMINAL",0,0,"⠴",,terminal_output
558
+ 558,81163,"TERMINAL",0,0,"⠦",,terminal_output
559
+ 559,81286,"TERMINAL",0,0,"⠧\r\nadded 260 packages, and audited 261 packages in 13s\r\n⠧\r\n⠧74 packages are looking for funding\r\n⠧ run `npm fund` for details\r\n⠧\r\nfound 0 vulnerabilities\r\n⠧\r\nYour extension crowd-pilot has been created!\r\n\r\nTo start editing with Visual Studio Code, use the following commands:\r\n\r\n code crowd-pilot\r\n\r\nOpen vsc-extension-quickstart.md inside the new extension for further instructions\r\non how to modify, test and publish your extension.\r\n\r\nFor more information, also visit http://code.visualstudio.com and follow us @code.\r\n\r\r\n\r\n? Do you want to open the new folder with Visual Studio Code? (Use arrow keys)\r\n❯ Open with `code` \r\n Skip [?25l",,terminal_output
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-84bc9952-c4b0-4456-bdc2-984faf53684f1751163593750-2025_06_28-19.19.55.196/source.csv ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,3,"tasks",0,0,"",Log,tab
3
+ 2,18,"utils/dataloader.py",0,0,"import functools\nimport jax\n\nimport tensorflow as tf\n\n# reserve GPU memory for JAX only if tensorflow is built with GPU support\ntf.config.experimental.set_visible_devices([], ""GPU"")\n\n\n# --- TensorFlow function for processing: slicing, normalization ---\ndef _tf_process_episode(episode_tensor, seq_len, image_h, image_w, image_c):\n """"""\n Processes a raw episode tensor in TensorFlow.\n Takes a full episode, extracts a random sequence, and normalizes it.\n Args:\n episode_tensor: A TensorFlow tensor representing a full video episode.\n Expected shape: (dynamic_length, image_h, image_w, image_c)\n Expected dtype: e.g., tf.uint8 (raw pixel values)\n seq_len: The desired length of the sub-sequence to extract.\n image_h: The height of each frame.\n image_w: The width of each frame.\n image_c: The number of channels in each frame.\n Returns:\n A TensorFlow tensor representing the processed video sequence.\n Shape: (seq_len, image_h, image_w, image_c)\n Dtype: tf.float32 (normalized pixel values)\n """"""\n current_episode_len = tf.shape(episode_tensor)[0]\n\n max_start_idx = current_episode_len - seq_len\n\n start_idx = tf.random.uniform(\n shape=(), minval=0, maxval=max_start_idx + 1, dtype=tf.int32\n )\n\n seq = episode_tensor[start_idx : start_idx + seq_len]\n\n seq = tf.cast(seq, tf.float32) / 255.0\n\n # Ensure the final shape is statically known for batching.\n # tf.reshape is robust, but tf.ensure_shape or set_shape can also be used if confident.\n processed_sequence = tf.reshape(seq, [seq_len, image_h, image_w, image_c])\n\n return processed_sequence\n\n\ndef _parse_tfrecord_fn(example_proto, image_h, image_w, image_c):\n feature_description = {\n ""height"": tf.io.FixedLenFeature([], tf.int64),\n ""width"": tf.io.FixedLenFeature([], tf.int64),\n ""channels"": tf.io.FixedLenFeature([], tf.int64),\n ""sequence_length"": tf.io.FixedLenFeature([], tf.int64),\n ""raw_video"": tf.io.FixedLenFeature([], tf.string),\n }\n example = tf.io.parse_single_example(example_proto, feature_description)\n\n video_shape = (example[""sequence_length""], image_h, image_w, image_c)\n\n episode_tensor = tf.io.decode_raw(example[""raw_video""], out_type=tf.uint8)\n episode_tensor = tf.reshape(episode_tensor, video_shape)\n\n episode_tensor = tf.ensure_shape(episode_tensor, [None, image_h, image_w, image_c])\n return episode_tensor\n\n\ndef get_dataloader(\n tfrecord_paths: list[str], # List of TFRecord file paths\n seq_len: int,\n global_batch_size: int,\n image_h: int,\n image_w: int,\n image_c: int,\n shuffle_buffer_size: int = 1000,\n num_parallel_calls: int = tf.data.AUTOTUNE,\n seed: int = 42,\n):\n """"""\n Creates a tf.data.Dataset pipeline from TFRecord files.\n """"""\n if not tfrecord_paths:\n raise ValueError(""tfrecord_paths list cannot be empty."")\n\n process_id = jax.process_index()\n num_processes = jax.process_count()\n\n assert (\n global_batch_size % num_processes == 0\n ), ""Global batch size {global_batch_size} \\n must be divisible by the number of JAX processes {num_processes} for proper sharding.""\n per_process_batch_size = global_batch_size // num_processes\n\n dataset = tf.data.TFRecordDataset(\n tfrecord_paths, num_parallel_reads=tf.data.AUTOTUNE\n )\n\n dataset = dataset.shard(num_shards=num_processes, index=process_id)\n\n # (f.srambical) NOTE: For TFRecords, it's often good to have a large shuffle buffer.\n if shuffle_buffer_size > 0:\n dataset = dataset.shuffle(\n buffer_size=shuffle_buffer_size, seed=seed, reshuffle_each_iteration=True\n )\n parse_fn = functools.partial(\n _parse_tfrecord_fn, image_h=image_h, image_w=image_w, image_c=image_c\n )\n dataset = dataset.map(parse_fn, num_parallel_calls=num_parallel_calls)\n\n tf_process_fn = functools.partial(\n _tf_process_episode,\n seq_len=seq_len,\n image_h=image_h,\n image_w=image_w,\n image_c=image_c,\n )\n dataset = dataset.map(tf_process_fn, num_parallel_calls=num_parallel_calls)\n\n dataset = dataset.repeat(None)\n dataset = dataset.batch(per_process_batch_size, drop_remainder=True)\n dataset = dataset.prefetch(tf.data.AUTOTUNE)\n\n return dataset.as_numpy_iterator()\n",python,tab
4
+ 3,40,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
5
+ 4,41,"utils/dataloader.py",0,0,"",python,tab
6
+ 5,1469,"utils/dataloader.py",0,0,"",python,selection_command
7
+ 6,3873,"train_tokenizer.py",0,0,"from dataclasses import dataclass, field\nimport os\nimport time\n\nimport einops\nfrom flax.training import orbax_utils\nfrom flax.training.train_state import TrainState\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax\nfrom orbax.checkpoint import PyTreeCheckpointer\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\n\nfrom models.tokenizer import TokenizerVQVAE\nfrom utils.dataloader import get_dataloader\n\nts = int(time.time())\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 300_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = ""data_tfrecords/coinrun""\n checkpoint: str = """"\n # Optimization\n vq_beta: float = 0.25\n batch_size: int = 48\n min_lr: float = 3e-4\n max_lr: float = 3e-4\n warmup_steps: int = 10000\n # Tokenizer\n model_dim: int = 512\n latent_dim: int = 32\n num_latents: int = 1024\n patch_size: int = 4\n num_blocks: int = 8\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.01\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_tokenizer""\n tags: list[str] = field(default_factory=lambda: [""tokenizer""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_gradients: bool = False\n\n\nargs = tyro.cli(Args)\n\n\ndef tokenizer_loss_fn(params, state, inputs):\n # --- Compute loss ---\n outputs = state.apply_fn(\n params,\n inputs,\n training=True,\n rngs={""params"": inputs[""rng""], ""dropout"": inputs[""dropout_rng""]},\n )\n mse = jnp.square(inputs[""videos""] - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = inputs[""videos""].clip(0, 1).reshape(-1, *inputs[""videos""].shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = pix.psnr(gt, recon).mean()\n ssim = pix.ssim(gt, recon).mean()\n _, index_counts = jnp.unique_counts(\n jnp.ravel(outputs[""indices""]), size=args.num_latents, fill_value=0\n )\n codebook_usage = (index_counts != 0).mean()\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=codebook_usage,\n )\n return loss, (outputs[""recon""], metrics)\n\n\n@jax.jit\ndef train_step(state, inputs):\n grad_fn = jax.value_and_grad(tokenizer_loss_fn, has_aux=True, allow_int=True)\n (loss, (recon, metrics)), grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n if args.log_gradients:\n metrics[""encoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""encoder""]\n )\n metrics[""vq_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""vq""]\n )\n metrics[""decoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""decoder""]\n )\n return state, loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.PRNGKey(args.seed)\n if args.log and jax.process_index() == 0:\n wandb.init(\n entity=args.entity,\n project=args.project,\n name=args.name,\n tags=args.tags,\n group=""debug"",\n config=args\n )\n\n # --- Initialize model ---\n tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n )\n rng, _rng = jax.random.split(rng)\n image_shape = (args.image_height, args.image_width, args.image_channels)\n inputs = dict(\n videos=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len, *image_shape),\n dtype=jnp.float32,\n ),\n )\n init_params = tokenizer.init(_rng, inputs)\n\n # --- Initialize optimizer ---\n lr_schedule = optax.warmup_cosine_decay_schedule(\n args.min_lr, args.max_lr, args.warmup_steps, args.num_steps\n )\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4)\n train_state = TrainState.create(apply_fn=tokenizer.apply, params=init_params, tx=tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n train_state = jax.device_put(train_state, replicated_sharding)\n\n # --- Load checkpoint ---\n step = 0\n if args.checkpoint:\n restore_target = {""model"": train_state}\n restore_args = orbax_utils.restore_args_from_target(restore_target)\n train_state.params[""params""].update(\n PyTreeCheckpointer()\n .restore(args.checkpoint, item=restore_target, restore_args=restore_args)[\n ""model""\n ]\n .params[""params""]\n )\n # Assume checkpoint is of the form tokenizer_<timestamp>_<step>\n step += int(args.checkpoint.split(""_"")[-1])\n\n # --- TRAIN LOOP ---\n tfrecord_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".tfrecord"")\n ]\n dataloader = get_dataloader(\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n tfrecord_files,\n args.seq_len,\n args.batch_size,\n *image_shape,\n )\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng, _rng_dropout = jax.random.split(rng, 3)\n\n videos_sharding = NamedSharding(\n mesh, PartitionSpec(""data"", None, None, None, None)\n )\n videos = jax.make_array_from_process_local_data(videos_sharding, videos)\n\n inputs = dict(videos=videos, rng=_rng, dropout_rng=_rng_dropout)\n start_time = time.time()\n train_state, loss, recon, metrics = train_step(train_state, inputs)\n elapsed_time = (time.time() - start_time) * 1000\n print(f""Step {step}, loss: {loss}, step time: {elapsed_time}ms"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n ""step_time_ms"": elapsed_time,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0]\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n if step % args.log_checkpoint_interval == 0:\n ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,\n )\n if step >= args.num_steps:\n break\n",python,tab
8
+ 7,5631,"models/tokenizer.py",0,0,"from typing import Dict, Any, Tuple\n\nimport flax.linen as nn\n\nfrom utils.preprocess import patchify, unpatchify\nfrom utils.nn import STTransformer, VectorQuantizer\n\n\nclass TokenizerVQVAE(nn.Module):\n """"""ST-ViVit VQ-VAE""""""\n\n in_dim: int\n model_dim: int\n latent_dim: int\n num_latents: int\n patch_size: int\n num_blocks: int\n num_heads: int\n dropout: float\n codebook_dropout: float\n\n def setup(self):\n self.encoder = STTransformer(\n self.model_dim,\n self.latent_dim,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n )\n self.vq = VectorQuantizer(\n self.latent_dim,\n self.num_latents,\n self.codebook_dropout,\n )\n self.out_dim = self.in_dim * self.patch_size**2\n self.decoder = STTransformer(\n self.model_dim,\n self.out_dim,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n )\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n H, W = batch[""videos""].shape[2:4]\n outputs = self.vq_encode(batch[""videos""], training)\n recon = self.decoder(outputs[""z_q""]) # (B, T, H_down * W_down, C)\n recon = nn.sigmoid(recon)\n outputs[""recon""] = unpatchify(recon, self.patch_size, H, W)\n return outputs\n\n def vq_encode(self, videos: Any, training: bool = True) -> Dict[str, Any]:\n # --- Preprocess + encode ---\n B, T = videos.shape[:2]\n x = patchify(videos, self.patch_size)\n N = x.shape[2]\n x = self.encoder(x) # (B, T, N, E)\n\n # --- Vector quantize ---\n x = x.reshape(B * T * N, self.latent_dim)\n z_q, z, emb, indices = self.vq(x, training)\n z_q = z_q.reshape(B, T, N, self.latent_dim)\n indices = indices.reshape(B, T, N)\n return dict(z_q=z_q, z=z, emb=emb, indices=indices)\n\n def decode(self, indices: Any, video_hw: Tuple[int, int]):\n z = self.vq.codebook[indices]\n recon = self.decoder(z)\n recon = nn.sigmoid(recon)\n return unpatchify(recon, self.patch_size, *video_hw)\n",python,tab
9
+ 8,6632,"models/tokenizer.py",1009,0,"",python,selection_command
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-971339fb-c463-429e-a956-f7bb98fdea341755623101217-2025_08_19-19.05.03.822/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-978ed4eb-d9c9-4380-b981-e501087459181750623968304-2025_06_22-13.26.11.394/source.csv ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,13,"utils/dataloader.py",0,0,"from cgi import test\nimport functools\nimport jax\n\nimport tensorflow as tf\n\n\n# --- TensorFlow function for processing: slicing, normalization ---\ndef _tf_process_episode(episode_tensor, seq_len, image_h, image_w, image_c):\n """"""\n Processes a raw episode tensor in TensorFlow.\n Takes a full episode, extracts a random sequence, and normalizes it.\n Args:\n episode_tensor: A TensorFlow tensor representing a full video episode.\n Expected shape: (dynamic_length, image_h, image_w, image_c)\n Expected dtype: e.g., tf.uint8 (raw pixel values)\n seq_len: The desired length of the sub-sequence to extract.\n image_h: The height of each frame.\n image_w: The width of each frame.\n image_c: The number of channels in each frame.\n Returns:\n A TensorFlow tensor representing the processed video sequence.\n Shape: (seq_len, image_h, image_w, image_c)\n Dtype: tf.float32 (normalized pixel values)\n """"""\n current_episode_len = tf.shape(episode_tensor)[0]\n\n max_start_idx = current_episode_len - seq_len\n\n start_idx = tf.random.uniform(\n shape=(), minval=0, maxval=max_start_idx + 1, dtype=tf.int32\n )\n\n seq = episode_tensor[start_idx : start_idx + seq_len]\n\n seq = tf.cast(seq, tf.float32) / 255.0\n\n # Ensure the final shape is statically known for batching.\n # tf.reshape is robust, but tf.ensure_shape or set_shape can also be used if confident.\n processed_sequence = tf.reshape(seq, [seq_len, image_h, image_w, image_c])\n\n return processed_sequence\n\n\ndef _parse_tfrecord_fn(example_proto, image_h, image_w, image_c):\n feature_description = {\n ""height"": tf.io.FixedLenFeature([], tf.int64),\n ""width"": tf.io.FixedLenFeature([], tf.int64),\n ""channels"": tf.io.FixedLenFeature([], tf.int64),\n ""sequence_length"": tf.io.FixedLenFeature([], tf.int64),\n ""raw_video"": tf.io.FixedLenFeature([], tf.string),\n }\n example = tf.io.parse_single_example(example_proto, feature_description)\n\n video_shape = (example[""sequence_length""], image_h, image_w, image_c)\n\n episode_tensor = tf.io.decode_raw(example[""raw_video""], out_type=tf.uint8)\n episode_tensor = tf.reshape(episode_tensor, video_shape)\n\n episode_tensor = tf.ensure_shape(episode_tensor, [None, image_h, image_w, image_c])\n return episode_tensor\n\n\ndef get_dataloader(\n tfrecord_paths: list[str], # List of TFRecord file paths\n seq_len: int,\n global_batch_size: int,\n image_h: int,\n image_w: int,\n image_c: int,\n shuffle_buffer_size: int = 1000,\n num_parallel_calls: int = tf.data.AUTOTUNE,\n cache_processed_data: bool = True,\n seed: int = 42,\n):\n """"""\n Creates a tf.data.Dataset pipeline from TFRecord files.\n """"""\n if not tfrecord_paths:\n raise ValueError(""tfrecord_paths list cannot be empty."")\n\n process_id = jax.process_index()\n num_processes = jax.process_count()\n\n assert global_batch_size % num_processes == 0, ""Global batch size {global_batch_size} \\n must be divisible by the number of JAX processes {num_processes} for proper sharding.""\n per_process_batch_size = global_batch_size // num_processes\n\n dataset = tf.data.TFRecordDataset(\n tfrecord_paths, num_parallel_reads=tf.data.AUTOTUNE\n )\n \n dataset = dataset.shard(num_shards=num_processes, index=process_id)\n\n # (f.srambical) NOTE: For TFRecords, it's often good to have a large shuffle buffer.\n if shuffle_buffer_size > 0:\n dataset = dataset.shuffle(\n buffer_size=shuffle_buffer_size, seed=seed, reshuffle_each_iteration=True\n )\n parse_fn = functools.partial(\n _parse_tfrecord_fn, image_h=image_h, image_w=image_w, image_c=image_c\n )\n dataset = dataset.map(parse_fn, num_parallel_calls=num_parallel_calls)\n\n dataset = dataset.cache() if cache_processed_data else dataset\n\n tf_process_fn = functools.partial(\n _tf_process_episode,\n seq_len=seq_len,\n image_h=image_h,\n image_w=image_w,\n image_c=image_c,\n )\n dataset = dataset.map(tf_process_fn, num_parallel_calls=num_parallel_calls)\n\n dataset = dataset.repeat(None)\n dataset = dataset.batch(per_process_batch_size, drop_remainder=True)\n dataset = dataset.prefetch(tf.data.AUTOTUNE)\n\n return dataset.as_numpy_iterator()\n",python,tab
3
+ 2,27,"tasks",0,0,"",Log,tab
4
+ 3,36,"utils/dataloader.py",0,0,"",python,tab
5
+ 4,56,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
6
+ 5,869,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"1:26:11 PM [info] Activating crowd-code\n1:26:11 PM [info] Recording started\n",Log,content
7
+ 6,39651,"utils/dataloader.py",0,0,"",python,tab
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-9deb94c8-8483-468e-9e97-468dc0f8f81a1766266039649-2025_12_20-22.27.31.492/source.csv ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,3,"crates/cli/src/main.rs",0,0,"//! CLI tool for serializing crowd-pilot IDE interaction data.\n//!\n//! This tool processes CSV session files and outputs JSONL format suitable for\n//! NeMo SFT training. It uses the HuggingFace tokenizers Rust library for\n//! accurate token counting.\n\nuse std::path::PathBuf;\n\nuse clap::Parser;\nuse tokenizers::Tokenizer as HfTokenizer;\n\nuse crowd_pilot_serializer_core::{\n pipeline::{PipelineConfig, PipelineResult},\n process_all_sessions, write_jsonl_output, Tokenizer,\n};\n\n/// Serialize crowd-pilot CSV sessions to NeMo JSONL format.\n#[derive(Parser, Debug)]\n#[command(name = ""crowd-pilot-serialize"")]\n#[command(author, version, about, long_about = None)]\nstruct Args {\n /// Root directory containing CSV session files\n #[arg(long)]\n csv_root: PathBuf,\n\n /// Output directory for JSONL files\n #[arg(long)]\n output_dir: PathBuf,\n\n /// HuggingFace tokenizer model name or path\n #[arg(long)]\n tokenizer: String,\n\n /// Maximum tokens per conversation chunk\n #[arg(long, default_value = ""8192"")]\n max_tokens_per_conversation: usize,\n\n /// Maximum tokens per message\n #[arg(long, default_value = ""2048"")]\n max_tokens_per_message: usize,\n\n /// Minimum messages required to keep a conversation\n #[arg(long, default_value = ""5"")]\n min_conversation_messages: usize,\n\n /// Viewport radius (lines above/below cursor)\n #[arg(long, default_value = ""10"")]\n viewport_radius: usize,\n\n /// Coalesce radius for grouping nearby edits\n #[arg(long, default_value = ""5"")]\n coalesce_radius: usize,\n\n /// Fraction of sessions for validation (0.0-1.0)\n #[arg(long, default_value = ""0.1"")]\n val_ratio: f64,\n\n /// Custom system prompt (optional)\n #[arg(long)]\n system_prompt: Option<String>,\n}\n\nconst DEFAULT_SYSTEM_PROMPT: &str = r#""You are a helpful assistant that can interact multiple times with a computer shell to solve programming tasks.\nYour response must contain exactly ONE bash code block with ONE command (or commands connected with && or ||).\n\nFormat your response as shown in <format_example>.\n\n<format_example>\n```bash\nyour_command_here\n```\n</format_example>\n\nFailure to follow these rules will cause your response to be rejected.""#;\n\n/// Wrapper around HuggingFace tokenizers for token counting and truncation.\n///\n/// This uses the Rust-native tokenizers library, which is `Send + Sync`\n/// and enables true parallel tokenization without the Python GIL.\nstruct RustTokenizer {\n inner: HfTokenizer,\n}\n\nimpl RustTokenizer {\n /// Load a HuggingFace tokenizer from a model name or path.\n fn load(model_name: &str) -> Result<Self, Box<dyn std::error::Error>> {\n let inner = HfTokenizer::from_pretrained(model_name, None)\n .map_err(|e| e as Box<dyn std::error::Error>)?;\n Ok(Self { inner })\n }\n}\n\nimpl Tokenizer for RustTokenizer {\n fn count_tokens(&self, text: &str) -> usize {\n self.inner\n .encode(text, false)\n .expect(""Failed to encode text with tokenizer"")\n .get_ids()\n .len()\n }\n\n fn truncate_to_max_tokens(&self, text: &str, max_tokens: usize) -> String {\n let encoding = self.inner\n .encode(text, false)\n .expect(""Failed to encode text with tokenizer"");\n \n let ids = encoding.get_ids();\n if ids.len() <= max_tokens {\n return text.to_string();\n }\n \n let truncated_ids: Vec<u32> = ids[..max_tokens].to_vec();\n self.inner\n .decode(&truncated_ids, true)\n .expect(""Failed to decode truncated tokens"")\n }\n}\n\nfn main() -> Result<(), Box<dyn std::error::Error>> {\n let args = Args::parse();\n\n println!(""Loading tokenizer from {}..."", args.tokenizer);\n let tokenizer = RustTokenizer::load(&args.tokenizer)?;\n\n let config = PipelineConfig {\n max_tokens_per_conversation: args.max_tokens_per_conversation,\n max_tokens_per_message: args.max_tokens_per_message,\n min_conversation_messages: args.min_conversation_messages,\n viewport_radius: args.viewport_radius,\n coalesce_radius: args.coalesce_radius,\n val_ratio: args.val_ratio,\n };\n\n println!(""Processing CSV files from {:?}..."", args.csv_root);\n let session_results = process_all_sessions(\n &args.csv_root,\n &tokenizer,\n &config,\n )?;\n\n let total_sessions = session_results.len();\n println!(""Processed {} sessions"", total_sessions);\n\n let system_prompt = args.system_prompt.as_deref().unwrap_or(DEFAULT_SYSTEM_PROMPT);\n\n println!(""Writing output to {:?}..."", args.output_dir);\n let result: PipelineResult = write_jsonl_output(\n session_results,\n &args.output_dir,\n args.val_ratio,\n system_prompt,\n )?;\n\n let metadata_path = args.output_dir.join(""metadata.json"");\n let metadata = serde_json::json!({\n ""config"": {\n ""csv_root"": args.csv_root.to_string_lossy(),\n ""output_dir"": args.output_dir.to_string_lossy(),\n ""tokenizer"": args.tokenizer,\n ""max_tokens_per_conversation"": args.max_tokens_per_conversation,\n ""max_tokens_per_message"": args.max_tokens_per_message,\n ""min_conversation_messages"": args.min_conversation_messages,\n ""viewport_radius"": args.viewport_radius,\n ""coalesce_radius"": args.coalesce_radius,\n ""val_ratio"": args.val_ratio,\n },\n ""counts"": {\n ""total_sessions"": result.total_sessions,\n ""total_conversations"": result.total_conversations,\n ""train_conversations"": result.train_conversations,\n ""val_conversations"": result.val_conversations,\n },\n ""stats"": {\n ""total_messages"": result.total_messages,\n ""total_tokens"": result.total_tokens,\n ""avg_messages_per_conversation"": if result.total_conversations > 0 {\n result.total_messages as f64 / result.total_conversations as f64\n } else {\n 0.0\n },\n ""avg_tokens_per_conversation"": if result.total_conversations > 0 {\n result.total_tokens as f64 / result.total_conversations as f64\n } else {\n 0.0\n },\n },\n ""files"": {\n ""train_path"": args.output_dir.join(""training.jsonl"").to_string_lossy(),\n ""val_path"": args.output_dir.join(""validation.jsonl"").to_string_lossy(),\n },\n });\n std::fs::write(&metadata_path, serde_json::to_string_pretty(&metadata)?)?;\n\n println!(""\n[summary]"");\n println!("" Total sessions processed: {}"", result.total_sessions);\n println!("" Train conversations: {}"", result.train_conversations);\n println!("" Val conversations: {}"", result.val_conversations);\n println!("" Total messages: {}"", result.total_messages);\n println!("" Total tokens: {}"", result.total_tokens);\n println!("" Output: {:?}/{{training,validation}}.jsonl"", args.output_dir);\n println!("" Metadata: {:?}"", metadata_path);\n\n Ok(())\n}\n",rust,tab
3
+ 2,220,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"10:27:31 PM [info] Activating crowd-code\n10:27:31 PM [info] Recording started\n10:27:31 PM [info] Initializing git provider using file system watchers...\n",Log,tab
4
+ 3,246,"extension-output-pdoom-org.crowd-code-#1-crowd-code",153,0,"10:27:31 PM [info] Git repository found\n10:27:31 PM [info] Git provider initialized successfully\n10:27:31 PM [info] Initial git state: [object Object]\n",Log,content
5
+ 4,1927,"crates/cli/src/main.rs",0,0,"",rust,tab
6
+ 5,3595,"TERMINAL",0,0,"",,terminal_focus
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-a046d89c-9bee-4791-a53b-dd74376ff2861755421590581-2025_08_17-11.06.44.343/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-af9d387a-db29-4e3d-9a10-63e0995d4e191758702840888-2025_09_24-10.34.06.824/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-b00cd52f-686b-4cad-89ec-cf5dcdc287a11753702370531-2025_07_28-13.32.59.505/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-b656ac92-3ae1-4979-b90f-67b013ce79bf1762368361483-2025_11_05-19.46.06.776/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-b8ef8518-bf52-489e-b2a9-5e402fd02c471760857710761-2025_10_19-09.08.39.428/source.csv ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,1,"crowd_code_player/replay_file.py",0,0,"import pandas as pd\nimport curses\nimport time\nimport argparse\n\ndef offset_to_yx(content, offset):\n """"""Converts a 1D string offset to 2D (y, x) coordinates.""""""\n # Ensure offset is within the bounds of the content length\n offset = min(len(content), int(offset))\n \n # Find the line number by counting newlines before the offset\n y = content.count('\n', 0, offset)\n \n # Find the column number by finding the last newline before the offset\n last_newline_pos = content.rfind('\n', 0, offset)\n if last_newline_pos == -1:\n x = offset\n else:\n x = offset - last_newline_pos - 1\n \n return y, x\n\ndef apply_change(content, offset, length, new_text):\n """"""Applies a text change to the content string.""""""\n content = str(content)\n new_text = str(new_text) if pd.notna(new_text) else """"\n offset, length = int(offset), int(length)\n \n # Convert literal \n and \r characters to actual newlines and carriage returns\n new_text = new_text.replace('\\n', '\n').replace('\\r', '\r')\n \n if offset > len(content):\n content += ' ' * (offset - len(content)) # Pad if offset is out of bounds\n \n return content[:offset] + new_text + content[offset + length:]\n\ndef replay_trace(stdscr, filepath, speed_factor, long_pause_threshold=120000):\n """"""Main function to replay the coding trace in the terminal.""""""\n # --- Curses Setup ---\n curses.curs_set(0) # We'll draw our own cursor\n stdscr.nodelay(1)\n curses.start_color()\n curses.use_default_colors()\n curses.init_pair(1, curses.COLOR_WHITE, -1) # For status bar\n curses.init_pair(2, curses.COLOR_BLACK, curses.COLOR_WHITE) # For our cursor\n\n # --- Data Loading ---\n try:\n df = pd.read_csv(filepath).sort_values('Time').reset_index(drop=True)\n except FileNotFoundError:\n print(f""Error: The file '{filepath}' was not found."")\n return\n\n # --- State Management ---\n file_states = {}\n scroll_states = {} # Tracks the top-line for each file's viewport\n active_file = None\n paused = False\n \n # --- Main Replay Loop ---\n for i in range(len(df)):\n # --- Handle User Input for Playback Control ---\n key = stdscr.getch()\n if key == ord('q'): break\n if key == ord(' '): paused = not paused\n if key == curses.KEY_UP: speed_factor = min(100, speed_factor * 1.5)\n if key == curses.KEY_DOWN: speed_factor = max(0.1, speed_factor / 1.5)\n \n # Handle Paused State\n if paused:\n height, width = stdscr.getmaxyx()\n stdscr.addstr(height - 1, 0, ""PAUSED"".ljust(width - 1), curses.A_REVERSE)\n stdscr.refresh()\n while paused:\n time.sleep(0.1)\n key = stdscr.getch()\n if key == ord(' '): paused = False\n elif key == ord('q'): return\n\n # --- Process Event ---\n event = df.iloc[i]\n active_file = event['File']\n \n # Initialize state for new files\n if active_file not in file_states:\n file_states[active_file] = """"\n scroll_states[active_file] = 0\n\n \n # Apply content change based on event type\n if active_file == ""TERMINAL"":\n # For terminal, just append text and add a newline\n terminal_text = str(event['Text']) if pd.notna(event['Text']) else """"\n # Convert literal \n and \r characters to actual newlines and carriage returns\n terminal_text = terminal_text.replace('\\n', '\n').replace('\\r', '\r')\n file_states[active_file] += terminal_text + '\n'\n else:\n file_states[active_file] = apply_change(\n file_states[active_file], event['RangeOffset'], \n event['RangeLength'], event['Text']\n )\n \n # --- Calculate Cursor and Scrolling ---\n content = file_states[active_file]\n cursor_y, cursor_x = offset_to_yx(content, event['RangeOffset'])\n scroll_y = scroll_states[active_file]\n height, width = stdscr.getmaxyx()\n visible_height = height - 2 # Account for status bars\n\n # Adjust scroll to keep cursor in view\n if active_file == ""TERMINAL"":\n # For terminal, always scroll to bottom to show latest content\n lines = content.split('\n')\n total_lines = len(lines)\n if total_lines > visible_height:\n scroll_y = max(0, total_lines - visible_height)\n else:\n # For regular files, keep cursor in view\n if cursor_y < scroll_y:\n scroll_y = cursor_y\n elif cursor_y >= scroll_y + visible_height:\n scroll_y = cursor_y - visible_height + 1\n \n scroll_states[active_file] = scroll_y\n\n # --- Render to Screen ---\n stdscr.clear()\n \n # Display file content with scrolling\n lines = content.split('\n')\n for j in range(visible_height):\n line_idx = scroll_y + j\n if line_idx < len(lines):\n stdscr.addstr(j, 0, lines[line_idx][:width - 1])\n \n # Draw our custom cursor\n display_y = cursor_y - scroll_y\n if 0 <= display_y < visible_height and 0 <= cursor_x < width:\n # Ensure we don't try to draw on a non-existent character\n line_len = len(lines[cursor_y]) if cursor_y < len(lines) else 0\n char_to_draw_under = lines[cursor_y][cursor_x] if cursor_x < line_len else "" ""\n stdscr.attron(curses.color_pair(2))\n stdscr.addstr(display_y, cursor_x, char_to_draw_under)\n stdscr.attroff(curses.color_pair(2))\n\n # Status Bar\n status_bar_text = f""File: {active_file} | Time: {event['Time']/1000:.1f}s | Event: {event['Type']} | Speed: {speed_factor:.1f}x""\n stdscr.attron(curses.color_pair(1) | curses.A_REVERSE)\n stdscr.addstr(height - 2, 0, status_bar_text.ljust(width - 1))\n stdscr.attroff(curses.color_pair(1) | curses.A_REVERSE)\n\n # Help Text\n help_text = ""PAUSE/PLAY [space] | FASTER [↑] | SLOWER [↓] | QUIT [q]""\n stdscr.addstr(height - 1, 0, help_text)\n\n stdscr.refresh()\n\n # --- Wait for Next Event ---\n if i + 1 < len(df):\n time_delta_ms = df.iloc[i+1]['Time'] - event['Time']\n sleep_duration_s = max(0, time_delta_ms / 1000.0)\n \n # Check for long pauses\n if time_delta_ms > long_pause_threshold:\n # Display long pause message\n height, width = stdscr.getmaxyx()\n pause_message = ""Long pause detected. User might be googling, thinking or might have gone for a coffee...""\n stdscr.addstr(height - 3, 0, pause_message.ljust(width - 1), curses.A_REVERSE)\n stdscr.refresh()\n time.sleep(1) # Show message for 1 seconds\n stdscr.clear()\n else:\n time.sleep(sleep_duration_s / speed_factor)\n\nif __name__ == ""__main__"":\n parser = argparse.ArgumentParser(description=""Replay coding traces from a CSV file in the terminal."")\n parser.add_argument(""filepath"", help=""The path to the source CSV file."")\n parser.add_argument(""--speed"", type=float, default=20.0, help=""Initial playback speed multiplier."")\n parser.add_argument(""--long_pause_threshold"", type=int, default=120000, help=""Threshold for long pause in milliseconds."")\n args = parser.parse_args()\n\n curses.wrapper(replay_trace, args.filepath, args.speed, args.long_pause_threshold)",python,tab
3
+ 2,70,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"9:08:39 AM [info] Activating crowd-code\n9:08:39 AM [info] Recording started\n9:08:39 AM [info] Initializing git provider using file system watchers...\n9:08:39 AM [info] Git repository found\n9:08:39 AM [info] Git provider initialized successfully\n",Log,tab
4
+ 3,207,"extension-output-pdoom-org.crowd-code-#1-crowd-code",245,0,"9:08:39 AM [info] Initial git state: [object Object]\n",Log,content
5
+ 4,13185962,"crowd_code_player/replay_file.py",0,0,"",python,tab
6
+ 5,13191919,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
7
+ 6,13194346,"crowd_code_player/replay_file.py",0,0,"",python,tab
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-c00a9926-ed80-4aa8-b582-d6f932b7ac281754468444238-2025_08_06-10.20.46.993/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-c3986074-5718-42fa-9de0-e87adef2d7e21764449343379-2025_11_29-21.49.07.420/source.csv ADDED
@@ -0,0 +1,204 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,12,"Untitled-1",0,0,"",plaintext,tab
3
+ 2,92,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"9:49:07 PM [info] Activating crowd-code\n9:49:07 PM [info] Recording started\n9:49:07 PM [info] Initializing git provider using file system watchers...\n9:49:07 PM [info] No workspace folder found\n",Log,tab
4
+ 3,705,"Untitled-1",0,0,"",plaintext,tab
5
+ 4,2176,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
6
+ 5,4365,"Untitled-1",0,0,"",plaintext,tab
7
+ 6,8845,"Untitled-1",0,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
8
+ 7,10173,"Untitled-1",46,0,"",plaintext,selection_command
9
+ 8,10427,"Untitled-1",39,0,"",plaintext,selection_command
10
+ 9,10455,"Untitled-1",32,0,"",plaintext,selection_command
11
+ 10,10488,"Untitled-1",0,0,"",plaintext,selection_command
12
+ 11,10947,"Untitled-1",0,0,"\n",plaintext,content
13
+ 12,11669,"Untitled-1",0,0,"O",plaintext,content
14
+ 13,11671,"Untitled-1",1,0,"",plaintext,selection_keyboard
15
+ 14,12236,"Untitled-1",0,1,"",plaintext,content
16
+ 15,12414,"Untitled-1",0,0,"\n",plaintext,content
17
+ 16,12853,"Untitled-1",0,0,"",plaintext,selection_command
18
+ 17,17664,"Untitled-1",34,13,"/* crowd-pilot: replacement */\nREPLACED LINE 1\nREPLACED LINE 2",plaintext,content
19
+ 18,18633,"Untitled-1",97,30,"",plaintext,content
20
+ 19,20554,"Untitled-1",1,0,"",plaintext,selection_command
21
+ 20,20772,"Untitled-1",2,0,"",plaintext,selection_command
22
+ 21,21489,"TERMINAL",0,0,"undefinedfranzsrambical@MBF6N9WFVKFV ~ % echo ""Hello World""",,terminal_command
23
+ 22,21490,"TERMINAL",0,0,"]633;CHello World\r\n% \r \r",,terminal_output
24
+ 23,24938,"Untitled-1",1,0,"",plaintext,selection_command
25
+ 24,25473,"Untitled-1",0,0,"",plaintext,selection_command
26
+ 25,30419,"Untitled-1",97,0,"",plaintext,selection_command
27
+ 26,36417,"Untitled-1",97,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
28
+ 27,38953,"Untitled-1",143,0,"",plaintext,selection_command
29
+ 28,39115,"Untitled-1",136,0,"",plaintext,selection_command
30
+ 29,39248,"Untitled-1",129,0,"",plaintext,selection_command
31
+ 30,39405,"Untitled-1",97,0,"",plaintext,selection_command
32
+ 31,39570,"Untitled-1",81,0,"",plaintext,selection_command
33
+ 32,39728,"Untitled-1",65,0,"",plaintext,selection_command
34
+ 33,47262,"Untitled-1",129,13,"/* crowd-pilot: replacement */\nREPLACED LINE 1\nREPLACED LINE 2",plaintext,content
35
+ 34,48120,"Untitled-1",192,30,"",plaintext,content
36
+ 35,49086,"TERMINAL",0,0,"echo ""Hello World""",,terminal_command
37
+ 36,49086,"TERMINAL",0,0,"]633;CHello World\r\n% \r \r",,terminal_output
38
+ 37,50060,"Untitled-1",192,0,"",plaintext,selection_command
39
+ 38,51373,"Untitled-1",192,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
40
+ 39,52859,"Untitled-1",238,0,"",plaintext,selection_command
41
+ 40,53106,"Untitled-1",231,0,"",plaintext,selection_command
42
+ 41,53133,"Untitled-1",224,0,"",plaintext,selection_command
43
+ 42,53388,"Untitled-1",192,0,"",plaintext,selection_command
44
+ 43,53563,"Untitled-1",176,0,"",plaintext,selection_command
45
+ 44,53737,"Untitled-1",160,0,"",plaintext,selection_command
46
+ 45,65878,"Untitled-1",224,13,"/* crowd-pilot: replacement */\nREPLACED LINE 1\nREPLACED LINE 2",plaintext,content
47
+ 46,66782,"Untitled-1",287,30,"",css,content
48
+ 47,67390,"TERMINAL",0,0,"echo ""Hello World""",,terminal_command
49
+ 48,67391,"TERMINAL",0,0,"]633;CHello World\r\n% \r \r",,terminal_output
50
+ 49,67836,"Untitled-1",287,0,"",css,selection_command
51
+ 50,70690,"Untitled-1",271,0,"",css,selection_command
52
+ 51,70936,"Untitled-1",255,0,"",css,selection_command
53
+ 52,70976,"Untitled-1",224,0,"",css,selection_command
54
+ 53,70999,"Untitled-1",192,0,"",css,selection_command
55
+ 54,71031,"Untitled-1",176,0,"",css,selection_command
56
+ 55,71278,"Untitled-1",160,0,"",css,selection_command
57
+ 56,71700,"Untitled-1",129,0,"",css,selection_command
58
+ 57,72094,"Untitled-1",97,0,"",css,selection_command
59
+ 58,72256,"Untitled-1",81,0,"",css,selection_command
60
+ 59,72438,"Untitled-1",65,0,"",css,selection_command
61
+ 60,72570,"Untitled-1",34,0,"",css,selection_command
62
+ 61,72771,"Untitled-1",2,0,"",css,selection_command
63
+ 62,72926,"Untitled-1",1,0,"",css,selection_command
64
+ 63,73090,"Untitled-1",0,0,"",css,selection_command
65
+ 64,73571,"Untitled-1",1,0,"",css,selection_command
66
+ 65,74166,"Untitled-1",2,0,"",css,selection_command
67
+ 66,74827,"Untitled-1",34,0,"",css,selection_command
68
+ 67,75075,"Untitled-1",65,0,"",css,selection_command
69
+ 68,75109,"Untitled-1",81,0,"",css,selection_command
70
+ 69,75139,"Untitled-1",97,0,"",css,selection_command
71
+ 70,75506,"Untitled-1",129,0,"",css,selection_command
72
+ 71,75854,"Untitled-1",160,0,"",css,selection_command
73
+ 72,76053,"Untitled-1",176,0,"",css,selection_command
74
+ 73,76235,"Untitled-1",192,0,"",css,selection_command
75
+ 74,76444,"Untitled-1",224,0,"",css,selection_command
76
+ 75,76621,"Untitled-1",255,0,"",css,selection_command
77
+ 76,77943,"Untitled-1",224,0,"",css,selection_command
78
+ 77,78146,"Untitled-1",192,0,"",css,selection_command
79
+ 78,78560,"Untitled-1",176,0,"",css,selection_command
80
+ 79,78735,"Untitled-1",160,0,"",css,selection_command
81
+ 80,78995,"Untitled-1",129,0,"",css,selection_command
82
+ 81,79020,"Untitled-1",97,0,"",css,selection_command
83
+ 82,79051,"Untitled-1",81,0,"",css,selection_command
84
+ 83,79090,"Untitled-1",65,0,"",css,selection_command
85
+ 84,79115,"Untitled-1",34,0,"",css,selection_command
86
+ 85,79532,"Untitled-1",2,0,"",css,selection_command
87
+ 86,79700,"Untitled-1",1,0,"",css,selection_command
88
+ 87,79888,"Untitled-1",0,0,"",css,selection_command
89
+ 88,84218,"Untitled-1",1,0,"",css,selection_command
90
+ 89,84471,"Untitled-1",2,0,"",css,selection_command
91
+ 90,84502,"Untitled-1",34,0,"",css,selection_command
92
+ 91,84533,"Untitled-1",65,0,"",css,selection_command
93
+ 92,84567,"Untitled-1",81,0,"",css,selection_command
94
+ 93,84655,"Untitled-1",65,0,"",css,selection_command
95
+ 94,84918,"Untitled-1",34,0,"",css,selection_command
96
+ 95,84950,"Untitled-1",2,0,"",css,selection_command
97
+ 96,84979,"Untitled-1",1,0,"",css,selection_command
98
+ 97,85012,"Untitled-1",0,0,"",css,selection_command
99
+ 98,85430,"Untitled-1",2,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",css,content
100
+ 99,94236,"Untitled-1",34,13,"/* crowd-pilot: replacement */\nREPLACED LINE 1\nREPLACED LINE 2",css,content
101
+ 100,95288,"Untitled-1",97,93,"",css,content
102
+ 101,96532,"TERMINAL",0,0,"echo ""Hello World""",,terminal_command
103
+ 102,96532,"TERMINAL",0,0,"]633;CHello World\r\n% \r \r",,terminal_output
104
+ 103,99561,"Untitled-1",319,0,"",css,selection_command
105
+ 104,383997,"Untitled-1",0,0,"",css,selection_command
106
+ 105,386194,"Untitled-1",2,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",css,content
107
+ 106,387706,"Untitled-1",34,13,"/* crowd-pilot: replacement */\nREPLACED LINE 1\nREPLACED LINE 2",css,content
108
+ 107,388546,"Untitled-1",97,93,"",css,content
109
+ 108,389467,"TERMINAL",0,0,"echo ""Hello World""",,terminal_command
110
+ 109,389468,"TERMINAL",0,0,"]633;CHello World\r\n% \r \r",,terminal_output
111
+ 110,389985,"Untitled-1",351,0,"",css,selection_command
112
+ 111,392009,"Untitled-1",0,0,"",css,selection_command
113
+ 112,435636,"Untitled-1",1,0,"",css,selection_command
114
+ 113,435939,"Untitled-1",2,0,"",css,selection_command
115
+ 114,436120,"Untitled-1",34,0,"",css,selection_command
116
+ 115,436241,"Untitled-1",65,0,"",css,selection_command
117
+ 116,436494,"Untitled-1",81,0,"",css,selection_command
118
+ 117,436528,"Untitled-1",97,0,"",css,selection_command
119
+ 118,436560,"Untitled-1",113,0,"",css,selection_command
120
+ 119,436597,"Untitled-1",129,0,"",css,selection_command
121
+ 120,436621,"Untitled-1",145,0,"",css,selection_command
122
+ 121,436655,"Untitled-1",161,0,"",css,selection_command
123
+ 122,436692,"Untitled-1",193,0,"",css,selection_command
124
+ 123,436726,"Untitled-1",224,0,"",css,selection_command
125
+ 124,436760,"Untitled-1",240,0,"",css,selection_command
126
+ 125,436794,"Untitled-1",256,0,"",css,selection_command
127
+ 126,436829,"Untitled-1",288,0,"",css,selection_command
128
+ 127,436866,"Untitled-1",319,0,"",css,selection_command
129
+ 128,436898,"Untitled-1",335,0,"",css,selection_command
130
+ 129,436929,"Untitled-1",351,0,"",css,selection_command
131
+ 130,437242,"Untitled-1",335,0,"",css,selection_command
132
+ 131,437422,"Untitled-1",319,0,"",css,selection_command
133
+ 132,437591,"Untitled-1",288,0,"",css,selection_command
134
+ 133,437819,"Untitled-1",256,0,"",css,selection_command
135
+ 134,438453,"Untitled-1",288,0,"",css,selection_command
136
+ 135,438606,"Untitled-1",319,0,"",css,selection_command
137
+ 136,438781,"Untitled-1",335,0,"",css,selection_command
138
+ 137,438935,"Untitled-1",351,0,"",css,selection_command
139
+ 138,441157,"Untitled-1",335,0,"",css,selection_command
140
+ 139,442706,"Untitled-1",351,0,"",css,selection_command
141
+ 140,443502,"Untitled-1",335,0,"",css,selection_command
142
+ 141,445599,"Untitled-1",319,0,"",css,selection_command
143
+ 142,447662,"Untitled-1",288,0,"",css,selection_command
144
+ 143,448304,"Untitled-1",319,0,"",css,selection_command
145
+ 144,448484,"Untitled-1",335,0,"",css,selection_command
146
+ 145,448663,"Untitled-1",351,0,"",css,selection_command
147
+ 146,449171,"Untitled-1",335,0,"",css,selection_command
148
+ 147,449503,"Untitled-1",319,0,"",css,selection_command
149
+ 148,449855,"Untitled-1",288,0,"",css,selection_command
150
+ 149,450818,"Untitled-1",256,0,"",css,selection_command
151
+ 150,451253,"Untitled-1",240,0,"",css,selection_command
152
+ 151,451650,"Untitled-1",256,0,"",css,selection_command
153
+ 152,451816,"Untitled-1",288,0,"",css,selection_command
154
+ 153,451973,"Untitled-1",319,0,"",css,selection_command
155
+ 154,452123,"Untitled-1",335,0,"",css,selection_command
156
+ 155,452255,"Untitled-1",351,0,"",css,selection_command
157
+ 156,452635,"Untitled-1",335,0,"",css,selection_command
158
+ 157,453104,"Untitled-1",319,0,"",css,selection_command
159
+ 158,516245,"Untitled-1",335,0,"",css,selection_command
160
+ 159,516400,"Untitled-1",351,0,"",css,selection_command
161
+ 160,516870,"Untitled-1",335,0,"",css,selection_command
162
+ 161,536501,"Untitled-1",319,0,"",css,selection_command
163
+ 162,539788,"Untitled-1",335,0,"",css,selection_command
164
+ 163,539801,"Untitled-1",319,0,"",css,selection_command
165
+ 164,547248,"Untitled-1",288,0,"",css,selection_command
166
+ 165,549881,"Untitled-1",256,0,"",css,selection_command
167
+ 166,550560,"Untitled-1",288,0,"",css,selection_command
168
+ 167,557888,"Untitled-1",256,0,"",css,selection_command
169
+ 168,564388,"Untitled-1",240,0,"",css,selection_command
170
+ 169,574820,"Untitled-1",256,0,"",css,selection_command
171
+ 170,575146,"Untitled-1",288,0,"",css,selection_command
172
+ 171,575472,"Untitled-1",319,0,"",css,selection_command
173
+ 172,576513,"Untitled-1",335,0,"",css,selection_command
174
+ 173,577382,"Untitled-1",319,0,"",css,selection_command
175
+ 174,577855,"Untitled-1",335,0,"",css,selection_command
176
+ 175,578213,"Untitled-1",319,0,"",css,selection_command
177
+ 176,578928,"Untitled-1",288,0,"",css,selection_command
178
+ 177,580100,"Untitled-1",319,0,"",css,selection_command
179
+ 178,580270,"Untitled-1",335,0,"",css,selection_command
180
+ 179,580437,"Untitled-1",351,0,"",css,selection_command
181
+ 180,580914,"Untitled-1",335,0,"",css,selection_command
182
+ 181,581367,"Untitled-1",319,0,"",css,selection_command
183
+ 182,581740,"Untitled-1",288,0,"",css,selection_command
184
+ 183,582266,"Untitled-1",256,0,"",css,selection_command
185
+ 184,583188,"Untitled-1",288,0,"",css,selection_command
186
+ 185,583354,"Untitled-1",319,0,"",css,selection_command
187
+ 186,583503,"Untitled-1",335,0,"",css,selection_command
188
+ 187,583637,"Untitled-1",351,0,"",css,selection_command
189
+ 188,583986,"Untitled-1",335,0,"",css,selection_command
190
+ 189,584420,"Untitled-1",319,0,"",css,selection_command
191
+ 190,584868,"Untitled-1",288,0,"",css,selection_command
192
+ 191,585942,"Untitled-1",319,0,"",css,selection_command
193
+ 192,586107,"Untitled-1",335,0,"",css,selection_command
194
+ 193,586249,"Untitled-1",351,0,"",css,selection_command
195
+ 194,586567,"Untitled-1",335,0,"",css,selection_command
196
+ 195,587983,"Untitled-1",319,0,"",css,selection_command
197
+ 196,588349,"Untitled-1",288,0,"",css,selection_command
198
+ 197,589742,"Untitled-1",319,0,"",css,selection_command
199
+ 198,590406,"Untitled-1",288,0,"",css,selection_command
200
+ 199,590871,"Untitled-1",319,0,"",css,selection_command
201
+ 200,591603,"Untitled-1",288,0,"",css,selection_command
202
+ 201,592418,"Untitled-1",256,0,"",css,selection_command
203
+ 202,597588,"Untitled-1",240,0,"",css,selection_command
204
+ 203,599719,"Untitled-1",224,0,"",css,selection_command
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-c8c1182e-9d15-45c9-9466-b9340fc3403b1754557906412-2025_08_07-11.11.49.271/source.csv ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,2,"sample.py",0,0,"from dataclasses import dataclass\nimport time\nimport os\nimport optax\n\nimport dm_pix as pix\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\nimport numpy as np\nimport orbax.checkpoint as ocp\nfrom PIL import Image, ImageDraw\nimport tyro\nfrom flax import nnx\n\nfrom genie import Genie\nfrom utils.dataloader import get_dataloader\n\n\n@dataclass\nclass Args:\n # Experiment\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = ""data/coinrun_episodes""\n checkpoint: str = """"\n # Sampling\n batch_size: int = 1\n maskgit_steps: int = 25\n temperature: float = 1.0\n sample_argmax: bool = True\n start_frame: int = 0\n # Tokenizer checkpoint\n tokenizer_dim: int = 512\n tokenizer_ffn_dim: int = 2048\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 4\n tokenizer_num_heads: int = 8\n # LAM checkpoint\n lam_dim: int = 512\n lam_ffn_dim: int = 2048\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 4\n lam_num_heads: int = 8\n # Dynamics checkpoint\n dyna_dim: int = 512\n dyna_ffn_dim: int = 2048\n dyna_num_blocks: int = 6\n dyna_num_heads: int = 8\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n\n\nargs = tyro.cli(Args)\n\nif __name__ == ""__main__"":\n """"""\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n """"""\n jax.distributed.initialize()\n\n rng = jax.random.key(args.seed)\n\n # --- Load Genie checkpoint ---\n rngs = nnx.Rngs(rng)\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n lam_ffn_dim=args.lam_ffn_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=False,\n # Dynamics\n dyna_dim=args.dyna_dim,\n dyna_ffn_dim=args.dyna_ffn_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n )\n\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n step_format_fixed_length=6,\n )\n checkpoint_manager = ocp.CheckpointManager(\n args.checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n dummy_tx = optax.adamw(\n learning_rate=optax.linear_schedule(0.0001, 0.0001, 10000),\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n dummy_optimizer = nnx.Optimizer(genie, dummy_tx)\n\n abstract_optimizer = nnx.eval_shape(lambda: dummy_optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(dummy_optimizer, restored_optimizer_state)\n\n # --- Define sampling function ---\n def _sampling_fn(model: Genie, batch: dict) -> jax.Array:\n """"""Runs Genie.sample with pre-defined generation hyper-parameters.""""""\n return model.sample(\n batch,\n args.seq_len,\n args.maskgit_steps,\n args.temperature,\n args.sample_argmax,\n )\n\n # --- Define autoregressive sampling loop ---\n @nnx.jit\n def _autoreg_sample(rng, video_batch_BSHWC, action_batch_E):\n input_video_BTHWC = video_batch_BSHWC[:, :args.start_frame]\n rng, _rng = jax.random.split(rng)\n batch = dict(videos=input_video_BTHWC, latent_actions=action_batch_E, rng=_rng)\n generated_vid_BSHWC = _sampling_fn(genie, batch)\n return generated_vid_BSHWC\n\n # --- Get video + latent actions ---\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n args.batch_size,\n args.image_height,\n args.image_width,\n args.image_channels,\n # We don't use workers in order to avoid grain shutdown issues (https://github.com/google/grain/issues/398)\n num_workers=0,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n dataloader = iter(dataloader)\n video_batch_BSHWC = next(dataloader)\n gt_video = jnp.asarray(video_batch_BSHWC, dtype=jnp.float32) / 255.0\n video_batch_BSHWC = gt_video.astype(args.dtype)\n # Get latent actions for all videos in the batch\n batch = dict(videos=video_batch_BSHWC)\n action_batch_E = genie.vq_encode(batch, training=False)\n\n # --- Sample + evaluate video ---\n recon_video_BSHWC = _autoreg_sample(rng, video_batch_BSHWC, action_batch_E)\n recon_video_BSHWC = recon_video_BSHWC.astype(jnp.float32)\n gt = gt_video[:, : recon_video_BSHWC.shape[1]].clip(0, 1).reshape(-1, *gt_video.shape[2:])\n recon = recon_video_BSHWC.clip(0, 1).reshape(-1, *recon_video_BSHWC.shape[2:])\n ssim = jnp.asarray(\n pix.ssim(gt[:, args.start_frame:], recon[:, args.start_frame:])\n ).mean()\n print(f""SSIM: {ssim}"")\n\n # --- Construct video ---\n true_videos = (gt_video * 255).astype(np.uint8)\n pred_videos = (recon_video_BSHWC * 255).astype(np.uint8)\n video_comparison = np.zeros((2, *recon_video_BSHWC.shape), dtype=np.uint8)\n video_comparison[0] = true_videos[:, : args.seq_len]\n video_comparison[1] = pred_videos\n frames = einops.rearrange(video_comparison, ""n b t h w c -> t (b h) (n w) c"")\n\n # --- Save video ---\n imgs = [Image.fromarray(img) for img in frames]\n # Write actions on each frame, on each row (i.e., for each video in the batch, on the GT row)\n B, S, _, _, _ = video_batch_BSHWC.shape\n action_batch_BSm11 = jnp.reshape(action_batch_E, (B, S-1, 1))\n for t, img in enumerate(imgs[1:]):\n d = ImageDraw.Draw(img)\n for row in range(action_batch_BSm11.shape[0]):\n action = action_batch_BSm11[row, t, 0]\n y_offset = row * video_batch_BSHWC.shape[2] + 2\n d.text((2, y_offset), f""{action}"", fill=255)\n imgs[0].save(\n f""generation_{time.time()}.gif"",\n save_all=True,\n append_images=imgs[1:],\n duration=250,\n loop=0,\n )\n",python,tab
3
+ 2,54,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"11:11:49 AM [info] Activating crowd-code\n11:11:49 AM [info] Recording started\n11:11:49 AM [info] Initializing git provider using file system watchers...\n11:11:49 AM [info] Git repository found\n11:11:49 AM [info] Git provider initialized successfully\n",Log,tab
4
+ 3,94,"extension-output-pdoom-org.crowd-code-#1-crowd-code",250,0,"11:11:49 AM [info] Initial git state: [object Object]\n",Log,content
5
+ 4,910,"sample.py",0,0,"",python,tab
6
+ 5,2878,"train_tokenizer.py",0,0,"from dataclasses import dataclass, field\nimport os\nfrom typing import cast\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom models.tokenizer import TokenizerVQVAE\nfrom utils.dataloader import get_dataloader\nfrom utils.lr_utils import get_lr_schedule\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 300_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n vq_beta: float = 0.25\n batch_size: int = 48\n init_lr: float = 0.0\n max_lr: float = 3e-4\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 20000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n warmup_steps: int = 10000\n # Tokenizer\n model_dim: int = 512\n ffn_dim: int = 2048\n latent_dim: int = 32\n num_latents: int = 1024\n patch_size: int = 4\n num_blocks: int = 4\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.01\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_tokenizer""\n tags: list[str] = field(default_factory=lambda: [""tokenizer""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n wandb_id: str = """"\n use_flash_attention: bool = True\n\n\nargs = tyro.cli(Args)\n\n\ndef tokenizer_loss_fn(\n model: TokenizerVQVAE, inputs: dict\n) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n # --- Compute loss ---\n # FIXME (f.srambical): Can we even do native int8 training without casting the video at all?\n # FIXME (f.srambical): If the tokenizer is the reason for the dynamics model being memory-bound,\n # should we at least train the tokenizer natively in int8?\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n model.train()\n outputs = model(inputs, training=True)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n mse = jnp.square(gt - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n _, index_counts = jnp.unique_counts(\n jnp.ravel(outputs[""indices""]), size=args.num_latents, fill_value=0\n )\n codebook_usage = (index_counts != 0).mean()\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=codebook_usage,\n )\n return loss, (outputs[""recon""], metrics)\n\n\n@nnx.jit\ndef train_step(\n tokenizer: TokenizerVQVAE, optimizer: nnx.Optimizer, inputs: dict\n) -> tuple[jax.Array, jax.Array, dict]:\n def loss_fn(model: TokenizerVQVAE) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n return tokenizer_loss_fn(model, inputs)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\n tokenizer\n )\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""encoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""encoder""]\n )\n metrics[""vq_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""vq""]\n )\n metrics[""decoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""decoder""]\n )\n return loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n ffn_dim=args.ffn_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n )\n\n _, params, _ = nnx.split(tokenizer, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n optimizer = nnx.Optimizer(tokenizer, tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n\n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n\n # --- TRAIN LOOP ---\n dataloader = (\n jax.make_array_from_process_local_data(videos_sharding, elem)\n for elem in grain_iterator\n )\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n inputs = dict(videos=videos)\n loss, recon, metrics = train_step(tokenizer, optimizer, inputs)\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n grain_iterator # type: ignore\n ),\n ),\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n",python,tab
7
+ 6,3975,"train_tokenizer.py",0,0,"",python,selection_command
8
+ 7,4795,"train_tokenizer.py",1999,0,"",python,selection_command
9
+ 8,6363,"train_tokenizer.py",2022,0,"",python,selection_command
10
+ 9,6621,"train_tokenizer.py",2062,0,"",python,selection_command
11
+ 10,6655,"train_tokenizer.py",2109,0,"",python,selection_command
12
+ 11,6681,"train_tokenizer.py",2136,0,"",python,selection_command
13
+ 12,6713,"train_tokenizer.py",2233,0,"",python,selection_command
14
+ 13,6747,"train_tokenizer.py",2334,0,"",python,selection_command
15
+ 14,6781,"train_tokenizer.py",2397,0,"",python,selection_command
16
+ 15,6816,"train_tokenizer.py",2463,0,"",python,selection_command
17
+ 16,6853,"train_tokenizer.py",2508,0,"",python,selection_command
18
+ 17,6885,"train_tokenizer.py",2526,0,"",python,selection_command
19
+ 18,7087,"train_tokenizer.py",2569,0,"",python,selection_command
20
+ 19,7227,"train_tokenizer.py",2629,0,"",python,selection_command
21
+ 20,7429,"train_tokenizer.py",2680,0,"",python,selection_command
22
+ 21,7577,"train_tokenizer.py",2765,0,"",python,selection_command
23
+ 22,10526,"train_tokenizer.py",2680,0,"",python,selection_command
24
+ 23,10678,"train_tokenizer.py",2629,0,"",python,selection_command
25
+ 24,17735,"train_lam.py",0,0,"from dataclasses import dataclass, field\nimport os\nfrom typing import cast\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom models.lam import LatentActionModel\nfrom utils.dataloader import get_dataloader\nfrom utils.lr_utils import get_lr_schedule\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n vq_beta: float = 0.25\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n warmup_steps: int = 5000\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n vq_reset_thresh: int = 50\n # LAM\n model_dim: int = 512\n ffn_dim: int = 2048\n latent_dim: int = 32\n num_latents: int = 6\n patch_size: int = 16\n num_blocks: int = 4\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.0\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_lam""\n tags: list[str] = field(default_factory=lambda: [""lam""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_checkpoint_keep_period: int = 20000\n wandb_id: str = """"\n use_flash_attention: bool = True\n\n\nargs = tyro.cli(Args)\n\n\ndef lam_loss_fn(\n model: LatentActionModel, inputs: dict\n) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:\n # --- Compute loss ---\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n model.train()\n outputs = model(inputs, training=True)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=(index_counts != 0).mean(),\n )\n return loss, (outputs[""recon""], index_counts, metrics)\n\n\n@nnx.jit\ndef train_step(\n lam: LatentActionModel,\n optimizer: nnx.Optimizer,\n inputs: dict,\n action_last_active: jax.Array,\n rng: jax.Array,\n) -> tuple[jax.Array, jax.Array, jax.Array, dict]:\n def loss_fn(\n model: LatentActionModel,\n ) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:\n return lam_loss_fn(model, inputs)\n\n # --- Update model ---\n (loss, (recon, idx_counts, metrics)), grads = nnx.value_and_grad(\n loss_fn, has_aux=True\n )(lam)\n optimizer.update(grads)\n\n # --- Reset inactive latent actions ---\n codebook = lam.vq.codebook\n num_codes = len(codebook)\n active_codes = idx_counts != 0.0\n action_last_active = jnp.where(active_codes, 0, action_last_active + 1)\n p_code = active_codes / active_codes.sum()\n reset_idxs = jax.random.choice(rng, num_codes, shape=(num_codes,), p=p_code)\n do_reset = action_last_active >= args.vq_reset_thresh\n new_codebook = jnp.where(\n jnp.expand_dims(do_reset, -1), codebook[reset_idxs], codebook.value\n )\n lam.vq.codebook.value = new_codebook\n action_last_active = jnp.where(do_reset, 0, action_last_active)\n return loss, recon, action_last_active, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n ffn_dim=args.ffn_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n )\n\n # Count parameters\n _, params, _ = nnx.split(lam, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n optimizer = nnx.Optimizer(lam, tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n\n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n\n # --- TRAIN LOOP ---\n dataloader = (\n jax.make_array_from_process_local_data(videos_sharding, elem)\n for elem in grain_iterator\n )\n print(f""Starting training from step {step}..."")\n action_last_active = jnp.zeros(args.num_latents, dtype=jnp.int32)\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng = jax.random.split(rng)\n\n inputs = dict(videos=videos, rng=_rng)\n rng, _rng = jax.random.split(rng)\n loss, recon, action_last_active, metrics = train_step(\n lam, optimizer, inputs, action_last_active, _rng\n )\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0, 1:].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n grain_iterator # type: ignore\n ),\n ),\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n",python,tab
26
+ 25,18531,"train_lam.py",0,0,"",python,selection_command
27
+ 26,19228,"train_lam.py",1966,0,"",python,selection_command
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-cb2b7623-a503-48be-abf1-496fd7fc45d81755596347163-2025_08_19-11.39.13.701/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-ccf7a99f-954c-4a69-b08a-cef496899ce51762421470970-2025_11_06-10.31.14.52/source.csv ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 2,130,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"10:31:14 AM [info] Activating crowd-code\n10:31:14 AM [info] Recording started\n10:31:14 AM [info] Initializing git provider using file system watchers...\n10:31:14 AM [info] No workspace folder found\n",Log,tab
3
+ 3,1806,"Untitled-1",0,0,"",plaintext,tab
4
+ 4,3476,"TERMINAL",0,0,"Test",,terminal_focus
5
+ 5,3506,"Untitled-1",0,0,"hello world\n",plaintext,content
6
+ 6,3805,"TERMINAL",0,0,"echo VSCode test",,terminal_command
7
+ 7,3805,"TERMINAL",0,0,"]633;CVSCode test\r\n% \r \r",,terminal_output
8
+ 8,13647,"Untitled-1",0,0,"",plaintext,selection_command
9
+ 9,13650,"Untitled-1",0,0,"hello world\n",plaintext,content
10
+ 10,13684,"TERMINAL",0,0,"echo VSCode test",,terminal_command
11
+ 11,13688,"TERMINAL",0,0,"]633;CVSCode test\r\n% \r \r",,terminal_output
12
+ 12,16138,"Untitled-1",24,0,"",plaintext,selection_command
13
+ 13,16622,"Untitled-1",0,0,"",plaintext,selection_command
14
+ 14,16624,"Untitled-1",0,0,"hello world\n",plaintext,content
15
+ 15,16651,"TERMINAL",0,0,"echo VSCode test",,terminal_command
16
+ 16,16651,"TERMINAL",0,0,"]633;CVSCode test\r\n% \r \r",,terminal_output
17
+ 17,205627,"Untitled-1",24,0,"",plaintext,selection_command
18
+ 18,205769,"Untitled-1",36,0,"",plaintext,selection_command
19
+ 19,206566,"Untitled-1",0,0,"",plaintext,selection_command
20
+ 20,206569,"Untitled-1",0,0,"hello world\n",plaintext,content
21
+ 21,206607,"TERMINAL",0,0,"echo VSCode test",,terminal_command
22
+ 22,206608,"TERMINAL",0,0,"]633;CVSCode test\r\n% \r \r",,terminal_output
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-ce6b4bab-9c2d-4454-8220-eaa00cd55d321764450093947-2025_11_29-22.01.38.36/source.csv ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,11,"Untitled-1",0,0,"",plaintext,tab
3
+ 2,134,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"10:01:38 PM [info] Activating crowd-code\n10:01:38 PM [info] Recording started\n10:01:38 PM [info] Initializing git provider using file system watchers...\n10:01:38 PM [info] No workspace folder found\n",Log,tab
4
+ 3,1072,"Untitled-1",0,0,"",plaintext,tab
5
+ 4,3158,"TERMINAL",0,0,"Test",,terminal_focus
6
+ 5,3166,"Untitled-1",0,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
7
+ 6,5589,"Untitled-1",46,0,"",plaintext,selection_command
8
+ 7,5722,"Untitled-1",39,0,"",plaintext,selection_command
9
+ 8,5972,"Untitled-1",32,0,"",plaintext,selection_command
10
+ 9,6003,"Untitled-1",0,0,"",plaintext,selection_command
11
+ 10,7472,"Untitled-1",0,0,"\n",plaintext,content
12
+ 11,8492,"Untitled-1",0,0,"\n",plaintext,content
13
+ 12,9454,"Untitled-1",1,0,"",plaintext,selection_command
14
+ 13,9489,"Untitled-1",0,0,"",plaintext,selection_command
15
+ 14,14861,"Untitled-1",1,0,"",plaintext,selection_command
16
+ 15,15105,"Untitled-1",2,0,"",plaintext,selection_command
17
+ 16,15139,"Untitled-1",34,0,"",plaintext,selection_command
18
+ 17,15172,"Untitled-1",41,0,"",plaintext,selection_command
19
+ 18,15205,"Untitled-1",48,0,"",plaintext,selection_command
20
+ 19,15238,"Untitled-1",78,0,"",plaintext,selection_command
21
+ 20,15602,"Untitled-1",78,0,"\n",plaintext,content
22
+ 21,16370,"Untitled-1",79,0,"\n",plaintext,content
23
+ 22,17205,"Untitled-1",79,0,"",plaintext,selection_command
24
+ 23,17290,"Untitled-1",78,0,"",plaintext,selection_command
25
+ 24,17418,"Untitled-1",48,0,"",plaintext,selection_command
26
+ 25,17552,"Untitled-1",41,0,"",plaintext,selection_command
27
+ 26,17726,"Untitled-1",34,0,"",plaintext,selection_command
28
+ 27,18089,"Untitled-1",2,0,"",plaintext,selection_command
29
+ 28,19275,"Untitled-1",1,0,"",plaintext,selection_command
30
+ 29,20757,"Untitled-1",2,0,"",plaintext,selection_command
31
+ 30,21095,"Untitled-1",34,0,"",plaintext,selection_command
32
+ 31,21352,"Untitled-1",41,0,"",plaintext,selection_command
33
+ 32,21388,"Untitled-1",48,0,"",plaintext,selection_command
34
+ 33,21411,"Untitled-1",78,0,"",plaintext,selection_command
35
+ 34,21445,"Untitled-1",79,0,"",plaintext,selection_command
36
+ 35,21481,"Untitled-1",80,0,"",plaintext,selection_command
37
+ 36,21915,"Untitled-1",79,0,"",plaintext,selection_command
38
+ 37,22072,"Untitled-1",78,0,"",plaintext,selection_command
39
+ 38,22240,"Untitled-1",48,0,"",plaintext,selection_command
40
+ 39,22419,"Untitled-1",41,0,"",plaintext,selection_command
41
+ 40,22601,"Untitled-1",34,0,"",plaintext,selection_command
42
+ 41,22770,"Untitled-1",2,0,"",plaintext,selection_command
43
+ 42,22925,"Untitled-1",1,0,"",plaintext,selection_command
44
+ 43,23803,"Untitled-1",0,0,"",plaintext,selection_command
45
+ 44,25988,"Untitled-1",1,0,"",plaintext,selection_command
46
+ 45,26208,"Untitled-1",0,0,"",plaintext,selection_command
47
+ 46,34354,"Untitled-1",34,13,"/* crowd-pilot: replacement */\nREPLACED LINE 1\nREPLACED LINE 2",plaintext,content
48
+ 47,35131,"Untitled-1",97,32,"",plaintext,content
49
+ 48,36660,"TERMINAL",0,0,"echo ""Hello World""",,terminal_command
50
+ 49,36661,"TERMINAL",0,0,"]633;CHello World\r\n% \r \r",,terminal_output
51
+ 50,38263,"Untitled-1",97,0,"",plaintext,selection_command
52
+ 51,41690,"Untitled-1",81,0,"",plaintext,selection_command
53
+ 52,41723,"Untitled-1",65,0,"",plaintext,selection_command
54
+ 53,43175,"Untitled-1",81,0,"",plaintext,selection_command
55
+ 54,43454,"Untitled-1",65,0,"",plaintext,selection_command
56
+ 55,43625,"Untitled-1",34,0,"",plaintext,selection_command
57
+ 56,43785,"Untitled-1",2,0,"",plaintext,selection_command
58
+ 57,44036,"Untitled-1",34,0,"",plaintext,selection_command
59
+ 58,44242,"Untitled-1",65,0,"",plaintext,selection_command
60
+ 59,45767,"Untitled-1",34,0,"",plaintext,selection_command
61
+ 60,46072,"Untitled-1",2,0,"",plaintext,selection_command
62
+ 61,53580,"Untitled-1",34,0,"",plaintext,selection_command
63
+ 62,54846,"Untitled-1",2,0,"",plaintext,selection_command
64
+ 63,56678,"Untitled-1",34,0,"",plaintext,selection_command
65
+ 64,62597,"Untitled-1",2,0,"",plaintext,selection_command
66
+ 65,62735,"Untitled-1",1,0,"",plaintext,selection_command
67
+ 66,62897,"Untitled-1",0,0,"",plaintext,selection_command
68
+ 67,68697,"Untitled-1",1,0,"",plaintext,selection_command
69
+ 68,69740,"Untitled-1",2,0,"",plaintext,selection_command
70
+ 69,69985,"Untitled-1",34,0,"",plaintext,selection_command
71
+ 70,70020,"Untitled-1",65,0,"",plaintext,selection_command
72
+ 71,70055,"Untitled-1",81,0,"",plaintext,selection_command
73
+ 72,70087,"Untitled-1",97,0,"",plaintext,selection_command
74
+ 73,70891,"Untitled-1",81,0,"",plaintext,selection_command
75
+ 74,86518,"Untitled-1",97,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
76
+ 75,88609,"Untitled-1",65,0,"",plaintext,selection_command
77
+ 76,89392,"Untitled-1",34,0,"",plaintext,selection_command
78
+ 77,89775,"Untitled-1",2,0,"",plaintext,selection_command
79
+ 78,90495,"Untitled-1",34,0,"",plaintext,selection_command
80
+ 79,90841,"Untitled-1",2,0,"",plaintext,selection_command
81
+ 80,91262,"Untitled-1",34,0,"",plaintext,selection_command
82
+ 81,91473,"Untitled-1",65,0,"",plaintext,selection_command
83
+ 82,92310,"Untitled-1",34,0,"",plaintext,selection_command
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-d5c0f30b-efff-4cba-9421-06a8c78914011759343086476-2025_10_01-20.24.54.587/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-dbb18611-0352-4268-ab16-fb8c7e2b82741763481981825-2025_11_18-17.06.23.992/source.csv ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 2,76,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"5:06:23 PM [info] Activating crowd-code\n5:06:23 PM [info] Recording started\n5:06:23 PM [info] Initializing git provider using file system watchers...\n5:06:24 PM [info] Git repository found\n5:06:24 PM [info] Git provider initialized successfully\n",Log,tab
3
+ 3,183,"extension-output-pdoom-org.crowd-code-#1-crowd-code",245,0,"5:06:24 PM [info] Initial git state: [object Object]\n",Log,content
4
+ 4,4028,"nemo_run/run/api.py",0,0,"# SPDX-FileCopyrightText: Copyright (c) 2024 NVIDIA CORPORATION & AFFILIATES. All rights reserved.\n# SPDX-License-Identifier: Apache-2.0\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import List, Optional, Union\n\nfrom fiddle import Buildable\n\nfrom nemo_run.config import Script, get_type_namespace\nfrom nemo_run.core.execution.base import Executor\nfrom nemo_run.run.experiment import Experiment\nfrom nemo_run.run.plugin import ExperimentPlugin as Plugin\nfrom nemo_run.run.task import direct_run_fn\n\n\ndef run(\n fn_or_script: Union[Buildable, Script],\n executor: Optional[Executor] = None,\n plugins: Optional[Union[Plugin, List[Plugin]]] = None,\n name: str = """",\n dryrun: bool = False,\n direct: bool = False,\n detach: bool = False,\n tail_logs: bool = True,\n log_level: str = ""INFO"",\n):\n """"""\n Runs a single configured function on the specified executor.\n If no executor is specified, it runs the run.Partial function directly\n i.e. equivalent to calling the python function directly.\n\n Examples\n --------\n .. code-block:: python\n\n import nemo_run as run\n\n # Run it directly in the same process\n run.run(configured_fn)\n\n # Do a dryrun\n run.run(configured_fn, dryrun=True)\n\n # Specify a custom executor\n local_executor = LocalExecutor()\n run.run(configured_fn, executor=local_executor)\n\n slurm_executor = run.SlurmExecutor(...)\n run.run(configured_fn, executor=slurm_executor)\n\n """"""\n if not isinstance(fn_or_script, (Buildable, Script)):\n raise TypeError(f""Need a configured Buildable or run.Script. Got {fn_or_script}."")\n\n if direct or executor is None:\n direct_run_fn(fn_or_script, dryrun=dryrun)\n return\n\n if plugins:\n plugins = [plugins] if not isinstance(plugins, list) else plugins\n\n if getattr(fn_or_script, ""is_lazy"", False):\n fn_or_script = fn_or_script.resolve()\n\n default_name = (\n fn_or_script.get_name()\n if isinstance(fn_or_script, Script)\n else get_type_namespace(fn_or_script.__fn_or_cls__)\n )\n name = name or default_name\n with Experiment(title=name, executor=executor, log_level=log_level) as exp:\n exp.add(fn_or_script, tail_logs=tail_logs, plugins=plugins, name=name)\n if dryrun:\n exp.dryrun()\n return\n\n exp.run(detach=detach)\n",python,tab
5
+ 5,13165,"nemo_run/run/api.py",1007,0,"",python,selection_mouse
6
+ 6,14823,"nemo_run/run/api.py",2399,0,"",python,selection_keyboard
7
+ 7,14897,"nemo_run/run/api.py",2895,0,"",python,selection_keyboard
8
+ 8,15761,"nemo_run/run/api.py",2864,0,"",python,selection_command
9
+ 9,16012,"nemo_run/run/api.py",2863,0,"",python,selection_command
10
+ 10,16045,"nemo_run/run/api.py",2844,0,"",python,selection_command
11
+ 11,16078,"nemo_run/run/api.py",2819,0,"",python,selection_command
12
+ 12,16111,"nemo_run/run/api.py",2800,0,"",python,selection_command
13
+ 13,16372,"nemo_run/run/api.py",2721,0,"",python,selection_command
14
+ 14,16582,"nemo_run/run/api.py",2641,0,"",python,selection_command
15
+ 15,20164,"nemo_run/run/api.py",2645,0,"",python,selection_command
16
+ 16,27084,"nemo_run/run/api.py",2725,0,"",python,selection_command
17
+ 17,27290,"nemo_run/run/api.py",2729,0,"",python,selection_command
18
+ 18,27463,"nemo_run/run/api.py",2732,0,"",python,selection_command
19
+ 19,28096,"nemo_run/run/api.py",2733,0,"",python,selection_command
20
+ 20,28484,"nemo_run/run/experiment.py",0,0,"# SPDX-FileCopyrightText: Copyright (c) 2024 NVIDIA CORPORATION & AFFILIATES. All rights reserved.\n# SPDX-License-Identifier: Apache-2.0\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport contextvars\nimport copy\nimport importlib.util\nimport inspect\nimport json\nimport os\nimport pprint\nimport shutil\nimport sys\nimport time\nimport traceback\nfrom concurrent.futures import Future, ThreadPoolExecutor, as_completed\nfrom pathlib import Path\nfrom typing import Optional, Type, Union\n\nimport fiddle as fdl\nimport networkx as nx\nimport rich\nfrom fiddle._src import daglish, diffing\nfrom rich.console import Group\nfrom rich.live import Live\nfrom rich.panel import Panel\nfrom rich.progress import BarColumn, Progress, SpinnerColumn, TaskID, TimeElapsedColumn\nfrom rich.progress import Task as RichTask\nfrom rich.syntax import Syntax\nfrom torchx.specs.api import AppState\n\nimport nemo_run as run\nfrom nemo_run.config import (\n Config,\n ConfigurableMixin,\n Partial,\n Script,\n get_nemorun_home,\n get_type_namespace,\n)\nfrom nemo_run.core.execution.base import Executor\nfrom nemo_run.core.execution.dgxcloud import DGXCloudExecutor\nfrom nemo_run.core.execution.docker import DockerExecutor\nfrom nemo_run.core.execution.lepton import LeptonExecutor\nfrom nemo_run.core.execution.local import LocalExecutor\nfrom nemo_run.core.execution.skypilot import SkypilotExecutor\nfrom nemo_run.core.execution.skypilot_jobs import SkypilotJobsExecutor\nfrom nemo_run.core.execution.slurm import SlurmExecutor\nfrom nemo_run.core.frontend.console.api import CONSOLE, configure_logging, deconfigure_logging\nfrom nemo_run.core.serialization.zlib_json import ZlibJSONSerializer\nfrom nemo_run.core.tunnel.client import SSHTunnel, Tunnel\nfrom nemo_run.core.tunnel.rsync import rsync\nfrom nemo_run.run.job import Job, JobGroup\nfrom nemo_run.run.plugin import ExperimentPlugin\nfrom nemo_run.run.torchx_backend.runner import get_runner\nfrom nemo_run.run.utils import TeeStdoutStderr\n\n_current_experiment: contextvars.ContextVar[""Experiment""] = contextvars.ContextVar(\n ""nemo_current_experiment""\n)\n\n\nclass DummyConsole:\n """"""A dummy console that mimics rich.console.Console but does nothing.""""""\n\n def __getattr__(self, name):\n """"""Return a no-op function for any attribute access.""""""\n\n def no_op(*args, **kwargs):\n pass\n\n return no_op\n\n\nclass Experiment(ConfigurableMixin):\n """"""\n A context manager to launch and manage multiple runs, all using pure Python.\n\n run.Experiment provides researchers with\n a simple and flexible way to create and manage their ML experiments.\n\n Building on the core blocks of nemo_run,\n the Experiment can be used as an umbrella under which a user can\n launch different configured functions on multiple remote clusters.\n\n The Experiment takes care of storing the run metadata,\n launching it on the specified cluster, and syncing the logs and artifacts.\n\n Additionally, the Experiment also provides management tools to easily inspect and reproduce past experiments.\n Some of the use-cases that it enables are listed below:\n\n 1. Check the status and logs of a past experiment\n 2. Reconstruct a past experiment and relaunch it after some changes\n 3. Compare different runs of the same experiment.\n\n This API allows users to programmatically define their experiments.\n To get a glance of the flexibility provided, here are some use cases\n which can be supported by the Experiment in just a few lines of code.\n\n 1. Launch a benchmarking run on different GPUs at the same time in parallel\n 2. Launch a sequential data processing pipeline on a CPU heavy cluster\n 3. Launch hyperparameter grid search runs on a single cluster in parallel\n 4. Launch hyperparameter search runs distributed across all available clusters\n\n The design is heavily inspired from `XManager <https://github.com/google-deepmind/xmanager/blob/main/docs/xm_launch_api_principles.md>`_.\n\n Under the hood, the Experiment metadata is stored in the local filesystem\n inside a user specified directory controlled by get_nemorun_home() env var.\n We will explore making the metadata more persistent in the future.\n\n .. note::\n `Experiment.add` and `Experiment.run` methods inside Experiment can currently only be used within its context manager.\n\n Examples\n --------\n .. code-block:: python\n\n # An experiment that runs a pre-configured training example\n # on multiple GPU specific clusters (A100 and H100 shown here) in parallel using torchrun\n # Assumes that example_to_run is pre-configured using run.Partial\n with run.Experiment(""example-multiple-gpus"", executor=""h100_cluster"") as exp:\n # Set up the run on H100\n # Setting up a single task is identical to setting up a single run outside the experiment\n h100_cluster: run.SlurmExecutor = exp.executor.clone()\n h100_cluster.nodes = 2\n\n # torchrun manages the processes on a single node\n h100_cluster.ntasks_per_node = 1\n h100_cluster.gpus_per_task = 8\n\n h100_cluster.packager.subpath = ""subpath/to/your/code/repo""\n h100_cluster.launcher = ""torchrun""\n\n exp.add(\n ""example_h100"",\n fn=example_to_run,\n tail_logs=True,\n executor=h100_cluster,\n )\n\n # Set up the run on A100\n a100_cluster: run.Config[SlurmExecutor] = h100_cluster.clone()\n a100_cluster.tunnel = run.Config(\n SSHTunnel,\n host=os.environ[""A100_HOST""],\n user=""your_user_in_cluster"",\n identity=""path_to_your_ssh_key""\n )\n\n exp.add(\n ""example_a100"",\n fn=example_to_run,\n tail_logs=True,\n executor=a100_cluster,\n )\n\n # Runs all the task in the experiment.\n # By default, all tasks will be run in parallel if all different executors support parallel execution.\n # You can set sequential=True to run the tasks sequentially.\n exp.run()\n\n # Upon exiting the context manager, the Experiment will automatically wait for all tasks to complete,\n # and optionally tail logs for tasks that have tail_logs=True.\n # A detach mode (if the executors support it) will be available soon.\n # Once all tasks have completed,\n # the Experiment will display a status table and clean up resources like ssh tunnels.\n\n # You can also manage the experiment at a later point in time\n exp = run.Experiment.from_title(""example-multiple-gpus"")\n exp.status()\n exp.logs(task_id=""example_a100"")\n\n """"""\n\n GOODBYE_MESSAGE_PYTHON = """"""\n# The experiment was run with the following tasks: {tasks}\n# You can inspect and reconstruct this experiment at a later point in time using:\nexperiment = run.Experiment.from_id(""{exp_id}"")\nexperiment.status() # Gets the overall status\nexperiment.logs(""{tasks[0]}"") # Gets the log for the provided task\nexperiment.cancel(""{tasks[0]}"") # Cancels the provided task if still running\n""""""\n\n GOODBYE_MESSAGE_BASH = """"""\n# You can inspect this experiment at a later point in time using the CLI as well:\nnemo experiment status {exp_id}\nnemo experiment logs {exp_id} 0\nnemo experiment cancel {exp_id} 0\n""""""\n _PARALLEL_SUPPORTED_EXECUTORS = (\n SlurmExecutor,\n LocalExecutor,\n SkypilotExecutor,\n SkypilotJobsExecutor,\n DockerExecutor,\n DGXCloudExecutor,\n LeptonExecutor,\n )\n _DETACH_SUPPORTED_EXECUTORS = (\n SlurmExecutor,\n SkypilotExecutor,\n SkypilotJobsExecutor,\n DGXCloudExecutor,\n LeptonExecutor,\n )\n _DEPENDENCY_SUPPORTED_EXECUTORS = (SlurmExecutor,)\n _RUNNER_DEPENDENT_EXECUTORS = (LocalExecutor,)\n _CONFIG_FILE = ""_CONFIG""\n _VERSION_FILE = ""_VERSION""\n _TASK_FILE = ""_TASKS""\n _DONE_FILE = ""_DONE""\n _TUNNELS_FILE = ""_TUNNELS""\n _current_experiment_token: Optional[contextvars.Token]\n\n @classmethod\n def catalog(\n cls: Type[""Experiment""],\n title: str = """",\n ) -> list[str]:\n """"""\n List all experiments inside get_nemorun_home(), optionally with the provided title.\n """"""\n parent_dir = os.path.join(get_nemorun_home(), ""experiments"", title)\n return _get_sorted_dirs(parent_dir)\n\n @classmethod\n def _from_config(cls: Type[""Experiment""], exp_dir: str) -> ""Experiment"":\n id = os.path.basename(exp_dir)\n with open(os.path.join(exp_dir, cls._CONFIG_FILE), ""r"") as f:\n config = f.read()\n\n if not config:\n raise ValueError(f""Experiment {id} not found."")\n\n serializer = ZlibJSONSerializer()\n cfg: Config[""Experiment""] = fdl.cast(Config, serializer.deserialize(config))\n if ""id"" not in cfg.__arguments__:\n cfg.id = id\n\n cfg._reconstruct = True\n\n exp: ""Experiment"" = fdl.build(cfg)\n exp._jobs = exp._load_jobs()\n try:\n exp.tunnels = exp._load_tunnels()\n except Exception as e:\n exp.console.log(\n f""Exception {e} loading tunnels for experiment {id}, will continue without loading tunnels.""\n )\n\n return exp\n\n @classmethod\n def from_id(\n cls: Type[""Experiment""],\n id: str,\n ) -> ""Experiment"":\n """"""\n Reconstruct an experiment with the specified id.\n """"""\n title, _, _ = id.rpartition(""_"")\n parent_dir = os.path.join(get_nemorun_home(), ""experiments"", title)\n exp_dir = os.path.join(parent_dir, id)\n\n assert os.path.isdir(exp_dir), f""Experiment {id} not found.""\n\n exp = cls._from_config(exp_dir)\n return exp\n\n @classmethod\n def from_title(\n cls: Type[""Experiment""],\n title: str,\n ) -> ""Experiment"":\n """"""\n Reconstruct an experiment with the specified title.\n """"""\n parent_dir = os.path.join(get_nemorun_home(), ""experiments"", title)\n exp_dir = _get_latest_dir(parent_dir)\n\n assert os.path.isdir(exp_dir), f""Experiment {id} not found.""\n\n exp = cls._from_config(exp_dir)\n return exp\n\n def __init__(\n self,\n title: str,\n executor: Executor | None = None, # type: ignore\n id: str | None = None,\n log_level: str = ""INFO"",\n _reconstruct: bool = False,\n jobs: list[Job | JobGroup] | None = None,\n base_dir: str | None = None,\n clean_mode: bool = False,\n enable_goodbye_message: bool = True,\n threadpool_workers: int = 16,\n skip_status_at_exit: bool = False,\n serialize_metadata_for_scripts: bool = True,\n ) -> None:\n """"""\n Initializes an experiment run by creating its metadata directory and saving the experiment config.\n\n Args:\n title: Title or name for the experiment\n executor: Any executor that subclasses run.Executor and is supported by NeMo-Run.\n This will be used as the default executor for tasks if an explicit one is not specified.\n Users can also clone this and make task specific executor changes.\n id (Optional): Unique id for the experiment run.\n If not specified, will be set automatically based on the current timestamp.\n log_level: Set log level for the experiment. Defaults to WARN.\n _reconstruct: Generally, the user does not need to specify this flag.\n This is only set to True when using run.Experiment.from_dir.\n clean_mode: If True, disables all console output (logs, progress bars, etc.). Defaults to False.\n enable_goodbye_message: if True, prints goodbye message after submitting job. Defaults to True.\n """"""\n configure_logging(level=log_level)\n self._reconstruct = _reconstruct\n if _reconstruct:\n assert id, ""Cannot reconstruct an experiment without id.""\n\n self._title = title\n self._id = id or f""{title}_{int(time.time())}""\n self._enable_goodbye_message = enable_goodbye_message\n self._threadpool_workers = threadpool_workers\n self._skip_status_at_exit = skip_status_at_exit\n self._serialize_metadata_for_scripts = serialize_metadata_for_scripts\n\n base_dir = str(base_dir or get_nemorun_home())\n self._exp_dir = os.path.join(base_dir, ""experiments"", title, self._id)\n\n self.log_level = log_level\n self._runner = get_runner(component_defaults=None, experiment=self)\n\n if not _reconstruct:\n self.executor = executor if executor else LocalExecutor()\n else:\n assert isinstance(executor, Executor)\n self.executor = executor\n\n self._jobs: list[Job | JobGroup] = jobs or []\n self.tunnels: dict[str, Tunnel] = {}\n self.console = CONSOLE\n self.clean_mode = clean_mode\n if self.clean_mode:\n self.console = DummyConsole()\n self._launched = False\n self._live_progress = None\n self._current_experiment_token = None\n\n def to_config(self) -> Config:\n return Config(\n self.__class__,\n title=self._title,\n id=self._id,\n executor=self.executor.to_config(),\n log_level=self.log_level,\n clean_mode=self.clean_mode,\n threadpool_workers=self._threadpool_workers,\n enable_goodbye_message=self._enable_goodbye_message,\n skip_status_at_exit=self._skip_status_at_exit,\n serialize_metadata_for_scripts=self._serialize_metadata_for_scripts,\n )\n\n def _save_experiment(self, exist_ok: bool = False):\n os.makedirs(self._exp_dir, exist_ok=exist_ok)\n self._save_config()\n\n def _save_config(self):\n with open(os.path.join(self._exp_dir, self.__class__._CONFIG_FILE), ""w+"") as f:\n f.write(ZlibJSONSerializer().serialize(self.to_config()))\n\n with open(os.path.join(self._exp_dir, self.__class__._VERSION_FILE), ""w+"") as f:\n f.write(f""{run.__version__}\n"")\n\n def _save_tunnels(self):\n serializer = ZlibJSONSerializer()\n serialized_tunnels = {\n k: serializer.serialize(v.to_config()) for k, v in self.tunnels.items()\n }\n with open(os.path.join(self._exp_dir, self.__class__._TUNNELS_FILE), ""w+"") as f:\n json.dump(serialized_tunnels, f)\n\n def _load_tunnels(self) -> dict[str, Tunnel]:\n with open(os.path.join(self._exp_dir, self.__class__._TUNNELS_FILE)) as f:\n serialized_tunnels = json.load(f)\n serializer = ZlibJSONSerializer()\n return {k: fdl.build(serializer.deserialize(v)) for k, v in serialized_tunnels.items()}\n\n def _save_jobs(self):\n serialized_jobs = list(map(lambda job: job.serialize(), self.jobs))\n with open(os.path.join(self._exp_dir, self.__class__._TASK_FILE), ""w+"") as f:\n json.dump(serialized_jobs, f)\n\n if ""__main__"" in sys.modules:\n main_module = sys.modules[""__main__""]\n try:\n with open(os.path.join(self._exp_dir, ""__main__.py""), ""w+"") as f:\n f.write(inspect.getsource(main_module))\n except TypeError:\n ...\n\n def _load_jobs(self) -> list[Job | JobGroup]:\n with open(os.path.join(self._exp_dir, self._TASK_FILE)) as f:\n serialized_jobs = json.load(f)\n\n serializer = ZlibJSONSerializer()\n jobs = []\n for job_cfg, task_cfg in serialized_jobs:\n job_cfg = serializer.deserialize(job_cfg)\n\n job: Job | JobGroup = fdl.build(job_cfg)\n if isinstance(job, Job):\n job.task = task_cfg # type: ignore\n elif isinstance(job, JobGroup):\n job.tasks = task_cfg # type: ignore\n else:\n raise ValueError(f""Unknown task type: {task_cfg.__fn_or_cls__}"")\n\n jobs.append(job)\n\n return jobs\n\n def _prepare(self, exist_ok: bool = False):\n self._save_experiment(exist_ok=exist_ok)\n\n for job in self.jobs:\n job.prepare(serialize_metadata_for_scripts=self._serialize_metadata_for_scripts)\n\n self._save_jobs()\n\n def _add_single_job(\n self,\n task: Union[Partial, Script],\n executor: Executor,\n name: str = """",\n plugins: Optional[list[ExperimentPlugin]] = None,\n tail_logs: bool = False,\n dependencies: Optional[list[str]] = None,\n ) -> str:\n if isinstance(task, Script):\n default_name = task.get_name()\n else:\n default_name = get_type_namespace(task.__fn_or_cls__)\n\n reuse_job_dir = True if name else False\n name = name or default_name\n if any(map(lambda job: job.id == name, self.jobs)):\n task_id = f""{name}_{len(self.jobs)}""\n else:\n task_id = name\n\n self._validate_task(task_info=task_id, task=task)\n\n executor = executor.clone()\n executor.assign(\n self._id,\n self._exp_dir,\n task_id=task_id,\n task_dir=name if reuse_job_dir else task_id,\n )\n\n cloned = copy.deepcopy(task) if isinstance(task, Script) else task.clone()\n job = Job(\n id=task_id,\n task=cloned,\n executor=executor,\n plugins=plugins,\n tail_logs=tail_logs,\n dependencies=dependencies or [],\n )\n plugins = plugins or []\n for plugin in plugins:\n plugin.assign(self._id)\n plugin.setup(cloned, executor)\n\n self._jobs.append(job)\n return job.id\n\n def _add_job_group(\n self,\n tasks: list[Partial | Script],\n executor: list[Executor] | Executor,\n name: str,\n plugins: Optional[list[ExperimentPlugin]] = None,\n tail_logs: bool = False,\n dependencies: Optional[list[str]] = None,\n ) -> str:\n if any(map(lambda task: task.id == name, self.jobs)):\n task_id = f""{name}_{len(self.jobs)}""\n else:\n task_id = name\n\n for i, _task in enumerate(tasks):\n self._validate_task(task_info=f""Job Group: {task_id}, job index: {i}"", task=_task)\n\n executors = executor if isinstance(executor, list) else [executor]\n cloned_executors = []\n for executor in executors:\n new_executor = executor.clone()\n cloned_executors.append(new_executor)\n new_executor.assign(self._id, self._exp_dir, task_id, task_dir=name)\n\n cloned_tasks = []\n for task in tasks:\n cloned_task = copy.deepcopy(task) if isinstance(task, Script) else task.clone()\n cloned_tasks.append(cloned_task)\n\n job_group = JobGroup(\n id=task_id,\n tasks=cloned_tasks,\n executors=cloned_executors,\n plugins=plugins,\n tail_logs=tail_logs,\n dependencies=dependencies or [],\n )\n plugins = plugins or []\n for plugin in plugins:\n for i, task in enumerate(cloned_tasks):\n _executor = job_group.executors if job_group._merge else job_group.executors[i] # type: ignore\n assert isinstance(_executor, Executor)\n plugin.setup(task, _executor)\n\n self._jobs.append(job_group)\n return job_group.id\n\n def _validate_task(self, task_info: str, task: Union[Partial, Script]) -> None:\n valid = True\n message = """"\n if isinstance(task, Partial):\n serializer = ZlibJSONSerializer()\n serialized = serializer.serialize(task)\n deserialized = serializer.deserialize(serialized)\n diff = diffing.build_diff(deserialized, task)\n diff = {\n daglish.path_str(d.target): (d.new_value if hasattr(d, ""new_value"") else None) # type: ignore\n for d in diff.changes\n }\n if deserialized != task:\n valid = False\n message += f""""""\nDeserialized task does not match original task. The following paths in your task need to be wrapped in `run.Config` or `run.Partial`:\n\n{pprint.PrettyPrinter(indent=4).pformat(diff)}\n\nFor more information about `run.Config` and `run.Partial`, please refer to https://github.com/NVIDIA-NeMo/Run/blob/main/docs/source/guides/configuration.md.\n""""""\n if not valid:\n raise RuntimeError(f""Failed to validate task {task_info}.\n{message}"")\n\n def add(\n self,\n task: Union[Partial, Script] | list[Union[Partial, Script]],\n executor: Executor | list[Executor] | None = None,\n name: str = """",\n plugins: Optional[list[ExperimentPlugin]] = None,\n tail_logs: bool = False,\n dependencies: Optional[list[str]] = None,\n ) -> str:\n """"""\n Add a configured function along with its executor config to the experiment.\n """"""\n assert _current_experiment.get(None) == self, (\n ""Using Experiment without it's context manager is not permitted.""\n )\n\n job_ids = set([job.id for job in self.jobs])\n for dep in dependencies or []:\n assert dep in job_ids, f""Dependency {dep} not found.""\n\n executor = executor or self.executor\n if not isinstance(task, list):\n assert executor and isinstance(executor, Executor)\n job_id = self._add_single_job(\n task,\n executor,\n name,\n plugins=plugins,\n tail_logs=tail_logs,\n dependencies=dependencies.copy() if dependencies else None,\n )\n else:\n assert name, ""name is required for task group.""\n job_id = self._add_job_group(\n task,\n executor,\n name,\n plugins=plugins,\n tail_logs=tail_logs,\n dependencies=dependencies.copy() if dependencies else None,\n )\n\n return job_id\n\n def dryrun(self, log: bool = True, exist_ok: bool = False, delete_exp_dir: bool = True):\n """"""\n Logs the raw scripts that will be executed for each task.\n """"""\n if log:\n self.console.log(f""[bold magenta]Experiment {self._id} dryrun..."")\n\n self._prepare(exist_ok=exist_ok)\n\n for job in self.jobs:\n if isinstance(job, Job):\n if log:\n self.console.log(f""[bold magenta]Task {job.id}\n"")\n elif isinstance(job, JobGroup):\n if log:\n self.console.log(f""[bold magenta]Task Group {job.id}\n"")\n job.launch(wait=False, runner=self._runner, dryrun=True, direct=False, log_dryrun=log)\n\n if delete_exp_dir:\n shutil.rmtree(self._exp_dir)\n\n def run(\n self,\n sequential: bool = False,\n detach: bool = False,\n tail_logs: bool = False,\n direct: bool = False,\n ):\n """"""\n Runs all the tasks in the experiment.\n\n By default, all tasks are run in parallel.\n\n If sequential=True, all tasks will be run one after the other.\n The order is based on the order in which they were added.\n\n Parallel mode only works if all executors in the experiment support it.\n Currently, all executors support parallel mode.\n\n In sequential mode, if all executor supports dependencies, then all tasks will be scheduled at once\n by specifying the correct dependencies to each task.\n Otherwise, the experiment.run call will block and each task that is scheduled will be executed sequentially.\n In this particular case, we cannot guarantee the state of the experiment if the process exits in the middle.\n\n Currently, only the slurm executor supports dependencies.\n\n Args:\n sequential: If True, runs all tasks sequentially in the order they were added. Defaults to False.\n detach: If True, detaches from the process after launching the tasks. Only supported for Slurm and Skypilot. Defaults to False.\n tail_logs: If True, tails logs from all tasks in the experiment. If False, relies on task specific setting. Defaults to False.\n direct: If True, runs all tasks in the experiment sequentially in the same process. Note that if direct=True, then sequential also will be True. Defaults to False.\n """"""\n assert _current_experiment.get(None) == self, (\n ""Using Experiment without it's context manager is not permitted.""\n )\n\n if self._launched:\n self.console.log(""[bold magenta]Experiment already running..."")\n return\n\n if self._reconstruct:\n self.console.log(""[bold magenta]Experiment in inspection mode..."")\n return\n\n # Prepare experiment before running\n\n # in case of multi-node execution with LocalExecutor+torchrun+slurm, run only on first rank\n if int(os.getenv(""SLURM_PROCID"", 0)) == 0:\n self._prepare()\n\n if direct:\n self.console.log(\n ""[bold magenta]Running the experiment with direct=True. ""\n ""This will launch all jobs sequentially in the same process.""\n )\n if not self.jobs:\n self.console.log(""[bold red]No jobs to run in this experiment."")\n return\n\n assert all(map(lambda job: isinstance(job, Job), self.jobs)), (\n ""Jobs in this experiment contain JobGroup which cannot be run directly for now.""\n )\n\n assert all(map(lambda job: not job.dependencies, self.jobs)), (\n ""Jobs in this experiment contain dependencies which cannot be run directly for now.""\n )\n\n for job in self.jobs:\n assert isinstance(job, Job)\n with TeeStdoutStderr(\n os.path.join(job.executor.job_dir, f""log_{job.id}_direct_run.out"")\n ):\n job.launch(wait=True, direct=True, runner=self._runner)\n\n self._save_jobs()\n self._launched = any(map(lambda job: job.launched, self.jobs))\n self._direct = True\n return\n\n executors = set()\n for job in self.jobs:\n if isinstance(job, Job):\n executors.add(job.executor.__class__)\n elif isinstance(job, JobGroup):\n if isinstance(job.executors, list):\n for executor in job.executors:\n executors.add(executor.__class__)\n else:\n executors.add(job.executors.__class__)\n\n if detach and any(map(lambda x: x not in self._DETACH_SUPPORTED_EXECUTORS, executors)):\n self.console.log(\n ""[bold red] Cannot detach from this experiment. Please keep it running until completion.""\n )\n detach = False\n\n is_dag = any(map(lambda job: len(job.dependencies) > 0, self.jobs))\n assert not (is_dag and sequential), (\n ""Jobs in this experiment have dependencies, they cannot be run sequentially. Set sequential=False.""\n )\n\n if sequential:\n for i in range(1, len(self.jobs)):\n self.jobs[i].dependencies.append(self.jobs[i - 1].id)\n\n self.dryrun(log=False, exist_ok=True, delete_exp_dir=False)\n for tunnel in self.tunnels.values():\n if isinstance(tunnel, SSHTunnel):\n tunnel.connect()\n assert tunnel.session, f""SSH tunnel {tunnel.key} failed to connect.""\n rsync(tunnel.session, self._exp_dir, os.path.dirname(tunnel.job_dir))\n\n symlink_cmds = []\n for packaging_job in tunnel.packaging_jobs.values():\n if packaging_job.symlink:\n symlink_cmds.append(packaging_job.symlink_cmd())\n\n if symlink_cmds:\n tunnel.run("" && "".join(symlink_cmds))\n\n self._save_tunnels()\n\n return self._run_dag(detach=detach, tail_logs=tail_logs, executors=executors)\n\n def _run_dag(self, detach: bool, tail_logs: bool, executors: set[Executor]):\n job_map = {job.id: job for job in self._jobs}\n graph = nx.DiGraph()\n job_ids = set([job.id for job in self.jobs])\n for job in self.jobs:\n graph.add_node(job.id, job=job)\n for dep in job.dependencies:\n assert dep in job_ids, f""Dependency {dep} not found in job list {job_ids}.""\n graph.add_edge(dep, job.id)\n\n assert nx.is_directed_acyclic_graph(graph), ""Jobs have cyclic dependencies.""\n order = [sorted(generation) for generation in nx.topological_generations(graph)]\n add_deps = False\n if len(order) > 1:\n if all(map(lambda x: x in self._DEPENDENCY_SUPPORTED_EXECUTORS, executors)):\n wait = False\n add_deps = True\n self.detach = detach\n else:\n wait = True\n if len(self.jobs) > 1:\n self.console.log(\n f""[bold cyan]Dependencies not supported for atleast one of {executors}.""\n ""All jobs will be run one after the other based on their dependencies, please keep the process alive.""\n )\n if detach:\n self.console.log(\n ""[bold red] Cannot detach from this experiment. Please keep it running until completion.""\n )\n else:\n # All jobs will be executed in parallel\n assert all(map(lambda x: x in self._PARALLEL_SUPPORTED_EXECUTORS, executors)), (\n f""Parallel mode not supported for atleast one of {executors}. Set sequential=True.""\n )\n wait = False\n self.detach = detach\n\n for level in order:\n # Launch jobs in this level concurrently since they are independent\n\n def _set_context(ctx: contextvars.Context):\n for var, value in ctx.items():\n var.set(value)\n\n ctx = contextvars.copy_context()\n\n def _launch(node: str):\n job: Job | JobGroup = job_map[node]\n self.console.log(f""[bold cyan]Launching job {job.id} for experiment {self._title}"")\n if tail_logs:\n job.tail_logs = True\n\n try:\n if add_deps:\n deps = []\n for dep_id in job.dependencies:\n dep = job_map[dep_id]\n handle = dep.handle\n assert dep.launched and handle, (\n f""Dependency {dep.id} for {job.id} not yet launched.""\n )\n deps.append(handle)\n\n job.executor.dependencies = deps # type: ignore\n\n job.launch(wait=False, runner=self._runner)\n return job\n\n except Exception as e:\n self.console.log(f""Error running job {job.id}: {e}"")\n raise e\n\n launched_jobs: list[Job | JobGroup] = []\n with ThreadPoolExecutor(\n initializer=_set_context, initargs=(ctx,), max_workers=self._threadpool_workers\n ) as pool:\n futures = [pool.submit(_launch, node) for node in level]\n for future in as_completed(futures):\n launched_jobs.append(future.result())\n\n if wait:\n self._wait_for_jobs(jobs=launched_jobs)\n\n self._save_jobs()\n self._launched = any(map(lambda job: job.launched, self.jobs))\n self._waited = wait\n\n def _wait_for_jobs(self, jobs: list[Job | JobGroup]):\n def set_context(context: contextvars.Context):\n for var, value in context.items():\n var.set(value)\n\n context = contextvars.copy_context()\n with ThreadPoolExecutor(initializer=set_context, initargs=(context,)) as executor:\n futures: dict[Future, Job | JobGroup] = {}\n for job in jobs:\n if isinstance(job, Job):\n handle_exists = job.handle\n else:\n handle_exists = len(job.handles) > 0 and all(job.handles)\n\n if job.launched and handle_exists:\n self._initialize_live_progress()\n self._add_progress(job=job)\n future = executor.submit(\n job.wait,\n runner=self._runner\n if isinstance(\n job.executor,\n self._RUNNER_DEPENDENT_EXECUTORS,\n )\n else get_runner(),\n )\n futures[future] = job\n\n for future in as_completed(futures.keys()):\n job = futures[future]\n try:\n future.result()\n self._update_progress(job, job.state)\n except Exception as e:\n self.console.log(f""Exception while waiting for Job {job.id}: {e}"")\n self.console.log(*traceback.format_exception(e))\n self._update_progress(job, AppState.UNKNOWN)\n finally:\n job.cleanup()\n\n def _initialize_tunnels(self, extract_from_executors: bool = False):\n if extract_from_executors:\n for job in self.jobs:\n if (\n isinstance(job.executor, SlurmExecutor)\n and job.executor.tunnel.key not in self.tunnels\n ):\n self.tunnels[job.executor.tunnel.key] = job.executor.tunnel\n\n for tunnel in self.tunnels.values():\n if isinstance(tunnel, SSHTunnel):\n tunnel.connect()\n assert tunnel.session, f""SSH tunnel {tunnel.key} failed to connect.""\n\n def status(self, return_dict: bool = False) -> Optional[dict[str, dict[str, str]]]:\n """"""\n Prints a table specifying the status of all tasks.\n\n .. note::\n status is not supported for local executor\n and the status for a task using the local executor\n will be listed as UNKNOWN in most cases\n """"""\n _set_current_experiment = False\n if not self._current_experiment_token:\n _current_experiment.set(self)\n _set_current_experiment = True\n\n def _get_job_info_and_dict(\n idx: int, job: Job | JobGroup\n ) -> tuple[list[str], dict[str, str]]:\n job_info = []\n job_info.append(f""[bold green]Task {idx}[/bold green]: [bold orange1]{job.id}"")\n job_info.append(\n f""- [bold green]Status[/bold green]: {str(job.status(runner=self._runner))}""\n )\n job_info.append(f""- [bold green]Executor[/bold green]: {job.executor.info()}"")\n\n try:\n _, _, path_str = job.handle.partition(""://"")\n path = path_str.split(""/"")\n app_id = path[1]\n except Exception:\n app_id = """"\n\n job_info.append(f""- [bold green]Job id[/bold green]: {app_id}"")\n directory_info = [\n ""- [bold green]Local Directory[/bold green]: "" + job.executor.job_dir,\n ]\n job_dict = {\n ""name"": job.id,\n ""status"": job.status(runner=self._runner),\n ""executor"": job.executor.info(),\n ""job_id"": app_id,\n ""handle"": job.handle,\n ""local_dir"": job.executor.job_dir,\n }\n\n if isinstance(job.executor, SlurmExecutor) and isinstance(\n job.executor.tunnel, SSHTunnel\n ):\n directory_info.extend(\n [\n ""- [bold green]Remote Directory[/bold green]: ""\n + os.path.join(\n job.executor.tunnel.job_dir,\n Path(job.executor.job_dir).name,\n ),\n ]\n )\n job_dict[""remote_dir""] = os.path.join(\n job.executor.tunnel.job_dir,\n Path(job.executor.job_dir).name,\n )\n job_info.extend(directory_info)\n return job_info, job_dict\n\n self._initialize_tunnels(extract_from_executors=True)\n try:\n result_dict = {}\n job_infos: list[Group | None] = [None] * len(self.jobs)\n\n # Parallelize IO-bound status retrieval across jobs\n def _collect(arg):\n idx, job = arg\n job_info, job_dict = _get_job_info_and_dict(idx, job)\n return idx, job.id, job_info, job_dict\n\n # Propagate context variables to worker threads so helpers that rely on them keep working\n def _set_context(ctx: contextvars.Context):\n for var, value in ctx.items():\n var.set(value)\n\n ctx = contextvars.copy_context()\n with ThreadPoolExecutor(\n initializer=_set_context, initargs=(ctx,), max_workers=self._threadpool_workers\n ) as pool:\n futures = [pool.submit(_collect, (idx, job)) for idx, job in enumerate(self.jobs)]\n for future in as_completed(futures):\n idx, job_id, job_info, job_dict = future.result()\n job_infos[idx] = Group(*job_info)\n result_dict[job_id] = job_dict\n\n # Remove potential None slots (should not occur)\n job_infos = [ji for ji in job_infos if ji is not None]\n\n if return_dict:\n return result_dict\n\n self.console.print()\n self.console.print(\n f""[bold green]Experiment Status for[/bold green] [bold orange1]{self._id}"",\n new_line_start=True,\n )\n for job_info in job_infos:\n self.console.print(job_info, soft_wrap=True, new_line_start=True, highlight=False)\n self.console.print()\n finally:\n if _set_current_experiment and self._current_experiment_token:\n _current_experiment.reset(self._current_experiment_token)\n self._current_experiment_token = None\n\n def cancel(self, job_id: str):\n """"""\n Cancels an existing job if still running.\n """"""\n _set_current_experiment = False\n if not self._current_experiment_token:\n _current_experiment.set(self)\n _set_current_experiment = True\n\n self.console.log(f""[bold cyan]Cancelling {job_id} if still running"")\n try:\n job = next(filter(lambda x: x.id == job_id, self.jobs))\n job.cancel(runner=self._runner)\n except StopIteration:\n self.console.log(f""[bold red]Job {job_id} not found"")\n except Exception as e:\n self.console.log(f""[bold red]Failed to cancel {job_id}\nError: {e}\n"")\n self.console.log(*traceback.format_exception(e))\n finally:\n if _set_current_experiment and self._current_experiment_token:\n _current_experiment.reset(self._current_experiment_token)\n self._current_experiment_token = None\n\n def logs(self, job_id: str, regex: str | None = None):\n """"""\n Prints the logs of the specified job_id, optionally filtered by regex.\n """"""\n _set_current_experiment = False\n if not self._current_experiment_token:\n _current_experiment.set(self)\n _set_current_experiment = True\n\n self.console.log(f""[bold cyan]Fetching logs for {job_id}"")\n try:\n job = next(filter(lambda x: x.id == job_id, self.jobs))\n if isinstance(job, Job) and job.handle.endswith(""direct_run""):\n self.console.log(""This job was run with direct=True."")\n self.console.log(\n f""Logs may be present in task directory at:\n[bold]{job.executor.job_dir}.""\n )\n return\n\n try:\n job.logs(runner=self._runner, regex=regex)\n except Exception as e:\n self.console.log(f""[bold red]Failed to get logs for {job_id}\nError: {e}\n"")\n self.console.log(\n f""Logs may be present in job directory at:\n[bold]{job.executor.job_dir}.""\n )\n except StopIteration:\n self.console.log(f""[bold red]Job {job_id} not found"")\n finally:\n if _set_current_experiment and self._current_experiment_token:\n _current_experiment.reset(self._current_experiment_token)\n self._current_experiment_token = None\n\n def reset(self) -> ""Experiment"":\n """"""\n Resets an experiment to make it ready for a relaunch.\n Only works if the current experiment run has already been launched.\n """"""\n if not self._reconstruct and not os.path.isfile(\n os.path.join(self._exp_dir, self._DONE_FILE)\n ):\n self.console.log(\n f""[bold magenta]Experiment {self._id} has not run yet, skipping reset...""\n )\n return self\n\n old_id, old_exp_dir, old_launched = self._id, self._exp_dir, self._launched\n self._id = f""{self._title}_{int(time.time())}""\n self._exp_dir = os.path.join(get_nemorun_home(), ""experiments"", self._title, self._id)\n self._launched = False\n self._live_progress = None\n\n jobs = self._jobs\n self._jobs = []\n serializer = ZlibJSONSerializer()\n _set_current_experiment = False\n if not self._current_experiment_token:\n _current_experiment.set(self)\n _set_current_experiment = True\n\n try:\n if ""__external_main__"" not in sys.modules:\n maybe_load_external_main(old_exp_dir)\n\n for job in jobs:\n if isinstance(job, Job):\n if isinstance(job.task, str):\n _task = serializer.deserialize(job.task)\n if _task.__fn_or_cls__ == Script:\n job.task = fdl.build(_task)\n else:\n job.task = _task # type: ignore\n\n self.add(\n job.task,\n job.executor,\n name=job.id,\n tail_logs=job.tail_logs,\n )\n else:\n if isinstance(job.tasks, str):\n tasks = serializer.deserialize(job.tasks)\n job.tasks = [\n fdl.build(task) if task.__fn_or_cls__ == Script else task\n for task in tasks\n ]\n\n self.add(\n job.tasks,\n job.executors,\n name=job.id,\n tail_logs=job.tail_logs,\n )\n except Exception as e:\n self.console.log(\n f""[bold magenta]Failed resetting Experiment {self._id} due to error: {e}""\n )\n # Double check exp dir is unchanged\n new_path = os.path.join(get_nemorun_home(), ""experiments"", self._title, self._id)\n if self._exp_dir == new_path and new_path != old_exp_dir:\n shutil.rmtree(self._exp_dir)\n\n self._id = old_id\n self._exp_dir = old_exp_dir\n self._launched = old_launched\n self._jobs = self._load_jobs()\n finally:\n if _set_current_experiment and self._current_experiment_token:\n _current_experiment.reset(self._current_experiment_token)\n self._current_experiment_token = None\n\n self._reconstruct = False\n return self\n\n def _initialize_live_progress(self):\n if not self._live_progress:\n # Disable live progress if we are tailing logs for any task\n # as tty output consistency can not be guaranteed as of now\n if self.clean_mode or any(map(lambda job: job.tail_logs, self.jobs)):\n return\n\n assert isinstance(self.console, rich.console.Console)\n self._progress = Progress(\n ""{task.description}"",\n SpinnerColumn(),\n BarColumn(bar_width=None),\n TimeElapsedColumn(),\n )\n self._exp_panel = Panel(\n self._progress,\n title=f""[b]{self._id}"",\n padding=(1, 3),\n )\n self._task_progress: dict[str, TaskID] = {}\n self._live_progress = Live(self._exp_panel, console=self.console, refresh_per_second=10)\n self._live_progress.start(refresh=True)\n\n def _add_progress(self, job: Job | JobGroup):\n if self._live_progress:\n self._task_progress[job.id] = self._progress.add_task(\n f""[bold green]{job.id}"", total=None\n )\n\n def _update_progress(self, job: Job | JobGroup, state: AppState):\n if self._live_progress:\n color = ""[bold green]"" if state == AppState.SUCCEEDED else ""[bold red]""\n task_progress_id = self._task_progress[job.id]\n self._progress.stop_task(task_progress_id)\n self._progress.update(\n task_progress_id,\n description=f""{color}{job.id} {state}"",\n )\n progress_task: RichTask = self._progress._tasks[task_progress_id]\n progress_task.finished_time = progress_task.elapsed\n progress_task.completed = progress_task.elapsed or 0.0\n progress_task.total = progress_task.elapsed\n\n self._progress.refresh()\n\n def _cleanup(self, tunnels: bool = True):\n if tunnels and hasattr(self, ""tunnels""):\n for tunnel in self.tunnels.values():\n try:\n tunnel.cleanup()\n except Exception:\n ...\n\n self._runner.close()\n\n if (\n _current_experiment is not None\n and _current_experiment.get(None)\n and self._current_experiment_token\n ):\n _current_experiment.reset(self._current_experiment_token)\n self._current_experiment_token = None\n\n def __enter__(self) -> ""Experiment"":\n self._current_experiment_token = _current_experiment.set(self)\n self.console.rule(\n f""[bold magenta]Entering Experiment {self._title} with id: {self._id}"",\n )\n return self\n\n def __exit__(self, exc_type, exc_value, tb):\n try:\n if hasattr(self, ""detach"") and self.detach:\n self.console.rule(f""[bold magenta]Detaching from Experiment {self._id}."")\n self.console.log(\n ""Task specific cleanup won't be run.\n""\n ""Ephemeral logs and artifacts may be lost."",\n )\n\n if self._launched and not self._skip_status_at_exit:\n self.status()\n return\n\n if self._launched:\n if hasattr(self, ""_direct"") and self._direct:\n self.console.rule(\n f""[bold magenta]Direct run Experiment {self._id}"",\n )\n if not self._skip_status_at_exit:\n self.status()\n return\n\n if hasattr(self, ""_waited"") and self._waited:\n self.console.rule(\n f""[bold magenta]Done waiting for Experiment {self._id}"",\n )\n if not self._skip_status_at_exit:\n self.status()\n return\n\n self.console.rule(\n f""[bold magenta]Waiting for Experiment {self._id} to finish"",\n )\n if not self._skip_status_at_exit:\n self.status()\n\n self._wait_for_jobs(jobs=self.jobs)\n finally:\n if self._live_progress:\n self._live_progress.stop()\n\n self._cleanup(tunnels=False)\n if self._launched:\n Path(os.path.join(self._exp_dir, self._DONE_FILE)).touch()\n if self._enable_goodbye_message:\n self.console.print(\n Syntax(\n self.GOODBYE_MESSAGE_PYTHON.format(\n exp_id=self._id,\n tasks=list(map(lambda job: job.id, self.jobs)),\n ),\n ""python"",\n theme=os.environ.get(""NEMO_RUN_CODE_THEME"", ""monokai""),\n )\n )\n self.console.print(\n Syntax(\n self.GOODBYE_MESSAGE_BASH.format(\n exp_id=self._id,\n tasks=list(map(lambda job: job.id, self.jobs)),\n ),\n ""shell"",\n theme=os.environ.get(""NEMO_RUN_CODE_THEME"", ""monokai""),\n )\n )\n\n def _repr_svg_(self):\n return self.to_config()._repr_svg_()\n\n def __del__(self):\n try:\n deconfigure_logging()\n self._cleanup()\n except Exception:\n pass\n\n @property\n def jobs(self) -> list[Job | JobGroup]:\n return Jobs(self._jobs)\n\n @jobs.setter\n def jobs(self, jobs: list[Job | JobGroup]):\n self._jobs = jobs\n\n @property\n def tasks(self) -> list[Config]:\n serializer = ZlibJSONSerializer()\n\n for job in self._jobs:\n if isinstance(job, Job):\n if isinstance(job.task, str):\n _task = serializer.deserialize(job.task)\n if _task.__fn_or_cls__ == Script:\n job.task = fdl.build(_task)\n else:\n job.task = _task # type: ignore\n else:\n if isinstance(job.tasks, str):\n tasks = serializer.deserialize(job.tasks)\n job.tasks = [\n fdl.build(task) if task.__fn_or_cls__ == Script else task for task in tasks\n ]\n\n return Tasks((job.task if isinstance(job, Job) else job.tasks) for job in self._jobs)\n\n\nclass Tasks(list, ConfigurableMixin): ...\n\n\nclass Jobs(list, ConfigurableMixin): ...\n\n\ndef _get_latest_dir(path) -> str:\n dirs = [d for d in os.listdir(path) if os.path.isdir(os.path.join(path, d))]\n latest_dir = max(dirs, key=lambda d: os.path.getctime(os.path.join(path, d)))\n return os.path.join(path, latest_dir)\n\n\ndef _get_sorted_dirs(path: str) -> list[str]:\n if not os.path.exists(path):\n return []\n dirs = [d for d in os.listdir(path) if os.path.isdir(os.path.join(path, d))]\n dirs = sorted(dirs, key=lambda d: os.path.getctime(os.path.join(path, d)))\n return list(dirs)\n\n\n_LOADED_MAINS = set()\n\n\ndef maybe_load_external_main(exp_dir: str):\n main_file = Path(exp_dir) / ""__main__.py""\n if main_file.exists() and main_file not in _LOADED_MAINS:\n _LOADED_MAINS.add(main_file)\n\n spec = importlib.util.spec_from_file_location(""__external_main__"", main_file)\n if spec is not None and spec.loader is not None:\n new_main_module = importlib.util.module_from_spec(spec)\n spec.loader.exec_module(new_main_module)\n\n if ""__external_main__"" not in sys.modules:\n sys.modules[""__external_main__""] = new_main_module\n else:\n external = sys.modules[""__external_main__""]\n for attr in dir(new_main_module):\n if not attr.startswith(""__""):\n setattr(external, attr, getattr(new_main_module, attr))\n\n existing_main = sys.modules[""__main__""]\n for attr in dir(new_main_module):\n if not attr.startswith(""__""):\n setattr(existing_main, attr, getattr(new_main_module, attr))\n",python,tab
21
+ 21,28495,"nemo_run/run/experiment.py",21140,0,"",python,selection_command
22
+ 22,47563,"nemo_run/run/experiment.py",22052,0,"",python,selection_command
23
+ 23,53173,"nemo_run/run/experiment.py",16851,0,"",python,selection_command
24
+ 24,70240,"nemo_run/run/experiment.py",22052,0,"",python,selection_command
25
+ 25,71372,"nemo_run/run/experiment.py",22088,0,"",python,selection_command
26
+ 26,71498,"nemo_run/run/experiment.py",22114,0,"",python,selection_command
27
+ 27,71849,"nemo_run/run/experiment.py",22136,0,"",python,selection_command
28
+ 28,72202,"nemo_run/run/experiment.py",22132,0,"",python,selection_command
29
+ 29,72685,"nemo_run/run/experiment.py",21295,0,"",python,selection_command
30
+ 30,76876,"nemo_run/run/api.py",0,0,"",python,tab
31
+ 31,80718,"nemo_run/run/api.py",2794,0,"",python,selection_command
32
+ 32,82805,"nemo_run/run/api.py",2789,0,"",python,selection_command
33
+ 33,83021,"nemo_run/run/api.py",2667,0,"",python,selection_command
34
+ 34,83368,"nemo_run/run/api.py",2620,0,"",python,selection_command
35
+ 35,83690,"nemo_run/run/api.py",2613,0,"",python,selection_command
36
+ 36,84822,"nemo_run/run/api.py",2618,0,"",python,selection_command
37
+ 37,85003,"nemo_run/run/api.py",2620,0,"",python,selection_command
38
+ 38,85153,"nemo_run/run/api.py",2625,0,"",python,selection_command
39
+ 39,85303,"nemo_run/run/api.py",2628,0,"",python,selection_command
40
+ 40,88095,"nemo_run/run/api.py",2625,0,"",python,selection_command
41
+ 41,88282,"nemo_run/run/api.py",2620,0,"",python,selection_command
42
+ 42,88484,"nemo_run/run/api.py",2618,0,"",python,selection_command
43
+ 43,88864,"nemo_run/run/api.py",2620,0,"",python,selection_command
44
+ 44,91096,"nemo_run/run/api.py",2613,0,"",python,selection_command
45
+ 45,91737,"nemo_run/run/api.py",1166,0,"",python,selection_command
46
+ 46,97798,"nemo_run/run/api.py",2613,0,"",python,selection_command
47
+ 47,100932,"nemo_run/run/experiment.py",0,0,"",python,tab
48
+ 48,104509,"nemo_run/run/experiment.py",22132,0,"",python,selection_command
49
+ 49,106747,"nemo_run/run/experiment.py",22052,0,"",python,selection_command
50
+ 50,106932,"nemo_run/run/experiment.py",16851,0,"",python,selection_command
51
+ 51,107728,"nemo_run/run/experiment.py",16876,0,"",python,selection_command
52
+ 52,107828,"nemo_run/run/experiment.py",16890,0,"",python,selection_command
53
+ 53,107969,"nemo_run/run/experiment.py",16928,0,"",python,selection_command
54
+ 54,108097,"nemo_run/run/experiment.py",16956,0,"",python,selection_command
55
+ 55,108538,"nemo_run/run/experiment.py",17320,0,"",python,selection_command
56
+ 56,108874,"nemo_run/run/experiment.py",17344,0,"",python,selection_command
57
+ 57,109553,"nemo_run/run/experiment.py",17320,0,"",python,selection_command
58
+ 58,110770,"nemo_run/run/experiment.py",17296,0,"",python,selection_command
59
+ 59,114285,"nemo_run/run/experiment.py",17344,0,"",python,selection_command
60
+ 60,125656,"nemo_run/run/experiment.py",17380,0,"",python,selection_command
61
+ 61,125803,"nemo_run/run/experiment.py",17440,0,"",python,selection_command
62
+ 62,126049,"nemo_run/run/experiment.py",17444,0,"",python,selection_command
63
+ 63,128524,"nemo_run/run/experiment.py",17493,0,"",python,selection_command
64
+ 64,128766,"nemo_run/run/experiment.py",17507,0,"",python,selection_command
65
+ 65,128800,"nemo_run/run/experiment.py",17522,0,"",python,selection_command
66
+ 66,128832,"nemo_run/run/experiment.py",17535,0,"",python,selection_command
67
+ 67,128866,"nemo_run/run/experiment.py",17581,0,"",python,selection_command
68
+ 68,128899,"nemo_run/run/experiment.py",17594,0,"",python,selection_command
69
+ 69,128933,"nemo_run/run/experiment.py",17630,0,"",python,selection_command
70
+ 70,128967,"nemo_run/run/experiment.py",17655,0,"",python,selection_command
71
+ 71,129000,"nemo_run/run/experiment.py",17677,0,"",python,selection_command
72
+ 72,129297,"nemo_run/run/experiment.py",17655,0,"",python,selection_command
73
+ 73,129437,"nemo_run/run/experiment.py",17630,0,"",python,selection_command
74
+ 74,129588,"nemo_run/run/experiment.py",17594,0,"",python,selection_command
75
+ 75,129897,"nemo_run/run/experiment.py",17630,0,"",python,selection_command
76
+ 76,130148,"nemo_run/run/experiment.py",17655,0,"",python,selection_command
77
+ 77,130179,"nemo_run/run/experiment.py",17677,0,"",python,selection_command
78
+ 78,130211,"nemo_run/run/experiment.py",17704,0,"",python,selection_command
79
+ 79,130368,"nemo_run/run/experiment.py",17733,0,"",python,selection_command
80
+ 80,130772,"nemo_run/run/experiment.py",17741,0,"",python,selection_command
81
+ 81,130937,"nemo_run/run/experiment.py",17742,0,"",python,selection_command
82
+ 82,143983,"nemo_run/run/experiment.py",17721,56," task_dir=name if reuse_job_dir else task_id,",python,selection_command
83
+ 83,260984,"nemo_run/run/experiment.py",17742,0,"",python,selection_command
84
+ 84,358190,"nemo_run/config.py",0,0,"# SPDX-FileCopyrightText: Copyright (c) 2024 NVIDIA CORPORATION & AFFILIATES. All rights reserved.\n# SPDX-License-Identifier: Apache-2.0\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom __future__ import annotations\n\nimport copy\nimport dataclasses\nimport inspect\nimport os\nimport re\nimport sys\nimport typing\nfrom pathlib import Path\nfrom types import MappingProxyType\nfrom typing import Any, Callable, Generic, Iterable, Optional, Type, TypeVar, Union, get_args\n\nimport fiddle as fdl\nimport fiddle._src.experimental.dataclasses as fdl_dc\nimport graphviz\nfrom fiddle._src import config, daglish, daglish_extensions\nfrom fiddle._src.casting import register_supported_cast\nfrom fiddle._src.config import TypeOrCallableProducingT\nfrom fiddle.graphviz import render, render_diff\nfrom typing_extensions import Annotated, ParamSpec, Self\n\nimport nemo_run.exceptions as run_exceptions\n\nParams = ParamSpec(""Params"")\nReturnType = TypeVar(""ReturnType"")\n\n_T = TypeVar(""_T"")\n_BuildableT = TypeVar(""_BuildableT"", bound=fdl.Buildable)\n\nRECURSIVE_TYPES = (typing.Union, typing.Optional)\n_NEMORUN_HOME = os.environ.get(""NEMORUN_HOME"", os.path.expanduser(""~/.nemo_run""))\nRUNDIR_NAME = ""nemo_run""\nRUNDIR_SPECIAL_NAME = ""/$nemo_run""\nSCRIPTS_DIR = ""scripts""\n\n# Metadata keys\nUSE_WITH_RAY_CLUSTER_KEY = ""use_with_ray_cluster""\n\n\ndef get_nemorun_home() -> str:\n """"""\n Get the current NEMORUN_HOME directory path.\n\n Returns:\n The path to the NEMORUN_HOME directory.\n """"""\n return _NEMORUN_HOME\n\n\ndef set_nemorun_home(path: str) -> None:\n """"""\n Set the NEMORUN_HOME directory path.\n\n Args:\n path: The new path for NEMORUN_HOME.\n """"""\n global _NEMORUN_HOME\n _NEMORUN_HOME = os.path.expanduser(path)\n\n\ndef get_type_namespace(typ: Type | Callable) -> str:\n """"""\n Get the namespace of a type or callable.\n\n Args:\n typ: The type or callable to get the namespace for.\n\n Returns:\n A string representing the namespace of the type or callable.\n\n Examples:\n >>> class MyClass:\n ... pass\n >>> get_type_namespace(MyClass)\n 'your_module.MyClass'\n """"""\n module = typ.__module__\n if module == ""__main__"":\n # Get the filename without extension\n main_module = sys.modules[""__main__""]\n filename = os.path.basename(main_module.__file__)\n module = os.path.splitext(filename)[0]\n\n if isinstance(typ, fdl.Buildable):\n typ = typ.__fn_or_cls__\n\n _name = getattr(typ, ""__qualname__"", str(typ))\n if _name.startswith(""ForwardRef""):\n _name = _name.split(""."")[-1]\n return f""{module}.{_name}""\n\n\ndef get_underlying_types(type_hint: typing.Any) -> typing.Set[typing.Type]:\n if isinstance(type_hint, typing._GenericAlias): # type: ignore\n if str(type_hint).startswith(""typing.Annotated""):\n origin = type_hint.__origin__\n if hasattr(origin, ""__origin__""):\n origin = origin.__origin__\n else:\n origin = type_hint.__origin__\n if origin in RECURSIVE_TYPES:\n types = set()\n for arg in type_hint.__args__:\n types.update(get_underlying_types(arg))\n return types\n return {type_hint}\n\n\ndef from_dict(raw_data: dict | list | str | float | int | bool, cls: Type[_T]) -> _T:\n if isinstance(raw_data, dict):\n underlying_types = get_underlying_types(cls)\n underlying_types = [tp for tp in underlying_types if tp is not type(None)]\n assert len(underlying_types) == 1, (\n f""Unable to load {cls}. Nested union types are not currently supported.""\n )\n cls = underlying_types[0] # type: ignore\n\n if dataclasses.is_dataclass(cls):\n fields_dict = {\n f.name: from_dict(raw_data.get(f.name), f.type) # type: ignore\n for f in dataclasses.fields(cls)\n if f.init\n }\n return cls(**fields_dict) # type: ignore\n elif isinstance(raw_data, list):\n return [from_dict(item, cls.__args__[0]) for item in raw_data] # type: ignore\n else:\n return raw_data # type: ignore\n\n\ndef set_value(cfg: config.Buildable, key: str, value: Any) -> None:\n """"""Set an attribute's value.\n\n Args:\n cfg: A `fdl.Buildable` whose attribute is to be overridden.\n assignment: String representing attribute's override expression. Of the form\n `attribute=value`.\n """"""\n *parents, last = _parse_path(key)\n\n walk = typing.cast(Any, cfg)\n try:\n for parent in parents:\n walk = parent.follow(walk)\n except Exception as e:\n raise run_exceptions.SetValueError(f'Invalid path ""{key}"".') from e\n\n try:\n if isinstance(last, daglish.Attr):\n setattr(walk, last.name, value)\n elif isinstance(last, daglish.Key):\n walk[last.key] = value\n else:\n raise run_exceptions.SetValueError(f""Unexpected path element {last}."")\n except Exception as e:\n raise run_exceptions.SetValueError(f'Could not set ""{key}"" to ""{value}"".') from e\n\n\nclass _CloneAndFNMixin:\n def clone(self):\n """"""Returns a deep clone of the object.""""""\n return copy.deepcopy(self)\n\n def walk(self: _BuildableT, **kwargs) -> _BuildableT: # type: ignore\n """"""\n Recursively applies a transformation function to attributes within the configuration object\n and its children that match the keys provided in kwargs. Attributes not listed in kwargs\n are not modified.\n\n Args:\n **kwargs (dict): A dictionary where keys are attribute names and values are functions\n that take the current attribute value and return a new value.\n\n Returns\n -------\n Config: A new Config instance with selectively modified attributes.\n\n Examples\n --------\n >>> config = Config(model=ModelConfig(seq_length=128))\n >>> new_config = config.walk(seq_length=lambda cfg: cfg.seq_length * 2)\n >>> new_config.model.seq_length\n 256\n """"""\n return _try_set_all(self, _walk=True, **kwargs)\n\n def broadcast(self: _BuildableT, **kwargs) -> _BuildableT: # type: ignore\n """"""\n Sets new values to attributes within the configuration object and its children that match\n the keys provided in kwargs. Attributes not listed in kwargs are not modified.\n\n Args:\n **kwargs (dict): A dictionary where keys are attribute names and values are the new\n values to be set.\n\n Returns\n -------\n Config: A new Config instance with selectively updated attributes.\n\n Examples\n --------\n >>> config = Config(model=ModelConfig(tensor_model_parallel_size=1))\n >>> new_config = config.broadcast(tensor_model_parallel_size=2)\n >>> new_config.model.tensor_model_parallel_size\n 2\n """"""\n return _try_set_all(self, **kwargs)\n\n\nclass _VisualizeMixin:\n def visualize(self, **kwargs) -> graphviz.Graph:\n return render(self, **kwargs)\n\n def diff(self, old: Self, trim=True, **kwargs):\n return render_diff(old=old, new=self, trim=trim, **kwargs)\n\n def save_config_img(self, path_str: str) -> None:\n """"""\n Saves the configuration to a file.\n\n Args:\n path (str): The file path where the configuration should be saved.\n fdl_fn (Partial): The function descriptor library function to save.\n\n Example:\n >>> save_config_img(""path/to/dir"", some_fdl_fn)\n """"""\n path: Path = Path(path_str)\n\n if not path.suffix:\n path = path / ""config.png""\n elif path.suffix != "".png"":\n raise ValueError(""The file extension must be .png"")\n\n path.parent.mkdir(parents=True, exist_ok=True)\n\n with path.open(""wb"") as f:\n f.write(self.visualize().pipe(""png""))\n\n def _repr_svg_(self):\n """"""Special method used by Jupyter to represent an object as SVG.\n\n Returns\n -------\n str: SVG representation of the Config object if Graphviz can render it.\n If Graphviz rendering fails or is not available, it returns None.\n """"""\n try:\n # Attempt to render using Graphviz and return SVG representation\n return self.visualize().pipe(format=""svg"").decode(""utf-8"")\n except Exception as e:\n # If rendering fails, log the exception or handle it as needed\n print(f""Graphviz rendering failed: {e}"")\n return self.__repr__()\n\n\nclass Config(Generic[_T], fdl.Config[_T], _CloneAndFNMixin, _VisualizeMixin):\n """"""\n Wrapper around fdl.Config with nemo_run specific functionality.\n See `fdl.Config <https://fiddle.readthedocs.io/en/latest/api_reference/core.html#config>`_ for more.\n """"""\n\n def __init__(\n self,\n fn_or_cls: Union[fdl.Buildable[_T], TypeOrCallableProducingT[_T]],\n *args,\n bind_args: bool = True,\n **kwargs,\n ):\n # Handle dict types by converting to _kwargs_to_dict function\n if fn_or_cls == {} or (hasattr(fn_or_cls, ""__origin__"") and fn_or_cls.__origin__ is dict):\n fn_or_cls = dict # type: ignore\n bind_args = False\n\n new_kwargs = kwargs\n if bind_args and not isinstance(fn_or_cls, fdl.Buildable):\n try:\n new_kwargs = _bind_args(fn_or_cls, *args, **kwargs)\n except Exception:\n new_kwargs = kwargs\n\n super().__init__(fn_or_cls, *args, **new_kwargs)\n\n @classmethod\n def __unflatten__(\n cls,\n values: Iterable[Any],\n metadata: config.BuildableTraverserMetadata,\n ):\n # If this is a dictionary config, reconstruct it with the arguments\n if metadata.fn_or_cls is dict:\n return cls(**metadata.arguments(values))\n return super().__unflatten__(values, metadata)\n\n\nclass Partial(Generic[_T], fdl.Partial[_T], _CloneAndFNMixin, _VisualizeMixin):\n """"""\n Wrapper around fdl.Partial with nemo_run specific functionality.\n See `fdl.Partial <https://fiddle.readthedocs.io/en/latest/api_reference/core.html#partial>`_ for more.\n """"""\n\n def __init__(\n self,\n fn_or_cls: Union[fdl.Buildable[_T], TypeOrCallableProducingT[_T]],\n *args,\n bind_args: bool = True,\n **kwargs,\n ):\n new_kwargs = kwargs\n if bind_args and not isinstance(fn_or_cls, fdl.Buildable):\n try:\n new_kwargs = _bind_args(fn_or_cls, **kwargs)\n except Exception:\n new_kwargs = kwargs\n\n super().__init__(fn_or_cls, *args, **new_kwargs)\n\n\nregister_supported_cast(fdl.Config, Config)\nregister_supported_cast(fdl.Partial, Partial)\nregister_supported_cast(Config, Config)\nregister_supported_cast(Partial, Partial)\n\n\nclass ConfigurableMixin(_VisualizeMixin):\n """"""\n A mixin class that provides configuration and visualization functionality.\n\n This mixin adds methods for converting objects to Config instances,\n visualizing configurations, and comparing configurations.\n\n For classes that are not dataclasses, the `to_config` method needs to be\n overridden to provide custom conversion logic to Config instances.\n """"""\n\n def diff(self, old: Self, trim=True, **kwargs):\n """"""\n Generate a visual difference between this configuration and an old one.\n\n Args:\n old (Self): The old configuration to compare against.\n trim (bool, optional): Whether to trim unchanged parts. Defaults to True.\n **kwargs: Additional arguments to pass to render_diff.\n\n Returns:\n graphviz.Digraph: A graph representing the differences between configurations.\n """"""\n return render_diff(old=old.to_config(), new=self.to_config(), trim=trim, **kwargs)\n\n def to_config(self) -> Config[Self]:\n """"""\n Convert the current object to a Config instance.\n\n This method automatically converts dataclasses to Config instances.\n For classes that are not dataclasses, this method needs to be overridden\n to provide custom conversion logic.\n\n Returns:\n Config: A Config representation of the current object.\n\n Raises:\n NotImplementedError: If the object type cannot be converted to Config\n or if the method is not overridden for non-dataclass types.\n\n Note:\n For classes that are not dataclasses, you must override this method\n to define how the object should be converted to a Config instance.\n """"""\n if dataclasses.is_dataclass(self):\n try:\n return fdl.cast(\n Config, fdl_dc.convert_dataclasses_to_configs(self, allow_post_init=True)\n )\n except Exception as e:\n raise NotImplementedError(\n f""Cannot convert type {type(self)} to Config"",\n f""Please implement a method `to_config` on {type(self)}."",\n ) from e\n elif isinstance(self, (list, tuple, dict)):\n return self # type: ignore\n else:\n raise NotImplementedError(\n f""Cannot convert type {type(self)} to Config. ""\n f""Please override the `to_config` method for {type(self)}.""\n )\n\n def _repr_svg_(self):\n """"""\n Generate an SVG representation of the object for Jupyter notebooks.\n\n Returns:\n str: SVG representation of the object if it can be rendered,\n otherwise returns the string representation.\n """"""\n if isinstance(self, (list, tuple, dict)):\n try:\n return render(self).pipe(format=""svg"").decode(""utf-8"")\n except Exception as e:\n print(f""Graphviz rendering failed: {e}"")\n return self.__repr__()\n\n return self.to_config()._repr_svg_()\n\n\n@dataclasses.dataclass\nclass Script(ConfigurableMixin):\n """"""\n Dataclass to configure raw scripts.\n\n Examples:\n\n .. code-block:: python\n\n file_based_script = run.Script(""./scripts/echo.sh"")\n\n inline_script = run.Script(\n inline=\""\""\""\n env\n echo ""Hello 1""\n echo ""Hello 2""\n \""\""\""\n )\n\n """"""\n\n #: Path to your script\n path: str = """"\n #: Inline contents of the script. Either path or inline needs to be set.\n inline: str = """"\n #: Args to pass to your scripts, only applicable when path is set.\n args: list[str] = dataclasses.field(default_factory=list)\n #: Environment variables to set when running the script.\n env: dict[str, str] = dataclasses.field(default_factory=dict)\n #: Entrypoint to use, defaults to bash.\n entrypoint: str = ""bash""\n #: Whether to use ``python -m`` when executing via python.\n m: bool = False\n\n metadata: dict[str, Any] = dataclasses.field(default_factory=dict)\n\n def __post_init__(self):\n assert self.path or self.inline\n assert self.entrypoint, ""Need to provide an entrypoint for script.""\n if self.m:\n assert ""python"" in self.entrypoint, ""-m can only be used with python""\n\n def get_name(self):\n if self.inline:\n name = self.inline.strip()[:10]\n return re.sub(""[^0-9a-zA-Z]+"", ""_"", name)\n else:\n return os.path.basename(self.path)\n\n def to_command(\n self, with_entrypoint: bool = False, filename: Optional[str] = None, is_local: bool = False\n ) -> list[str]:\n if self.inline:\n if filename:\n os.makedirs(os.path.dirname(filename), exist_ok=True)\n with open(filename, ""w"") as f:\n f.write(""#!/usr/bin/bash\n"" + self.inline)\n\n if is_local:\n cmd = [filename]\n else:\n cmd = [os.path.join(f""/{RUNDIR_NAME}"", SCRIPTS_DIR, Path(filename).name)]\n\n if with_entrypoint:\n cmd = [self.entrypoint] + cmd\n\n return cmd\n\n inline = self.inline.replace('""', '\\""')\n cmd = [""-c"", f'""{inline}""']\n if with_entrypoint:\n cmd = [self.entrypoint] + cmd\n\n return cmd\n\n args = [self.path] + self.args\n if self.m:\n cmd = [""-m""] + args\n else:\n cmd = args\n\n if with_entrypoint:\n if self.entrypoint:\n cmd = [self.entrypoint] + cmd\n else:\n raise ValueError(""Cannot use with_entrypoint=True without specifying entrypoint"")\n\n return cmd\n\n\n# A type alias for an optional type that is annotated with a Config.\n# This is useful for when you want to specify a type is Optional but\n# always want to provide a default config.\nOptionalDefaultConfig = Annotated[Optional[_T], Config[_T]]\nOptionalDefaultPartial = Annotated[Optional[_T], Partial[_T]]\n\n\ndef _parse_path(path: str) -> daglish.Path:\n """"""Parses a path into a list of either attributes or index lookups.""""""\n if not path.startswith(""["") and not path.startswith("".""):\n path = f"".{path}"" # Add a leading `.` to make parsing work properly.\n\n return daglish_extensions.parse_path(path)\n\n\ndef _bind_args(\n fn_or_cls: TypeOrCallableProducingT,\n *fn_args: fdl.Config | str | Callable,\n **fn_kwargs: fdl.Config | str | Callable,\n) -> dict[str, fdl.Config | str | Callable]:\n sig = inspect.signature(fn_or_cls)\n params = sig.parameters\n\n if set(fn_kwargs) > set(params):\n raise TypeError(\n f""{set(fn_kwargs) - set(params)} does not exist as args in {fn_or_cls.__module__}:{fn_or_cls.__name__}. Please remove them.""\n )\n\n final_args = _construct_args(fn_or_cls, params, fn_kwargs)\n final_args = fn_kwargs | final_args\n sig.bind(*fn_args, **final_args)\n return final_args\n\n\ndef _construct_args(\n fn_or_cls: TypeOrCallableProducingT,\n params: MappingProxyType[str, inspect.Parameter],\n kwargs: dict[str, fdl.Config | str | Callable],\n):\n final_args = {}\n\n primitive = [str, float, int, bool, bytes]\n primitive.extend([Optional[t] for t in primitive])\n\n for name, parameter in params.items():\n arg = kwargs.get(name, None)\n\n if arg:\n if dataclasses.is_dataclass(arg):\n final_args[name] = fdl.cast(\n Config,\n fdl_dc.convert_dataclasses_to_configs(arg, allow_post_init=True),\n )\n else:\n final_args[name] = arg\n elif str(parameter.annotation).startswith(""typing.Annotated""):\n args = get_args(parameter.annotation)\n if str(args[0]).startswith(""typing.Optional"") and len(args) > 1:\n cfg_type = get_args(args[0])[0]\n buildable = args[1].__origin__\n if issubclass(buildable, fdl.Buildable):\n final_args[name] = buildable(cfg_type)\n\n return final_args\n\n\nConfigT = TypeVar(""ConfigT"", Config, Partial)\n\n\ndef _try_set_all(config: _BuildableT, _walk: bool = False, **kwargs) -> _BuildableT:\n for key, val in kwargs.items():\n if hasattr(config, key):\n _val = val(config) if _walk else val\n setattr(config, key, _val)\n\n for attr_name in dir(config):\n try:\n if hasattr(config, attr_name):\n attr = getattr(config, attr_name)\n if isinstance(attr, (fdl.Config, fdl.Partial)):\n _try_set_all(attr, _walk=_walk, **kwargs)\n except ValueError:\n pass\n\n return config\n",python,tab
85
+ 85,365930,"nemo_run/config.py",1809,0,"",python,selection_command
86
+ 86,367240,"nemo_run/config.py",1840,0,"",python,selection_command
87
+ 87,367377,"nemo_run/config.py",1848,0,"",python,selection_command
88
+ 88,367914,"nemo_run/config.py",1852,0,"",python,selection_command
89
+ 89,368072,"nemo_run/config.py",1856,0,"",python,selection_command
90
+ 90,368252,"nemo_run/config.py",1864,0,"",python,selection_command
91
+ 91,368766,"nemo_run/config.py",1893,0,"",python,selection_command
92
+ 92,368885,"nemo_run/config.py",1905,0,"",python,selection_command
93
+ 93,369025,"nemo_run/config.py",1927,0,"",python,selection_command
94
+ 94,369176,"nemo_run/config.py",1961,0,"",python,selection_command
95
+ 95,369508,"nemo_run/config.py",1983,0,"",python,selection_command
96
+ 96,371249,"nemo_run/config.py",1570,0,"",python,selection_command
97
+ 97,597772,"nemo_run/config.py",1652,0,"",python,selection_command
98
+ 98,598468,"nemo_run/config.py",1570,0,"",python,selection_command
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-e1878deb-64ca-4d0d-aaf1-5699f0eabd291756538623447-2025_08_30-08.23.51.174/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-ea3b8a60-f42f-4cc5-977e-bdd5170b12601763540725942-2025_11_19-09.25.37.711/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-f0382786-979c-4a6d-8e9b-f5977f18eb4f1753726151187-2025_07_28-20.09.13.67/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-f1861737-541e-498a-8209-7a659cb20b861757609758410-2025_09_11-18.56.00.140/source.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 2,96,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"6:56:00 PM [info] Activating crowd-code\n6:56:00 PM [info] Recording started\n6:56:00 PM [info] Initializing git provider using file system watchers...\n6:56:00 PM [info] Git repository found\n6:56:00 PM [info] Git provider initialized successfully\n6:56:00 PM [info] Initial git state: [object Object]\n",Log,tab
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-f6f489a9-a5c9-44d3-8bfa-e950be10e08d1763478651049-2025_11_18-16.10.57.25/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-faba6583-b2c9-4b94-9ba6-9f240428520a1750722089894-2025_06_23-23.49.28.299/source.csv ADDED
@@ -0,0 +1,151 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type90,608191,"src/extension.ts",9984,0,"A",typescript,content
2
+ 91,608196,"src/extension.ts",9985,0,"",typescript,selection_keyboard
3
+ 92,608218,"src/extension.ts",9985,0,"P",typescript,content
4
+ 93,608220,"src/extension.ts",9986,0,"",typescript,selection_keyboard
5
+ 94,608231,"src/extension.ts",9986,0,"I",typescript,content
6
+ 95,608233,"src/extension.ts",9987,0,"",typescript,selection_keyboard
7
+ 96,608439,"src/extension.ts",9987,0,"=",typescript,content
8
+ 97,608442,"src/extension.ts",9988,0,"",typescript,selection_keyboard
9
+ 98,609356,"src/extension.ts",9988,0,"f",typescript,content
10
+ 99,609363,"src/extension.ts",9989,0,"",typescript,selection_keyboard
11
+ 100,609425,"src/extension.ts",9989,0,"o",typescript,content
12
+ 101,609429,"src/extension.ts",9990,0,"",typescript,selection_keyboard
13
+ 102,609547,"src/extension.ts",9990,0,"o",typescript,content
14
+ 103,609551,"src/extension.ts",9991,0,"",typescript,selection_keyboard
15
+ 104,610051,"src/extension.ts",9990,0,"",typescript,selection_command
16
+ 105,610641,"src/extension.ts",9984,7,"",typescript,content
17
+ 106,610907,"src/extension.ts",9984,0,"A",typescript,content
18
+ 107,610908,"src/extension.ts",9985,0,"",typescript,selection_keyboard
19
+ 108,610957,"src/extension.ts",9985,0,"P",typescript,content
20
+ 109,610960,"src/extension.ts",9986,0,"",typescript,selection_keyboard
21
+ 110,610983,"src/extension.ts",9986,0,"I",typescript,content
22
+ 111,610986,"src/extension.ts",9987,0,"",typescript,selection_keyboard
23
+ 112,611972,"src/extension.ts",9986,0,"",typescript,selection_command
24
+ 113,612398,"src/extension.ts",9984,0,"",typescript,selection_command
25
+ 114,613099,"src/extension.ts",9984,3,"",typescript,content
26
+ 115,613505,"src/extension.ts",9984,0,"A",typescript,content
27
+ 116,613506,"src/extension.ts",9985,0,"",typescript,selection_keyboard
28
+ 117,613558,"src/extension.ts",9985,0,"P",typescript,content
29
+ 118,613562,"src/extension.ts",9986,0,"",typescript,selection_keyboard
30
+ 119,613614,"src/extension.ts",9986,0,"I",typescript,content
31
+ 120,613618,"src/extension.ts",9987,0,"",typescript,selection_keyboard
32
+ 121,613842,"src/extension.ts",9987,0,"=",typescript,content
33
+ 122,613844,"src/extension.ts",9988,0,"",typescript,selection_keyboard
34
+ 123,614226,"src/extension.ts",9988,0,"f",typescript,content
35
+ 124,614230,"src/extension.ts",9989,0,"",typescript,selection_keyboard
36
+ 125,614281,"src/extension.ts",9989,0,"o",typescript,content
37
+ 126,614414,"src/extension.ts",9990,0,"o",typescript,content
38
+ 127,614418,"src/extension.ts",9991,0,"",typescript,selection_keyboard
39
+ 128,614676,"src/extension.ts",9990,0,"",typescript,selection_command
40
+ 129,615398,"src/extension.ts",9984,7,"",typescript,content
41
+ 130,632977,"src/extension.ts",9984,0,"\n",typescript,content160,689584,"src/extension.ts",9984,0,"\n",typescript,content
42
+ 161,689722,"src/extension.ts",9985,0,"\n",typescript,content
43
+ 162,689854,"src/extension.ts",9986,0,"\n",typescript,content
44
+ 163,689985,"src/extension.ts",9987,0,"\n",typescript,content
45
+ 164,690120,"src/extension.ts",9988,0,"\n",typescript,content
46
+ 165,690244,"src/extension.ts",9989,0,"\n",typescript,content195,720758,"src/extension.ts",9990,1,"",typescript,content
47
+ 196,720908,"src/extension.ts",9989,1,"",typescript,content
48
+ 197,721051,"src/extension.ts",9988,1,"",typescript,content
49
+ 198,721198,"src/extension.ts",9987,1,"",typescript,content
50
+ 199,721347,"src/extension.ts",9986,1,"",typescript,content
51
+ 200,721495,"src/extension.ts",9985,1,"",typescript,content
52
+ 201,721627,"src/extension.ts",9984,1,"",typescript,content
53
+ 202,721775,"src/extension.ts",9984,0,"\n",typescript,content
54
+ 203,721906,"src/extension.ts",9985,0,"\n",typescript,content
55
+ 204,722150,"src/extension.ts",9986,0,"\n",typescript,content
56
+ 205,722190,"src/extension.ts",9987,0,"\n",typescript,content
57
+ 206,722219,"src/extension.ts",9988,0,"\n",typescript,content
58
+ 207,722259,"src/extension.ts",9989,0,"\n",typescript,content
59
+ 208,722283,"src/extension.ts",9990,0,"\n",typescript,content
60
+ 209,722320,"src/extension.ts",9991,0,"\n",typescript,content
61
+ 210,722350,"src/extension.ts",9992,0,"\n",typescript,content
62
+ 211,722387,"src/extension.ts",9993,0,"\n",typescript,content
63
+ 212,722421,"src/extension.ts",9994,0,"\n",typescript,content
64
+ 213,722454,"src/extension.ts",9995,0,"\n",typescript,content
65
+ 214,722489,"src/extension.ts",9996,0,"\n",typescript,content
66
+ 215,722520,"src/extension.ts",9997,0,"\n",typescript,content
67
+ 216,722557,"src/extension.ts",9998,0,"\n",typescript,content
68
+ 217,722587,"src/extension.ts",9999,0,"\n",typescript,content
69
+ 218,722622,"src/extension.ts",10000,0,"\n",typescript,content
70
+ 219,722654,"src/extension.ts",10001,0,"\n",typescript,content
71
+ 220,722688,"src/extension.ts",10002,0,"\n",typescript,content
72
+ 221,722800,"src/extension.ts",10002,1,"",typescript,content
73
+ 222,723054,"src/extension.ts",10001,1,"",typescript,content
74
+ 223,723084,"src/extension.ts",10000,1,"",typescript,content
75
+ 224,723119,"src/extension.ts",9999,1,"",typescript,content254,744309,"src/extension.ts",9990,1,"",typescript,content
76
+ 255,744454,"src/extension.ts",9989,1,"",typescript,content
77
+ 256,744592,"src/extension.ts",9988,1,"",typescript,content
78
+ 257,744729,"src/extension.ts",9987,1,"",typescript,content
79
+ 258,744874,"src/extension.ts",9986,1,"",typescript,content
80
+ 259,745025,"src/extension.ts",9985,1,"",typescript,content
81
+ 260,745342,"src/extension.ts",9984,1,"",typescript,content
82
+ 261,745611,"src/extension.ts",9984,0,"\n",typescript,content
83
+ 262,745787,"src/extension.ts",9985,0,"\n",typescript,content
84
+ 263,745943,"src/extension.ts",9986,0,"\n",typescript,content
85
+ 264,746077,"src/extension.ts",9987,0,"\n",typescript,content
86
+ 265,746223,"src/extension.ts",9988,0,"\n",typescript,content
87
+ 266,746357,"src/extension.ts",9989,0,"\n",typescript,content
88
+ 267,746479,"src/extension.ts",9990,0,"\n",typescript,content
89
+ 268,746689,"src/extension.ts",9990,1,"",typescript,content
90
+ 269,746852,"src/extension.ts",9989,1,"",typescript,content299,1440083,"src/extension.ts",9984,0,"\n",typescript,content
91
+ 300,1440491,"src/extension.ts",9985,0,"\n",typescript,content
92
+ 301,1440637,"src/extension.ts",9986,0,"\n",typescript,content
93
+ 302,1440791,"src/extension.ts",9987,0,"\n",typescript,content
94
+ 303,1440934,"src/extension.ts",9988,0,"\n",typescript,content
95
+ 304,1441072,"src/extension.ts",9989,0,"\n",typescript,content
96
+ 305,1441279,"src/extension.ts",9989,1,"",typescript,content
97
+ 306,1441454,"src/extension.ts",9988,1,"",typescript,content
98
+ 307,1441595,"src/extension.ts",9987,1,"",typescript,content
99
+ 308,1441758,"src/extension.ts",9986,1,"",typescript,content
100
+ 309,1441888,"src/extension.ts",9985,1,"",typescript,content
101
+ 310,1442037,"src/extension.ts",9984,1,"",typescript,content
102
+ 311,1474476,"src/extension.ts",9984,0,"\n",typescript,content341,1511212,"src/extension.ts",9983,0,"\n",typescript,content
103
+ 342,1511454,"src/extension.ts",9984,0,"\n",typescript,content
104
+ 343,1511587,"src/extension.ts",9985,0,"\n",typescript,content
105
+ 344,1511731,"src/extension.ts",9986,0,"\n",typescript,content
106
+ 345,1511864,"src/extension.ts",9987,0,"\n",typescript,content
107
+ 346,1511996,"src/extension.ts",9988,0,"\n",typescript,content
108
+ 347,1512115,"src/extension.ts",9989,0,"\n",typescript,content
109
+ 348,1512251,"src/extension.ts",9990,0,"\n",typescript,content
110
+ 349,1512384,"src/extension.ts",9991,0,"\n",typescript,content379,1535334,"src/extension.ts",9983,0,"\n",typescript,content
111
+ 380,1535574,"src/extension.ts",9984,0,"\n",typescript,content
112
+ 381,1535825,"src/extension.ts",9985,0,"\n",typescript,content
113
+ 382,1535861,"src/extension.ts",9986,0,"\n",typescript,content
114
+ 383,1535890,"src/extension.ts",9987,0,"\n",typescript,content
115
+ 384,1535926,"src/extension.ts",9988,0,"\n",typescript,content414,1662008,"src/extension.ts",9983,0,"\n",typescript,content
116
+ 415,1662468,"src/extension.ts",9984,0,"\n",typescript,content
117
+ 416,1662714,"src/extension.ts",9985,0,"\n",typescript,content
118
+ 417,1662748,"src/extension.ts",9986,0,"\n",typescript,content
119
+ 418,1662778,"src/extension.ts",9987,0,"\n",typescript,content
120
+ 419,1662813,"src/extension.ts",9988,0,"\n",typescript,content
121
+ 420,1662845,"src/extension.ts",9989,0,"\n",typescript,content
122
+ 421,1662880,"src/extension.ts",9990,0,"\n",typescript,content
123
+ 422,1662913,"src/extension.ts",9991,0,"\n",typescript,content
124
+ 423,1662949,"src/extension.ts",9992,0,"\n",typescript,content
125
+ 424,1663179,"src/extension.ts",9993,0,"\n",typescript,content
126
+ 425,1663429,"src/extension.ts",9993,1,"",typescript,content
127
+ 426,1663685,"src/extension.ts",9992,1,"",typescript,content
128
+ 427,1663709,"src/extension.ts",9991,1,"",typescript,content
129
+ 428,1663747,"src/extension.ts",9990,1,"",typescript,content
130
+ 429,1663775,"src/extension.ts",9989,1,"",typescript,content
131
+ 430,1663813,"src/extension.ts",9988,1,"",typescript,content
132
+ 431,1663842,"src/extension.ts",9987,1,"",typescript,content
133
+ 432,1664057,"src/extension.ts",9986,1,"",typescript,content
134
+ 433,1664229,"src/extension.ts",9985,1,"",typescript,content
135
+ 434,1664513,"src/extension.ts",9984,1,"",typescript,content
136
+ 435,1674303,"src/extension.ts",9984,0,"K",typescript,content
137
+ 436,1674306,"src/extension.ts",9985,0,"",typescript,selection_keyboard
138
+ 437,1674445,"src/extension.ts",9985,0,"E",typescript,content
139
+ 438,1674448,"src/extension.ts",9986,0,"",typescript,selection_keyboard
140
+ 439,1675238,"src/extension.ts",9986,0,"Y",typescript,content
141
+ 440,1675248,"src/extension.ts",9987,0,"",typescript,selection_keyboard
142
+ 441,1675504,"src/extension.ts",9987,0,"=",typescript,content
143
+ 442,1675506,"src/extension.ts",9988,0,"",typescript,selection_keyboard
144
+ 443,1675690,"src/extension.ts",9988,0,"f",typescript,content
145
+ 444,1675691,"src/extension.ts",9989,0,"",typescript,selection_keyboard
146
+ 445,1675750,"src/extension.ts",9989,0,"o",typescript,content
147
+ 446,1675754,"src/extension.ts",9990,0,"",typescript,selection_keyboard
148
+ 447,1675900,"src/extension.ts",9990,0,"o",typescript,content
149
+ 448,1675904,"src/extension.ts",9991,0,"",typescript,selection_keyboard
150
+ 449,1676387,"src/extension.ts",9990,0,"",typescript,selection_command
151
+ 450,1676937,"src/extension.ts",9983,8,"",typescript,content
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-fc31d59a-d5b8-4dc1-8760-333f6ac9e55c1763046484160-2025_11_13-16.08.05.840/source.csv ADDED
The diff for this file is too large to render. See raw diff