mihirma commited on
Commit
cba778b
·
verified ·
1 Parent(s): cd5df33

Add files using upload-large-folder tool

Browse files
Files changed (18) hide show
  1. 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-0e6e3fa0-317b-4a46-bdb4-752f8e86cd181758538014008-2025_09_22-12.47.12.63/source.csv +0 -0
  2. 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-1505f3d0-0cb4-4cc0-84bf-678810d0ac8f1757148592235-2025_09_06-10.49.56.658/source.csv +17 -0
  3. 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-1f52a67d-2cca-4352-b39c-6cd9f3effae01765648091845-2025_12_13-18.48.27.228/source.csv +66 -0
  4. 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-2790c5c3-9c96-49ac-ab9e-481a8033e7cd1757945282635-2025_09_15-16.08.09.625/source.csv +0 -0
  5. 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-9543d8e2-2376-4957-873e-df7016d502961763465687199-2025_11_18-12.34.49.442/source.csv +245 -0
  6. 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-ac4f665d-1bc1-467a-98b3-5da2178968731760857710663-2025_10_19-09.08.50.386/source.csv +6 -0
  7. 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-c877f8c1-c8e0-4a7b-8720-40bb4df915221754138206219-2025_08_02-14.36.55.608/source.csv +0 -0
  8. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-14f0662f-0032-43e8-be9f-8e53d6f150ad1758635869882-2025_09_23-15.57.52.616/source.csv +0 -0
  9. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-1fa9b85d-3794-4f3b-b7a0-5170b7d2faaa1762362332596-2025_11_05-18.05.39.648/source.csv +19 -0
  10. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-4b7a193b-6fd0-48b6-a605-a1ce6ba179221764439942282-2025_11_29-19.12.25.371/source.csv +10 -0
  11. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-549b7320-9591-428d-ba6a-d3d8fb65a0001764416891089-2025_11_29-12.48.19.667/source.csv +1006 -0
  12. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-8eaf046e-99c4-4091-a85d-91e359564aa51756825908437-2025_09_02-17.11.50.753/source.csv +0 -0
  13. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-9543d8e2-2376-4957-873e-df7016d502961763465687199-2025_11_18-12.34.49.442/source.csv +226 -0
  14. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-a7b808c2-b1d0-43a0-a38c-8b82cd2886711764488770794-2025_11_30-08.46.18.897/source.csv +32 -0
  15. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-aeed47b9-f6ef-4272-b0ca-0c15ab4c25021758266694991-2025_09_19-09.25.04.660/source.csv +151 -0
  16. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-d9cdf338-0ddd-4679-853a-6d7bdf2b18581751046137722-2025_06_27-10.42.19.354/source.csv +167 -0
  17. 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-ef5ea013-ac2b-459c-8783-a7b025d58a391754900011518-2025_08_11-10.13.34.249/source.csv +0 -0
  18. 507ab0ec0dfe0c18ad7778dd15e072f92367194c94623114de802c8ed9c52e20/crowd-code-b9559366-0d71-4ceb-9b37-1d3a0cf03cd61750867779082-2025_06_25-18.09.57.465/source.csv +64 -0
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-0e6e3fa0-317b-4a46-bdb4-752f8e86cd181758538014008-2025_09_22-12.47.12.63/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-1505f3d0-0cb4-4cc0-84bf-678810d0ac8f1757148592235-2025_09_06-10.49.56.658/source.csv ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,1,"train_tokenizer.py",0,0,"import os\n\nos.environ.setdefault(""XLA_PYTHON_CLIENT_MEM_FRACTION"", ""0.98"")\n\nfrom dataclasses import dataclass, field\nfrom typing import cast, Optional\n\nimport einops\nimport itertools\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom models.tokenizer import TokenizerVQVAE\nfrom utils.dataloader import get_dataloader\nfrom utils.train_utils import (\n get_lr_schedule,\n count_parameters_by_component,\n print_mem_stats,\n print_compiled_memory_stats,\n print_compiled_cost_analysis,\n)\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 300_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n vq_beta: float = 0.25\n batch_size: int = 48\n init_lr: float = 0.0\n max_lr: float = 3e-4\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 20000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n warmup_steps: int = 10000\n # Tokenizer\n model_dim: int = 512\n ffn_dim: int = 2048\n latent_dim: int = 32\n num_latents: int = 1024\n patch_size: int = 4\n num_blocks: int = 4\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.01\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_tokenizer""\n tags: list[str] = field(default_factory=lambda: [""tokenizer""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n wandb_id: str = """"\n use_flash_attention: bool = True\n\n\ndef build_model(args: Args, rng: jax.Array) -> tuple[TokenizerVQVAE, jax.Array]:\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n return (\n TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n ffn_dim=args.ffn_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n ),\n rng,\n )\n\n\ndef build_optimizer(\n model: TokenizerVQVAE, args: Args\n) -> tuple[nnx.Optimizer, optax.Schedule]:\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.param_dtype, # moments in full precision\n )\n optimizer = nnx.Optimizer(model, tx)\n return optimizer, lr_schedule\n\n\ndef build_mesh_and_sharding(\n num_devices: int,\n) -> tuple[Mesh, NamedSharding, NamedSharding]:\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n return mesh, replicated_sharding, videos_sharding\n\n\ndef shard_optimizer_states(\n optimizer: nnx.Optimizer, replicated_sharding: NamedSharding\n) -> None:\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n\ndef build_dataloader(args: Args) -> grain.DataLoaderIterator:\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n return grain_iterator\n\n\ndef build_checkpoint_manager(args: Args) -> ocp.CheckpointManager:\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n return checkpoint_manager\n\n\ndef restore_checkpoint_if_needed(\n args: Args,\n checkpoint_manager: ocp.CheckpointManager,\n optimizer: nnx.Optimizer,\n grain_iterator: grain.DataLoaderIterator,\n restore_step: Optional[int] = None,\n) -> tuple[int, nnx.Optimizer, grain.DataLoaderIterator]:\n step = 0\n if restore_step is None:\n restore_step = checkpoint_manager.latest_step()\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n restore_step,\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n grain_iterator = restored[""dataloader_state""]\n step = restore_step or 0\n print(f""Restored dataloader and model state from step {step}"")\n return step, optimizer, grain_iterator\n\n\ndef main(args: Args) -> None:\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n tokenizer, rng = build_model(args, rng)\n\n _, params, _ = nnx.split(tokenizer, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n optimizer, lr_schedule = build_optimizer(tokenizer, args)\n del tokenizer\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n mesh, replicated_sharding, videos_sharding = build_mesh_and_sharding(num_devices)\n\n shard_optimizer_states(optimizer, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n checkpoint_manager = build_checkpoint_manager(args)\n\n # --- Create DataLoaderIterator from dataloader ---\n grain_iterator = build_dataloader(args)\n\n # --- Restore checkpoint ---\n step, optimizer, grain_iterator = restore_checkpoint_if_needed(\n args, checkpoint_manager, optimizer, grain_iterator\n )\n\n # --- Define loss and train step (close over args) ---\n def tokenizer_loss_fn(\n model: TokenizerVQVAE, inputs: dict\n ) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n model.train()\n outputs = model(inputs, training=True)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n mse = jnp.square(gt - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n gt_clipped = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt_clipped, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt_clipped, recon)).mean()\n _, index_counts = jnp.unique_counts(\n jnp.ravel(outputs[""indices""]), size=args.num_latents, fill_value=0\n )\n codebook_usage = (index_counts != 0).mean()\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=codebook_usage,\n )\n return loss, (outputs[""recon""], metrics)\n\n @nnx.jit(donate_argnums=0)\n def train_step(\n optimizer: nnx.Optimizer, inputs: dict\n ) -> tuple[jax.Array, jax.Array, dict]:\n def loss_fn(model: TokenizerVQVAE) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n return tokenizer_loss_fn(model, inputs)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\n optimizer.model\n )\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""encoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""encoder""]\n )\n metrics[""vq_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""vq""]\n )\n metrics[""decoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""decoder""]\n )\n return loss, recon, metrics\n\n # --- TRAIN LOOP ---\n dataloader = (\n jax.make_array_from_process_local_data(videos_sharding, elem)\n for elem in grain_iterator\n )\n if jax.process_index() == 0:\n first_videos = next(dataloader)\n sample_inputs = dict(videos=first_videos)\n compiled = train_step.lower(optimizer, sample_inputs).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training\n dataloader = itertools.chain([first_videos], dataloader)\n print(f""Starting training from step {step}..."")\n first_step = step\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n inputs = dict(videos=videos)\n loss, recon, metrics = train_step(optimizer, inputs)\n if step == first_step:\n print_mem_stats(""After params initialized"")\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n grain_iterator # type: ignore\n ),\n ),\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n\n\nif __name__ == ""__main__"":\n args = tyro.cli(Args)\n main(args)\n",python,tab
3
+ 2,113,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"10:49:56 AM [info] Activating crowd-code\n10:49:56 AM [info] Recording started\n10:49:56 AM [info] Initializing git provider using file system watchers...\n10:49:56 AM [info] Git repository found\n10:49:56 AM [info] Git provider initialized successfully\n",Log,tab
4
+ 3,224,"extension-output-pdoom-org.crowd-code-#1-crowd-code",250,0,"10:49:56 AM [info] Initial git state: [object Object]\n",Log,content
5
+ 4,188000455,"train_tokenizer.py",0,0,"",python,tab
6
+ 5,188012251,"train_tokenizer.py",0,0,"Switched from branch 'main' to 'demo-notebook'",python,git_branch_checkout
7
+ 6,188015548,"README.md",0,0,"<h1 align=""center"">🧞‍♀️ Jasmine: A simple, performant and scalable JAX-based world modeling codebase 🧞‍♀️</h1>\n\n<p align=""center"">\n <a href= ""https://github.com/FLAIROx/jafar/blob/main/LICENSE"">\n <img src=""https://img.shields.io/badge/license-Apache2.0-blue.svg"" /></a>\n <a href= ""https://github.com/psf/black"">\n <img src=""https://img.shields.io/badge/code%20style-black-000000.svg"" /></a>\n</p>\n\nJasmine is a production-ready JAX-based world modeling codebase. It currently implements the high-level architecture of [Genie: Generative Interactive Environments](https://arxiv.org/abs/2402.15391) (Bruce et al., 2024) with [MaskGIT](https://arxiv.org/abs/2202.04200) (Chang et al., 2022), as well as an autoregressive (causal) baseline. A diffusion baseline is coming soon.\n\nJasmine scales from single hosts to hundreds of xPUs thanks to XLA and strives to be an easily hackable, batteries-included foundation for world modeling research.\n\n<h2 name=""overview"" id=""overview"">Overview</h2>\n\n- Asynchronous & distributed checkpointing thanks to [orbax.checkpoint](https://github.com/google/orbax)\n - Jasmine also supports mixing and matching hardware topologies (e.g. train on four nodes, load the checkpoint on a single node)\n- Optimized dataloading thanks to [Grain](https://github.com/google/grain)\n - Dataloading scales with the number of processes (i.e. nodes/xPUs)\n- Checkpointing of model weights, optimizer and dataloader states\n- Full reproducibility with **identical** training curves (thanks to seeded dataloading and training, and [JAX' approach to pseudo random numbers](https://docs.jax.dev/en/latest/random-numbers.html))\n- Automatic checkpoint deletion/retention according to specified retention policy thanks to `orbax.checkpoint.CheckpointManager`\n- Mixed precision training using `bfloat16`\n - `int8` training is on the roadmap via [aqt](https://github.com/google/aqt)\n- FlashAttention thanks to [cuDNN SDPA](https://github.com/jax-ml/jax/blob/a155c5a9997924170e0067d552351a9833c12c11/jax/_src/cudnn/fused_attention_stablehlo.py#L842)\n- Frame-level KV cache resets for accelerated spatiotemporal attention in causal baseline (still in PR)\n- Activation checkpointing (even onto host memory if desired)\n- DDP (changing to FSDP requires changing **a single line of code**)\n- WSD learning rate schedule\n - No need to retrain from scratch if you want to train for longer\n- Index-shuffling during dataloading\n- Google-native stack\n - https://github.com/google/orbax for checkpointing\n - https://github.com/google/grain for dataloading\n - https://github.com/google-deepmind/dm_pix for image manipulation\n - https://github.com/google/array_record as the data format\n- Easy model inspection thanks to [treescope](https://github.com/google-deepmind/treescope)\n- Easy model surgery thanks to the new [flax.nnx](https://flax.readthedocs.io/en/latest/migrating/linen_to_nnx.html) API\n- [Shape suffixes](https://medium.com/@NoamShazeer/shape-suffixes-good-coding-style-f836e72e24fd) throughout the repository\n\n<h2 name=""start"" id=""start"">Setup 🧗</h2>\n\nJasmine requires `python 3.10`, `jax 0.6.2`, and `flax 0.10.7`. To install the requirements, run:\n\n```bash\npip install -r requirements.txt\npre-commit install\n```\n\n---\n\n<h2 name=""dataset"" id=""dataset"">Dataset 📂</h2>\n\nYou can either download our preprocessed dataset from [Hugging Face](https://huggingface.co/datasets/p-doom/open_ai_minecraft_arrayrecords_chunked) or preprocess [OpenAI's VPT dataset](https://github.com/openai/Video-Pre-Training) manually.\n\n### Option 1: Use Preprocessed Dataset (Recommended)\n\nThe easiest way to get started is to download our preprocessed dataset from Hugging Face. This script will handle downloading and extracting it:\n\n```bash\nbash input_pipeline/download/download_array_records.sh\n```\n\n---\n\n### Option 2: Manual Download & Preprocessing of OpenAI's VPT Dataset\n\nIf you prefer to use the raw VPT dataset from OpenAI and preprocess it yourself, follow these steps:\n\n1. **Download index files:**\n This will download the initial index file:\n\n ```bash\n bash input_pipeline/download/openai/download_index_files.sh\n ```\n\n2. **Download from all index files:**\n This may take a long time depending on your bandwidth:\n\n ```bash\n python input_pipeline/download/openai/download_videos.py --index_file_path data/open_ai_index_files/all_7xx_Apr_6.json\n python input_pipeline/download/openai/download_videos.py --index_file_path data/open_ai_index_files/all_8xx_Jun_29.json\n python input_pipeline/download/openai/download_videos.py --index_file_path data/open_ai_index_files/all_9xx_Jun_29.json\n python input_pipeline/download/openai/download_videos.py --index_file_path data/open_ai_index_files/all_10xx_Jun_29.json\n ```\n\n3. **Preprocess videos into ArrayRecords:**\n For efficient distributed training, convert the raw videos into the arrayrecord format (make sure to have [ffmpeg](https://github.com/FFmpeg/FFmpeg) installed on your machine):\n\n ```bash\n python input_pipeline/preprocess/video_to_array_records.py\n ```\n\n> **Note:** This is a large dataset and may take considerable time and storage to download and process.\n\n\n<h2 name=""train"" id=""train"">Quick Start 🚀 </h2>\n\nGenie has three components: a [video tokenizer](models/tokenizer.py), a [latent action model](models/lam.py), and a [dynamics model](models/dynamics.py). Each of these components are trained separately, however, the dynamics model requires a pre-trained video tokenizer (and latent action model).\n\nTo train the video tokenizer, run:\n\n```bash\npython train_tokenizer.py --ckpt_dir <path>\n```\n\nTo train the latent action model, run:\n\n```bash\npython train_lam.py --ckpt_dir <path>\n```\n\nOnce the tokenizer and LAM are trained, the dynamics model can be trained with:\n\n```bash\npython train_dynamics.py --tokenizer_checkpoint <path> --lam_checkpoint <path>\n```\n\nLogging with `wandb` is supported. To enable logging, set the `WANDB_API_KEY` environment variable or run:\n\n```bash\nwandb login\n```\n\nTraining can then be logged by setting the `--log` flag:\n\n```bash\npython train_tokenizer.py --log --entity <wandb-entity> --project <wandb-project>\n```\n\n<h2 name=""cite"" id=""cite"">Citing 📜 </h2>\n\nJasmine was built by [Mihir Mahajan](https://maharajamihir.github.io/), [Alfred Nguyen](https://avocadoali.github.io/) and [Franz Srambical](https://srambical.fr/), but started as a fork of [Jafar](https://github.com/flairox/jafar), built by [Matthew Jackson](https://matthewtjackson.com) and [Timon Willi](https://www.timonwilli.com).\n\nIf you use Jasmine in your work, please cite us, Jafar, and the original Genie paper as follows:\n\n```\n@article{\n mahajan2025jasmine,\n title={Jasmine: A simple, performant and scalable JAX-based world modeling codebase},\n author={Mihir Mahajan and Alfred Nguyen and Franz Srambical and Stefan Bauer},\n journal = {p(doom) blog},\n year={2025},\n url={https://pdoom.org/jasmine.html},\n note = {https://pdoom.org/blog.html}\n}\n```\n```\n@inproceedings{\n willi2024jafar,\n title={Jafar: An Open-Source Genie Reimplemention in Jax},\n author={Timon Willi and Matthew Thomas Jackson and Jakob Nicolaus Foerster},\n booktitle={First Workshop on Controllable Video Generation @ ICML 2024},\n year={2024},\n url={https://openreview.net/forum?id=ZZGaQHs9Jb}\n}\n```\n```\n@inproceedings{\n bruce2024genie,\n title={Genie: Generative Interactive Environments},\n author={Jake Bruce and Michael D Dennis and Ashley Edwards and Jack Parker-Holder and Yuge Shi and Edward Hughes and Matthew Lai and Aditi Mavalankar and Richie Steigerwald and Chris Apps and Yusuf Aytar and Sarah Maria Elisabeth Bechtle and Feryal Behbahani and Stephanie C.Y. Chan and Nicolas Heess and Lucy Gonzalez and Simon Osindero and Sherjil Ozair and Scott Reed and Jingwei Zhang and Konrad Zolna and Jeff Clune and Nando de Freitas and Satinder Singh and Tim Rockt{\""a}schel},\n booktitle={Forty-first International Conference on Machine Learning},\n year={2024},\n url={https://openreview.net/forum?id=bJbSbJskOS}\n}\n```\n",markdown,tab
8
+ 7,188019111,"README.md",2726,0,"",markdown,selection_command
9
+ 8,188019403,"README.md",2810,0,"\n",markdown,content
10
+ 9,188019614,"README.md",2811,0," - Modularized training script for easy inspection using notebooks ([demo notebook](https://colab.research.google.com/drive/1zHkciFIZxXloJgue9F5LtFlA0m00rJIf?usp=sharing))",markdown,content
11
+ 10,188019618,"README.md",2985,0,"",markdown,selection_keyboard
12
+ 11,188020348,"README.md",2984,0,"",markdown,selection_command
13
+ 12,188020659,"README.md",2815,0,"",markdown,selection_command
14
+ 13,188020891,"README.md",2811,4,"",markdown,content
15
+ 14,188033002,"README.md",0,0,"<h1 align=""center"">🧞‍♀️ Jasmine: A simple, performant and scalable JAX-based world modeling codebase 🧞‍♀️</h1>\n\n<p align=""center"">\n <a href= ""https://github.com/FLAIROx/jafar/blob/main/LICENSE"">\n <img src=""https://img.shields.io/badge/license-Apache2.0-blue.svg"" /></a>\n <a href= ""https://github.com/psf/black"">\n <img src=""https://img.shields.io/badge/code%20style-black-000000.svg"" /></a>\n</p>\n\nJasmine is a production-ready JAX-based world modeling codebase. It currently implements the high-level architecture of [Genie: Generative Interactive Environments](https://arxiv.org/abs/2402.15391) (Bruce et al., 2024) with [MaskGIT](https://arxiv.org/abs/2202.04200) (Chang et al., 2022), as well as an autoregressive (causal) baseline. A diffusion baseline is coming soon.\n\nJasmine scales from single hosts to hundreds of xPUs thanks to XLA and strives to be an easily hackable, batteries-included foundation for world modeling research.\n\n<h2 name=""overview"" id=""overview"">Overview</h2>\n\n- Asynchronous & distributed checkpointing thanks to [orbax.checkpoint](https://github.com/google/orbax)\n - Jasmine also supports mixing and matching hardware topologies (e.g. train on four nodes, load the checkpoint on a single node)\n- Optimized dataloading thanks to [Grain](https://github.com/google/grain)\n - Dataloading scales with the number of processes (i.e. nodes/xPUs)\n- Checkpointing of model weights, optimizer and dataloader states\n- Full reproducibility with **identical** training curves (thanks to seeded dataloading and training, and [JAX' approach to pseudo random numbers](https://docs.jax.dev/en/latest/random-numbers.html))\n- Automatic checkpoint deletion/retention according to specified retention policy thanks to `orbax.checkpoint.CheckpointManager`\n- Mixed precision training using `bfloat16`\n - `int8` training is on the roadmap via [aqt](https://github.com/google/aqt)\n- FlashAttention thanks to [cuDNN SDPA](https://github.com/jax-ml/jax/blob/a155c5a9997924170e0067d552351a9833c12c11/jax/_src/cudnn/fused_attention_stablehlo.py#L842)\n- Frame-level KV cache resets for accelerated spatiotemporal attention in causal baseline (still in PR)\n- Activation checkpointing (even onto host memory if desired)\n- DDP (changing to FSDP requires changing **a single line of code**)\n- WSD learning rate schedule\n - No need to retrain from scratch if you want to train for longer\n- Index-shuffling during dataloading\n- Google-native stack\n - https://github.com/google/orbax for checkpointing\n - https://github.com/google/grain for dataloading\n - https://github.com/google-deepmind/dm_pix for image manipulation\n - https://github.com/google/array_record as the data format\n- Easy model inspection thanks to [treescope](https://github.com/google-deepmind/treescope)\n- Modularized training script for easy inspection using notebooks ([demo notebook](https://colab.research.google.com/drive/1zHkciFIZxXloJgue9F5LtFlA0m00rJIf?usp=sharing))\n- Easy model surgery thanks to the new [flax.nnx](https://flax.readthedocs.io/en/latest/migrating/linen_to_nnx.html) API\n- [Shape suffixes](https://medium.com/@NoamShazeer/shape-suffixes-good-coding-style-f836e72e24fd) throughout the repository\n\n<h2 name=""start"" id=""start"">Setup 🧗</h2>\n\nJasmine requires `python 3.10`, `jax 0.6.2`, and `flax 0.10.7`. To install the requirements, run:\n\n```bash\npip install -r requirements.txt\npre-commit install\n```\n\n---\n\n<h2 name=""dataset"" id=""dataset"">Dataset 📂</h2>\n\nYou can either download our preprocessed dataset from [Hugging Face](https://huggingface.co/datasets/p-doom/open_ai_minecraft_arrayrecords_chunked) or preprocess [OpenAI's VPT dataset](https://github.com/openai/Video-Pre-Training) manually.\n\n### Option 1: Use Preprocessed Dataset (Recommended)\n\nThe easiest way to get started is to download our preprocessed dataset from Hugging Face. This script will handle downloading and extracting it:\n\n```bash\nbash input_pipeline/download/download_array_records.sh\n```\n\n---\n\n### Option 2: Manual Download & Preprocessing of OpenAI's VPT Dataset\n\nIf you prefer to use the raw VPT dataset from OpenAI and preprocess it yourself, follow these steps:\n\n1. **Download index files:**\n This will download the initial index file:\n\n ```bash\n bash input_pipeline/download/openai/download_index_files.sh\n ```\n\n2. **Download from all index files:**\n This may take a long time depending on your bandwidth:\n\n ```bash\n python input_pipeline/download/openai/download_videos.py --index_file_path data/open_ai_index_files/all_7xx_Apr_6.json\n python input_pipeline/download/openai/download_videos.py --index_file_path data/open_ai_index_files/all_8xx_Jun_29.json\n python input_pipeline/download/openai/download_videos.py --index_file_path data/open_ai_index_files/all_9xx_Jun_29.json\n python input_pipeline/download/openai/download_videos.py --index_file_path data/open_ai_index_files/all_10xx_Jun_29.json\n ```\n\n3. **Preprocess videos into ArrayRecords:**\n For efficient distributed training, convert the raw videos into the arrayrecord format (make sure to have [ffmpeg](https://github.com/FFmpeg/FFmpeg) installed on your machine):\n\n ```bash\n python input_pipeline/preprocess/video_to_array_records.py\n ```\n\n> **Note:** This is a large dataset and may take considerable time and storage to download and process.\n\n\n<h2 name=""train"" id=""train"">Quick Start 🚀 </h2>\n\nGenie has three components: a [video tokenizer](models/tokenizer.py), a [latent action model](models/lam.py), and a [dynamics model](models/dynamics.py). Each of these components are trained separately, however, the dynamics model requires a pre-trained video tokenizer (and latent action model).\n\nTo train the video tokenizer, run:\n\n```bash\npython train_tokenizer.py --ckpt_dir <path>\n```\n\nTo train the latent action model, run:\n\n```bash\npython train_lam.py --ckpt_dir <path>\n```\n\nOnce the tokenizer and LAM are trained, the dynamics model can be trained with:\n\n```bash\npython train_dynamics.py --tokenizer_checkpoint <path> --lam_checkpoint <path>\n```\n\nLogging with `wandb` is supported. To enable logging, set the `WANDB_API_KEY` environment variable or run:\n\n```bash\nwandb login\n```\n\nTraining can then be logged by setting the `--log` flag:\n\n```bash\npython train_tokenizer.py --log --entity <wandb-entity> --project <wandb-project>\n```\n\n<h2 name=""cite"" id=""cite"">Citing 📜 </h2>\n\nJasmine was built by [Mihir Mahajan](https://maharajamihir.github.io/), [Alfred Nguyen](https://avocadoali.github.io/) and [Franz Srambical](https://srambical.fr/), but started as a fork of [Jafar](https://github.com/flairox/jafar), built by [Matthew Jackson](https://matthewtjackson.com) and [Timon Willi](https://www.timonwilli.com).\n\nIf you use Jasmine in your work, please cite us, Jafar, and the original Genie paper as follows:\n\n```\n@article{\n mahajan2025jasmine,\n title={Jasmine: A simple, performant and scalable JAX-based world modeling codebase},\n author={Mihir Mahajan and Alfred Nguyen and Franz Srambical and Stefan Bauer},\n journal = {p(doom) blog},\n year={2025},\n url={https://pdoom.org/jasmine.html},\n note = {https://pdoom.org/blog.html}\n}\n```\n```\n@inproceedings{\n willi2024jafar,\n title={Jafar: An Open-Source Genie Reimplemention in Jax},\n author={Timon Willi and Matthew Thomas Jackson and Jakob Nicolaus Foerster},\n booktitle={First Workshop on Controllable Video Generation @ ICML 2024},\n year={2024},\n url={https://openreview.net/forum?id=ZZGaQHs9Jb}\n}\n```\n```\n@inproceedings{\n bruce2024genie,\n title={Genie: Generative Interactive Environments},\n author={Jake Bruce and Michael D Dennis and Ashley Edwards and Jack Parker-Holder and Yuge Shi and Edward Hughes and Matthew Lai and Aditi Mavalankar and Richie Steigerwald and Chris Apps and Yusuf Aytar and Sarah Maria Elisabeth Bechtle and Feryal Behbahani and Stephanie C.Y. Chan and Nicolas Heess and Lucy Gonzalez and Simon Osindero and Sherjil Ozair and Scott Reed and Jingwei Zhang and Konrad Zolna and Jeff Clune and Nando de Freitas and Satinder Singh and Tim Rockt{\""a}schel},\n booktitle={Forty-first International Conference on Machine Learning},\n year={2024},\n url={https://openreview.net/forum?id=bJbSbJskOS}\n}\n```\n",markdown,tab
16
+ 15,188033071,"README.md",2811,0,"",markdown,selection_command
17
+ 16,189015678,"README.md",0,0,"",markdown,tab
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-1f52a67d-2cca-4352-b39c-6cd9f3effae01765648091845-2025_12_13-18.48.27.228/source.csv ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,3,"crates/cli/src/main.rs",0,0,"//! CLI tool for serializing crowd-pilot IDE interaction data.\n//!\n//! This tool processes CSV session files and outputs JSONL format suitable for\n//! NeMo SFT training. It uses an embedded Python interpreter to load HuggingFace\n//! tokenizers for accurate token counting.\n\nuse std::path::PathBuf;\n\nuse clap::Parser;\nuse pyo3::prelude::*;\nuse pyo3::types::PyModule;\n\nuse crowd_pilot_serializer_core::{\n pipeline::{PipelineConfig, PipelineResult},\n process_all_sessions, write_jsonl_output, Tokenizer,\n};\n\n/// Serialize crowd-pilot CSV sessions to NeMo JSONL format.\n#[derive(Parser, Debug)]\n#[command(name = ""crowd-pilot-serialize"")]\n#[command(author, version, about, long_about = None)]\nstruct Args {\n /// Root directory containing CSV session files\n #[arg(long)]\n csv_root: PathBuf,\n\n /// Output directory for JSONL files\n #[arg(long)]\n output_dir: PathBuf,\n\n /// HuggingFace tokenizer model name or path\n #[arg(long)]\n tokenizer: String,\n\n /// Maximum tokens per conversation chunk\n #[arg(long, default_value = ""8192"")]\n max_tokens_per_conversation: usize,\n\n /// Maximum tokens per message\n #[arg(long, default_value = ""2048"")]\n max_tokens_per_message: usize,\n\n /// Minimum messages required to keep a conversation\n #[arg(long, default_value = ""5"")]\n min_conversation_messages: usize,\n\n /// Viewport radius (lines above/below cursor)\n #[arg(long, default_value = ""10"")]\n viewport_radius: usize,\n\n /// Coalesce radius for grouping nearby edits\n #[arg(long, default_value = ""5"")]\n coalesce_radius: usize,\n\n /// Fraction of sessions for validation (0.0-1.0)\n #[arg(long, default_value = ""0.1"")]\n val_ratio: f64,\n\n /// Custom system prompt (optional)\n #[arg(long)]\n system_prompt: Option<String>,\n}\n\nconst DEFAULT_SYSTEM_PROMPT: &str = r#""You are a helpful assistant that can interact multiple times with a computer shell to solve programming tasks.\nYour response must contain exactly ONE bash code block with ONE command (or commands connected with && or ||).\n\nFormat your response as shown in <format_example>.\n\n<format_example>\n```bash\nyour_command_here\n```\n</format_example>\n\nFailure to follow these rules will cause your response to be rejected.""#;\n\n/// Wrapper around Python tokenizer for exact token counting and truncation.\nstruct PythonTokenizer {\n tokenizer: Py<PyAny>,\n}\n\nimpl PythonTokenizer {\n /// Load a HuggingFace tokenizer.\n fn load(model_name: &str) -> PyResult<Self> {\n Python::with_gil(|py| {\n let transformers = PyModule::import(py, ""transformers"")?;\n let auto_tokenizer = transformers.getattr(""AutoTokenizer"")?;\n let tokenizer = auto_tokenizer.call_method1(""from_pretrained"", (model_name,))?;\n Ok(Self {\n tokenizer: tokenizer.into(),\n })\n })\n }\n}\n\nimpl Tokenizer for PythonTokenizer {\n fn count_tokens(&self, text: &str) -> usize {\n Python::with_gil(|py| {\n let tokenizer = self.tokenizer.as_ref(py);\n let tokens = tokenizer\n .call_method1(""encode"", (text,))\n .expect(""Failed to encode text with tokenizer"");\n tokens.len().unwrap()\n })\n }\n\n fn truncate_to_max_tokens(&self, text: &str, max_tokens: usize) -> String {\n Python::with_gil(|py| {\n let tokenizer = self.tokenizer.as_ref(py);\n let kwargs = pyo3::types::PyDict::new(py);\n kwargs.set_item(""max_length"", max_tokens).unwrap();\n kwargs.set_item(""truncation"", true).unwrap();\n \n let tokens = tokenizer\n .call_method(""encode"", (text,), Some(kwargs))\n .expect(""Failed to encode text with tokenizer"");\n \n tokenizer\n .call_method1(""decode"", (tokens,))\n .expect(""Failed to decode tokens"")\n .extract()\n .unwrap()\n })\n }\n}\n\nfn main() -> Result<(), Box<dyn std::error::Error>> {\n let args = Args::parse();\n\n println!(""Loading tokenizer from {}..."", args.tokenizer);\n let tokenizer = PythonTokenizer::load(&args.tokenizer)?;\n\n let config = PipelineConfig {\n max_tokens_per_conversation: args.max_tokens_per_conversation,\n max_tokens_per_message: args.max_tokens_per_message,\n min_conversation_messages: args.min_conversation_messages,\n viewport_radius: args.viewport_radius,\n coalesce_radius: args.coalesce_radius,\n val_ratio: args.val_ratio,\n };\n\n println!(""Processing CSV files from {:?}..."", args.csv_root);\n let session_results = process_all_sessions(\n &args.csv_root,\n &tokenizer,\n &config,\n )?;\n\n let total_sessions = session_results.len();\n println!(""Processed {} sessions"", total_sessions);\n\n let system_prompt = args.system_prompt.as_deref().unwrap_or(DEFAULT_SYSTEM_PROMPT);\n\n println!(""Writing output to {:?}..."", args.output_dir);\n let result: PipelineResult = write_jsonl_output(\n session_results,\n &args.output_dir,\n args.val_ratio,\n system_prompt,\n )?;\n\n let metadata_path = args.output_dir.join(""metadata.json"");\n let metadata = serde_json::json!({\n ""config"": {\n ""csv_root"": args.csv_root.to_string_lossy(),\n ""output_dir"": args.output_dir.to_string_lossy(),\n ""tokenizer"": args.tokenizer,\n ""max_tokens_per_conversation"": args.max_tokens_per_conversation,\n ""max_tokens_per_message"": args.max_tokens_per_message,\n ""min_conversation_messages"": args.min_conversation_messages,\n ""viewport_radius"": args.viewport_radius,\n ""coalesce_radius"": args.coalesce_radius,\n ""val_ratio"": args.val_ratio,\n },\n ""counts"": {\n ""total_sessions"": result.total_sessions,\n ""total_conversations"": result.total_conversations,\n ""train_conversations"": result.train_conversations,\n ""val_conversations"": result.val_conversations,\n },\n ""stats"": {\n ""total_messages"": result.total_messages,\n ""total_tokens"": result.total_tokens,\n ""avg_messages_per_conversation"": if result.total_conversations > 0 {\n result.total_messages as f64 / result.total_conversations as f64\n } else {\n 0.0\n },\n ""avg_tokens_per_conversation"": if result.total_conversations > 0 {\n result.total_tokens as f64 / result.total_conversations as f64\n } else {\n 0.0\n },\n },\n ""files"": {\n ""train_path"": args.output_dir.join(""training.jsonl"").to_string_lossy(),\n ""val_path"": args.output_dir.join(""validation.jsonl"").to_string_lossy(),\n },\n });\n std::fs::write(&metadata_path, serde_json::to_string_pretty(&metadata)?)?;\n\n println!(""\n[summary]"");\n println!("" Total sessions processed: {}"", result.total_sessions);\n println!("" Train conversations: {}"", result.train_conversations);\n println!("" Val conversations: {}"", result.val_conversations);\n println!("" Total messages: {}"", result.total_messages);\n println!("" Total tokens: {}"", result.total_tokens);\n println!("" Output: {:?}/{{training,validation}}.jsonl"", args.output_dir);\n println!("" Metadata: {:?}"", metadata_path);\n\n Ok(())\n}\n\n",rust,tab
3
+ 2,248,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"6:48:27 PM [info] Activating crowd-code\n6:48:27 PM [info] Recording started\n6:48:27 PM [info] Initializing git provider using file system watchers...\n6:48:27 PM [info] Git repository found\n",Log,tab
4
+ 3,345,"extension-output-pdoom-org.crowd-code-#1-crowd-code",189,0,"6:48:27 PM [info] Git provider initialized successfully\n6:48:27 PM [info] Initial git state: [object Object]\n",Log,content
5
+ 4,3836,"TERMINAL",0,0,"",,terminal_focus
6
+ 5,3838,"crates/cli/src/main.rs",0,0,"",rust,tab
7
+ 6,5667,"TERMINAL",0,0,"squeue",,terminal_command
8
+ 7,5680,"TERMINAL",0,0,"]633;C JOBID USER PARTITION NODES CPUS ST SUBMIT_TIME START_TIME TIME TIME_LIMIT NODELIST(REASON)\r\n 36313 xiao.liu interacti 1 128 R 2025-12-13T12:56:48 2025-12-13T12:56:48 5:51:44 23:59:00 hai005\r\n 36303 xiao.liu interacti 1 128 R 2025-12-12T23:25:52 2025-12-12T23:25:52 19:22:40 23:59:00 hai006\r\n 36317 mihir.maha standard 1 10 R 2025-12-13T15:27:10 2025-12-13T15:27:10 3:21:22 1-00:00:00 hai004\r\n 36314 xiao.liu standard 1 128 R 2025-12-13T13:27:57 2025-12-13T15:13:56 3:34:36 23:59:00 hai007\r\n 36304 nishant.ku standard 3 624 R 2025-12-13T07:56:43 2025-12-13T07:56:43 10:51:49 1-00:00:00 hai[001-003]\r\n]0;franz.srambical@hai-login1:~/crowd-pilot-serializer",,terminal_output
9
+ 8,8439,"TERMINAL",0,0,"bash",,terminal_focus
10
+ 9,9983,"TERMINAL",0,0,"bash",,terminal_focus
11
+ 10,27972,"TERMINAL",0,0,"source .venv/bin/activate",,terminal_command
12
+ 11,27988,"TERMINAL",0,0,"]633;C]0;franz.srambical@hai-login1:~/crowd-pilot-serializer",,terminal_output
13
+ 12,30366,"TERMINAL",0,0,"cargo clean",,terminal_command
14
+ 13,30367,"TERMINAL",0,0,"]633;C",,terminal_output
15
+ 14,30559,"TERMINAL",0,0," Cleaning [ ] 0.00% \r",,terminal_output
16
+ 15,30651,"TERMINAL",0,0," Cleaning [ ] 0.63% \r",,terminal_output
17
+ 16,30739,"TERMINAL",0,0," Cleaning [ ] 1.28% \r",,terminal_output
18
+ 17,30856,"TERMINAL",0,0," Cleaning [ ] 1.91% \r",,terminal_output
19
+ 18,31022,"TERMINAL",0,0," Cleaning [ ] 2.58% \r",,terminal_output
20
+ 19,31127,"TERMINAL",0,0," Cleaning [ ] 3.21% \r",,terminal_output
21
+ 20,31249,"TERMINAL",0,0," Cleaning [ ] 3.89% \r Cleaning [> ] 4.58% \r",,terminal_output
22
+ 21,31341,"TERMINAL",0,0," Cleaning [> ] 5.30% \r",,terminal_output
23
+ 22,31473,"TERMINAL",0,0," Cleaning [> ] 6.09% \r",,terminal_output
24
+ 23,31558,"TERMINAL",0,0," Cleaning [> ] 6.97% \r",,terminal_output
25
+ 24,31713,"TERMINAL",0,0," Cleaning [> ] 7.82% \r",,terminal_output
26
+ 25,31775,"TERMINAL",0,0," Cleaning [=> ] 8.70% \r",,terminal_output
27
+ 26,31883,"TERMINAL",0,0," Cleaning [=> ] 9.57% \r",,terminal_output
28
+ 27,32024,"TERMINAL",0,0," Cleaning [=> ] 10.43% \r",,terminal_output
29
+ 28,32160,"TERMINAL",0,0," Cleaning [=> ] 11.28% \r",,terminal_output
30
+ 29,32237,"TERMINAL",0,0," Cleaning [==> ] 12.13% \r",,terminal_output
31
+ 30,32307,"TERMINAL",0,0," Cleaning [==> ] 12.99% \r",,terminal_output
32
+ 31,32403,"TERMINAL",0,0," Cleaning [==> ] 13.82% \r",,terminal_output
33
+ 32,32476,"TERMINAL",0,0," Cleaning [==> ] 14.65% \r",,terminal_output
34
+ 33,32598,"TERMINAL",0,0," Cleaning [==> ] 15.51% \r",,terminal_output
35
+ 34,32705,"TERMINAL",0,0," Cleaning [===> ] 16.36% \r",,terminal_output
36
+ 35,32785,"TERMINAL",0,0," Cleaning [===> ] 17.21% \r",,terminal_output
37
+ 36,32907,"TERMINAL",0,0," Cleaning [===> ] 18.07% \r",,terminal_output
38
+ 37,32989,"TERMINAL",0,0," Cleaning [===> ] 18.92% \r",,terminal_output
39
+ 38,33122,"TERMINAL",0,0," Cleaning [===> ] 19.78% \r",,terminal_output
40
+ 39,33221,"TERMINAL",0,0," Cleaning [====> ] 20.63% \r",,terminal_output
41
+ 40,33352,"TERMINAL",0,0," Cleaning [====> ] 21.48% \r",,terminal_output
42
+ 41,33419,"TERMINAL",0,0," Cleaning [====> ] 22.16% \r",,terminal_output
43
+ 42,33518,"TERMINAL",0,0," Cleaning [====> ] 22.79% \r",,terminal_output
44
+ 43,33637,"TERMINAL",0,0," Cleaning [====> ] 23.44% \r",,terminal_output
45
+ 44,33812,"TERMINAL",0,0," Cleaning [=====> ] 24.07% \r",,terminal_output
46
+ 45,33894,"TERMINAL",0,0," Cleaning [=====> ] 24.67% \r Cleaning [=====> ] 25.37% \r",,terminal_output
47
+ 46,34019,"TERMINAL",0,0," Cleaning [=====> ] 26.02% \r",,terminal_output
48
+ 47,34217,"TERMINAL",0,0," Cleaning [=====> ] 26.72% \r Cleaning [=====> ] 27.33% \r",,terminal_output
49
+ 48,34297,"TERMINAL",0,0," Cleaning [======> ] 28.02% \r",,terminal_output
50
+ 49,34407,"TERMINAL",0,0," Cleaning [======> ] 28.74% \r",,terminal_output
51
+ 50,34526,"TERMINAL",0,0," Cleaning [======> ] 29.42% \r",,terminal_output
52
+ 51,34743,"TERMINAL",0,0," Cleaning [======> ] 30.09% \r",,terminal_output
53
+ 52,34746,"TERMINAL",0,0," Cleaning [======> ] 30.76% \r",,terminal_output
54
+ 53,34813,"TERMINAL",0,0," Cleaning [======> ] 31.42% \r",,terminal_output
55
+ 54,34910,"TERMINAL",0,0," Cleaning [=======> ] 32.07% \r",,terminal_output
56
+ 55,35053,"TERMINAL",0,0," Cleaning [=======> ] 32.72% \r",,terminal_output
57
+ 56,35073,"TERMINAL",0,0," Cleaning [=======> ] 33.33% \rerror: failed to remove file `/fast/home/franz.srambical/crowd-pilot-serializer/target/debug/deps/.nfs000000012d584f9d00000407`\r\n\r\nCaused by:\r\n Device or resource busy (os error 16)\r\n\r\nError: failed to remove directory `/fast/home/franz.srambical/crowd-pilot-serializer/target/debug/deps`\r\n\r\nCaused by:\r\n Device or resource busy (os error 16)\r\n]0;franz.srambical@hai-login1:~/crowd-pilot-serializer",,terminal_output
58
+ 57,41510,"TERMINAL",0,0,"bash",,terminal_focus
59
+ 58,43518,"TERMINAL",0,0,"exit",,terminal_command
60
+ 59,44223,"TERMINAL",0,0,"]633;Cexit\r\n",,terminal_output
61
+ 60,44226,"TERMINAL",0,0,"bash",,terminal_focus
62
+ 61,52525,"TERMINAL",0,0,"rm -rf target/",,terminal_command
63
+ 62,52574,"TERMINAL",0,0,"]633;Crm: cannot remove 'target/debug/deps/.nfs000000012d584f9d00000407': Device or resource busy\r\nrm: cannot remove 'target/debug/deps/.nfs000000012ce413230000040b': Device or resource busy\r\nrm: cannot remove 'target/debug/deps/.nfs000000012cf90e1000000408': Device or resource busy\r\nrm: cannot remove 'target/debug/deps/.nfs000000012cb3183b0000040a': Device or resource busy\r\nrm: cannot remove 'target/debug/deps/.nfs000000012c0f154900000406': Device or resource busy\r\nrm: cannot remove 'target/debug/deps/.nfs000000012c166c5400000409': Device or resource busy\r\nrm: cannot remove 'target/debug/deps/.nfs000000012d1819c400000405': Device or resource busy\r\n",,terminal_output
64
+ 63,56197,"TERMINAL",0,0,"^C\r\n]0;franz.srambical@hai-login1:~/crowd-pilot-serializer",,terminal_output
65
+ 64,59615,"TERMINAL",0,0,"squeue",,terminal_command
66
+ 65,59616,"TERMINAL",0,0,"]633;C JOBID USER PARTITION NODES CPUS ST SUBMIT_TIME START_TIME TIME TIME_LIMIT NODELIST(REASON)\r\n 36313 xiao.liu interacti 1 128 R 2025-12-13T12:56:48 2025-12-13T12:56:48 5:52:38 23:59:00 hai005\r\n 36303 xiao.liu interacti 1 128 R 2025-12-12T23:25:52 2025-12-12T23:25:52 19:23:34 23:59:00 hai006\r\n 36317 mihir.maha standard 1 10 R 2025-12-13T15:27:10 2025-12-13T15:27:10 3:22:16 1-00:00:00 hai004\r\n 36314 xiao.liu standard 1 128 R 2025-12-13T13:27:57 2025-12-13T15:13:56 3:35:30 23:59:00 hai007\r\n 36304 nishant.ku standard 3 624 R 2025-12-13T07:56:43 2025-12-13T07:56:43 10:52:43 1-00:00:00 hai[001-003]\r\n]0;franz.srambical@hai-login1:~/crowd-pilot-serializer",,terminal_output
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-2790c5c3-9c96-49ac-ab9e-481a8033e7cd1757945282635-2025_09_15-16.08.09.625/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-9543d8e2-2376-4957-873e-df7016d502961763465687199-2025_11_18-12.34.49.442/source.csv ADDED
@@ -0,0 +1,245 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 2,67,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"12:34:49 PM [info] Activating crowd-code\n12:34:49 PM [info] Recording started\n12:34:49 PM [info] Initializing git provider using file system watchers...\n12:34:49 PM [info] Git repository found\n12:34:49 PM [info] Git provider initialized successfully\n",Log,tab
3
+ 3,116,"extension-output-pdoom-org.crowd-code-#1-crowd-code",250,0,"12:34:49 PM [info] Initial git state: [object Object]\n",Log,content
4
+ 4,2975,"extension-output-pdoom-org.crowd-code-#1-crowd-code",304,0,"",Log,selection_mouse
5
+ 5,13030,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,tab
6
+ 6,20469,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",64,0,"",python,selection_command
7
+ 7,20661,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",66,0,"",python,selection_command
8
+ 8,20694,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",132,0,"",python,selection_command
9
+ 9,20728,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",199,0,"",python,selection_command
10
+ 10,20761,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",241,0,"",python,selection_command
11
+ 11,20794,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",243,0,"",python,selection_command
12
+ 12,20828,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",292,0,"",python,selection_command
13
+ 13,20863,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",294,0,"",python,selection_command
14
+ 14,20904,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",364,0,"",python,selection_command
15
+ 15,20936,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",432,0,"",python,selection_command
16
+ 16,20962,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",507,0,"",python,selection_command
17
+ 17,20996,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",577,0,"",python,selection_command
18
+ 18,21030,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",610,0,"",python,selection_command
19
+ 19,21068,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",611,0,"",python,selection_command
20
+ 20,21097,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",639,0,"",python,selection_command
21
+ 21,21131,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",640,0,"",python,selection_command
22
+ 22,21166,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",671,0,"",python,selection_command
23
+ 23,21203,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",694,0,"",python,selection_command
24
+ 24,21233,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",707,0,"",python,selection_command
25
+ 25,21265,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",795,0,"",python,selection_command
26
+ 26,21297,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",796,0,"",python,selection_command
27
+ 27,21331,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",852,0,"",python,selection_command
28
+ 28,21364,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",914,0,"",python,selection_command
29
+ 29,21397,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",965,0,"",python,selection_command
30
+ 30,21431,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1047,0,"",python,selection_command
31
+ 31,21464,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1148,0,"",python,selection_command
32
+ 32,21497,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1245,0,"",python,selection_command
33
+ 33,21530,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1319,0,"",python,selection_command
34
+ 34,21565,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1369,0,"",python,selection_command
35
+ 35,21598,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1370,0,"",python,selection_command
36
+ 36,21631,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1393,0,"",python,selection_command
37
+ 37,21668,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1394,0,"",python,selection_command
38
+ 38,21701,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1395,0,"",python,selection_command
39
+ 39,21890,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1423,0,"",python,selection_command
40
+ 40,23050,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1427,0,"",python,selection_command
41
+ 41,23197,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1432,0,"",python,selection_command
42
+ 42,23489,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1435,0,"",python,selection_command
43
+ 43,23702,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1438,0,"",python,selection_command
44
+ 44,23912,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1441,0,"",python,selection_command
45
+ 45,24119,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1442,0,"",python,selection_command
46
+ 46,24327,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1448,0,"",python,selection_command
47
+ 47,24581,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1449,0,"",python,selection_command
48
+ 48,24945,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1451,0,"",python,selection_command
49
+ 49,25331,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1452,0,"",python,selection_command
50
+ 50,26280,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1423,0,"",python,selection_command
51
+ 51,26667,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1395,0,"",python,selection_command
52
+ 52,26920,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1394,0,"",python,selection_command
53
+ 53,26954,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1393,0,"",python,selection_command
54
+ 54,146021,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
55
+ 55,146654,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,tab
56
+ 56,203332,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2838,0,"",python,selection_keyboard
57
+ 57,203923,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",5124,0,"",python,selection_keyboard
58
+ 58,204398,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7126,0,"",python,selection_keyboard
59
+ 59,204829,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9057,0,"",python,selection_keyboard
60
+ 60,205408,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_keyboard
61
+ 61,210105,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7721,0,"",python,selection_keyboard
62
+ 62,210235,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",5922,0,"",python,selection_keyboard
63
+ 63,210382,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3490,0,"",python,selection_keyboard
64
+ 64,210771,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1906,0,"",python,selection_keyboard
65
+ 65,210944,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",64,0,"",python,selection_keyboard
66
+ 66,211099,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_keyboard
67
+ 67,214839,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10,0,"",python,selection_command
68
+ 68,2267039,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2294,0,"",python,selection_mouse
69
+ 69,2272461,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2114,0,"",python,selection_mouse
70
+ 70,3052561,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,10041,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,selection_command
71
+ 71,3052668,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_command
72
+ 72,3145153,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_command
73
+ 73,3148126,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7487,0,"",python,selection_command
74
+ 74,3149770,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7876,0,"",python,selection_command
75
+ 75,3150226,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",8235,0,"",python,selection_command
76
+ 76,3150898,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7876,0,"",python,selection_command
77
+ 77,3151966,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",8235,0,"",python,selection_command
78
+ 78,3161866,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9070,0,"",python,selection_command
79
+ 79,3164224,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9131,0,"",python,selection_command
80
+ 80,3166810,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9194,0,"",python,selection_command
81
+ 81,3167056,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9257,0,"",python,selection_command
82
+ 82,3167090,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9321,0,"",python,selection_command
83
+ 83,3167123,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9386,0,"",python,selection_command
84
+ 84,3167156,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9443,0,"",python,selection_command
85
+ 85,3167189,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9481,0,"",python,selection_command
86
+ 86,3167223,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9531,0,"",python,selection_command
87
+ 87,3168286,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9532,0,"",python,selection_command
88
+ 88,3168527,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9538,0,"",python,selection_command
89
+ 89,3168566,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9539,0,"",python,selection_command
90
+ 90,3168597,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9546,0,"",python,selection_command
91
+ 91,3168633,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9547,0,"",python,selection_command
92
+ 92,3168887,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9555,0,"",python,selection_command
93
+ 93,3169118,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
94
+ 94,3169292,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9583,0,"",python,selection_command
95
+ 95,3169900,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
96
+ 96,3170394,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
97
+ 97,3171165,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
98
+ 98,3171795,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
99
+ 99,3171945,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
100
+ 100,3172433,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
101
+ 101,3173957,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
102
+ 102,3174398,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
103
+ 103,3174801,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9682,0,"",python,selection_command
104
+ 104,3175123,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9746,0,"",python,selection_command
105
+ 105,3175686,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9682,0,"",python,selection_command
106
+ 106,3175821,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
107
+ 107,3175981,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
108
+ 108,3240534,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_command
109
+ 109,3241210,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1898,0,"",python,selection_keyboard
110
+ 110,3241373,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3420,0,"",python,selection_keyboard
111
+ 111,3242843,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3370,0,"",python,selection_command
112
+ 112,3243091,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3298,0,"",python,selection_command
113
+ 113,3243124,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3288,0,"",python,selection_command
114
+ 114,3243158,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3287,0,"",python,selection_command
115
+ 115,3243191,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3211,0,"",python,selection_command
116
+ 116,3243225,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3137,0,"",python,selection_command
117
+ 117,3243258,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3058,0,"",python,selection_command
118
+ 118,3243291,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3057,0,"",python,selection_command
119
+ 119,3243324,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2999,0,"",python,selection_command
120
+ 120,3243357,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2991,0,"",python,selection_command
121
+ 121,3243391,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2973,0,"",python,selection_command
122
+ 122,3243424,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2956,0,"",python,selection_command
123
+ 123,3243458,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2932,0,"",python,selection_command
124
+ 124,3243491,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2915,0,"",python,selection_command
125
+ 125,3243525,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2898,0,"",python,selection_command
126
+ 126,3243558,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2876,0,"",python,selection_command
127
+ 127,3243591,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2854,0,"",python,selection_command
128
+ 128,3243632,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2838,0,"",python,selection_command
129
+ 129,3243658,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2817,0,"",python,selection_command
130
+ 130,3243691,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2793,0,"",python,selection_command
131
+ 131,3243734,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2767,0,"",python,selection_command
132
+ 132,3243758,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2756,0,"",python,selection_command
133
+ 133,3243791,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2721,0,"",python,selection_command
134
+ 134,3243824,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2688,0,"",python,selection_command
135
+ 135,3243857,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2655,0,"",python,selection_command
136
+ 136,3243891,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2621,0,"",python,selection_command
137
+ 137,3243925,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2585,0,"",python,selection_command
138
+ 138,3243958,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2547,0,"",python,selection_command
139
+ 139,3243991,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2512,0,"",python,selection_command
140
+ 140,3244025,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2483,0,"",python,selection_command
141
+ 141,3244059,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2451,0,"",python,selection_command
142
+ 142,3244092,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2427,0,"",python,selection_command
143
+ 143,3244125,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2388,0,"",python,selection_command
144
+ 144,3286103,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2411,0,"",python,selection_command
145
+ 145,3296751,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2388,23,"",python,content
146
+ 146,3298859,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2388,0," sequence_parallelis",python,content
147
+ 147,3298862,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2388,0,"",python,selection_command
148
+ 148,3483131,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
149
+ 149,3484741,"TERMINAL",0,0,"",,terminal_focus
150
+ 150,3484752,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,tab
151
+ 151,3788287,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2427,0,"",python,selection_command
152
+ 152,3788457,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2428,0,"",python,selection_command
153
+ 153,3788946,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2429,0,"",python,selection_command
154
+ 154,3789583,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2427,0,"",python,selection_command
155
+ 155,4549882,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,10041,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,selection_command
156
+ 156,4550471,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_command
157
+ 157,4559897,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,10041,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,selection_command
158
+ 158,4560021,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_command
159
+ 159,4560890,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,content
160
+ 160,4583906,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18686,0,"",python,selection_mouse
161
+ 161,4583914,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18685,0,"",python,selection_command
162
+ 162,4584529,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18699,0,"",python,selection_command
163
+ 163,4585394,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18743,0,"",python,selection_command
164
+ 164,4585610,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18752,0,"",python,selection_command
165
+ 165,4591082,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,tab
166
+ 166,4591147,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",64,0,"",python,selection_command
167
+ 167,4595172,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,10041,"",python,content
168
+ 168,4597038,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,tab
169
+ 169,5454427,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9372,0,"",python,selection_mouse
170
+ 170,5454796,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_command
171
+ 171,5455466,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",64,0,"",python,selection_command
172
+ 172,5455793,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1906,0,"",python,selection_keyboard
173
+ 173,5456269,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3490,0,"",python,selection_keyboard
174
+ 174,5456884,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3420,0,"",python,selection_command
175
+ 175,5457139,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3370,0,"",python,selection_command
176
+ 176,5457172,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3298,0,"",python,selection_command
177
+ 177,5457202,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3288,0,"",python,selection_command
178
+ 178,5457232,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3287,0,"",python,selection_command
179
+ 179,5457265,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3211,0,"",python,selection_command
180
+ 180,5457298,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3137,0,"",python,selection_command
181
+ 181,5457331,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3058,0,"",python,selection_command
182
+ 182,5457367,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3057,0,"",python,selection_command
183
+ 183,5457402,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2999,0,"",python,selection_command
184
+ 184,5457431,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2991,0,"",python,selection_command
185
+ 185,5457465,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2973,0,"",python,selection_command
186
+ 186,5457497,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2956,0,"",python,selection_command
187
+ 187,5457533,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2932,0,"",python,selection_command
188
+ 188,5457565,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2915,0,"",python,selection_command
189
+ 189,5457598,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2898,0,"",python,selection_command
190
+ 190,5457631,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2876,0,"",python,selection_command
191
+ 191,5457665,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2854,0,"",python,selection_command
192
+ 192,5457699,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2838,0,"",python,selection_command
193
+ 193,5457732,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2817,0,"",python,selection_command
194
+ 194,5457764,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2793,0,"",python,selection_command
195
+ 195,5457799,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2767,0,"",python,selection_command
196
+ 196,5457833,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2756,0,"",python,selection_command
197
+ 197,5457866,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2721,0,"",python,selection_command
198
+ 198,5457899,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2688,0,"",python,selection_command
199
+ 199,5457931,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2655,0,"",python,selection_command
200
+ 200,5457964,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2621,0,"",python,selection_command
201
+ 201,5457999,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2585,0,"",python,selection_command
202
+ 202,5458033,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2547,0,"",python,selection_command
203
+ 203,5458067,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2512,0,"",python,selection_command
204
+ 204,5458102,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2483,0,"",python,selection_command
205
+ 205,5458137,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2451,0,"",python,selection_command
206
+ 206,5459402,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2483,0,"",python,selection_command
207
+ 207,5459593,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2512,0,"",python,selection_command
208
+ 208,5459748,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2547,0,"",python,selection_command
209
+ 209,5460152,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2551,0,"",python,selection_command
210
+ 210,5461248,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2589,0,"",python,selection_command
211
+ 211,5461431,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2625,0,"",python,selection_command
212
+ 212,5461588,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2659,0,"",python,selection_command
213
+ 213,5461782,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2692,0,"",python,selection_command
214
+ 214,5462187,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2725,0,"",python,selection_command
215
+ 215,5462694,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2692,0,"",python,selection_command
216
+ 216,5462868,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2659,0,"",python,selection_command
217
+ 217,5463112,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2625,0,"",python,selection_command
218
+ 218,5463146,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2589,0,"",python,selection_command
219
+ 219,5463184,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2551,0,"",python,selection_command
220
+ 220,5463214,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2516,0,"",python,selection_command
221
+ 221,5463329,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2487,0,"",python,selection_command
222
+ 222,8520791,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,10041,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,selection_command
223
+ 223,8520860,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_command
224
+ 224,8685736,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2797,0,"",python,selection_command
225
+ 225,8686757,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2813,0,"",python,selection_command
226
+ 226,8686946,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2814,0,"",python,selection_command
227
+ 227,10758031,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2813,0,"",python,selection_command
228
+ 228,10758522,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2797,0,"",python,selection_command
229
+ 229,10759094,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2813,0,"",python,selection_command
230
+ 230,10759461,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2797,0,"",python,selection_command
231
+ 231,10762705,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2771,0,"",python,selection_command
232
+ 232,10880161,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2797,0,"",python,selection_command
233
+ 233,14746153,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9275,0,"",python,selection_mouse
234
+ 234,14750022,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9465,0,"",python,selection_mouse
235
+ 235,14757418,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9932,0,"",python,selection_mouse
236
+ 236,14758360,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9464,0,"",python,selection_mouse
237
+ 237,15922202,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_command
238
+ 238,15924565,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1008,0,"",python,selection_command
239
+ 239,15925176,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1016,0,"",python,selection_command
240
+ 240,15925334,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1023,0,"",python,selection_command
241
+ 241,15925510,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1047,0,"",python,selection_command
242
+ 242,15925934,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1023,0,"",python,selection_command
243
+ 243,15926409,"nemo/collections/llm/recipes/finetune_default.py",0,0,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import TYPE_CHECKING, Any, Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\n\nimport nemo.lightning as nl\nfrom nemo.collections import llm\nfrom nemo.collections.llm.gpt.data.packed_sequence import PackedSequenceSpecs\nfrom nemo.collections.llm.peft import DoRA, LoRA\nfrom nemo.collections.llm.recipes.log.default import tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.precision.mixed_precision import bf16_mixed\nfrom nemo.lightning.pytorch.callbacks import PEFT\nfrom nemo.utils.exp_manager import TimingCallback\n\nif TYPE_CHECKING:\n from lightning.pytorch.loggers import TensorBoardLogger, WandbLogger\n\nTokenizerType = Any\n\n\ndef default_finetune_recipe(\n model: run.Config[pl.LightningModule],\n resume_path: str,\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n packed_sequence: bool = False, # once packing recipe is well tested, change this default to true\n tokenizer: Optional[TokenizerType] = ""model"",\n) -> run.Partial:\n """"""\n Create a default fine-tuning recipe for any model.\n\n This function sets up a template for a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n\n Args:\n model (run.Config[pl.LightningModule]): Configuration for a NeMo model.\n resume_path (str): Path to the Huggingface model or pretrained distributed checkpoint for resume\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n packed_sequence (bool): Whether to use packed sequence.\n tokenizer (Optional[TokenizerType]): Tokenizer setting to be applied. Can be 'data' or 'model'\n or an instance of TokenizerSpec.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n See usages of this recipe for further details.\n """"""\n if packed_sequence:\n datamodule = run.Config(\n llm.SquadDataModule,\n seq_length=2048,\n global_batch_size=8,\n micro_batch_size=1,\n packed_sequence_specs=PackedSequenceSpecs(packed_sequence_size=2048),\n )\n else:\n datamodule = run.Config(llm.SquadDataModule, seq_length=2048, global_batch_size=128, micro_batch_size=1)\n recipe = run.Partial(\n llm.finetune,\n model=model,\n trainer=default_finetune_trainer(\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n ),\n data=datamodule,\n log=default_finetune_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(max_lr=1e-4, min_lr=0, warmup_steps=50, adam_beta2=0.98),\n resume=nemo_resume(resume_path),\n tokenizer=tokenizer,\n )\n\n return recipe\n\n\ndef default_finetune_trainer(\n tensor_parallelism=1,\n pipeline_parallelism=1,\n pipeline_parallelism_type=torch.bfloat16,\n virtual_pipeline_parallelism=None,\n context_parallelism=1,\n sequence_parallelism=False,\n num_nodes=1,\n num_gpus_per_node=8,\n max_steps=1000,\n limit_test_batches=None,\n limit_val_batches=None,\n val_check_interval=30,\n):\n """"""\n Create a default fine-tuning trainer for any model.\n\n This function sets up a template for strategy and trainer.\n\n Args:\n See docstrings of MegatronStrategy and Trainer.\n\n Returns:\n run.Config: Config for a finetuning trainer.\n\n See usages of this in recipes for further details.\n """"""\n strategy = run.Config(\n nl.MegatronStrategy,\n tensor_model_parallel_size=tensor_parallelism,\n pipeline_model_parallel_size=pipeline_parallelism,\n pipeline_dtype=pipeline_parallelism_type,\n virtual_pipeline_model_parallel_size=virtual_pipeline_parallelism,\n context_parallel_size=context_parallelism,\n sequence_parallel=sequence_parallelism,\n gradient_as_bucket_view=True,\n ckpt_load_strictness=""log_all"",\n )\n\n trainer = run.Config(\n nl.Trainer,\n accelerator=""gpu"",\n accumulate_grad_batches=1,\n devices=num_gpus_per_node,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=1,\n max_steps=max_steps,\n num_nodes=num_nodes,\n plugins=bf16_mixed(),\n strategy=strategy,\n use_distributed_sampler=False,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n )\n\n return trainer\n\n\ndef default_finetune_log(\n dir: Optional[str] = None,\n name: str = ""default"",\n tensorboard_logger: Optional[run.Config['TensorBoardLogger']] = None,\n wandb_logger: Optional[run.Config['WandbLogger']] = None,\n) -> run.Config[nl.NeMoLogger]:\n """"""\n Create a default fine-tuning logger for any model.\n\n This function sets up a template for ModelCheckpoint and NeMoLogger.\n\n Args:\n See docstrings of ModelCheckpoint and NeMoLogger.\n\n Returns:\n run.Config: Config for a finetuning NeMoLogger.\n\n See usages of this in recipes for further details.\n """"""\n\n ckpt = run.Config(\n nl.ModelCheckpoint,\n save_last=""link"",\n save_top_k=2,\n every_n_train_steps=50,\n filename=""{model_name}--{val_loss:.2f}-{step}-{consumed_samples}"",\n )\n\n return run.Config(\n nl.NeMoLogger,\n ckpt=ckpt,\n name=name,\n tensorboard=tensorboard_logger,\n wandb=wandb_logger,\n log_dir=dir,\n )\n\n\ndef nemo_resume(model_id: str) -> run.Config[nl.AutoResume]:\n """"""\n Configure automatic resumption from a NeMo checkpoint converted from Huggingface for\n https://huggingface.co/{model_id}.\n\n This NeMo checkpoint should be converted from Huggingface beforehand, using nemo.collections.llm.import_ckpt.\n When converting the checkpoint, the NeMo checkpoint will be saved in NEMO_HOME (set to ~/.cache/nemo by default).\n\n This function sets up the configuration to resume training from path nemo://{model_id}.\n This translates to the full path {NEMO_HOME}/models/{model_id}.\n\n Args:\n model_id (str): Path to the Huggingface model or pretrained distributed checkpoint for resume\n\n Returns:\n run.Config[nl.AutoResume]: Configuration for resuming from NeMo checkpoint.\n """"""\n return run.Config(\n nl.AutoResume,\n restore_config=run.Config(nl.RestoreConfig, path=f""nemo://{model_id}""),\n )\n\n\n@run.cli.factory(name='lora')\ndef lora() -> run.Config[PEFT]:\n """"""\n Factory function to create a LoRA configuration.\n\n Returns:\n run.Config[PEFT]: Configuration for the LoRA class.\n\n Examples:\n CLI usage:\n $ nemo llm finetune -f llama3_8b peft=lora\n\n Python API usage:\n >>> lora_config = lora()\n >>> print(lora_config)\n """"""\n return run.Config(LoRA)\n\n\n@run.cli.factory(name='dora')\ndef dora() -> run.Config[PEFT]:\n """"""\n Factory function to create a DoRA configuration.\n\n Returns:\n run.Config[PEFT]: Configuration for the DoRA class.\n\n Examples:\n CLI usage:\n $ nemo llm finetune -f llama3_8b peft=dora\n\n Python API usage:\n >>> dora_config = dora()\n >>> print(dora_config)\n """"""\n return run.Config(DoRA)\n",python,tab
244
+ 244,15926417,"nemo/collections/llm/recipes/finetune_default.py",1382,0,"",python,selection_command
245
+ 245,15929448,"nemo/collections/llm/recipes/finetune_default.py",3372,0,"",python,selection_keyboard
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-ac4f665d-1bc1-467a-98b3-5da2178968731760857710663-2025_10_19-09.08.50.386/source.csv ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,3,"slurm/dev/franz/berlin/coinrun/sample/maskgit/sample_mila_submission_case_study_vanilla.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=24:00:00\n#SBATCH --cpus-per-task=8\n#SBATCH --gres=gpu:1\n#SBATCH --output=/fast/project/HFMI_SynergyUnit/jafar_ws/logs/franz/coinrun/dynamics_sample/%x_%j.log\n#SBATCH --error=/fast/project/HFMI_SynergyUnit/jafar_ws/logs/franz/coinrun/dynamics_sample/%x_%j.log\n#SBATCH --job-name=coinrun_sample_maskgit_mila_submission_case_study_vanilla\n\n# Activate virtual environment\nsource .venv/bin/activate\n\narray_records_dir=""/fast/project/HFMI_SynergyUnit/jafar_ws/data/coinrun/array_records_10M_npy_arr_rec/array_record/test""\nCHECKPOINT_PATH=""/fast/project/HFMI_SynergyUnit/jafar_ws/checkpoints/coinrun/dynamics/dynamics_case_study_dataset_10M_30031""\n\ncurrent_branch=$(git rev-parse --abbrev-ref HEAD)\nif [ ""$current_branch"" != ""main"" ]; then\n echo ""This script must be run from the main branch. Current branch is $current_branch. Exiting.""\n exit 1\nfi\n\necho ""Sampling from checkpoint: $CHECKPOINT_PATH""\n\nsrun python jasmine/sample.py \\n --seed=1 \\n --maskgit_steps=1 \\n --tokenizer_ffn_dim=512 \\n --tokenizer_num_blocks=8 \\n --dyna_ffn_dim=512 \\n --dyna_num_blocks=12 \\n --output_dir=gifs/dynamics_case_study_dataset_10M_vanilla \\n --checkpoint $CHECKPOINT_PATH \\n --data_dir=$array_records_dir \\n --seq_len=16 \\n --batch_size=32 \\n --patch_size=4 \\n --start_frame=4 \\n --image_height=64 \\n --image_width=64 \\n --dyna_type=maskgit\n",shellscript,tab
3
+ 2,262,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"9:08:50 AM [info] Activating crowd-code\n9:08:50 AM [info] Recording started\n9:08:50 AM [info] Initializing git provider using file system watchers...\n9:08:50 AM [info] Git repository found\n9:08:50 AM [info] Git provider initialized successfully\n9:08:50 AM [info] Initial git state: [object Object]\n",Log,tab
4
+ 3,4665,"TERMINAL",0,0,"",,terminal_command
5
+ 4,11496,"TERMINAL",0,0,"",,terminal_command
6
+ 5,106853,"TERMINAL",0,0,"",,terminal_command
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-c877f8c1-c8e0-4a7b-8720-40bb4df915221754138206219-2025_08_02-14.36.55.608/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-14f0662f-0032-43e8-be9f-8e53d6f150ad1758635869882-2025_09_23-15.57.52.616/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-1fa9b85d-3794-4f3b-b7a0-5170b7d2faaa1762362332596-2025_11_05-18.05.39.648/source.csv ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,3,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",0,0,"/*---------------------------------------------------------------------------------------------\n * Copyright (c) Microsoft Corporation. All rights reserved.\n * Licensed under the MIT License. See License.txt in the project root for license information.\n *--------------------------------------------------------------------------------------------*/\n\n/** @jsxRuntime automatic */\n/** @jsxImportSource ../../../../prompt/jsx-runtime/ */\nimport { CopilotContentExclusionManager, StatusBarEvent } from '../../contentExclusion/contentExclusionManager';\nimport { ICompletionsContextService } from '../../context';\nimport { logger, LogTarget } from '../../logger';\n\nimport { IInstantiationService, ServicesAccessor } from '../../../../../../../util/vs/platform/instantiation/common/instantiation';\nimport { ICompletionsTelemetryService } from '../../../../bridge/src/completionsTelemetryServiceBridge';\nimport { DataPipe, VirtualPrompt } from '../../../../prompt/src/components/virtualPrompt';\nimport { TokenizerName } from '../../../../prompt/src/tokenization';\nimport { CancellationToken, Position } from '../../../../types/src';\nimport { CompletionState } from '../../completionState';\nimport { telemetryException, TelemetryWithExp } from '../../telemetry';\nimport { TextDocumentContents } from '../../textDocument';\nimport { CodeSnippets } from '../components/codeSnippets';\nimport { CompletionsContext } from '../components/completionsContext';\nimport { CompletionsPromptOk, CompletionsPromptRenderer } from '../components/completionsPromptRenderer';\nimport { ContextProviderBridge } from '../components/contextProviderBridge';\nimport { CurrentFile } from '../components/currentFile';\nimport { DocumentMarker } from '../components/marker';\nimport { RecentEdits } from '../components/recentEdits';\nimport { SimilarFiles } from '../components/similarFiles';\nimport { splitContextCompletionsPrompt } from '../components/splitContextPrompt';\nimport { SplitContextPromptRenderer } from '../components/splitContextPromptRenderer';\nimport { Traits } from '../components/traits';\nimport {\n\tContextProviderTelemetry,\n\tmatchContextItems,\n\tResolvedContextItem,\n\ttelemetrizeContextItems,\n\tuseContextProviderAPI,\n} from '../contextProviderRegistry';\nimport { getCodeSnippetsFromContextItems } from '../contextProviders/codeSnippets';\nimport {\n\tCodeSnippetWithId,\n\tSupportedContextItemWithId,\n\tTraitWithId,\n} from '../contextProviders/contextItemSchemas';\nimport { getTraitsFromContextItems, ReportTraitsTelemetry } from '../contextProviders/traits';\nimport { componentStatisticsToPromptMatcher, ContextProviderStatistics } from '../contextProviderStatistics';\nimport {\n\t_contextTooShort,\n\t_copilotContentExclusion,\n\t_promptCancelled,\n\t_promptError,\n\tgetPromptOptions,\n\tMIN_PROMPT_CHARS,\n\tPromptResponse,\n\ttrimLastLine,\n} from '../prompt';\nimport { isIncludeNeighborFilesActive } from '../similarFiles/neighborFiles';\nimport {\n\tCompletionsPromptFactory,\n\tCompletionsPromptOptions,\n\tPromptOpts,\n} from './completionsPromptFactory';\n\nexport type CompletionRequestDocument = TextDocumentContents;\n\nexport type CompletionRequestData = {\n\tdocument: CompletionRequestDocument;\n\tposition: Position;\n\ttelemetryData: TelemetryWithExp;\n\tcancellationToken?: CancellationToken;\n\t// see inlineCompletions data param\n\tdata?: unknown;\n\t// Context provider items\n\ttraits?: TraitWithId[];\n\tcodeSnippets?: CodeSnippetWithId[];\n\tturnOffSimilarFiles?: boolean;\n\tsuffixMatchThreshold?: number;\n\tmaxPromptTokens: number;\n\ttokenizer?: TokenizerName;\n};\n\nexport function isCompletionRequestData(data: unknown): data is CompletionRequestData {\n\tif (!data || typeof data !== 'object') { return false; }\n\n\tconst req = data as Partial<CompletionRequestData>;\n\n\t// Check document\n\tif (!req.document) { return false; }\n\n\t// Check position\n\tif (!req.position) { return false; }\n\tif (req.position.line === undefined) { return false; }\n\tif (req.position.character === undefined) { return false; }\n\n\t// Check telemetryData\n\tif (!req.telemetryData) { return false; }\n\n\treturn true;\n}\n\nexport enum PromptOrdering {\n\tDefault = 'default',\n\tSplitContext = 'splitContext',\n}\n\ntype DeclarativePromptFunction = typeof defaultCompletionsPrompt;\ntype AvailableDeclarativePrompts = {\n\t[K in PromptOrdering]: {\n\t\tpromptFunction: DeclarativePromptFunction;\n\t\trenderer: typeof CompletionsPromptRenderer;\n\t};\n};\n\nconst availableDeclarativePrompts: AvailableDeclarativePrompts = {\n\t[PromptOrdering.Default]: {\n\t\tpromptFunction: defaultCompletionsPrompt,\n\t\trenderer: CompletionsPromptRenderer,\n\t},\n\t[PromptOrdering.SplitContext]: {\n\t\tpromptFunction: splitContextCompletionsPrompt,\n\t\trenderer: SplitContextPromptRenderer,\n\t},\n};\n\n// The weights mimic the PromptPriorityList from prompt/src/wishlist.ts\nfunction defaultCompletionsPrompt(accessor: ServicesAccessor) {\n\tconst ctx = accessor.get(ICompletionsContextService);\n\treturn (\n\t\t<>\n\t\t\t<CompletionsContext>\n\t\t\t\t<DocumentMarker ctx={ctx} weight={0.7} />\n\t\t\t\t<Traits weight={0.6} />\n\t\t\t\t<CodeSnippets ctx={ctx} weight={0.9} />\n\t\t\t\t<SimilarFiles ctx={ctx} weight={0.8} />\n\t\t\t\t<RecentEdits ctx={ctx} weight={0.99} />\n\t\t\t</CompletionsContext>\n\t\t\t<CurrentFile weight={1} />\n\t\t</>\n\t);\n}\n\n// Exported for testing\nexport class ComponentsCompletionsPromptFactory implements CompletionsPromptFactory {\n\tprivate virtualPrompt: VirtualPrompt;\n\tprivate pipe: DataPipe;\n\tprivate renderer: CompletionsPromptRenderer;\n\tprivate promptOrdering: PromptOrdering;\n\tprivate logTarget;\n\n\tconstructor(\n\t\tvirtualPrompt: VirtualPrompt | undefined = undefined,\n\t\tordering: PromptOrdering | undefined = undefined,\n\t\t@ICompletionsContextService private readonly ctx: ICompletionsContextService,\n\t\t@IInstantiationService private readonly instantiationService: IInstantiationService,\n\t\t@ICompletionsTelemetryService private readonly completionsTelemetryService: ICompletionsTelemetryService,\n\t) {\n\t\tthis.logTarget = this.ctx.get(LogTarget);\n\t\tthis.promptOrdering = ordering ?? PromptOrdering.Default;\n\t\tthis.virtualPrompt = virtualPrompt ?? new VirtualPrompt(this.completionsPrompt());\n\t\tthis.pipe = this.virtualPrompt.createPipe();\n\t\tthis.renderer = this.getRenderer();\n\t}\n\n\tasync prompt(opts: CompletionsPromptOptions, cancellationToken?: CancellationToken): Promise<PromptResponse> {\n\t\ttry {\n\t\t\treturn await this.createPromptUnsafe(opts, cancellationToken);\n\t\t} catch (e) {\n\t\t\treturn this.errorPrompt(e as Error);\n\t\t}\n\t}\n\n\tasync createPromptUnsafe(\n\t\t{ completionId, completionState, telemetryData, promptOpts }: CompletionsPromptOptions,\n\t\tcancellationToken?: CancellationToken\n\t): Promise<PromptResponse> {\n\t\tconst { maxPromptLength, suffixPercent, suffixMatchThreshold } = this.instantiationService.invokeFunction(getPromptOptions,\n\t\t\ttelemetryData,\n\t\t\tcompletionState.textDocument.detectedLanguageId\n\t\t);\n\n\t\tconst failFastPrompt = await this.failFastPrompt(\n\t\t\tcompletionState.textDocument,\n\t\t\tcompletionState.position,\n\t\t\tsuffixPercent,\n\t\t\tcancellationToken\n\t\t);\n\t\tif (failFastPrompt) {\n\t\t\treturn failFastPrompt;\n\t\t}\n\n\t\t// TODO: Prompt ordering changes are triggered by ExP changes.\n\t\t// TODO@benibenj remove this as its always true (except in tests)\n\t\tconst promptOrdering = promptOpts?.separateContext ? PromptOrdering.SplitContext : PromptOrdering.Default;\n\t\tthis.setPromptOrdering(promptOrdering);\n\n\t\tconst start = performance.now();\n\n\t\tconst { traits, codeSnippets, turnOffSimilarFiles, resolvedContextItems } = await this.resolveContext(\n\t\t\tcompletionId,\n\t\t\tcompletionState,\n\t\t\ttelemetryData,\n\t\t\tcancellationToken,\n\t\t\tpromptOpts\n\t\t);\n\n\t\tawait this.updateComponentData(\n\t\t\tcompletionState.textDocument,\n\t\t\tcompletionState.position,\n\t\t\ttraits,\n\t\t\tcodeSnippets,\n\t\t\ttelemetryData,\n\t\t\tturnOffSimilarFiles,\n\t\t\tmaxPromptLength,\n\t\t\tcancellationToken,\n\t\t\tpromptOpts,\n\t\t\tsuffixMatchThreshold,\n\t\t\tpromptOpts?.tokenizer\n\t\t);\n\n\t\tif (cancellationToken?.isCancellationRequested) {\n\t\t\treturn _promptCancelled;\n\t\t}\n\n\t\tconst snapshot = this.virtualPrompt.snapshot(cancellationToken);\n\t\tconst snapshotStatus = snapshot.status;\n\t\tif (snapshotStatus === 'cancelled') {\n\t\t\treturn _promptCancelled;\n\t\t} else if (snapshotStatus === 'error') {\n\t\t\treturn this.errorPrompt(snapshot.error);\n\t\t}\n\n\t\tconst rendered = this.renderer.render(\n\t\t\tsnapshot.snapshot!,\n\t\t\t{\n\t\t\t\tdelimiter: '\n',\n\t\t\t\ttokenizer: promptOpts?.tokenizer,\n\t\t\t\tpromptTokenLimit: maxPromptLength,\n\t\t\t\tsuffixPercent: suffixPercent,\n\t\t\t\tlanguageId: completionState.textDocument.detectedLanguageId,\n\t\t\t},\n\t\t\tcancellationToken\n\t\t);\n\t\tif (rendered.status === 'cancelled') {\n\t\t\treturn _promptCancelled;\n\t\t} else if (rendered.status === 'error') {\n\t\t\treturn this.errorPrompt(rendered.error);\n\t\t}\n\n\t\tconst [prefix, trailingWs] = trimLastLine(rendered.prefix);\n\t\tconst renderedTrimmed = { ...rendered, prefix };\n\n\t\tlet contextProvidersTelemetry: ContextProviderTelemetry[] | undefined = undefined;\n\t\tconst languageId = completionState.textDocument.detectedLanguageId;\n\t\tif (this.instantiationService.invokeFunction(useContextProviderAPI, languageId, telemetryData)) {\n\t\t\tconst promptMatcher = componentStatisticsToPromptMatcher(rendered.metadata.componentStatistics);\n\t\t\tthis.ctx\n\t\t\t\t.get(ContextProviderStatistics)\n\t\t\t\t.getStatisticsForCompletion(completionId)\n\t\t\t\t.computeMatch(promptMatcher);\n\t\t\tcontextProvidersTelemetry = telemetrizeContextItems(this.ctx, completionId, resolvedContextItems);\n\t\t\t// To support generating context provider metrics of completion in COffE.\n\t\t\tlogger.debug(this.logTarget, `Context providers telemetry: '${JSON.stringify(contextProvidersTelemetry)}'`);\n\t\t}\n\t\tconst end = performance.now();\n\t\tthis.resetIfEmpty(rendered);\n\t\treturn this.successPrompt(renderedTrimmed, end, start, trailingWs, contextProvidersTelemetry);\n\t}\n\n\tprivate async updateComponentData(\n\t\ttextDocument: CompletionRequestDocument,\n\t\tposition: Position,\n\t\ttraits: TraitWithId[] | undefined,\n\t\tcodeSnippets: CodeSnippetWithId[] | undefined,\n\t\ttelemetryData: TelemetryWithExp,\n\t\tturnOffSimilarFiles: boolean,\n\t\tmaxPromptLength: number,\n\t\tcancellationToken?: CancellationToken,\n\t\topts: PromptOpts = {},\n\t\tsuffixMatchThreshold?: number,\n\t\ttokenizer?: TokenizerName\n\t) {\n\t\tconst completionRequestData = this.createRequestData(\n\t\t\ttextDocument,\n\t\t\tposition,\n\t\t\ttelemetryData,\n\t\t\tcancellationToken,\n\t\t\topts,\n\t\t\tmaxPromptLength,\n\t\t\ttraits,\n\t\t\tcodeSnippets,\n\t\t\tturnOffSimilarFiles,\n\t\t\tsuffixMatchThreshold,\n\t\t\ttokenizer\n\t\t);\n\t\tawait this.pipe.pump(completionRequestData);\n\t}\n\n\tprivate async resolveContext(\n\t\tcompletionId: string,\n\t\tcompletionState: CompletionState,\n\t\ttelemetryData: TelemetryWithExp,\n\t\tcancellationToken?: CancellationToken,\n\t\topts: PromptOpts = {}\n\t): Promise<{\n\t\ttraits: TraitWithId[] | undefined;\n\t\tcodeSnippets: CodeSnippetWithId[] | undefined;\n\t\tturnOffSimilarFiles: boolean;\n\t\tresolvedContextItems: ResolvedContextItem[];\n\t}> {\n\t\tlet resolvedContextItems: ResolvedContextItem[] = [];\n\t\tlet traits: TraitWithId[] | undefined;\n\t\tlet codeSnippets: CodeSnippetWithId[] | undefined;\n\t\tlet turnOffSimilarFiles = false;\n\t\tif (this.instantiationService.invokeFunction(useContextProviderAPI, completionState.textDocument.detectedLanguageId, telemetryData)) {\n\t\t\tresolvedContextItems = await this.ctx.get(ContextProviderBridge).resolution(completionId);\n\t\t\tconst { textDocument } = completionState;\n\t\t\t// Turn off neighboring files if:\n\t\t\t// - it's not explicitly enabled via EXP flag\n\t\t\t// - there are matched context providers\n\t\t\tconst matchedContextItems = resolvedContextItems.filter(matchContextItems);\n\t\t\tif (!this.instantiationService.invokeFunction(similarFilesEnabled, textDocument.detectedLanguageId, matchedContextItems, telemetryData)) {\n\t\t\t\tturnOffSimilarFiles = true;\n\t\t\t}\n\n\t\t\ttraits = await this.instantiationService.invokeFunction(getTraitsFromContextItems, completionId, matchedContextItems);\n\t\t\tvoid this.instantiationService.invokeFunction(ReportTraitsTelemetry,\n\t\t\t\t`contextProvider.traits`,\n\t\t\t\ttraits,\n\t\t\t\ttextDocument.detectedLanguageId,\n\t\t\t\ttextDocument.detectedLanguageId, // TextDocumentContext does not have clientLanguageId\n\t\t\t\ttelemetryData\n\t\t\t);\n\n\t\t\tcodeSnippets = await this.instantiationService.invokeFunction(getCodeSnippetsFromContextItems,\n\t\t\t\tcompletionId,\n\t\t\t\tmatchedContextItems,\n\t\t\t\ttextDocument.detectedLanguageId\n\t\t\t);\n\t\t}\n\t\treturn { traits, codeSnippets, turnOffSimilarFiles, resolvedContextItems };\n\t}\n\n\tprivate async failFastPrompt(\n\t\ttextDocument: TextDocumentContents,\n\t\tposition: Position,\n\t\tsuffixPercent: number,\n\t\tcancellationToken: CancellationToken | undefined\n\t) {\n\t\tif (cancellationToken?.isCancellationRequested) {\n\t\t\treturn _promptCancelled;\n\t\t}\n\t\tif (\n\t\t\t(\n\t\t\t\tawait this.ctx\n\t\t\t\t\t.get(CopilotContentExclusionManager)\n\t\t\t\t\t.evaluate(textDocument.uri, textDocument.getText(), StatusBarEvent.UPDATE)\n\t\t\t).isBlocked\n\t\t) {\n\t\t\treturn _copilotContentExclusion;\n\t\t}\n\n\t\tconst eligibleChars = suffixPercent > 0 ? textDocument.getText().length : textDocument.offsetAt(position);\n\t\tif (eligibleChars < MIN_PROMPT_CHARS) {\n\t\t\t// Too short context\n\t\t\treturn _contextTooShort;\n\t\t}\n\t}\n\n\tprivate createRequestData(\n\t\ttextDocument: CompletionRequestDocument,\n\t\tposition: Position,\n\t\ttelemetryData: TelemetryWithExp,\n\t\tcancellationToken: CancellationToken | undefined,\n\t\topts: PromptOpts,\n\t\tmaxPromptLength: number,\n\t\ttraits?: TraitWithId[],\n\t\tcodeSnippets?: CodeSnippetWithId[],\n\t\tturnOffSimilarFiles?: boolean,\n\t\tsuffixMatchThreshold?: number,\n\t\ttokenizer?: TokenizerName\n\t): CompletionRequestData {\n\t\treturn {\n\t\t\tdocument: textDocument,\n\t\t\tposition,\n\t\t\ttelemetryData,\n\t\t\tcancellationToken,\n\t\t\tdata: opts.data,\n\t\t\ttraits,\n\t\t\tcodeSnippets,\n\t\t\tturnOffSimilarFiles,\n\t\t\tsuffixMatchThreshold,\n\t\t\tmaxPromptTokens: maxPromptLength,\n\t\t\ttokenizer,\n\t\t};\n\t}\n\n\tprivate resetIfEmpty(rendered: CompletionsPromptOk) {\n\t\tif (rendered.prefix.length === 0 && rendered.suffix.length === 0) {\n\t\t\tthis.reset();\n\t\t}\n\t}\n\n\tprivate successPrompt(\n\t\trendered: CompletionsPromptOk,\n\t\tend: number,\n\t\tstart: number,\n\t\ttrailingWs: string,\n\t\tcontextProvidersTelemetry?: ContextProviderTelemetry[]\n\t): PromptResponse {\n\t\treturn {\n\t\t\ttype: 'prompt',\n\t\t\tprompt: {\n\t\t\t\tprefix: rendered.prefix,\n\t\t\t\tprefixTokens: rendered.prefixTokens,\n\t\t\t\tsuffix: rendered.suffix,\n\t\t\t\tsuffixTokens: rendered.suffixTokens,\n\t\t\t\tcontext: rendered.context,\n\t\t\t\tisFimEnabled: rendered.suffix.length > 0,\n\t\t\t},\n\t\t\tcomputeTimeMs: end - start,\n\t\t\ttrailingWs,\n\t\t\tneighborSource: new Map(),\n\t\t\tmetadata: rendered.metadata,\n\t\t\tcontextProvidersTelemetry,\n\t\t};\n\t}\n\n\tprivate errorPrompt(error: Error): PromptResponse {\n\t\ttelemetryException(this.completionsTelemetryService, error, 'PromptComponents.CompletionsPromptFactory');\n\t\tthis.reset();\n\t\treturn _promptError;\n\t}\n\n\tprivate reset() {\n\t\tthis.renderer = this.getRenderer();\n\t\tthis.virtualPrompt = new VirtualPrompt(this.completionsPrompt());\n\t\tthis.pipe = this.virtualPrompt.createPipe();\n\t}\n\n\tprivate setPromptOrdering(ordering: PromptOrdering) {\n\t\tif (this.promptOrdering !== ordering) {\n\t\t\tthis.promptOrdering = ordering;\n\t\t\tthis.reset();\n\t\t}\n\t}\n\n\tprivate completionsPrompt() {\n\t\tconst promptFunction =\n\t\t\tavailableDeclarativePrompts[this.promptOrdering]?.promptFunction ?? defaultCompletionsPrompt;\n\t\treturn this.instantiationService.invokeFunction(promptFunction);\n\t}\n\n\tprivate getRenderer() {\n\t\tconst promptInfo =\n\t\t\tavailableDeclarativePrompts[this.promptOrdering] ?? availableDeclarativePrompts[PromptOrdering.Default];\n\t\treturn new promptInfo.renderer();\n\t}\n}\n\n// Similar files is enabled if:\n// - the languageId is C/C++.\n// - it's explicitly enabled via EXP flag or config.\n// - no code snippets are provided (which includes the case when all providers error).\nfunction similarFilesEnabled(\n\taccessor: ServicesAccessor,\n\tdetectedLanguageId: string,\n\tmatchedContextItems: ResolvedContextItem<SupportedContextItemWithId>[],\n\ttelemetryData: TelemetryWithExp\n) {\n\tconst cppLanguageIds = ['cpp', 'c'];\n\tconst includeNeighboringFiles =\n\t\tisIncludeNeighborFilesActive(accessor, detectedLanguageId, telemetryData) || cppLanguageIds.includes(detectedLanguageId);\n\treturn (\n\t\tincludeNeighboringFiles || !matchedContextItems.some(ci => ci.data.some(item => item.type === 'CodeSnippet'))\n\t);\n}\n",typescriptreact,tab
3
+ 2,289,"tasks",0,0,"",Log,tab
4
+ 3,290,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",0,0,"",typescriptreact,tab
5
+ 4,343,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"6:05:39 PM [info] Activating crowd-code\n6:05:39 PM [info] Recording started\n6:05:39 PM [info] Initializing git provider using file system watchers...\n6:05:39 PM [info] Git repository found\n6:05:39 PM [info] Git provider initialized successfully\n6:05:39 PM [info] Initial git state: [object Object]\n",Log,tab
6
+ 5,12297,"extension-output-pdoom-org.crowd-code-#1-crowd-code",298,0,"",Log,selection_mouse
7
+ 6,13219,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",0,0,"",typescriptreact,tab
8
+ 7,347453,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",5114,0,"",typescriptreact,selection_command
9
+ 8,347571,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",5070,0,"",typescriptreact,selection_command
10
+ 9,1004602,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",5026,0,"",typescriptreact,selection_command
11
+ 10,1004729,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",4982,0,"",typescriptreact,selection_command
12
+ 11,1004853,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",4954,0,"",typescriptreact,selection_command
13
+ 12,1005293,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",4908,0,"",typescriptreact,selection_command
14
+ 13,2090202,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
15
+ 14,2092221,"TERMINAL",0,0,"",,terminal_focus
16
+ 15,2092222,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",0,0,"",typescriptreact,tab
17
+ 16,2093995,"TERMINAL",0,0,"cd ..",,terminal_command
18
+ 17,2094451,"TERMINAL",0,0,"ls",,terminal_command
19
+ 18,2094466,"TERMINAL",0,0,"]633;Ccleanrl crowd-code-player crowd-pilot jafar jax_cache maxtext npm-global oai-compatible-copilot sbatch-runner Stoix vscode-crowd-pilot-chat\r\n]0;franz.srambical@hai-login2:~",,terminal_output
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-4b7a193b-6fd0-48b6-a605-a1ce6ba179221764439942282-2025_11_29-19.12.25.371/source.csv ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 2,106,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"7:12:25 PM [info] Activating crowd-code\n7:12:25 PM [info] Recording started\n7:12:25 PM [info] Initializing git provider using file system watchers...\n7:12:25 PM [info] No workspace folder found\n",Log,tab
3
+ 3,2027,"extension-output-pdoom-org.crowd-code-#1-crowd-code",194,0,"7:12:27 PM [info] Retrying git provider initialization...\n7:12:27 PM [info] No workspace folder found\n",Log,content
4
+ 4,9498,"Untitled-1",0,0,"",plaintext,tab
5
+ 5,11149,"TERMINAL",0,0,"Test",,terminal_focus
6
+ 6,11153,"Untitled-1",0,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
7
+ 7,18176,"Untitled-1",46,0,"",plaintext,selection_command
8
+ 8,18275,"Untitled-1",39,0,"",plaintext,selection_command
9
+ 9,18422,"Untitled-1",32,0,"",plaintext,selection_command
10
+ 10,18611,"Untitled-1",0,0,"",plaintext,selection_command
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-549b7320-9591-428d-ba6a-d3d8fb65a0001764416891089-2025_11_29-12.48.19.667/source.csv ADDED
@@ -0,0 +1,1006 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 2,103,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"12:48:19 PM [info] Activating crowd-code\n12:48:19 PM [info] Recording started\n12:48:19 PM [info] Initializing git provider using file system watchers...\n12:48:19 PM [info] Git repository found\n12:48:19 PM [info] Git provider initialized successfully\n12:48:19 PM [info] Initial git state: [object Object]\n",Log,tab
3
+ 3,1736,"TERMINAL",0,0,"",,terminal_focus
4
+ 4,4333,"TERMINAL",0,0,"zsh",,terminal_focus
5
+ 5,20864,"TERMINAL",0,0,"cd webview-ui/sto",,terminal_command
6
+ 6,20866,"TERMINAL",0,0,"]633;Ccd: no such file or directory: webview-ui/sto\r\n% \r \r",,terminal_output
7
+ 7,22604,"TERMINAL",0,0,"cd webview-ui/sto",,terminal_command
8
+ 8,22605,"TERMINAL",0,0,"]633;Ccd: no such file or directory: webview-ui/sto\r\n% \r \r",,terminal_output
9
+ 9,28427,"TERMINAL",0,0,"cd webview-ui/.storybook",,terminal_command
10
+ 10,28428,"TERMINAL",0,0,"]633;C% \r \r",,terminal_output
11
+ 11,31719,"TERMINAL",0,0,"npm run storybook",,terminal_command
12
+ 12,31770,"TERMINAL",0,0,"]633;C",,terminal_output
13
+ 13,32548,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\nsh: storybook: command not found\r\n⠙% \r \r",,terminal_output
14
+ 14,201178,"TERMINAL",0,0,"cd ..",,terminal_command
15
+ 15,201179,"TERMINAL",0,0,"]633;C% \r \r",,terminal_output
16
+ 16,202166,"TERMINAL",0,0,"cd ..",,terminal_command
17
+ 17,202171,"TERMINAL",0,0,"]633;C% \r \r",,terminal_output
18
+ 18,208467,"TERMINAL",0,0,"npm install",,terminal_command
19
+ 19,208518,"TERMINAL",0,0,"]633;C",,terminal_output
20
+ 20,211999,"TERMINAL",0,0,"⠙",,terminal_output
21
+ 21,212258,"TERMINAL",0,0,"npm warn EBADENGINE Unsupported engine {\r\nnpm warn EBADENGINE package: '@sap/xsenv@6.0.0',\r\nnpm warn EBADENGINE required: { node: '^20.0.0 || ^22.0.0 || ^24.0.0' },\r\nnpm warn EBADENGINE current: { node: 'v23.11.0', npm: '10.9.2' }\r\nnpm warn EBADENGINE }\r\n⠹npm warn EBADENGINE Unsupported engine {\r\nnpm warn EBADENGINE package: 'opossum@9.0.0',\r\nnpm warn EBADENGINE required: { node: '^24 || ^22 || ^20' },\r\nnpm warn EBADENGINE current: { node: 'v23.11.0', npm: '10.9.2' }\r\nnpm warn EBADENGINE }\r\n⠹",,terminal_output
22
+ 22,212340,"TERMINAL",0,0,"⠸",,terminal_output
23
+ 23,212476,"TERMINAL",0,0,"⠼",,terminal_output
24
+ 24,212566,"TERMINAL",0,0,"⠴",,terminal_output
25
+ 25,212680,"TERMINAL",0,0,"⠦",,terminal_output
26
+ 26,212761,"TERMINAL",0,0,"⠧",,terminal_output
27
+ 27,212842,"TERMINAL",0,0,"⠇",,terminal_output
28
+ 28,212922,"TERMINAL",0,0,"⠏",,terminal_output
29
+ 29,213003,"TERMINAL",0,0,"⠋",,terminal_output
30
+ 30,213084,"TERMINAL",0,0,"⠙",,terminal_output
31
+ 31,213163,"TERMINAL",0,0,"⠹",,terminal_output
32
+ 32,213244,"TERMINAL",0,0,"⠸",,terminal_output
33
+ 33,213325,"TERMINAL",0,0,"⠼",,terminal_output
34
+ 34,213403,"TERMINAL",0,0,"⠴",,terminal_output
35
+ 35,213484,"TERMINAL",0,0,"⠦",,terminal_output
36
+ 36,213784,"TERMINAL",0,0,"⠧",,terminal_output
37
+ 37,213862,"TERMINAL",0,0,"⠇",,terminal_output
38
+ 38,213949,"TERMINAL",0,0,"⠏",,terminal_output
39
+ 39,214040,"TERMINAL",0,0,"⠋npm warn deprecated rimraf@3.0.2: Rimraf versions prior to v4 are no longer supported\r\n⠋npm warn deprecated rimraf@3.0.2: Rimraf versions prior to v4 are no longer supported\r\n⠋npm warn deprecated rimraf@3.0.2: Rimraf versions prior to v4 are no longer supported\r\n⠋",,terminal_output
40
+ 40,214240,"TERMINAL",0,0,"⠙npm warn deprecated glob@7.2.3: Glob versions prior to v9 are no longer supported\r\n⠙npm warn deprecated glob@7.2.3: Glob versions prior to v9 are no longer supported\r\n⠙npm warn deprecated glob@7.2.3: Glob versions prior to v9 are no longer supported\r\n⠙npm warn deprecated glob@7.2.3: Glob versions prior to v9 are no longer supported\r\n⠙",,terminal_output
41
+ 41,214321,"TERMINAL",0,0,"⠹",,terminal_output
42
+ 42,214401,"TERMINAL",0,0,"⠸",,terminal_output
43
+ 43,214483,"TERMINAL",0,0,"⠼",,terminal_output
44
+ 44,214563,"TERMINAL",0,0,"⠴",,terminal_output
45
+ 45,214643,"TERMINAL",0,0,"⠦",,terminal_output
46
+ 46,214722,"TERMINAL",0,0,"⠧",,terminal_output
47
+ 47,214803,"TERMINAL",0,0,"⠇",,terminal_output
48
+ 48,214882,"TERMINAL",0,0,"⠏",,terminal_output
49
+ 49,214963,"TERMINAL",0,0,"⠋",,terminal_output
50
+ 50,215043,"TERMINAL",0,0,"⠙",,terminal_output
51
+ 51,215122,"TERMINAL",0,0,"⠹",,terminal_output
52
+ 52,215200,"TERMINAL",0,0,"⠸",,terminal_output
53
+ 53,215288,"TERMINAL",0,0,"⠼",,terminal_output
54
+ 54,215364,"TERMINAL",0,0,"⠴",,terminal_output
55
+ 55,215443,"TERMINAL",0,0,"⠦",,terminal_output
56
+ 56,215524,"TERMINAL",0,0,"⠧",,terminal_output
57
+ 57,215604,"TERMINAL",0,0,"⠇",,terminal_output
58
+ 58,215684,"TERMINAL",0,0,"⠏",,terminal_output
59
+ 59,215765,"TERMINAL",0,0,"⠋",,terminal_output
60
+ 60,215845,"TERMINAL",0,0,"⠙",,terminal_output
61
+ 61,215925,"TERMINAL",0,0,"⠹",,terminal_output
62
+ 62,216004,"TERMINAL",0,0,"⠸",,terminal_output
63
+ 63,216085,"TERMINAL",0,0,"⠼",,terminal_output
64
+ 64,216164,"TERMINAL",0,0,"⠴",,terminal_output
65
+ 65,216254,"TERMINAL",0,0,"⠦",,terminal_output
66
+ 66,216326,"TERMINAL",0,0,"⠧",,terminal_output
67
+ 67,216408,"TERMINAL",0,0,"⠇",,terminal_output
68
+ 68,216487,"TERMINAL",0,0,"⠏",,terminal_output
69
+ 69,216567,"TERMINAL",0,0,"⠋",,terminal_output
70
+ 70,216647,"TERMINAL",0,0,"⠙",,terminal_output
71
+ 71,216728,"TERMINAL",0,0,"⠹",,terminal_output
72
+ 72,216809,"TERMINAL",0,0,"⠸",,terminal_output
73
+ 73,216888,"TERMINAL",0,0,"⠼",,terminal_output
74
+ 74,216971,"TERMINAL",0,0,"⠴",,terminal_output
75
+ 75,217050,"TERMINAL",0,0,"⠦",,terminal_output
76
+ 76,217132,"TERMINAL",0,0,"⠧",,terminal_output
77
+ 77,217211,"TERMINAL",0,0,"⠇",,terminal_output
78
+ 78,217294,"TERMINAL",0,0,"⠏",,terminal_output
79
+ 79,217371,"TERMINAL",0,0,"⠋",,terminal_output
80
+ 80,217452,"TERMINAL",0,0,"⠙",,terminal_output
81
+ 81,217532,"TERMINAL",0,0,"⠹",,terminal_output
82
+ 82,217613,"TERMINAL",0,0,"⠸",,terminal_output
83
+ 83,217694,"TERMINAL",0,0,"⠼",,terminal_output
84
+ 84,217774,"TERMINAL",0,0,"⠴",,terminal_output
85
+ 85,217856,"TERMINAL",0,0,"⠦",,terminal_output
86
+ 86,217936,"TERMINAL",0,0,"⠧",,terminal_output
87
+ 87,218018,"TERMINAL",0,0,"⠇",,terminal_output
88
+ 88,218098,"TERMINAL",0,0,"⠏",,terminal_output
89
+ 89,218176,"TERMINAL",0,0,"⠋",,terminal_output
90
+ 90,218258,"TERMINAL",0,0,"⠙",,terminal_output
91
+ 91,218339,"TERMINAL",0,0,"⠹",,terminal_output
92
+ 92,218427,"TERMINAL",0,0,"⠸",,terminal_output
93
+ 93,218501,"TERMINAL",0,0,"⠼",,terminal_output
94
+ 94,218582,"TERMINAL",0,0,"⠴",,terminal_output
95
+ 95,218662,"TERMINAL",0,0,"⠦",,terminal_output
96
+ 96,218742,"TERMINAL",0,0,"⠧",,terminal_output
97
+ 97,218821,"TERMINAL",0,0,"⠇",,terminal_output
98
+ 98,218900,"TERMINAL",0,0,"⠏",,terminal_output
99
+ 99,218982,"TERMINAL",0,0,"⠋",,terminal_output
100
+ 100,219063,"TERMINAL",0,0,"⠙",,terminal_output
101
+ 101,219142,"TERMINAL",0,0,"⠹",,terminal_output
102
+ 102,219223,"TERMINAL",0,0,"⠸",,terminal_output
103
+ 103,219304,"TERMINAL",0,0,"⠼",,terminal_output
104
+ 104,219385,"TERMINAL",0,0,"⠴",,terminal_output
105
+ 105,219464,"TERMINAL",0,0,"⠦",,terminal_output
106
+ 106,219544,"TERMINAL",0,0,"⠧",,terminal_output
107
+ 107,219625,"TERMINAL",0,0,"⠇",,terminal_output
108
+ 108,219704,"TERMINAL",0,0,"⠏",,terminal_output
109
+ 109,219785,"TERMINAL",0,0,"⠋",,terminal_output
110
+ 110,219866,"TERMINAL",0,0,"⠙",,terminal_output
111
+ 111,219946,"TERMINAL",0,0,"⠹",,terminal_output
112
+ 112,220024,"TERMINAL",0,0,"⠸",,terminal_output
113
+ 113,220107,"TERMINAL",0,0,"⠼",,terminal_output
114
+ 114,220188,"TERMINAL",0,0,"⠴",,terminal_output
115
+ 115,220268,"TERMINAL",0,0,"⠦",,terminal_output
116
+ 116,220349,"TERMINAL",0,0,"⠧",,terminal_output
117
+ 117,220431,"TERMINAL",0,0,"⠇",,terminal_output
118
+ 118,220509,"TERMINAL",0,0,"⠏",,terminal_output
119
+ 119,220591,"TERMINAL",0,0,"⠋",,terminal_output
120
+ 120,220668,"TERMINAL",0,0,"⠙",,terminal_output
121
+ 121,220751,"TERMINAL",0,0,"⠹",,terminal_output
122
+ 122,220831,"TERMINAL",0,0,"⠸",,terminal_output
123
+ 123,220911,"TERMINAL",0,0,"⠼",,terminal_output
124
+ 124,220991,"TERMINAL",0,0,"⠴",,terminal_output
125
+ 125,221072,"TERMINAL",0,0,"⠦",,terminal_output
126
+ 126,221151,"TERMINAL",0,0,"⠧",,terminal_output
127
+ 127,221232,"TERMINAL",0,0,"⠇",,terminal_output
128
+ 128,221314,"TERMINAL",0,0,"⠏",,terminal_output
129
+ 129,221396,"TERMINAL",0,0,"⠋",,terminal_output
130
+ 130,221474,"TERMINAL",0,0,"⠙",,terminal_output
131
+ 131,221556,"TERMINAL",0,0,"⠹",,terminal_output
132
+ 132,221634,"TERMINAL",0,0,"⠸",,terminal_output
133
+ 133,221715,"TERMINAL",0,0,"⠼",,terminal_output
134
+ 134,221794,"TERMINAL",0,0,"⠴",,terminal_output
135
+ 135,221881,"TERMINAL",0,0,"⠦",,terminal_output
136
+ 136,221958,"TERMINAL",0,0,"⠧",,terminal_output
137
+ 137,222038,"TERMINAL",0,0,"⠇",,terminal_output
138
+ 138,222119,"TERMINAL",0,0,"⠏",,terminal_output
139
+ 139,222202,"TERMINAL",0,0,"⠋",,terminal_output
140
+ 140,222281,"TERMINAL",0,0,"⠙",,terminal_output
141
+ 141,222361,"TERMINAL",0,0,"⠹",,terminal_output
142
+ 142,222441,"TERMINAL",0,0,"⠸",,terminal_output
143
+ 143,222522,"TERMINAL",0,0,"⠼",,terminal_output
144
+ 144,222603,"TERMINAL",0,0,"⠴",,terminal_output
145
+ 145,222683,"TERMINAL",0,0,"⠦",,terminal_output
146
+ 146,222763,"TERMINAL",0,0,"⠧",,terminal_output
147
+ 147,222847,"TERMINAL",0,0,"⠇",,terminal_output
148
+ 148,222944,"TERMINAL",0,0,"⠏",,terminal_output
149
+ 149,223014,"TERMINAL",0,0,"⠋",,terminal_output
150
+ 150,223095,"TERMINAL",0,0,"⠙",,terminal_output
151
+ 151,223175,"TERMINAL",0,0,"⠹",,terminal_output
152
+ 152,223255,"TERMINAL",0,0,"⠸",,terminal_output
153
+ 153,223335,"TERMINAL",0,0,"⠼",,terminal_output
154
+ 154,223417,"TERMINAL",0,0,"⠴",,terminal_output
155
+ 155,223497,"TERMINAL",0,0,"⠦",,terminal_output
156
+ 156,223577,"TERMINAL",0,0,"⠧",,terminal_output
157
+ 157,223664,"TERMINAL",0,0,"⠇",,terminal_output
158
+ 158,223739,"TERMINAL",0,0,"⠏",,terminal_output
159
+ 159,223820,"TERMINAL",0,0,"⠋",,terminal_output
160
+ 160,223900,"TERMINAL",0,0,"⠙",,terminal_output
161
+ 161,223982,"TERMINAL",0,0,"⠹",,terminal_output
162
+ 162,224062,"TERMINAL",0,0,"⠸",,terminal_output
163
+ 163,224142,"TERMINAL",0,0,"⠼",,terminal_output
164
+ 164,224221,"TERMINAL",0,0,"⠴",,terminal_output
165
+ 165,224304,"TERMINAL",0,0,"⠦",,terminal_output
166
+ 166,224385,"TERMINAL",0,0,"⠧",,terminal_output
167
+ 167,224465,"TERMINAL",0,0,"⠇",,terminal_output
168
+ 168,224545,"TERMINAL",0,0,"⠏",,terminal_output
169
+ 169,224624,"TERMINAL",0,0,"⠋",,terminal_output
170
+ 170,224703,"TERMINAL",0,0,"⠙",,terminal_output
171
+ 171,224796,"TERMINAL",0,0,"⠹",,terminal_output
172
+ 172,224874,"TERMINAL",0,0,"⠸",,terminal_output
173
+ 173,224954,"TERMINAL",0,0,"⠼",,terminal_output
174
+ 174,225033,"TERMINAL",0,0,"⠴",,terminal_output
175
+ 175,225113,"TERMINAL",0,0,"⠦",,terminal_output
176
+ 176,225193,"TERMINAL",0,0,"⠧",,terminal_output
177
+ 177,225275,"TERMINAL",0,0,"⠇",,terminal_output
178
+ 178,225354,"TERMINAL",0,0,"⠏",,terminal_output
179
+ 179,225434,"TERMINAL",0,0,"⠋",,terminal_output
180
+ 180,225514,"TERMINAL",0,0,"⠙",,terminal_output
181
+ 181,225594,"TERMINAL",0,0,"⠹",,terminal_output
182
+ 182,225677,"TERMINAL",0,0,"⠸",,terminal_output
183
+ 183,225757,"TERMINAL",0,0,"⠼",,terminal_output
184
+ 184,225837,"TERMINAL",0,0,"⠴",,terminal_output
185
+ 185,225918,"TERMINAL",0,0,"⠦",,terminal_output
186
+ 186,225996,"TERMINAL",0,0,"⠧",,terminal_output
187
+ 187,226078,"TERMINAL",0,0,"⠇",,terminal_output
188
+ 188,226158,"TERMINAL",0,0,"⠏",,terminal_output
189
+ 189,226237,"TERMINAL",0,0,"⠋",,terminal_output
190
+ 190,226319,"TERMINAL",0,0,"⠙",,terminal_output
191
+ 191,226396,"TERMINAL",0,0,"⠹",,terminal_output
192
+ 192,226478,"TERMINAL",0,0,"⠸",,terminal_output
193
+ 193,226559,"TERMINAL",0,0,"⠼",,terminal_output
194
+ 194,226637,"TERMINAL",0,0,"⠴",,terminal_output
195
+ 195,226721,"TERMINAL",0,0,"⠦",,terminal_output
196
+ 196,226800,"TERMINAL",0,0,"⠧",,terminal_output
197
+ 197,226880,"TERMINAL",0,0,"⠇",,terminal_output
198
+ 198,226960,"TERMINAL",0,0,"⠏",,terminal_output
199
+ 199,227039,"TERMINAL",0,0,"⠋",,terminal_output
200
+ 200,227137,"TERMINAL",0,0,"⠙",,terminal_output
201
+ 201,227200,"TERMINAL",0,0,"⠹",,terminal_output
202
+ 202,227279,"TERMINAL",0,0,"⠸",,terminal_output
203
+ 203,227362,"TERMINAL",0,0,"⠼",,terminal_output
204
+ 204,227442,"TERMINAL",0,0,"⠴",,terminal_output
205
+ 205,227532,"TERMINAL",0,0,"⠦",,terminal_output
206
+ 206,227612,"TERMINAL",0,0,"⠧",,terminal_output
207
+ 207,227696,"TERMINAL",0,0,"⠇",,terminal_output
208
+ 208,227774,"TERMINAL",0,0,"⠏",,terminal_output
209
+ 209,227859,"TERMINAL",0,0,"⠋",,terminal_output
210
+ 210,227946,"TERMINAL",0,0,"⠙",,terminal_output
211
+ 211,228025,"TERMINAL",0,0,"⠹",,terminal_output
212
+ 212,228107,"TERMINAL",0,0,"⠸",,terminal_output
213
+ 213,228186,"TERMINAL",0,0,"⠼",,terminal_output
214
+ 214,228267,"TERMINAL",0,0,"⠴",,terminal_output
215
+ 215,228347,"TERMINAL",0,0,"⠦",,terminal_output
216
+ 216,228427,"TERMINAL",0,0,"⠧",,terminal_output
217
+ 217,228508,"TERMINAL",0,0,"⠇",,terminal_output
218
+ 218,228586,"TERMINAL",0,0,"⠏",,terminal_output
219
+ 219,228667,"TERMINAL",0,0,"⠋",,terminal_output
220
+ 220,228747,"TERMINAL",0,0,"⠙",,terminal_output
221
+ 221,228827,"TERMINAL",0,0,"⠹",,terminal_output
222
+ 222,228906,"TERMINAL",0,0,"⠸",,terminal_output
223
+ 223,228987,"TERMINAL",0,0,"⠼",,terminal_output
224
+ 224,229067,"TERMINAL",0,0,"⠴",,terminal_output
225
+ 225,229147,"TERMINAL",0,0,"⠦",,terminal_output
226
+ 226,229226,"TERMINAL",0,0,"⠧",,terminal_output
227
+ 227,229306,"TERMINAL",0,0,"⠇",,terminal_output
228
+ 228,229452,"TERMINAL",0,0,"⠏",,terminal_output
229
+ 229,229562,"TERMINAL",0,0,"⠋",,terminal_output
230
+ 230,229645,"TERMINAL",0,0,"⠙",,terminal_output
231
+ 231,229725,"TERMINAL",0,0,"⠹",,terminal_output
232
+ 232,229802,"TERMINAL",0,0,"⠸",,terminal_output
233
+ 233,229888,"TERMINAL",0,0,"⠼",,terminal_output
234
+ 234,229968,"TERMINAL",0,0,"⠴",,terminal_output
235
+ 235,230048,"TERMINAL",0,0,"⠦",,terminal_output
236
+ 236,230129,"TERMINAL",0,0,"⠧",,terminal_output
237
+ 237,230211,"TERMINAL",0,0,"⠇",,terminal_output
238
+ 238,230294,"TERMINAL",0,0,"⠏",,terminal_output
239
+ 239,230373,"TERMINAL",0,0,"⠋",,terminal_output
240
+ 240,230452,"TERMINAL",0,0,"⠙",,terminal_output
241
+ 241,230532,"TERMINAL",0,0,"⠹",,terminal_output
242
+ 242,230615,"TERMINAL",0,0,"⠸",,terminal_output
243
+ 243,230693,"TERMINAL",0,0,"⠼",,terminal_output
244
+ 244,230774,"TERMINAL",0,0,"⠴",,terminal_output
245
+ 245,230854,"TERMINAL",0,0,"⠦",,terminal_output
246
+ 246,230933,"TERMINAL",0,0,"⠧",,terminal_output
247
+ 247,231014,"TERMINAL",0,0,"⠇",,terminal_output
248
+ 248,231093,"TERMINAL",0,0,"⠏",,terminal_output
249
+ 249,231174,"TERMINAL",0,0,"⠋",,terminal_output
250
+ 250,231253,"TERMINAL",0,0,"⠙",,terminal_output
251
+ 251,231333,"TERMINAL",0,0,"⠹",,terminal_output
252
+ 252,231413,"TERMINAL",0,0,"⠸",,terminal_output
253
+ 253,231492,"TERMINAL",0,0,"⠼",,terminal_output
254
+ 254,231573,"TERMINAL",0,0,"⠴",,terminal_output
255
+ 255,231653,"TERMINAL",0,0,"⠦",,terminal_output
256
+ 256,231732,"TERMINAL",0,0,"⠧",,terminal_output
257
+ 257,231813,"TERMINAL",0,0,"⠇",,terminal_output
258
+ 258,231894,"TERMINAL",0,0,"⠏",,terminal_output
259
+ 259,231973,"TERMINAL",0,0,"⠋",,terminal_output
260
+ 260,232054,"TERMINAL",0,0,"⠙",,terminal_output
261
+ 261,232134,"TERMINAL",0,0,"⠹",,terminal_output
262
+ 262,232213,"TERMINAL",0,0,"⠸",,terminal_output
263
+ 263,232293,"TERMINAL",0,0,"⠼",,terminal_output
264
+ 264,232373,"TERMINAL",0,0,"⠴",,terminal_output
265
+ 265,232453,"TERMINAL",0,0,"⠦",,terminal_output
266
+ 266,232534,"TERMINAL",0,0,"⠧",,terminal_output
267
+ 267,232613,"TERMINAL",0,0,"⠇",,terminal_output
268
+ 268,232694,"TERMINAL",0,0,"⠏",,terminal_output
269
+ 269,232773,"TERMINAL",0,0,"⠋",,terminal_output
270
+ 270,232854,"TERMINAL",0,0,"⠙",,terminal_output
271
+ 271,232939,"TERMINAL",0,0,"⠹",,terminal_output
272
+ 272,233013,"TERMINAL",0,0,"⠸",,terminal_output
273
+ 273,233094,"TERMINAL",0,0,"⠼",,terminal_output
274
+ 274,233173,"TERMINAL",0,0,"⠴",,terminal_output
275
+ 275,233253,"TERMINAL",0,0,"⠦",,terminal_output
276
+ 276,233334,"TERMINAL",0,0,"⠧",,terminal_output
277
+ 277,233415,"TERMINAL",0,0,"⠇",,terminal_output
278
+ 278,233494,"TERMINAL",0,0,"⠏",,terminal_output
279
+ 279,233574,"TERMINAL",0,0,"⠋",,terminal_output
280
+ 280,233655,"TERMINAL",0,0,"⠙",,terminal_output
281
+ 281,233735,"TERMINAL",0,0,"⠹",,terminal_output
282
+ 282,233816,"TERMINAL",0,0,"⠸",,terminal_output
283
+ 283,233895,"TERMINAL",0,0,"⠼",,terminal_output
284
+ 284,233974,"TERMINAL",0,0,"⠴",,terminal_output
285
+ 285,234055,"TERMINAL",0,0,"⠦",,terminal_output
286
+ 286,234134,"TERMINAL",0,0,"⠧",,terminal_output
287
+ 287,234219,"TERMINAL",0,0,"⠇",,terminal_output
288
+ 288,234299,"TERMINAL",0,0,"⠏",,terminal_output
289
+ 289,234378,"TERMINAL",0,0,"⠋",,terminal_output
290
+ 290,234458,"TERMINAL",0,0,"⠙",,terminal_output
291
+ 291,234539,"TERMINAL",0,0,"⠹",,terminal_output
292
+ 292,234618,"TERMINAL",0,0,"⠸",,terminal_output
293
+ 293,234716,"TERMINAL",0,0,"⠼",,terminal_output
294
+ 294,234793,"TERMINAL",0,0,"⠴",,terminal_output
295
+ 295,234872,"TERMINAL",0,0,"⠦",,terminal_output
296
+ 296,234952,"TERMINAL",0,0,"⠧",,terminal_output
297
+ 297,235033,"TERMINAL",0,0,"⠇",,terminal_output
298
+ 298,235112,"TERMINAL",0,0,"⠏",,terminal_output
299
+ 299,235194,"TERMINAL",0,0,"⠋",,terminal_output
300
+ 300,235274,"TERMINAL",0,0,"⠙",,terminal_output
301
+ 301,235352,"TERMINAL",0,0,"⠹",,terminal_output
302
+ 302,235434,"TERMINAL",0,0,"⠸",,terminal_output
303
+ 303,235513,"TERMINAL",0,0,"⠼",,terminal_output
304
+ 304,235594,"TERMINAL",0,0,"⠴",,terminal_output
305
+ 305,235673,"TERMINAL",0,0,"⠦",,terminal_output
306
+ 306,235752,"TERMINAL",0,0,"⠧",,terminal_output
307
+ 307,235833,"TERMINAL",0,0,"⠇",,terminal_output
308
+ 308,235914,"TERMINAL",0,0,"⠏",,terminal_output
309
+ 309,235994,"TERMINAL",0,0,"⠋",,terminal_output
310
+ 310,236073,"TERMINAL",0,0,"⠙",,terminal_output
311
+ 311,236154,"TERMINAL",0,0,"⠹",,terminal_output
312
+ 312,236260,"TERMINAL",0,0,"⠸",,terminal_output
313
+ 313,236339,"TERMINAL",0,0,"⠼",,terminal_output
314
+ 314,236419,"TERMINAL",0,0,"⠴",,terminal_output
315
+ 315,236500,"TERMINAL",0,0,"⠦",,terminal_output
316
+ 316,236581,"TERMINAL",0,0,"⠧",,terminal_output
317
+ 317,236668,"TERMINAL",0,0,"⠇",,terminal_output
318
+ 318,236741,"TERMINAL",0,0,"⠏",,terminal_output
319
+ 319,236824,"TERMINAL",0,0,"⠋",,terminal_output
320
+ 320,236902,"TERMINAL",0,0,"⠙",,terminal_output
321
+ 321,236984,"TERMINAL",0,0,"⠹",,terminal_output
322
+ 322,237064,"TERMINAL",0,0,"⠸",,terminal_output
323
+ 323,237145,"TERMINAL",0,0,"⠼",,terminal_output
324
+ 324,237225,"TERMINAL",0,0,"⠴",,terminal_output
325
+ 325,237305,"TERMINAL",0,0,"⠦",,terminal_output
326
+ 326,237386,"TERMINAL",0,0,"⠧",,terminal_output
327
+ 327,237467,"TERMINAL",0,0,"⠇",,terminal_output
328
+ 328,237549,"TERMINAL",0,0,"⠏",,terminal_output
329
+ 329,237628,"TERMINAL",0,0,"⠋",,terminal_output
330
+ 330,237709,"TERMINAL",0,0,"⠙",,terminal_output
331
+ 331,237788,"TERMINAL",0,0,"⠹",,terminal_output
332
+ 332,237871,"TERMINAL",0,0,"⠸",,terminal_output
333
+ 333,237953,"TERMINAL",0,0,"⠼",,terminal_output
334
+ 334,238063,"TERMINAL",0,0,"⠴",,terminal_output
335
+ 335,238116,"TERMINAL",0,0,"⠦",,terminal_output
336
+ 336,238198,"TERMINAL",0,0,"⠧",,terminal_output
337
+ 337,238280,"TERMINAL",0,0,"⠇",,terminal_output
338
+ 338,238359,"TERMINAL",0,0,"⠏",,terminal_output
339
+ 339,238440,"TERMINAL",0,0,"⠋",,terminal_output
340
+ 340,238522,"TERMINAL",0,0,"⠙",,terminal_output
341
+ 341,238604,"TERMINAL",0,0,"⠹",,terminal_output
342
+ 342,238685,"TERMINAL",0,0,"⠸",,terminal_output
343
+ 343,238765,"TERMINAL",0,0,"⠼",,terminal_output
344
+ 344,238845,"TERMINAL",0,0,"⠴",,terminal_output
345
+ 345,238928,"TERMINAL",0,0,"⠦",,terminal_output
346
+ 346,239008,"TERMINAL",0,0,"⠧",,terminal_output
347
+ 347,239088,"TERMINAL",0,0,"⠇",,terminal_output
348
+ 348,239171,"TERMINAL",0,0,"⠏",,terminal_output
349
+ 349,239252,"TERMINAL",0,0,"⠋",,terminal_output
350
+ 350,239333,"TERMINAL",0,0,"⠙",,terminal_output
351
+ 351,239413,"TERMINAL",0,0,"⠹",,terminal_output
352
+ 352,239494,"TERMINAL",0,0,"⠸",,terminal_output
353
+ 353,239573,"TERMINAL",0,0,"⠼",,terminal_output
354
+ 354,239655,"TERMINAL",0,0,"⠴",,terminal_output
355
+ 355,239736,"TERMINAL",0,0,"⠦",,terminal_output
356
+ 356,239817,"TERMINAL",0,0,"⠧",,terminal_output
357
+ 357,239898,"TERMINAL",0,0,"⠇",,terminal_output
358
+ 358,239978,"TERMINAL",0,0,"⠏",,terminal_output
359
+ 359,240059,"TERMINAL",0,0,"⠋",,terminal_output
360
+ 360,240140,"TERMINAL",0,0,"⠙",,terminal_output
361
+ 361,240222,"TERMINAL",0,0,"⠹",,terminal_output
362
+ 362,240303,"TERMINAL",0,0,"⠸",,terminal_output
363
+ 363,240384,"TERMINAL",0,0,"⠼",,terminal_output
364
+ 364,240464,"TERMINAL",0,0,"⠴",,terminal_output
365
+ 365,240544,"TERMINAL",0,0,"⠦",,terminal_output
366
+ 366,240627,"TERMINAL",0,0,"⠧",,terminal_output
367
+ 367,240707,"TERMINAL",0,0,"⠇",,terminal_output
368
+ 368,240789,"TERMINAL",0,0,"⠏",,terminal_output
369
+ 369,240867,"TERMINAL",0,0,"⠋",,terminal_output
370
+ 370,240950,"TERMINAL",0,0,"⠙",,terminal_output
371
+ 371,241029,"TERMINAL",0,0,"⠹",,terminal_output
372
+ 372,241110,"TERMINAL",0,0,"⠸",,terminal_output
373
+ 373,241192,"TERMINAL",0,0,"⠼",,terminal_output
374
+ 374,241273,"TERMINAL",0,0,"⠴",,terminal_output
375
+ 375,241356,"TERMINAL",0,0,"⠦",,terminal_output
376
+ 376,241434,"TERMINAL",0,0,"⠧",,terminal_output
377
+ 377,241515,"TERMINAL",0,0,"⠇",,terminal_output
378
+ 378,241595,"TERMINAL",0,0,"⠏",,terminal_output
379
+ 379,241676,"TERMINAL",0,0,"⠋",,terminal_output
380
+ 380,241759,"TERMINAL",0,0,"⠙",,terminal_output
381
+ 381,241841,"TERMINAL",0,0,"⠹",,terminal_output
382
+ 382,241921,"TERMINAL",0,0,"⠸",,terminal_output
383
+ 383,242003,"TERMINAL",0,0,"⠼",,terminal_output
384
+ 384,242082,"TERMINAL",0,0,"⠴",,terminal_output
385
+ 385,242163,"TERMINAL",0,0,"⠦",,terminal_output
386
+ 386,242244,"TERMINAL",0,0,"⠧",,terminal_output
387
+ 387,242325,"TERMINAL",0,0,"⠇",,terminal_output
388
+ 388,242406,"TERMINAL",0,0,"⠏",,terminal_output
389
+ 389,242487,"TERMINAL",0,0,"⠋",,terminal_output
390
+ 390,242568,"TERMINAL",0,0,"⠙",,terminal_output
391
+ 391,242648,"TERMINAL",0,0,"⠹",,terminal_output
392
+ 392,242732,"TERMINAL",0,0,"⠸",,terminal_output
393
+ 393,242813,"TERMINAL",0,0,"⠼",,terminal_output
394
+ 394,242892,"TERMINAL",0,0,"⠴",,terminal_output
395
+ 395,242971,"TERMINAL",0,0,"⠦",,terminal_output
396
+ 396,243051,"TERMINAL",0,0,"⠧",,terminal_output
397
+ 397,243135,"TERMINAL",0,0,"⠇",,terminal_output
398
+ 398,243215,"TERMINAL",0,0,"⠏",,terminal_output
399
+ 399,243295,"TERMINAL",0,0,"⠋",,terminal_output
400
+ 400,243377,"TERMINAL",0,0,"⠙",,terminal_output
401
+ 401,243457,"TERMINAL",0,0,"⠹",,terminal_output
402
+ 402,243540,"TERMINAL",0,0,"⠸",,terminal_output
403
+ 403,243621,"TERMINAL",0,0,"⠼",,terminal_output
404
+ 404,243700,"TERMINAL",0,0,"⠴",,terminal_output
405
+ 405,243781,"TERMINAL",0,0,"⠦",,terminal_output
406
+ 406,243864,"TERMINAL",0,0,"⠧",,terminal_output
407
+ 407,243946,"TERMINAL",0,0,"⠇",,terminal_output
408
+ 408,244025,"TERMINAL",0,0,"⠏",,terminal_output
409
+ 409,244106,"TERMINAL",0,0,"⠋",,terminal_output
410
+ 410,244186,"TERMINAL",0,0,"⠙",,terminal_output
411
+ 411,244267,"TERMINAL",0,0,"⠹",,terminal_output
412
+ 412,244348,"TERMINAL",0,0,"⠸",,terminal_output
413
+ 413,244436,"TERMINAL",0,0,"⠼",,terminal_output
414
+ 414,244515,"TERMINAL",0,0,"⠴",,terminal_output
415
+ 415,244597,"TERMINAL",0,0,"⠦",,terminal_output
416
+ 416,244678,"TERMINAL",0,0,"⠧",,terminal_output
417
+ 417,244768,"TERMINAL",0,0,"⠇",,terminal_output
418
+ 418,244846,"TERMINAL",0,0,"⠏",,terminal_output
419
+ 419,244927,"TERMINAL",0,0,"⠋",,terminal_output
420
+ 420,245007,"TERMINAL",0,0,"⠙",,terminal_output
421
+ 421,245089,"TERMINAL",0,0,"⠹",,terminal_output
422
+ 422,245170,"TERMINAL",0,0,"⠸",,terminal_output
423
+ 423,245251,"TERMINAL",0,0,"⠼",,terminal_output
424
+ 424,245332,"TERMINAL",0,0,"⠴",,terminal_output
425
+ 425,245411,"TERMINAL",0,0,"⠦",,terminal_output
426
+ 426,245495,"TERMINAL",0,0,"⠧",,terminal_output
427
+ 427,245573,"TERMINAL",0,0,"⠇",,terminal_output
428
+ 428,245656,"TERMINAL",0,0,"⠏",,terminal_output
429
+ 429,245735,"TERMINAL",0,0,"⠋",,terminal_output
430
+ 430,245818,"TERMINAL",0,0,"⠙",,terminal_output
431
+ 431,245898,"TERMINAL",0,0,"⠹",,terminal_output
432
+ 432,245979,"TERMINAL",0,0,"⠸",,terminal_output
433
+ 433,246061,"TERMINAL",0,0,"⠼",,terminal_output
434
+ 434,246142,"TERMINAL",0,0,"⠴",,terminal_output
435
+ 435,246222,"TERMINAL",0,0,"⠦",,terminal_output
436
+ 436,246312,"TERMINAL",0,0,"⠧",,terminal_output
437
+ 437,246384,"TERMINAL",0,0,"⠇",,terminal_output
438
+ 438,246467,"TERMINAL",0,0,"⠏",,terminal_output
439
+ 439,246549,"TERMINAL",0,0,"⠋",,terminal_output
440
+ 440,246635,"TERMINAL",0,0,"⠙",,terminal_output
441
+ 441,246716,"TERMINAL",0,0,"⠹",,terminal_output
442
+ 442,246798,"TERMINAL",0,0,"⠸",,terminal_output
443
+ 443,246879,"TERMINAL",0,0,"⠼",,terminal_output
444
+ 444,246961,"TERMINAL",0,0,"⠴",,terminal_output
445
+ 445,247041,"TERMINAL",0,0,"⠦",,terminal_output
446
+ 446,247122,"TERMINAL",0,0,"⠧",,terminal_output
447
+ 447,247201,"TERMINAL",0,0,"⠇",,terminal_output
448
+ 448,247282,"TERMINAL",0,0,"⠏",,terminal_output
449
+ 449,247362,"TERMINAL",0,0,"⠋",,terminal_output
450
+ 450,247444,"TERMINAL",0,0,"⠙",,terminal_output
451
+ 451,247525,"TERMINAL",0,0,"⠹",,terminal_output
452
+ 452,247607,"TERMINAL",0,0,"⠸",,terminal_output
453
+ 453,247686,"TERMINAL",0,0,"⠼",,terminal_output
454
+ 454,247768,"TERMINAL",0,0,"⠴",,terminal_output
455
+ 455,247846,"TERMINAL",0,0,"⠦",,terminal_output
456
+ 456,247929,"TERMINAL",0,0,"⠧",,terminal_output
457
+ 457,248011,"TERMINAL",0,0,"⠇",,terminal_output
458
+ 458,248096,"TERMINAL",0,0,"⠏",,terminal_output
459
+ 459,248177,"TERMINAL",0,0,"⠋",,terminal_output
460
+ 460,248257,"TERMINAL",0,0,"⠙",,terminal_output
461
+ 461,248340,"TERMINAL",0,0,"⠹",,terminal_output
462
+ 462,248420,"TERMINAL",0,0,"⠸",,terminal_output
463
+ 463,248499,"TERMINAL",0,0,"⠼",,terminal_output
464
+ 464,248580,"TERMINAL",0,0,"⠴",,terminal_output
465
+ 465,248663,"TERMINAL",0,0,"⠦",,terminal_output
466
+ 466,248743,"TERMINAL",0,0,"⠧",,terminal_output
467
+ 467,248824,"TERMINAL",0,0,"⠇",,terminal_output
468
+ 468,248906,"TERMINAL",0,0,"⠏",,terminal_output
469
+ 469,248986,"TERMINAL",0,0,"⠋",,terminal_output
470
+ 470,249068,"TERMINAL",0,0,"⠙",,terminal_output
471
+ 471,249149,"TERMINAL",0,0,"⠹",,terminal_output
472
+ 472,249244,"TERMINAL",0,0,"⠸",,terminal_output
473
+ 473,249326,"TERMINAL",0,0,"⠼",,terminal_output
474
+ 474,249407,"TERMINAL",0,0,"⠴",,terminal_output
475
+ 475,249488,"TERMINAL",0,0,"⠦",,terminal_output
476
+ 476,249568,"TERMINAL",0,0,"⠧",,terminal_output
477
+ 477,249655,"TERMINAL",0,0,"⠇",,terminal_output
478
+ 478,249767,"TERMINAL",0,0,"⠏",,terminal_output
479
+ 479,249840,"TERMINAL",0,0,"⠋",,terminal_output
480
+ 480,249912,"TERMINAL",0,0,"⠙",,terminal_output
481
+ 481,249993,"TERMINAL",0,0,"⠹",,terminal_output
482
+ 482,250073,"TERMINAL",0,0,"⠸",,terminal_output
483
+ 483,250155,"TERMINAL",0,0,"⠼",,terminal_output
484
+ 484,250235,"TERMINAL",0,0,"⠴",,terminal_output
485
+ 485,250316,"TERMINAL",0,0,"⠦",,terminal_output
486
+ 486,250398,"TERMINAL",0,0,"⠧",,terminal_output
487
+ 487,250479,"TERMINAL",0,0,"⠇",,terminal_output
488
+ 488,250558,"TERMINAL",0,0,"⠏",,terminal_output
489
+ 489,250642,"TERMINAL",0,0,"⠋",,terminal_output
490
+ 490,250720,"TERMINAL",0,0,"⠙",,terminal_output
491
+ 491,250800,"TERMINAL",0,0,"⠹",,terminal_output
492
+ 492,250881,"TERMINAL",0,0,"⠸",,terminal_output
493
+ 493,250963,"TERMINAL",0,0,"⠼",,terminal_output
494
+ 494,251048,"TERMINAL",0,0,"⠴",,terminal_output
495
+ 495,251123,"TERMINAL",0,0,"⠦",,terminal_output
496
+ 496,251205,"TERMINAL",0,0,"⠧",,terminal_output
497
+ 497,251286,"TERMINAL",0,0,"⠇",,terminal_output
498
+ 498,251367,"TERMINAL",0,0,"⠏",,terminal_output
499
+ 499,251447,"TERMINAL",0,0,"⠋",,terminal_output
500
+ 500,251528,"TERMINAL",0,0,"⠙",,terminal_output
501
+ 501,251609,"TERMINAL",0,0,"⠹",,terminal_output
502
+ 502,251689,"TERMINAL",0,0,"⠸",,terminal_output
503
+ 503,251771,"TERMINAL",0,0,"⠼",,terminal_output
504
+ 504,251850,"TERMINAL",0,0,"⠴",,terminal_output
505
+ 505,251932,"TERMINAL",0,0,"⠦",,terminal_output
506
+ 506,252013,"TERMINAL",0,0,"⠧",,terminal_output
507
+ 507,252091,"TERMINAL",0,0,"⠇",,terminal_output
508
+ 508,252173,"TERMINAL",0,0,"⠏",,terminal_output
509
+ 509,252255,"TERMINAL",0,0,"⠋",,terminal_output
510
+ 510,252335,"TERMINAL",0,0,"⠙",,terminal_output
511
+ 511,252415,"TERMINAL",0,0,"⠹",,terminal_output
512
+ 512,252497,"TERMINAL",0,0,"⠸",,terminal_output
513
+ 513,252579,"TERMINAL",0,0,"⠼",,terminal_output
514
+ 514,252659,"TERMINAL",0,0,"⠴",,terminal_output
515
+ 515,252740,"TERMINAL",0,0,"⠦",,terminal_output
516
+ 516,252819,"TERMINAL",0,0,"⠧",,terminal_output
517
+ 517,252899,"TERMINAL",0,0,"⠇",,terminal_output
518
+ 518,252981,"TERMINAL",0,0,"⠏",,terminal_output
519
+ 519,253062,"TERMINAL",0,0,"⠋",,terminal_output
520
+ 520,253144,"TERMINAL",0,0,"⠙",,terminal_output
521
+ 521,253224,"TERMINAL",0,0,"⠹",,terminal_output
522
+ 522,253305,"TERMINAL",0,0,"⠸",,terminal_output
523
+ 523,253386,"TERMINAL",0,0,"⠼",,terminal_output
524
+ 524,253469,"TERMINAL",0,0,"⠴",,terminal_output
525
+ 525,253549,"TERMINAL",0,0,"⠦",,terminal_output
526
+ 526,253628,"TERMINAL",0,0,"⠧",,terminal_output
527
+ 527,253710,"TERMINAL",0,0,"⠇",,terminal_output
528
+ 528,253791,"TERMINAL",0,0,"⠏",,terminal_output
529
+ 529,253870,"TERMINAL",0,0,"⠋",,terminal_output
530
+ 530,253953,"TERMINAL",0,0,"⠙",,terminal_output
531
+ 531,254033,"TERMINAL",0,0,"⠹",,terminal_output
532
+ 532,254113,"TERMINAL",0,0,"⠸",,terminal_output
533
+ 533,254194,"TERMINAL",0,0,"⠼",,terminal_output
534
+ 534,254276,"TERMINAL",0,0,"⠴",,terminal_output
535
+ 535,254357,"TERMINAL",0,0,"⠦",,terminal_output
536
+ 536,254435,"TERMINAL",0,0,"⠧",,terminal_output
537
+ 537,254516,"TERMINAL",0,0,"⠇",,terminal_output
538
+ 538,254598,"TERMINAL",0,0,"⠏",,terminal_output
539
+ 539,254678,"TERMINAL",0,0,"⠋",,terminal_output
540
+ 540,254759,"TERMINAL",0,0,"⠙",,terminal_output
541
+ 541,254840,"TERMINAL",0,0,"⠹",,terminal_output
542
+ 542,254920,"TERMINAL",0,0,"⠸",,terminal_output
543
+ 543,255002,"TERMINAL",0,0,"⠼",,terminal_output
544
+ 544,255083,"TERMINAL",0,0,"⠴",,terminal_output
545
+ 545,255162,"TERMINAL",0,0,"⠦",,terminal_output
546
+ 546,255248,"TERMINAL",0,0,"⠧",,terminal_output
547
+ 547,255326,"TERMINAL",0,0,"⠇",,terminal_output
548
+ 548,255408,"TERMINAL",0,0,"⠏",,terminal_output
549
+ 549,255487,"TERMINAL",0,0,"⠋",,terminal_output
550
+ 550,255572,"TERMINAL",0,0,"⠙",,terminal_output
551
+ 551,255648,"TERMINAL",0,0,"⠹",,terminal_output
552
+ 552,255729,"TERMINAL",0,0,"⠸",,terminal_output
553
+ 553,255811,"TERMINAL",0,0,"⠼",,terminal_output
554
+ 554,255894,"TERMINAL",0,0,"⠴",,terminal_output
555
+ 555,255976,"TERMINAL",0,0,"⠦",,terminal_output
556
+ 556,256056,"TERMINAL",0,0,"⠧",,terminal_output
557
+ 557,256141,"TERMINAL",0,0,"⠇",,terminal_output
558
+ 558,256218,"TERMINAL",0,0,"⠏",,terminal_output
559
+ 559,256297,"TERMINAL",0,0,"⠋",,terminal_output
560
+ 560,256380,"TERMINAL",0,0,"⠙",,terminal_output
561
+ 561,256475,"TERMINAL",0,0,"⠹",,terminal_output
562
+ 562,256559,"TERMINAL",0,0,"⠸",,terminal_output
563
+ 563,256632,"TERMINAL",0,0,"⠼",,terminal_output
564
+ 564,256712,"TERMINAL",0,0,"⠴",,terminal_output
565
+ 565,256795,"TERMINAL",0,0,"⠦",,terminal_output
566
+ 566,256874,"TERMINAL",0,0,"⠧",,terminal_output
567
+ 567,256956,"TERMINAL",0,0,"⠇",,terminal_output
568
+ 568,257036,"TERMINAL",0,0,"⠏",,terminal_output
569
+ 569,257117,"TERMINAL",0,0,"⠋",,terminal_output
570
+ 570,257205,"TERMINAL",0,0,"⠙",,terminal_output
571
+ 571,257281,"TERMINAL",0,0,"⠹",,terminal_output
572
+ 572,257360,"TERMINAL",0,0,"⠸",,terminal_output
573
+ 573,257441,"TERMINAL",0,0,"⠼",,terminal_output
574
+ 574,257521,"TERMINAL",0,0,"⠴",,terminal_output
575
+ 575,257601,"TERMINAL",0,0,"⠦",,terminal_output
576
+ 576,257682,"TERMINAL",0,0,"⠧",,terminal_output
577
+ 577,257764,"TERMINAL",0,0,"⠇",,terminal_output
578
+ 578,257845,"TERMINAL",0,0,"⠏",,terminal_output
579
+ 579,257927,"TERMINAL",0,0,"⠋",,terminal_output
580
+ 580,258008,"TERMINAL",0,0,"⠙",,terminal_output
581
+ 581,258091,"TERMINAL",0,0,"⠹",,terminal_output
582
+ 582,258171,"TERMINAL",0,0,"⠸",,terminal_output
583
+ 583,258249,"TERMINAL",0,0,"⠼",,terminal_output
584
+ 584,258334,"TERMINAL",0,0,"⠴",,terminal_output
585
+ 585,258431,"TERMINAL",0,0,"⠦",,terminal_output
586
+ 586,258506,"TERMINAL",0,0,"⠧",,terminal_output
587
+ 587,258584,"TERMINAL",0,0,"⠇",,terminal_output
588
+ 588,258666,"TERMINAL",0,0,"⠏",,terminal_output
589
+ 589,258758,"TERMINAL",0,0,"⠋",,terminal_output
590
+ 590,258826,"TERMINAL",0,0,"⠙",,terminal_output
591
+ 591,258906,"TERMINAL",0,0,"⠹",,terminal_output
592
+ 592,258991,"TERMINAL",0,0,"⠸",,terminal_output
593
+ 593,259069,"TERMINAL",0,0,"⠼",,terminal_output
594
+ 594,259150,"TERMINAL",0,0,"⠴",,terminal_output
595
+ 595,259231,"TERMINAL",0,0,"⠦",,terminal_output
596
+ 596,259296,"TERMINAL",0,0,"",,terminal_focus
597
+ 597,259404,"TERMINAL",0,0,"⠧⠇",,terminal_output
598
+ 598,259475,"TERMINAL",0,0,"⠏",,terminal_output
599
+ 599,259553,"TERMINAL",0,0,"⠋",,terminal_output
600
+ 600,259636,"TERMINAL",0,0,"⠙",,terminal_output
601
+ 601,259715,"TERMINAL",0,0,"⠹",,terminal_output
602
+ 602,259796,"TERMINAL",0,0,"⠸",,terminal_output
603
+ 603,259878,"TERMINAL",0,0,"⠼",,terminal_output
604
+ 604,259959,"TERMINAL",0,0,"⠴",,terminal_output
605
+ 605,260043,"TERMINAL",0,0,"⠦",,terminal_output
606
+ 606,260122,"TERMINAL",0,0,"⠧",,terminal_output
607
+ 607,260203,"TERMINAL",0,0,"⠇",,terminal_output
608
+ 608,260283,"TERMINAL",0,0,"⠏",,terminal_output
609
+ 609,260364,"TERMINAL",0,0,"⠋",,terminal_output
610
+ 610,260445,"TERMINAL",0,0,"⠙",,terminal_output
611
+ 611,260525,"TERMINAL",0,0,"⠹",,terminal_output
612
+ 612,260608,"TERMINAL",0,0,"⠸",,terminal_output
613
+ 613,260688,"TERMINAL",0,0,"⠼",,terminal_output
614
+ 614,260772,"TERMINAL",0,0,"⠴",,terminal_output
615
+ 615,260852,"TERMINAL",0,0,"⠦",,terminal_output
616
+ 616,260943,"TERMINAL",0,0,"⠧",,terminal_output
617
+ 617,261012,"TERMINAL",0,0,"⠇",,terminal_output
618
+ 618,261091,"TERMINAL",0,0,"⠏",,terminal_output
619
+ 619,261171,"TERMINAL",0,0,"⠋",,terminal_output
620
+ 620,261253,"TERMINAL",0,0,"⠙",,terminal_output
621
+ 621,261331,"TERMINAL",0,0,"⠹",,terminal_output
622
+ 622,261413,"TERMINAL",0,0,"⠸",,terminal_output
623
+ 623,261492,"TERMINAL",0,0,"⠼",,terminal_output
624
+ 624,261572,"TERMINAL",0,0,"⠴",,terminal_output
625
+ 625,261652,"TERMINAL",0,0,"⠦",,terminal_output
626
+ 626,261733,"TERMINAL",0,0,"⠧",,terminal_output
627
+ 627,261813,"TERMINAL",0,0,"⠇",,terminal_output
628
+ 628,261894,"TERMINAL",0,0,"⠏",,terminal_output
629
+ 629,261973,"TERMINAL",0,0,"⠋",,terminal_output
630
+ 630,262055,"TERMINAL",0,0,"⠙",,terminal_output
631
+ 631,262085,"TERMINAL",0,0,"cd webview-ui",,terminal_command
632
+ 632,262088,"TERMINAL",0,0,"]633;C% \r \r",,terminal_output
633
+ 633,262136,"TERMINAL",0,0,"⠹",,terminal_output
634
+ 634,262216,"TERMINAL",0,0,"⠸",,terminal_output
635
+ 635,262297,"TERMINAL",0,0,"⠼",,terminal_output
636
+ 636,262377,"TERMINAL",0,0,"⠴",,terminal_output
637
+ 637,262460,"TERMINAL",0,0,"⠦",,terminal_output
638
+ 638,262540,"TERMINAL",0,0,"⠧",,terminal_output
639
+ 639,262625,"TERMINAL",0,0,"⠇",,terminal_output
640
+ 640,262701,"TERMINAL",0,0,"⠏",,terminal_output
641
+ 641,262783,"TERMINAL",0,0,"⠋",,terminal_output
642
+ 642,262862,"TERMINAL",0,0,"⠙",,terminal_output
643
+ 643,262946,"TERMINAL",0,0,"⠹",,terminal_output
644
+ 644,263027,"TERMINAL",0,0,"⠸",,terminal_output
645
+ 645,263108,"TERMINAL",0,0,"⠼",,terminal_output
646
+ 646,263188,"TERMINAL",0,0,"⠴",,terminal_output
647
+ 647,263270,"TERMINAL",0,0,"⠦",,terminal_output
648
+ 648,263350,"TERMINAL",0,0,"⠧",,terminal_output
649
+ 649,263431,"TERMINAL",0,0,"⠇",,terminal_output
650
+ 650,263513,"TERMINAL",0,0,"⠏",,terminal_output
651
+ 651,263593,"TERMINAL",0,0,"⠋",,terminal_output
652
+ 652,263674,"TERMINAL",0,0,"⠙",,terminal_output
653
+ 653,263755,"TERMINAL",0,0,"⠹",,terminal_output
654
+ 654,263836,"TERMINAL",0,0,"⠸",,terminal_output
655
+ 655,263917,"TERMINAL",0,0,"⠼",,terminal_output
656
+ 656,263999,"TERMINAL",0,0,"⠴",,terminal_output
657
+ 657,264076,"TERMINAL",0,0,"⠦",,terminal_output
658
+ 658,264158,"TERMINAL",0,0,"⠧",,terminal_output
659
+ 659,264238,"TERMINAL",0,0,"⠇",,terminal_output
660
+ 660,264320,"TERMINAL",0,0,"⠏",,terminal_output
661
+ 661,264400,"TERMINAL",0,0,"⠋",,terminal_output
662
+ 662,264481,"TERMINAL",0,0,"⠙",,terminal_output
663
+ 663,264563,"TERMINAL",0,0,"⠹",,terminal_output
664
+ 664,264644,"TERMINAL",0,0,"⠸",,terminal_output
665
+ 665,264725,"TERMINAL",0,0,"⠼",,terminal_output
666
+ 666,264806,"TERMINAL",0,0,"⠴",,terminal_output
667
+ 667,264886,"TERMINAL",0,0,"⠦",,terminal_output
668
+ 668,264968,"TERMINAL",0,0,"⠧",,terminal_output
669
+ 669,265048,"TERMINAL",0,0,"⠇",,terminal_output
670
+ 670,265132,"TERMINAL",0,0,"⠏",,terminal_output
671
+ 671,265212,"TERMINAL",0,0,"⠋",,terminal_output
672
+ 672,265293,"TERMINAL",0,0,"⠙",,terminal_output
673
+ 673,265340,"TERMINAL",0,0,"npm run storybook",,terminal_command
674
+ 674,265373,"TERMINAL",0,0,"⠹",,terminal_output
675
+ 675,265391,"TERMINAL",0,0,"]633;C",,terminal_output
676
+ 676,265456,"TERMINAL",0,0,"⠸",,terminal_output
677
+ 677,265535,"TERMINAL",0,0,"⠼",,terminal_output
678
+ 678,265582,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\nsh: storybook: command not found\r\n⠙% \r \r",,terminal_output
679
+ 679,265617,"TERMINAL",0,0,"⠴",,terminal_output
680
+ 680,265697,"TERMINAL",0,0,"⠦",,terminal_output
681
+ 681,265780,"TERMINAL",0,0,"⠧",,terminal_output
682
+ 682,265860,"TERMINAL",0,0,"⠇",,terminal_output
683
+ 683,265943,"TERMINAL",0,0,"⠏",,terminal_output
684
+ 684,266022,"TERMINAL",0,0,"⠋",,terminal_output
685
+ 685,266104,"TERMINAL",0,0,"⠙",,terminal_output
686
+ 686,266184,"TERMINAL",0,0,"⠹",,terminal_output
687
+ 687,266264,"TERMINAL",0,0,"⠸",,terminal_output
688
+ 688,266344,"TERMINAL",0,0,"⠼",,terminal_output
689
+ 689,266427,"TERMINAL",0,0,"⠴",,terminal_output
690
+ 690,266508,"TERMINAL",0,0,"⠦",,terminal_output
691
+ 691,266589,"TERMINAL",0,0,"⠧",,terminal_output
692
+ 692,266670,"TERMINAL",0,0,"⠇",,terminal_output
693
+ 693,266751,"TERMINAL",0,0,"⠏",,terminal_output
694
+ 694,266833,"TERMINAL",0,0,"⠋",,terminal_output
695
+ 695,266915,"TERMINAL",0,0,"⠙",,terminal_output
696
+ 696,266995,"TERMINAL",0,0,"⠹",,terminal_output
697
+ 697,267076,"TERMINAL",0,0,"⠸",,terminal_output
698
+ 698,267154,"TERMINAL",0,0,"⠼",,terminal_output
699
+ 699,267238,"TERMINAL",0,0,"⠴",,terminal_output
700
+ 700,267319,"TERMINAL",0,0,"⠦",,terminal_output
701
+ 701,267400,"TERMINAL",0,0,"⠧",,terminal_output
702
+ 702,267480,"TERMINAL",0,0,"⠇",,terminal_output
703
+ 703,267559,"TERMINAL",0,0,"⠏",,terminal_output
704
+ 704,267774,"TERMINAL",0,0,"⠋",,terminal_output
705
+ 705,267845,"TERMINAL",0,0,"\r\n> claude-dev@3.38.3 prepare\r\n> husky\r\n\r\n⠋",,terminal_output
706
+ 706,268035,"TERMINAL",0,0,"⠙\r\nadded 1505 packages, and audited 1506 packages in 59s\r\n⠙\r\n⠙277 packages are looking for funding\r\n⠙ run `npm fund` for details\r\n⠙\r\nfound 0 vulnerabilities\r\n⠙% \r \r",,terminal_output
707
+ 707,270587,"TERMINAL",0,0,"zsh",,terminal_focus
708
+ 708,275890,"TERMINAL",0,0,"cd webview-ui",,terminal_command
709
+ 709,275896,"TERMINAL",0,0,"]633;C% \r \r",,terminal_output
710
+ 710,278185,"TERMINAL",0,0,"npm run storybook",,terminal_command
711
+ 711,278237,"TERMINAL",0,0,"]633;C",,terminal_output
712
+ 712,278317,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\nsh: storybook: command not found\r\n⠙% \r \r",,terminal_output
713
+ 713,282871,"TERMINAL",0,0,"npm install",,terminal_command
714
+ 714,282923,"TERMINAL",0,0,"]633;C",,terminal_output
715
+ 715,283584,"TERMINAL",0,0,"⠙",,terminal_output
716
+ 716,283850,"TERMINAL",0,0,"⠹",,terminal_output
717
+ 717,283971,"TERMINAL",0,0,"⠸",,terminal_output
718
+ 718,284051,"TERMINAL",0,0,"⠼",,terminal_output
719
+ 719,284140,"TERMINAL",0,0,"⠴",,terminal_output
720
+ 720,284240,"TERMINAL",0,0,"⠦",,terminal_output
721
+ 721,284501,"TERMINAL",0,0,"⠧",,terminal_output
722
+ 722,284569,"TERMINAL",0,0,"⠇",,terminal_output
723
+ 723,284653,"TERMINAL",0,0,"⠏",,terminal_output
724
+ 724,284737,"TERMINAL",0,0,"⠋",,terminal_output
725
+ 725,284817,"TERMINAL",0,0,"⠙",,terminal_output
726
+ 726,284900,"TERMINAL",0,0,"⠹",,terminal_output
727
+ 727,284984,"TERMINAL",0,0,"⠸",,terminal_output
728
+ 728,285063,"TERMINAL",0,0,"⠼",,terminal_output
729
+ 729,285144,"TERMINAL",0,0,"⠴",,terminal_output
730
+ 730,285227,"TERMINAL",0,0,"⠦",,terminal_output
731
+ 731,285306,"TERMINAL",0,0,"⠧",,terminal_output
732
+ 732,285385,"TERMINAL",0,0,"⠇",,terminal_output
733
+ 733,285465,"TERMINAL",0,0,"⠏",,terminal_output
734
+ 734,285545,"TERMINAL",0,0,"⠋",,terminal_output
735
+ 735,285626,"TERMINAL",0,0,"⠙",,terminal_output
736
+ 736,285705,"TERMINAL",0,0,"⠹",,terminal_output
737
+ 737,285785,"TERMINAL",0,0,"⠸",,terminal_output
738
+ 738,285866,"TERMINAL",0,0,"⠼",,terminal_output
739
+ 739,285946,"TERMINAL",0,0,"⠴",,terminal_output
740
+ 740,286027,"TERMINAL",0,0,"⠦",,terminal_output
741
+ 741,286107,"TERMINAL",0,0,"⠧",,terminal_output
742
+ 742,286187,"TERMINAL",0,0,"⠇",,terminal_output
743
+ 743,286300,"TERMINAL",0,0,"⠏",,terminal_output
744
+ 744,286374,"TERMINAL",0,0,"⠋",,terminal_output
745
+ 745,286452,"TERMINAL",0,0,"⠙",,terminal_output
746
+ 746,286534,"TERMINAL",0,0,"⠹",,terminal_output
747
+ 747,286613,"TERMINAL",0,0,"⠸",,terminal_output
748
+ 748,286694,"TERMINAL",0,0,"⠼",,terminal_output
749
+ 749,286774,"TERMINAL",0,0,"⠴",,terminal_output
750
+ 750,286854,"TERMINAL",0,0,"⠦",,terminal_output
751
+ 751,286934,"TERMINAL",0,0,"⠧",,terminal_output
752
+ 752,287014,"TERMINAL",0,0,"⠇",,terminal_output
753
+ 753,287095,"TERMINAL",0,0,"⠏",,terminal_output
754
+ 754,287175,"TERMINAL",0,0,"⠋",,terminal_output
755
+ 755,287256,"TERMINAL",0,0,"⠙",,terminal_output
756
+ 756,287336,"TERMINAL",0,0,"⠹",,terminal_output
757
+ 757,287415,"TERMINAL",0,0,"⠸",,terminal_output
758
+ 758,287496,"TERMINAL",0,0,"⠼",,terminal_output
759
+ 759,287576,"TERMINAL",0,0,"⠴",,terminal_output
760
+ 760,287656,"TERMINAL",0,0,"⠦",,terminal_output
761
+ 761,287734,"TERMINAL",0,0,"⠧",,terminal_output
762
+ 762,287817,"TERMINAL",0,0,"⠇",,terminal_output
763
+ 763,287928,"TERMINAL",0,0,"⠏npm warn deprecated @vscode/webview-ui-toolkit@1.4.0: This package has been deprecated, https://github.com/microsoft/vscode-webview-ui-toolkit/issues/561\r\n⠏",,terminal_output
764
+ 764,288009,"TERMINAL",0,0,"⠋",,terminal_output
765
+ 765,288091,"TERMINAL",0,0,"⠙",,terminal_output
766
+ 766,288171,"TERMINAL",0,0,"⠹",,terminal_output
767
+ 767,288249,"TERMINAL",0,0,"⠸",,terminal_output
768
+ 768,288329,"TERMINAL",0,0,"⠼",,terminal_output
769
+ 769,288411,"TERMINAL",0,0,"⠴",,terminal_output
770
+ 770,288491,"TERMINAL",0,0,"⠦",,terminal_output
771
+ 771,288572,"TERMINAL",0,0,"⠧",,terminal_output
772
+ 772,288653,"TERMINAL",0,0,"⠇",,terminal_output
773
+ 773,288734,"TERMINAL",0,0,"⠏",,terminal_output
774
+ 774,288814,"TERMINAL",0,0,"⠋",,terminal_output
775
+ 775,288894,"TERMINAL",0,0,"⠙",,terminal_output
776
+ 776,288976,"TERMINAL",0,0,"⠹",,terminal_output
777
+ 777,289057,"TERMINAL",0,0,"⠸",,terminal_output
778
+ 778,289135,"TERMINAL",0,0,"⠼",,terminal_output
779
+ 779,289218,"TERMINAL",0,0,"⠴",,terminal_output
780
+ 780,289298,"TERMINAL",0,0,"⠦",,terminal_output
781
+ 781,289378,"TERMINAL",0,0,"⠧",,terminal_output
782
+ 782,289456,"TERMINAL",0,0,"⠇",,terminal_output
783
+ 783,289537,"TERMINAL",0,0,"⠏",,terminal_output
784
+ 784,289617,"TERMINAL",0,0,"⠋",,terminal_output
785
+ 785,289697,"TERMINAL",0,0,"⠙",,terminal_output
786
+ 786,289776,"TERMINAL",0,0,"⠹",,terminal_output
787
+ 787,289858,"TERMINAL",0,0,"⠸",,terminal_output
788
+ 788,289937,"TERMINAL",0,0,"⠼",,terminal_output
789
+ 789,290016,"TERMINAL",0,0,"⠴",,terminal_output
790
+ 790,290098,"TERMINAL",0,0,"⠦",,terminal_output
791
+ 791,290176,"TERMINAL",0,0,"⠧",,terminal_output
792
+ 792,290258,"TERMINAL",0,0,"⠇",,terminal_output
793
+ 793,290339,"TERMINAL",0,0,"⠏",,terminal_output
794
+ 794,290418,"TERMINAL",0,0,"⠋",,terminal_output
795
+ 795,290496,"TERMINAL",0,0,"⠙",,terminal_output
796
+ 796,290578,"TERMINAL",0,0,"⠹",,terminal_output
797
+ 797,290659,"TERMINAL",0,0,"⠸",,terminal_output
798
+ 798,290738,"TERMINAL",0,0,"⠼",,terminal_output
799
+ 799,290820,"TERMINAL",0,0,"⠴",,terminal_output
800
+ 800,290900,"TERMINAL",0,0,"⠦",,terminal_output
801
+ 801,290982,"TERMINAL",0,0,"⠧",,terminal_output
802
+ 802,291061,"TERMINAL",0,0,"⠇",,terminal_output
803
+ 803,291142,"TERMINAL",0,0,"⠏",,terminal_output
804
+ 804,291222,"TERMINAL",0,0,"⠋",,terminal_output
805
+ 805,291303,"TERMINAL",0,0,"⠙",,terminal_output
806
+ 806,291381,"TERMINAL",0,0,"⠹",,terminal_output
807
+ 807,291463,"TERMINAL",0,0,"⠸",,terminal_output
808
+ 808,291542,"TERMINAL",0,0,"⠼",,terminal_output
809
+ 809,291623,"TERMINAL",0,0,"⠴",,terminal_output
810
+ 810,291702,"TERMINAL",0,0,"⠦",,terminal_output
811
+ 811,291781,"TERMINAL",0,0,"⠧",,terminal_output
812
+ 812,291865,"TERMINAL",0,0,"⠇",,terminal_output
813
+ 813,291944,"TERMINAL",0,0,"⠏",,terminal_output
814
+ 814,292025,"TERMINAL",0,0,"⠋",,terminal_output
815
+ 815,292106,"TERMINAL",0,0,"⠙",,terminal_output
816
+ 816,292186,"TERMINAL",0,0,"⠹",,terminal_output
817
+ 817,292266,"TERMINAL",0,0,"⠸",,terminal_output
818
+ 818,292349,"TERMINAL",0,0,"⠼",,terminal_output
819
+ 819,292426,"TERMINAL",0,0,"⠴",,terminal_output
820
+ 820,292506,"TERMINAL",0,0,"⠦",,terminal_output
821
+ 821,292585,"TERMINAL",0,0,"⠧",,terminal_output
822
+ 822,292667,"TERMINAL",0,0,"⠇",,terminal_output
823
+ 823,292747,"TERMINAL",0,0,"⠏",,terminal_output
824
+ 824,292828,"TERMINAL",0,0,"⠋",,terminal_output
825
+ 825,292908,"TERMINAL",0,0,"⠙",,terminal_output
826
+ 826,292999,"TERMINAL",0,0,"⠹",,terminal_output
827
+ 827,293078,"TERMINAL",0,0,"⠸",,terminal_output
828
+ 828,293160,"TERMINAL",0,0,"⠼",,terminal_output
829
+ 829,293239,"TERMINAL",0,0,"⠴",,terminal_output
830
+ 830,293320,"TERMINAL",0,0,"⠦",,terminal_output
831
+ 831,293399,"TERMINAL",0,0,"⠧",,terminal_output
832
+ 832,293481,"TERMINAL",0,0,"⠇",,terminal_output
833
+ 833,293558,"TERMINAL",0,0,"⠏",,terminal_output
834
+ 834,293639,"TERMINAL",0,0,"⠋",,terminal_output
835
+ 835,293720,"TERMINAL",0,0,"⠙",,terminal_output
836
+ 836,293802,"TERMINAL",0,0,"⠹",,terminal_output
837
+ 837,293880,"TERMINAL",0,0,"⠸",,terminal_output
838
+ 838,293960,"TERMINAL",0,0,"⠼",,terminal_output
839
+ 839,294040,"TERMINAL",0,0,"⠴",,terminal_output
840
+ 840,294121,"TERMINAL",0,0,"⠦",,terminal_output
841
+ 841,294202,"TERMINAL",0,0,"⠧",,terminal_output
842
+ 842,294283,"TERMINAL",0,0,"⠇",,terminal_output
843
+ 843,294363,"TERMINAL",0,0,"⠏",,terminal_output
844
+ 844,294441,"TERMINAL",0,0,"⠋",,terminal_output
845
+ 845,294523,"TERMINAL",0,0,"⠙",,terminal_output
846
+ 846,294621,"TERMINAL",0,0,"⠹",,terminal_output
847
+ 847,294701,"TERMINAL",0,0,"⠸",,terminal_output
848
+ 848,294780,"TERMINAL",0,0,"⠼",,terminal_output
849
+ 849,294861,"TERMINAL",0,0,"⠴",,terminal_output
850
+ 850,294943,"TERMINAL",0,0,"⠦",,terminal_output
851
+ 851,295021,"TERMINAL",0,0,"⠧",,terminal_output
852
+ 852,295102,"TERMINAL",0,0,"⠇",,terminal_output
853
+ 853,295186,"TERMINAL",0,0,"⠏",,terminal_output
854
+ 854,295286,"TERMINAL",0,0,"⠋",,terminal_output
855
+ 855,295359,"TERMINAL",0,0,"⠙",,terminal_output
856
+ 856,295440,"TERMINAL",0,0,"⠹",,terminal_output
857
+ 857,295521,"TERMINAL",0,0,"⠸",,terminal_output
858
+ 858,295601,"TERMINAL",0,0,"⠼",,terminal_output
859
+ 859,295682,"TERMINAL",0,0,"⠴",,terminal_output
860
+ 860,295763,"TERMINAL",0,0,"⠦",,terminal_output
861
+ 861,295844,"TERMINAL",0,0,"⠧",,terminal_output
862
+ 862,295926,"TERMINAL",0,0,"⠇",,terminal_output
863
+ 863,296013,"TERMINAL",0,0,"⠏",,terminal_output
864
+ 864,296086,"TERMINAL",0,0,"⠋",,terminal_output
865
+ 865,296170,"TERMINAL",0,0,"⠙",,terminal_output
866
+ 866,296251,"TERMINAL",0,0,"⠹",,terminal_output
867
+ 867,296330,"TERMINAL",0,0,"⠸",,terminal_output
868
+ 868,296411,"TERMINAL",0,0,"⠼",,terminal_output
869
+ 869,296492,"TERMINAL",0,0,"⠴",,terminal_output
870
+ 870,296572,"TERMINAL",0,0,"⠦",,terminal_output
871
+ 871,296651,"TERMINAL",0,0,"⠧",,terminal_output
872
+ 872,296732,"TERMINAL",0,0,"⠇",,terminal_output
873
+ 873,296815,"TERMINAL",0,0,"⠏",,terminal_output
874
+ 874,296891,"TERMINAL",0,0,"⠋",,terminal_output
875
+ 875,296970,"TERMINAL",0,0,"⠙",,terminal_output
876
+ 876,297050,"TERMINAL",0,0,"⠹",,terminal_output
877
+ 877,297130,"TERMINAL",0,0,"⠸",,terminal_output
878
+ 878,297211,"TERMINAL",0,0,"⠼",,terminal_output
879
+ 879,297291,"TERMINAL",0,0,"⠴",,terminal_output
880
+ 880,297374,"TERMINAL",0,0,"⠦",,terminal_output
881
+ 881,297453,"TERMINAL",0,0,"⠧",,terminal_output
882
+ 882,297533,"TERMINAL",0,0,"⠇",,terminal_output
883
+ 883,297613,"TERMINAL",0,0,"⠏",,terminal_output
884
+ 884,297694,"TERMINAL",0,0,"⠋",,terminal_output
885
+ 885,297775,"TERMINAL",0,0,"⠙",,terminal_output
886
+ 886,297853,"TERMINAL",0,0,"⠹",,terminal_output
887
+ 887,297934,"TERMINAL",0,0,"⠸",,terminal_output
888
+ 888,298014,"TERMINAL",0,0,"⠼",,terminal_output
889
+ 889,298094,"TERMINAL",0,0,"⠴",,terminal_output
890
+ 890,298173,"TERMINAL",0,0,"⠦",,terminal_output
891
+ 891,298254,"TERMINAL",0,0,"⠧",,terminal_output
892
+ 892,298334,"TERMINAL",0,0,"⠇",,terminal_output
893
+ 893,298415,"TERMINAL",0,0,"⠏",,terminal_output
894
+ 894,298494,"TERMINAL",0,0,"⠋",,terminal_output
895
+ 895,298573,"TERMINAL",0,0,"⠙",,terminal_output
896
+ 896,298656,"TERMINAL",0,0,"⠹",,terminal_output
897
+ 897,298737,"TERMINAL",0,0,"⠸",,terminal_output
898
+ 898,298817,"TERMINAL",0,0,"⠼",,terminal_output
899
+ 899,298899,"TERMINAL",0,0,"⠴",,terminal_output
900
+ 900,298979,"TERMINAL",0,0,"⠦",,terminal_output
901
+ 901,299058,"TERMINAL",0,0,"⠧",,terminal_output
902
+ 902,299139,"TERMINAL",0,0,"⠇",,terminal_output
903
+ 903,299220,"TERMINAL",0,0,"⠏",,terminal_output
904
+ 904,299313,"TERMINAL",0,0,"⠋",,terminal_output
905
+ 905,299394,"TERMINAL",0,0,"⠙",,terminal_output
906
+ 906,299475,"TERMINAL",0,0,"⠹",,terminal_output
907
+ 907,299555,"TERMINAL",0,0,"⠸",,terminal_output
908
+ 908,299638,"TERMINAL",0,0,"⠼",,terminal_output
909
+ 909,299717,"TERMINAL",0,0,"⠴",,terminal_output
910
+ 910,299800,"TERMINAL",0,0,"⠦",,terminal_output
911
+ 911,299879,"TERMINAL",0,0,"⠧",,terminal_output
912
+ 912,299961,"TERMINAL",0,0,"⠇",,terminal_output
913
+ 913,300041,"TERMINAL",0,0,"⠏",,terminal_output
914
+ 914,300122,"TERMINAL",0,0,"⠋",,terminal_output
915
+ 915,300204,"TERMINAL",0,0,"⠙",,terminal_output
916
+ 916,300284,"TERMINAL",0,0,"⠹",,terminal_output
917
+ 917,300366,"TERMINAL",0,0,"⠸",,terminal_output
918
+ 918,300447,"TERMINAL",0,0,"⠼",,terminal_output
919
+ 919,300528,"TERMINAL",0,0,"⠴",,terminal_output
920
+ 920,300610,"TERMINAL",0,0,"⠦",,terminal_output
921
+ 921,300691,"TERMINAL",0,0,"⠧",,terminal_output
922
+ 922,300773,"TERMINAL",0,0,"⠇",,terminal_output
923
+ 923,300989,"TERMINAL",0,0,"⠏",,terminal_output
924
+ 924,301052,"TERMINAL",0,0,"⠋\r\nadded 978 packages, and audited 984 packages in 18s\r\n⠋\r\n⠋187 packages are looking for funding\r\n⠋ run `npm fund` for details\r\n⠋\r\nfound 0 vulnerabilities\r\n⠋% \r \r",,terminal_output
925
+ 925,374417,"TERMINAL",0,0,"npm run storybook",,terminal_command
926
+ 926,374468,"TERMINAL",0,0,"]633;C",,terminal_output
927
+ 927,374749,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\n",,terminal_output
928
+ 928,375925,"TERMINAL",0,0,"storybook v9.1.7\r\n\r\n",,terminal_output
929
+ 929,376672,"TERMINAL",0,0,"(node:36013) ExperimentalWarning: Type Stripping is an experimental feature and might change at any time\r\n(Use `node --trace-warnings ...` to show where the warning was created)\r\n",,terminal_output
930
+ 930,376730,"TERMINAL",0,0,"Attention: Storybook now collects completely anonymous telemetry regarding usage. This information is used to shape Storybook's roadmap and prioritize features.\r\nYou can learn more, including how to opt-out if you'd not like to participate in this anonymous program, by visiting the following URL:\r\nhttps://storybook.js.org/telemetry\r\n\r\n",,terminal_output
931
+ 931,377210,"TERMINAL",0,0,"info => Starting manager..\r\n",,terminal_output
932
+ 932,377331,"TERMINAL",0,0,"No story files found for the specified pattern: src/**/*.mdx\r\n",,terminal_output
933
+ 933,377521,"TERMINAL",0,0,"info => Starting preview..\r\n",,terminal_output
934
+ 934,379138,"TERMINAL",0,0,"Building webview for vscode\r\n",,terminal_output
935
+ 935,380938,"TERMINAL",0,0,"info Using tsconfig paths for react-docgen\r\nThe `define` option contains an object with ""PATH"" for ""process.env"" key. It looks like you may have passed the entire `process.env` object to `define`, which can unintentionally expose all environment variables. This poses a security risk and is discouraged.\r\n",,terminal_output
936
+ 936,387648,"TERMINAL",0,0,"╭──────────────────────────────────────────────────────────────────────────────────────────╮\r\n│ │\r\n│ Storybook 9.1.7 for react-vite started │\r\n│ 312 ms for manager and 10 s for preview │\r\n│ │\r\n│ Local: http://localhost:6006/ │\r\n│ On your network: http://192.168.178.68:6006/ │\r\n│ │\r\n│ A new version (10.1.0) is available! │\r\n│ │\r\n│ Upgrade now: npx storybook@latest upgrade │\r\n│ │\r\n│ Read full changelog: https://github.com/storybookjs/storybook/blob/main/CHANGELOG.md │\r\n│ │\r\n╰──────────────────────────────────────────────────────────────────────────────────────────╯\r\n",,terminal_output
937
+ 937,390868,"TERMINAL",0,0,"[baseline-browser-mapping] The data in this module is over two months old. To ensure accurate Baseline data, please update: `npm i baseline-browser-mapping@latest -D`\r\n",,terminal_output
938
+ 938,395179,"TERMINAL",0,0,"12:54:54 PM [vite] (client) Pre-transform error: Failed to resolve import ""@shared/proto/cline/common"" from ""src/context/ClineAuthContext.tsx"". Does the file exist?\r\n Plugin: vite:import-analysis\r\n File: /Users/franzsrambical/Documents/pdoom/cline/webview-ui/src/context/ClineAuthContext.tsx:2:29\r\n 20 | import { jsxDEV as _jsxDEV } from ""react/jsx-dev-runtime"";\r\n 21 | var _s = $RefreshSig$(), _s1 = $RefreshSig$();\r\n 22 | import { EmptyRequest } from ""@shared/proto/cline/common"";\r\n | ^\r\n 23 | import deepEqual from ""fast-deep-equal"";\r\n 24 | import { createContext, useCallback, useContext, useEffect, useMemo, useState } from ""react"";\r\n12:54:54 PM [vite] (client) Pre-transform error: Failed to resolve import ""../services/grpc-client"" from ""src/context/ExtensionStateContext.tsx"". Does the file exist?\r\n Plugin: vite:import-analysis\r\n File: /Users/franzsrambical/Documents/pdoom/cline/webview-ui/src/context/ExtensionStateContext.tsx:31:91\r\n 34 | import { Environment } from ""../../../src/config"";\r\n 35 | import { basetenDefaultModelId, basetenModels, groqDefaultModelId, groqModels, openRouterDefaultModelId, openRouterDe...\r\n 36 | import { McpServiceClient, ModelsServiceClient, StateServiceClient, UiServiceClient } from ""../services/grpc-client"";\r\n | ^\r\n 37 | export const ExtensionStateContext = /*#__PURE__*/ createContext(undefined);\r\n 38 | export const ExtensionStateContextProvider = ({ children })=>{\r\n",,terminal_output
939
+ 939,395343,"TERMINAL",0,0,"12:54:54 PM [vite] (client) Pre-transform error: Failed to resolve import ""@shared/proto/cline/ui"" from ""../src/shared/proto-conversions/cline-message.ts"". Does the file exist?\r\n Plugin: vite:import-analysis\r\n File: /Users/franzsrambical/Documents/pdoom/cline/src/shared/proto-conversions/cline-message.ts:3:88\r\n 1 | import { ClineAsk, ClineMessageType, ClineSay } from ""@shared/proto/cline/ui"";\r\n | ^\r\n 2 | // Helper function to convert ClineAsk string to enum\r\n 3 | function convertClineAskToProtoEnum(ask) {\r\n12:54:54 PM [vite] (client) Pre-transform error: Failed to resolve import ""@shared/proto/cline/mcp"" from ""../src/shared/proto-conversions/mcp/mcp-server-conversion.ts"". Does the file exist?\r\n Plugin: vite:import-analysis\r\n File: /Users/franzsrambical/Documents/pdoom/cline/src/shared/proto-conversions/mcp/mcp-server-conversion.ts:7:7\r\n 1 | import { McpServerStatus } from ""@shared/proto/cline/mcp"";\r\n | ^\r\n 2 | // Helper to convert TS status to Proto enum\r\n 3 | function convertMcpStatusToProto(status) {\r\n12:54:54 PM [vite] (client) Pre-transform error: Failed to resolve import ""@shared/proto/cline/models"" from ""../src/shared/proto-conversions/models/typeConversion.ts"". Does the file exist?\r\n Plugin: vite:import-analysis\r\n File: /Users/franzsrambical/Documents/pdoom/cline/src/shared/proto-conversions/models/typeConversion.ts:8:7\r\n 1 | import { OpenRouterModelInfo, ThinkingConfig } from ""@shared/proto/cline/models"";\r\n | ^\r\n 2 | /**\r\n 3 | * Convert protobuf ThinkingConfig to application ThinkingConfig\r\n",,terminal_output
940
+ 940,398882,"TERMINAL",0,0,"\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n12:54:58 PM [vite] Internal server error: Failed to resolve import ""@shared/proto/cline/common"" from ""src/context/ClineAuthContext.tsx"". Does the file exist?\r\n Plugin: vite:import-analysis\r\n File: /Users/franzsrambical/Documents/pdoom/cline/webview-ui/src/context/ClineAuthContext.tsx:2:29\r\n 20 | import { jsxDEV as _jsxDEV } from ""react/jsx-dev-runtime"";\r\n 21 | var _s = $RefreshSig$(), _s1 = $RefreshSig$();\r\n 22 | import { EmptyRequest } from ""@shared/proto/cline/common"";\r\n | ^\r\n 23 | import deepEqual from ""fast-deep-equal"";\r\n 24 | import { createContext, useCallback, useContext, useEffect, useMemo, useState } from ""react"";\r\n at TransformPluginContext._formatLog (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:29618:43)\r\n at TransformPluginContext.error (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:29615:14)\r\n at normalizeUrl (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:27738:18)\r\n at async file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:27796:32\r\n at async Promise.all (index 3)\r\n at async TransformPluginContext.transform (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:27764:4)\r\n at async EnvironmentPluginContainer.transform (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:29416:14)\r\n at async loadAndTransform (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:23287:26)\r\n at async viteTransformMiddleware (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:25159:20)\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n12:54:58 PM [vite] Internal server error: Failed to resolve import ""../services/grpc-client"" from ""src/context/ExtensionStateContext.tsx"". Does the file exist?\r\n Plugin: vite:import-analysis\r\n File: /Users/franzsrambical/Documents/pdoom/cline/webview-ui/src/context/ExtensionStateContext.tsx:31:91\r\n 34 | import { Environment } from ""../../../src/config"";\r\n 35 | import { basetenDefaultModelId, basetenModels, groqDefaultModelId, groqModels, openRouterDefaultModelId, openRouterDe...\r\n 36 | import { McpServiceClient, ModelsServiceClient, StateServiceClient, UiServiceClient } from ""../services/grpc-client"";\r\n | ^\r\n 37 | export const ExtensionStateContext = /*#__PURE__*/ createContext(undefined);\r\n 38 | export const ExtensionStateContextProvider = ({ children })=>{\r\n at TransformPluginContext._formatLog (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:29618:43)\r\n at TransformPluginContext.error (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:29615:14)\r\n at normalizeUrl (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:27738:18)\r\n at async file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:27796:32\r\n at async Promise.all (index 17)\r\n at async TransformPluginContext.transform (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:27764:4)\r\n at async EnvironmentPluginContainer.transform (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:29416:14)\r\n at async loadAndTransform (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:23287:26)\r\n at async viteTransformMiddleware (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:25159:20)\r\n12:54:58 PM [vite] (client) Pre-transform error: Failed to resolve import ""@shared/proto/cline/mcp"" from ""../src/shared/proto-conversions/mcp/mcp-server-conversion.ts"". Does the file exist?\r\n Plugin: vite:import-analysis\r\n File: /Users/franzsrambical/Documents/pdoom/cline/src/shared/proto-conversions/mcp/mcp-server-conversion.ts:7:7\r\n 1 | import { McpServerStatus } from ""@shared/proto/cline/mcp"";\r\n | ^\r\n 2 | // Helper to convert TS status to Proto enum\r\n 3 | function convertMcpStatusToProto(status) {\r\n12:54:58 PM [vite] (client) Pre-transform error: Failed to resolve import ""@shared/proto/cline/models"" from ""../src/shared/proto-conversions/models/typeConversion.ts"". Does the file exist?\r\n Plugin: vite:import-analysis\r\n File: /Users/franzsrambical/Documents/pdoom/cline/src/shared/proto-conversions/models/typeConversion.ts:8:7\r\n 1 | import { OpenRouterModelInfo, ThinkingConfig } from ""@shared/proto/cline/models"";\r\n | ^\r\n 2 | /**\r\n 3 | * Convert protobuf ThinkingConfig to application ThinkingConfig\r\n12:54:58 PM [vite] (client) Pre-transform error: Failed to resolve import ""@shared/proto/cline/ui"" from ""../src/shared/proto-conversions/cline-message.ts"". Does the file exist?\r\n Plugin: vite:import-analysis\r\n File: /Users/franzsrambical/Documents/pdoom/cline/src/shared/proto-conversions/cline-message.ts:3:88\r\n 1 | import { ClineAsk, ClineMessageType, ClineSay } from ""@shared/proto/cline/ui"";\r\n | ^\r\n 2 | // Helper function to convert ClineAsk string to enum\r\n 3 | function convertClineAskToProtoEnum(ask) {\r\n",,terminal_output
941
+ 941,2981449,"TERMINAL",0,0,"\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n1:38:01 PM [vite] (client) page reload src/services/grpc-client.ts\r\n",,terminal_output
942
+ 942,2987755,"TERMINAL",0,0,"1:38:07 PM [vite] (client) ✨ new dependencies optimized: @bufbuild/protobuf/wire, uuid\r\n1:38:07 PM [vite] (client) ✨ optimized dependencies changed. reloading\r\n",,terminal_output
943
+ 943,2992777,"TERMINAL",0,0,"\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n1:38:12 PM [vite] (client) hmr update /src/index.css\r\n",,terminal_output
944
+ 944,3008542,"TERMINAL",0,0,"\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n1:38:28 PM [vite] (client) hmr update /src/index.css (x2)\r\n",,terminal_output
945
+ 945,3026313,"TERMINAL",0,0,"\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n1:38:45 PM [vite] (client) hmr update /src/index.css (x3)\r\n",,terminal_output
946
+ 946,3034948,"TERMINAL",0,0,"^C",,terminal_output
947
+ 947,3035530,"TERMINAL",0,0,"⠙% \r \r",,terminal_output
948
+ 948,3036908,"TERMINAL",0,0,"npm run storybook",,terminal_command
949
+ 949,3036960,"TERMINAL",0,0,"]633;C",,terminal_output
950
+ 950,3037124,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\n",,terminal_output
951
+ 951,3038143,"TERMINAL",0,0,"storybook v9.1.7\r\n\r\n",,terminal_output
952
+ 952,3038311,"TERMINAL",0,0,"(node:43008) ExperimentalWarning: Type Stripping is an experimental feature and might change at any time\r\n(Use `node --trace-warnings ...` to show where the warning was created)\r\n",,terminal_output
953
+ 953,3038770,"TERMINAL",0,0,"info => Starting manager..\r\n",,terminal_output
954
+ 954,3038852,"TERMINAL",0,0,"No story files found for the specified pattern: src/**/*.mdx\r\n",,terminal_output
955
+ 955,3039062,"TERMINAL",0,0,"info => Starting preview..\r\n",,terminal_output
956
+ 956,3040101,"TERMINAL",0,0,"Building webview for vscode\r\n",,terminal_output
957
+ 957,3042195,"TERMINAL",0,0,"info Using tsconfig paths for react-docgen\r\n=> Failed to build the preview\r\nReferenceError: __dirname is not defined\r\n at Object.viteFinal (./.storybook/main.ts:13:17)\r\n at ./node_modules/storybook/dist/common/index.cjs:23650:18\r\n at async createViteServer (./node_modules/@storybook/builder-vite/dist/index.js:92:25)\r\n at async Module.start (./node_modules/@storybook/builder-vite/dist/index.js:92:658)\r\n at async storybookDevServer (./node_modules/storybook/dist/core-server/index.cjs:42917:79)\r\n at async buildOrThrow (./node_modules/storybook/dist/core-server/index.cjs:39285:12)\r\n at async buildDevStandalone (./node_modules/storybook/dist/core-server/index.cjs:44170:78)\r\n at async withTelemetry (./node_modules/storybook/dist/core-server/index.cjs:42279:12)\r\n at async dev (./node_modules/storybook/dist/cli/bin/index.cjs:5905:3)\r\n at async r.<anonymous> (./node_modules/storybook/dist/cli/bin/index.cjs:6015:74)\r\n\r\nBroken build, fix the error above.\r\nYou may need to refresh the browser.\r\n\r\n[?25l? Would you like to help improve Storybook by sending anonymous crash reports? › (Y/n)",,terminal_output
958
+ 958,3056937,"TERMINAL",0,0,"✔ Would you like to help improve Storybook by sending anonymous crash reports? … no\r\n[?25h",,terminal_output
959
+ 959,3057351,"TERMINAL",0,0,"⠙% \r \r",,terminal_output
960
+ 960,3058995,"TERMINAL",0,0,"npm run storybook",,terminal_command
961
+ 961,3059046,"TERMINAL",0,0,"]633;C",,terminal_output
962
+ 962,3059206,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\n",,terminal_output
963
+ 963,3059534,"TERMINAL",0,0,"storybook v9.1.7\r\n\r\n",,terminal_output
964
+ 964,3059688,"TERMINAL",0,0,"(node:43148) ExperimentalWarning: Type Stripping is an experimental feature and might change at any time\r\n(Use `node --trace-warnings ...` to show where the warning was created)\r\n",,terminal_output
965
+ 965,3060001,"TERMINAL",0,0,"info => Starting manager..\r\nNo story files found for the specified pattern: src/**/*.mdx\r\n",,terminal_output
966
+ 966,3060120,"TERMINAL",0,0,"info => Starting preview..\r\n",,terminal_output
967
+ 967,3060243,"TERMINAL",0,0,"Building webview for vscode\r\n",,terminal_output
968
+ 968,3060516,"TERMINAL",0,0,"info Using tsconfig paths for react-docgen\r\n=> Failed to build the preview\r\nReferenceError: __dirname is not defined\r\n at Object.viteFinal (./.storybook/main.ts:13:17)\r\n at ./node_modules/storybook/dist/common/index.cjs:23650:18\r\n at async createViteServer (./node_modules/@storybook/builder-vite/dist/index.js:92:25)\r\n at async Module.start (./node_modules/@storybook/builder-vite/dist/index.js:92:658)\r\n at async storybookDevServer (./node_modules/storybook/dist/core-server/index.cjs:42917:79)\r\n at async buildOrThrow (./node_modules/storybook/dist/core-server/index.cjs:39285:12)\r\n at async buildDevStandalone (./node_modules/storybook/dist/core-server/index.cjs:44170:78)\r\n at async withTelemetry (./node_modules/storybook/dist/core-server/index.cjs:42279:12)\r\n at async dev (./node_modules/storybook/dist/cli/bin/index.cjs:5905:3)\r\n at async r.<anonymous> (./node_modules/storybook/dist/cli/bin/index.cjs:6015:74)\r\n\r\nBroken build, fix the error above.\r\nYou may need to refresh the browser.\r\n\r\n",,terminal_output
969
+ 969,3060956,"TERMINAL",0,0,"⠙% \r \r",,terminal_output
970
+ 970,3078567,"TERMINAL",0,0,"npm run storybook",,terminal_command
971
+ 971,3078619,"TERMINAL",0,0,"]633;C",,terminal_output
972
+ 972,3078759,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\n",,terminal_output
973
+ 973,3079085,"TERMINAL",0,0,"storybook v9.1.7\r\n\r\n",,terminal_output
974
+ 974,3079242,"TERMINAL",0,0,"(node:43265) ExperimentalWarning: Type Stripping is an experimental feature and might change at any time\r\n(Use `node --trace-warnings ...` to show where the warning was created)\r\n",,terminal_output
975
+ 975,3079508,"TERMINAL",0,0,"info => Starting manager..\r\nNo story files found for the specified pattern: src/**/*.mdx\r\n",,terminal_output
976
+ 976,3079609,"TERMINAL",0,0,"info => Starting preview..\r\n",,terminal_output
977
+ 977,3079857,"TERMINAL",0,0,"Building webview for vscode\r\n",,terminal_output
978
+ 978,3080137,"TERMINAL",0,0,"info Using tsconfig paths for react-docgen\r\n=> Failed to build the preview\r\nReferenceError: __dirname is not defined\r\n at Object.viteFinal (./.storybook/main.ts:13:17)\r\n at ./node_modules/storybook/dist/common/index.cjs:23650:18\r\n at async createViteServer (./node_modules/@storybook/builder-vite/dist/index.js:92:25)\r\n at async Module.start (./node_modules/@storybook/builder-vite/dist/index.js:92:658)\r\n at async storybookDevServer (./node_modules/storybook/dist/core-server/index.cjs:42917:79)\r\n at async buildOrThrow (./node_modules/storybook/dist/core-server/index.cjs:39285:12)\r\n at async buildDevStandalone (./node_modules/storybook/dist/core-server/index.cjs:44170:78)\r\n at async withTelemetry (./node_modules/storybook/dist/core-server/index.cjs:42279:12)\r\n at async dev (./node_modules/storybook/dist/cli/bin/index.cjs:5905:3)\r\n at async r.<anonymous> (./node_modules/storybook/dist/cli/bin/index.cjs:6015:74)\r\n\r\nBroken build, fix the error above.\r\nYou may need to refresh the browser.\r\n\r\n",,terminal_output
979
+ 979,3080559,"TERMINAL",0,0,"⠙% \r \r",,terminal_output
980
+ 980,3082050,"TERMINAL",0,0,"cd ..",,terminal_command
981
+ 981,3082051,"TERMINAL",0,0,"]633;C% \r \r",,terminal_output
982
+ 982,3087158,"TERMINAL",0,0,"npm run storybook",,terminal_command
983
+ 983,3087209,"TERMINAL",0,0,"]633;C",,terminal_output
984
+ 984,3087345,"TERMINAL",0,0,"\r\n> claude-dev@3.38.3 storybook\r\n> cd webview-ui && npm run storybook\r\n\r\n",,terminal_output
985
+ 985,3087427,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\n",,terminal_output
986
+ 986,3087717,"TERMINAL",0,0,"storybook v9.1.7\r\n\r\n",,terminal_output
987
+ 987,3087864,"TERMINAL",0,0,"(node:43453) ExperimentalWarning: Type Stripping is an experimental feature and might change at any time\r\n(Use `node --trace-warnings ...` to show where the warning was created)\r\n",,terminal_output
988
+ 988,3088122,"TERMINAL",0,0,"info => Starting manager..\r\nNo story files found for the specified pattern: src/**/*.mdx\r\n",,terminal_output
989
+ 989,3088223,"TERMINAL",0,0,"info => Starting preview..\r\n",,terminal_output
990
+ 990,3088362,"TERMINAL",0,0,"Building webview for vscode\r\n",,terminal_output
991
+ 991,3088601,"TERMINAL",0,0,"info Using tsconfig paths for react-docgen\r\n=> Failed to build the preview\r\nReferenceError: __dirname is not defined\r\n at Object.viteFinal (./.storybook/main.ts:13:17)\r\n at ./node_modules/storybook/dist/common/index.cjs:23650:18\r\n at async createViteServer (./node_modules/@storybook/builder-vite/dist/index.js:92:25)\r\n at async Module.start (./node_modules/@storybook/builder-vite/dist/index.js:92:658)\r\n at async storybookDevServer (./node_modules/storybook/dist/core-server/index.cjs:42917:79)\r\n at async buildOrThrow (./node_modules/storybook/dist/core-server/index.cjs:39285:12)\r\n at async buildDevStandalone (./node_modules/storybook/dist/core-server/index.cjs:44170:78)\r\n at async withTelemetry (./node_modules/storybook/dist/core-server/index.cjs:42279:12)\r\n at async dev (./node_modules/storybook/dist/cli/bin/index.cjs:5905:3)\r\n at async r.<anonymous> (./node_modules/storybook/dist/cli/bin/index.cjs:6015:74)\r\n\r\nBroken build, fix the error above.\r\nYou may need to refresh the browser.\r\n\r\n",,terminal_output
992
+ 992,3088992,"TERMINAL",0,0,"⠙⠙% \r \r",,terminal_output
993
+ 993,3161357,"TERMINAL",0,0,"cd webview-ui",,terminal_command
994
+ 994,3161357,"TERMINAL",0,0,"]633;C% \r \r",,terminal_output
995
+ 995,3162625,"TERMINAL",0,0,"npm run storybook",,terminal_command
996
+ 996,3162677,"TERMINAL",0,0,"]633;C",,terminal_output
997
+ 997,3162941,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\n",,terminal_output
998
+ 998,3163284,"TERMINAL",0,0,"storybook v9.1.7\r\n\r\n",,terminal_output
999
+ 999,3163477,"TERMINAL",0,0,"(node:43708) ExperimentalWarning: Type Stripping is an experimental feature and might change at any time\r\n(Use `node --trace-warnings ...` to show where the warning was created)\r\n",,terminal_output
1000
+ 1000,3163764,"TERMINAL",0,0,"info => Starting manager..\r\nNo story files found for the specified pattern: src/**/*.mdx\r\n",,terminal_output
1001
+ 1001,3163883,"TERMINAL",0,0,"info => Starting preview..\r\n",,terminal_output
1002
+ 1002,3164047,"TERMINAL",0,0,"Building webview for vscode\r\n",,terminal_output
1003
+ 1003,3164319,"TERMINAL",0,0,"info Using tsconfig paths for react-docgen\r\nThe `define` option contains an object with ""PATH"" for ""process.env"" key. It looks like you may have passed the entire `process.env` object to `define`, which can unintentionally expose all environment variables. This poses a security risk and is discouraged.\r\n",,terminal_output
1004
+ 1004,3169884,"TERMINAL",0,0,"╭──────────────────────────────────────────────────────────────────────────────────────────╮\r\n│ │\r\n│ Storybook 9.1.7 for react-vite started │\r\n│ 157 ms for manager and 5.95 s for preview │\r\n│ │\r\n│ Local: http://localhost:6006/ │\r\n│ On your network: http://192.168.178.68:6006/ │\r\n│ │\r\n│ A new version (10.1.0) is available! │\r\n│ │\r\n│ Upgrade now: npx storybook@latest upgrade │\r\n│ │\r\n│ Read full changelog: https://github.com/storybookjs/storybook/blob/main/CHANGELOG.md │\r\n│ │\r\n╰──────────────────────────────────────────────────────────────────────────────────────────╯\r\n",,terminal_output
1005
+ 1005,3170751,"TERMINAL",0,0,"[baseline-browser-mapping] The data in this module is over two months old. To ensure accurate Baseline data, please update: `npm i baseline-browser-mapping@latest -D`\r\n",,terminal_output
1006
+ 1006,3210633,"TERMINAL",0,0,"The request id ""/Users/franzsrambical/Documents/pdoom/cline/node_modules/@vscode/codicons/dist/codicon.ttf"" is outside of Vite serving allow list.\r\n\r\n- /Users/franzsrambical/Documents/pdoom/cline/webview-ui\r\n\r\nRefer to docs https://vite.dev/config/server-options.html#server-fs-allow for configurations and more details.\r\n\r\n",,terminal_output
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-8eaf046e-99c4-4091-a85d-91e359564aa51756825908437-2025_09_02-17.11.50.753/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-9543d8e2-2376-4957-873e-df7016d502961763465687199-2025_11_18-12.34.49.442/source.csv ADDED
@@ -0,0 +1,226 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 2,67,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"12:34:49 PM [info] Activating crowd-code\n12:34:49 PM [info] Recording started\n12:34:49 PM [info] Initializing git provider using file system watchers...\n12:34:49 PM [info] Git repository found\n12:34:49 PM [info] Git provider initialized successfully\n",Log,tab
3
+ 3,116,"extension-output-pdoom-org.crowd-code-#1-crowd-code",250,0,"12:34:49 PM [info] Initial git state: [object Object]\n",Log,content
4
+ 4,2975,"extension-output-pdoom-org.crowd-code-#1-crowd-code",304,0,"",Log,selection_mouse
5
+ 5,13030,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,tab
6
+ 6,20469,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",64,0,"",python,selection_command
7
+ 7,20661,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",66,0,"",python,selection_command
8
+ 8,20694,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",132,0,"",python,selection_command
9
+ 9,20728,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",199,0,"",python,selection_command
10
+ 10,20761,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",241,0,"",python,selection_command
11
+ 11,20794,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",243,0,"",python,selection_command
12
+ 12,20828,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",292,0,"",python,selection_command
13
+ 13,20863,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",294,0,"",python,selection_command
14
+ 14,20904,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",364,0,"",python,selection_command
15
+ 15,20936,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",432,0,"",python,selection_command
16
+ 16,20962,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",507,0,"",python,selection_command
17
+ 17,20996,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",577,0,"",python,selection_command
18
+ 18,21030,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",610,0,"",python,selection_command
19
+ 19,21068,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",611,0,"",python,selection_command
20
+ 20,21097,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",639,0,"",python,selection_command
21
+ 21,21131,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",640,0,"",python,selection_command
22
+ 22,21166,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",671,0,"",python,selection_command
23
+ 23,21203,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",694,0,"",python,selection_command
24
+ 24,21233,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",707,0,"",python,selection_command
25
+ 25,21265,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",795,0,"",python,selection_command
26
+ 26,21297,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",796,0,"",python,selection_command
27
+ 27,21331,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",852,0,"",python,selection_command
28
+ 28,21364,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",914,0,"",python,selection_command
29
+ 29,21397,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",965,0,"",python,selection_command
30
+ 30,21431,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1047,0,"",python,selection_command
31
+ 31,21464,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1148,0,"",python,selection_command
32
+ 32,21497,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1245,0,"",python,selection_command
33
+ 33,21530,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1319,0,"",python,selection_command
34
+ 34,21565,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1369,0,"",python,selection_command
35
+ 35,21598,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1370,0,"",python,selection_command
36
+ 36,21631,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1393,0,"",python,selection_command
37
+ 37,21668,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1394,0,"",python,selection_command
38
+ 38,21701,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1395,0,"",python,selection_command
39
+ 39,21890,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1423,0,"",python,selection_command
40
+ 40,23050,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1427,0,"",python,selection_command
41
+ 41,23197,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1432,0,"",python,selection_command
42
+ 42,23489,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1435,0,"",python,selection_command
43
+ 43,23702,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1438,0,"",python,selection_command
44
+ 44,23912,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1441,0,"",python,selection_command
45
+ 45,24119,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1442,0,"",python,selection_command
46
+ 46,24327,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1448,0,"",python,selection_command
47
+ 47,24581,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1449,0,"",python,selection_command
48
+ 48,24945,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1451,0,"",python,selection_command
49
+ 49,25331,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1452,0,"",python,selection_command
50
+ 50,26280,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1423,0,"",python,selection_command
51
+ 51,26667,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1395,0,"",python,selection_command
52
+ 52,26920,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1394,0,"",python,selection_command
53
+ 53,26954,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1393,0,"",python,selection_command
54
+ 54,146021,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
55
+ 55,146654,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,tab
56
+ 56,203332,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2838,0,"",python,selection_keyboard
57
+ 57,203923,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",5124,0,"",python,selection_keyboard
58
+ 58,204398,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7126,0,"",python,selection_keyboard
59
+ 59,204829,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9057,0,"",python,selection_keyboard
60
+ 60,205408,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_keyboard
61
+ 61,210105,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7721,0,"",python,selection_keyboard
62
+ 62,210235,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",5922,0,"",python,selection_keyboard
63
+ 63,210382,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3490,0,"",python,selection_keyboard
64
+ 64,210771,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1906,0,"",python,selection_keyboard
65
+ 65,210944,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",64,0,"",python,selection_keyboard
66
+ 66,211099,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_keyboard
67
+ 67,214839,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10,0,"",python,selection_command
68
+ 68,2267039,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2294,0,"",python,selection_mouse
69
+ 69,2272461,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2114,0,"",python,selection_mouse
70
+ 70,3052561,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,10041,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,selection_command
71
+ 71,3052668,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_command
72
+ 72,3145153,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_command
73
+ 73,3148126,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7487,0,"",python,selection_command
74
+ 74,3149770,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7876,0,"",python,selection_command
75
+ 75,3150226,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",8235,0,"",python,selection_command
76
+ 76,3150898,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7876,0,"",python,selection_command
77
+ 77,3151966,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",8235,0,"",python,selection_command
78
+ 78,3161866,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9070,0,"",python,selection_command
79
+ 79,3164224,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9131,0,"",python,selection_command
80
+ 80,3166810,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9194,0,"",python,selection_command
81
+ 81,3167056,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9257,0,"",python,selection_command
82
+ 82,3167090,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9321,0,"",python,selection_command
83
+ 83,3167123,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9386,0,"",python,selection_command
84
+ 84,3167156,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9443,0,"",python,selection_command
85
+ 85,3167189,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9481,0,"",python,selection_command
86
+ 86,3167223,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9531,0,"",python,selection_command
87
+ 87,3168286,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9532,0,"",python,selection_command
88
+ 88,3168527,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9538,0,"",python,selection_command
89
+ 89,3168566,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9539,0,"",python,selection_command
90
+ 90,3168597,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9546,0,"",python,selection_command
91
+ 91,3168633,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9547,0,"",python,selection_command
92
+ 92,3168887,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9555,0,"",python,selection_command
93
+ 93,3169118,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
94
+ 94,3169292,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9583,0,"",python,selection_command
95
+ 95,3169900,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
96
+ 96,3170394,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
97
+ 97,3171165,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
98
+ 98,3171795,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
99
+ 99,3171945,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
100
+ 100,3172433,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
101
+ 101,3173957,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
102
+ 102,3174398,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
103
+ 103,3174801,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9682,0,"",python,selection_command
104
+ 104,3175123,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9746,0,"",python,selection_command
105
+ 105,3175686,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9682,0,"",python,selection_command
106
+ 106,3175821,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
107
+ 107,3175981,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
108
+ 108,3240534,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_command
109
+ 109,3241210,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1898,0,"",python,selection_keyboard
110
+ 110,3241373,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3420,0,"",python,selection_keyboard
111
+ 111,3242843,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3370,0,"",python,selection_command
112
+ 112,3243091,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3298,0,"",python,selection_command
113
+ 113,3243124,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3288,0,"",python,selection_command
114
+ 114,3243158,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3287,0,"",python,selection_command
115
+ 115,3243191,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3211,0,"",python,selection_command
116
+ 116,3243225,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3137,0,"",python,selection_command
117
+ 117,3243258,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3058,0,"",python,selection_command
118
+ 118,3243291,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3057,0,"",python,selection_command
119
+ 119,3243324,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2999,0,"",python,selection_command
120
+ 120,3243357,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2991,0,"",python,selection_command
121
+ 121,3243391,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2973,0,"",python,selection_command
122
+ 122,3243424,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2956,0,"",python,selection_command
123
+ 123,3243458,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2932,0,"",python,selection_command
124
+ 124,3243491,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2915,0,"",python,selection_command
125
+ 125,3243525,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2898,0,"",python,selection_command
126
+ 126,3243558,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2876,0,"",python,selection_command
127
+ 127,3243591,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2854,0,"",python,selection_command
128
+ 128,3243632,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2838,0,"",python,selection_command
129
+ 129,3243658,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2817,0,"",python,selection_command
130
+ 130,3243691,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2793,0,"",python,selection_command
131
+ 131,3243734,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2767,0,"",python,selection_command
132
+ 132,3243758,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2756,0,"",python,selection_command
133
+ 133,3243791,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2721,0,"",python,selection_command
134
+ 134,3243824,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2688,0,"",python,selection_command
135
+ 135,3243857,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2655,0,"",python,selection_command
136
+ 136,3243891,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2621,0,"",python,selection_command
137
+ 137,3243925,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2585,0,"",python,selection_command
138
+ 138,3243958,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2547,0,"",python,selection_command
139
+ 139,3243991,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2512,0,"",python,selection_command
140
+ 140,3244025,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2483,0,"",python,selection_command
141
+ 141,3244059,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2451,0,"",python,selection_command
142
+ 142,3244092,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2427,0,"",python,selection_command
143
+ 143,3244125,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2388,0,"",python,selection_command
144
+ 144,3286103,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2411,0,"",python,selection_command
145
+ 145,3296751,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2388,23,"",python,content
146
+ 146,3298859,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2388,0," sequence_parallelis",python,content
147
+ 147,3298862,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2388,0,"",python,selection_command
148
+ 148,3483131,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
149
+ 149,3484741,"TERMINAL",0,0,"",,terminal_focus
150
+ 150,3484752,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,tab
151
+ 151,3788287,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2427,0,"",python,selection_command
152
+ 152,3788457,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2428,0,"",python,selection_command
153
+ 153,3788946,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2429,0,"",python,selection_command
154
+ 154,3789583,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2427,0,"",python,selection_command
155
+ 155,4549882,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,10041,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,selection_command
156
+ 156,4550471,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_command
157
+ 157,4559897,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,10041,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,selection_command
158
+ 158,4560021,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_command
159
+ 159,4560890,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,content
160
+ 160,4583906,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18686,0,"",python,selection_mouse
161
+ 161,4583914,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18685,0,"",python,selection_command
162
+ 162,4584529,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18699,0,"",python,selection_command
163
+ 163,4585394,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18743,0,"",python,selection_command
164
+ 164,4585610,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18752,0,"",python,selection_command
165
+ 165,4591082,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,tab
166
+ 166,4591147,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",64,0,"",python,selection_command
167
+ 167,4595172,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,10041,"",python,content
168
+ 168,4597038,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,tab
169
+ 169,5454427,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9372,0,"",python,selection_mouse
170
+ 170,5454796,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_command
171
+ 171,5455466,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",64,0,"",python,selection_command
172
+ 172,5455793,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1906,0,"",python,selection_keyboard
173
+ 173,5456269,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3490,0,"",python,selection_keyboard
174
+ 174,5456884,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3420,0,"",python,selection_command
175
+ 175,5457139,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3370,0,"",python,selection_command
176
+ 176,5457172,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3298,0,"",python,selection_command
177
+ 177,5457202,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3288,0,"",python,selection_command
178
+ 178,5457232,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3287,0,"",python,selection_command
179
+ 179,5457265,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3211,0,"",python,selection_command
180
+ 180,5457298,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3137,0,"",python,selection_command
181
+ 181,5457331,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3058,0,"",python,selection_command
182
+ 182,5457367,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3057,0,"",python,selection_command
183
+ 183,5457402,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2999,0,"",python,selection_command
184
+ 184,5457431,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2991,0,"",python,selection_command
185
+ 185,5457465,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2973,0,"",python,selection_command
186
+ 186,5457497,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2956,0,"",python,selection_command
187
+ 187,5457533,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2932,0,"",python,selection_command
188
+ 188,5457565,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2915,0,"",python,selection_command
189
+ 189,5457598,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2898,0,"",python,selection_command
190
+ 190,5457631,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2876,0,"",python,selection_command
191
+ 191,5457665,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2854,0,"",python,selection_command
192
+ 192,5457699,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2838,0,"",python,selection_command
193
+ 193,5457732,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2817,0,"",python,selection_command
194
+ 194,5457764,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2793,0,"",python,selection_command
195
+ 195,5457799,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2767,0,"",python,selection_command
196
+ 196,5457833,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2756,0,"",python,selection_command
197
+ 197,5457866,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2721,0,"",python,selection_command
198
+ 198,5457899,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2688,0,"",python,selection_command
199
+ 199,5457931,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2655,0,"",python,selection_command
200
+ 200,5457964,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2621,0,"",python,selection_command
201
+ 201,5457999,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2585,0,"",python,selection_command
202
+ 202,5458033,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2547,0,"",python,selection_command
203
+ 203,5458067,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2512,0,"",python,selection_command
204
+ 204,5458102,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2483,0,"",python,selection_command
205
+ 205,5458137,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2451,0,"",python,selection_command
206
+ 206,5459402,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2483,0,"",python,selection_command
207
+ 207,5459593,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2512,0,"",python,selection_command
208
+ 208,5459748,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2547,0,"",python,selection_command
209
+ 209,5460152,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2551,0,"",python,selection_command
210
+ 210,5461248,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2589,0,"",python,selection_command
211
+ 211,5461431,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2625,0,"",python,selection_command
212
+ 212,5461588,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2659,0,"",python,selection_command
213
+ 213,5461782,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2692,0,"",python,selection_command
214
+ 214,5462187,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2725,0,"",python,selection_command
215
+ 215,5462694,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2692,0,"",python,selection_command
216
+ 216,5462868,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2659,0,"",python,selection_command
217
+ 217,5463112,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2625,0,"",python,selection_command
218
+ 218,5463146,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2589,0,"",python,selection_command
219
+ 219,5463184,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2551,0,"",python,selection_command
220
+ 220,5463214,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2516,0,"",python,selection_command
221
+ 221,5463329,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2487,0,"",python,selection_command
222
+ 222,8520791,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,10041,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,selection_command
223
+ 223,8520860,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_command
224
+ 224,8685736,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2797,0,"",python,selection_command
225
+ 225,8686757,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2813,0,"",python,selection_command
226
+ 226,8686946,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2814,0,"",python,selection_command
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-a7b808c2-b1d0-43a0-a38c-8b82cd2886711764488770794-2025_11_30-08.46.18.897/source.csv ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,13,"Untitled-1",0,0,"",plaintext,tab
3
+ 2,77,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"8:46:18 AM [info] Activating crowd-code\n8:46:18 AM [info] Recording started\n8:46:18 AM [info] Initializing git provider using file system watchers...\n8:46:18 AM [info] No workspace folder found\n",Log,tab
4
+ 3,2028,"extension-output-pdoom-org.crowd-code-#1-crowd-code",194,0,"8:46:20 AM [info] Retrying git provider initialization...\n8:46:20 AM [info] No workspace folder found\n",Log,content
5
+ 4,2227,"Untitled-1",0,0,"",plaintext,tab
6
+ 5,3588,"TERMINAL",0,0,"Test",,terminal_focus
7
+ 6,3593,"Untitled-1",0,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
8
+ 7,4572,"Untitled-1",76,0,"/* crowd-pilot: replacement */\nREPLACED LINE 1\nREPLACED LINE 2",plaintext,content
9
+ 8,6374,"Untitled-1",122,0,"",plaintext,selection_command
10
+ 9,6522,"Untitled-1",91,0,"",plaintext,selection_command
11
+ 10,6765,"Untitled-1",122,0,"",plaintext,selection_command
12
+ 11,6932,"Untitled-1",138,0,"",plaintext,selection_command
13
+ 12,7497,"Untitled-1",123,15,"",plaintext,content
14
+ 13,8248,"TERMINAL",0,0,"echo ""Hello World""",,terminal_command
15
+ 14,8249,"TERMINAL",0,0,"]633;CHello World\r\n% \r \r",,terminal_output
16
+ 15,9755,"Untitled-1",107,0,"",plaintext,selection_command
17
+ 16,9857,"Untitled-1",76,0,"",plaintext,selection_command
18
+ 17,10024,"Untitled-1",46,0,"",plaintext,selection_command
19
+ 18,10659,"Untitled-1",39,0,"",plaintext,selection_command
20
+ 19,10826,"Untitled-1",32,0,"",plaintext,selection_command
21
+ 20,10955,"Untitled-1",0,0,"",plaintext,selection_command
22
+ 21,14633,"Untitled-1",0,31,"/* crowd-pilot: insert start */",plaintext,selection_command
23
+ 22,15287,"Untitled-1",0,38,"/* crowd-pilot: insert start */\nline A",plaintext,selection_command
24
+ 23,15426,"Untitled-1",0,45,"/* crowd-pilot: insert start */\nline A\nline B",plaintext,selection_command
25
+ 24,15551,"Untitled-1",0,75,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */",plaintext,selection_command
26
+ 25,15693,"Untitled-1",0,106,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n/* crowd-pilot: replacement */",plaintext,selection_command
27
+ 26,15820,"Untitled-1",0,122,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n/* crowd-pilot: replacement */\nREPLACED LINE 1",plaintext,selection_command
28
+ 27,15909,"Untitled-1",0,122,"",plaintext,content
29
+ 28,16827,"Untitled-1",1,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
30
+ 29,19025,"Untitled-1",0,1,"\n",plaintext,selection_command
31
+ 30,19300,"Untitled-1",0,77,"\n/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,selection_command
32
+ 31,19520,"Untitled-1",0,77,"",plaintext,content
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-aeed47b9-f6ef-4272-b0ca-0c15ab4c25021758266694991-2025_09_19-09.25.04.660/source.csv ADDED
@@ -0,0 +1,151 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,3,"cleanrl/ppo_atari_envpool.py",0,0,"# docs and experiment results can be found at https://docs.cleanrl.dev/rl-algorithms/ppo/#ppo_atari_envpoolpy\nimport os\nimport random\nimport time\nfrom collections import deque\nfrom dataclasses import dataclass\n\nimport envpool\nimport gym\nimport numpy as np\nimport torch\nimport torch.nn as nn\nimport torch.optim as optim\nimport tyro\nfrom torch.distributions.categorical import Categorical\nfrom torch.utils.tensorboard import SummaryWriter\n\n\n@dataclass\nclass Args:\n exp_name: str = os.path.basename(__file__)[: -len("".py"")]\n """"""the name of this experiment""""""\n seed: int = 1\n """"""seed of the experiment""""""\n torch_deterministic: bool = True\n """"""if toggled, `torch.backends.cudnn.deterministic=False`""""""\n cuda: bool = True\n """"""if toggled, cuda will be enabled by default""""""\n track: bool = False\n """"""if toggled, this experiment will be tracked with Weights and Biases""""""\n wandb_project_name: str = ""cleanRL""\n """"""the wandb's project name""""""\n wandb_entity: str = None\n """"""the entity (team) of wandb's project""""""\n capture_video: bool = False\n """"""whether to capture videos of the agent performances (check out `videos` folder)""""""\n\n # Algorithm specific arguments\n env_id: str = ""Breakout-v5""\n """"""the id of the environment""""""\n total_timesteps: int = 10000000\n """"""total timesteps of the experiments""""""\n learning_rate: float = 2.5e-4\n """"""the learning rate of the optimizer""""""\n num_envs: int = 8\n """"""the number of parallel game environments""""""\n num_steps: int = 128\n """"""the number of steps to run in each environment per policy rollout""""""\n anneal_lr: bool = True\n """"""Toggle learning rate annealing for policy and value networks""""""\n gamma: float = 0.99\n """"""the discount factor gamma""""""\n gae_lambda: float = 0.95\n """"""the lambda for the general advantage estimation""""""\n num_minibatches: int = 4\n """"""the number of mini-batches""""""\n update_epochs: int = 4\n """"""the K epochs to update the policy""""""\n norm_adv: bool = True\n """"""Toggles advantages normalization""""""\n clip_coef: float = 0.1\n """"""the surrogate clipping coefficient""""""\n clip_vloss: bool = True\n """"""Toggles whether or not to use a clipped loss for the value function, as per the paper.""""""\n ent_coef: float = 0.01\n """"""coefficient of the entropy""""""\n vf_coef: float = 0.5\n """"""coefficient of the value function""""""\n max_grad_norm: float = 0.5\n """"""the maximum norm for the gradient clipping""""""\n target_kl: float = None\n """"""the target KL divergence threshold""""""\n\n # to be filled in runtime\n batch_size: int = 0\n """"""the batch size (computed in runtime)""""""\n minibatch_size: int = 0\n """"""the mini-batch size (computed in runtime)""""""\n num_iterations: int = 0\n """"""the number of iterations (computed in runtime)""""""\n\n\nclass RecordEpisodeStatistics(gym.Wrapper):\n def __init__(self, env, deque_size=100):\n super().__init__(env)\n self.num_envs = getattr(env, ""num_envs"", 1)\n self.episode_returns = None\n self.episode_lengths = None\n\n def reset(self, **kwargs):\n observations = super().reset(**kwargs)\n self.episode_returns = np.zeros(self.num_envs, dtype=np.float32)\n self.episode_lengths = np.zeros(self.num_envs, dtype=np.int32)\n self.lives = np.zeros(self.num_envs, dtype=np.int32)\n self.returned_episode_returns = np.zeros(self.num_envs, dtype=np.float32)\n self.returned_episode_lengths = np.zeros(self.num_envs, dtype=np.int32)\n return observations\n\n def step(self, action):\n observations, rewards, dones, infos = super().step(action)\n self.episode_returns += infos[""reward""]\n self.episode_lengths += 1\n self.returned_episode_returns[:] = self.episode_returns\n self.returned_episode_lengths[:] = self.episode_lengths\n self.episode_returns *= 1 - infos[""terminated""]\n self.episode_lengths *= 1 - infos[""terminated""]\n infos[""r""] = self.returned_episode_returns\n infos[""l""] = self.returned_episode_lengths\n return (\n observations,\n rewards,\n dones,\n infos,\n )\n\n\ndef layer_init(layer, std=np.sqrt(2), bias_const=0.0):\n torch.nn.init.orthogonal_(layer.weight, std)\n torch.nn.init.constant_(layer.bias, bias_const)\n return layer\n\n\nclass Agent(nn.Module):\n def __init__(self, envs):\n super().__init__()\n self.network = nn.Sequential(\n layer_init(nn.Conv2d(4, 32, 8, stride=4)),\n nn.ReLU(),\n layer_init(nn.Conv2d(32, 64, 4, stride=2)),\n nn.ReLU(),\n layer_init(nn.Conv2d(64, 64, 3, stride=1)),\n nn.ReLU(),\n nn.Flatten(),\n layer_init(nn.Linear(64 * 7 * 7, 512)),\n nn.ReLU(),\n )\n self.actor = layer_init(nn.Linear(512, envs.single_action_space.n), std=0.01)\n self.critic = layer_init(nn.Linear(512, 1), std=1)\n\n def get_value(self, x):\n return self.critic(self.network(x / 255.0))\n\n def get_action_and_value(self, x, action=None):\n hidden = self.network(x / 255.0)\n logits = self.actor(hidden)\n probs = Categorical(logits=logits)\n if action is None:\n action = probs.sample()\n return action, probs.log_prob(action), probs.entropy(), self.critic(hidden)\n\n\nif __name__ == ""__main__"":\n args = tyro.cli(Args)\n args.batch_size = int(args.num_envs * args.num_steps)\n args.minibatch_size = int(args.batch_size // args.num_minibatches)\n args.num_iterations = args.total_timesteps // args.batch_size\n run_name = f""{args.env_id}__{args.exp_name}__{args.seed}__{int(time.time())}""\n if args.track:\n import wandb\n\n wandb.init(\n project=args.wandb_project_name,\n entity=args.wandb_entity,\n sync_tensorboard=True,\n config=vars(args),\n name=run_name,\n monitor_gym=True,\n save_code=True,\n )\n writer = SummaryWriter(f""runs/{run_name}"")\n writer.add_text(\n ""hyperparameters"",\n ""|param|value|\n|-|-|\n%s"" % (""\n"".join([f""|{key}|{value}|"" for key, value in vars(args).items()])),\n )\n\n # TRY NOT TO MODIFY: seeding\n random.seed(args.seed)\n np.random.seed(args.seed)\n torch.manual_seed(args.seed)\n torch.backends.cudnn.deterministic = args.torch_deterministic\n\n device = torch.device(""cuda"" if torch.cuda.is_available() and args.cuda else ""cpu"")\n\n # env setup\n envs = envpool.make(\n args.env_id,\n env_type=""gym"",\n num_envs=args.num_envs,\n episodic_life=True,\n reward_clip=True,\n seed=args.seed,\n )\n envs.num_envs = args.num_envs\n envs.single_action_space = envs.action_space\n envs.single_observation_space = envs.observation_space\n envs = RecordEpisodeStatistics(envs)\n assert isinstance(envs.action_space, gym.spaces.Discrete), ""only discrete action space is supported""\n\n agent = Agent(envs).to(device)\n optimizer = optim.Adam(agent.parameters(), lr=args.learning_rate, eps=1e-5)\n\n # ALGO Logic: Storage setup\n obs = torch.zeros((args.num_steps, args.num_envs) + envs.single_observation_space.shape).to(device)\n actions = torch.zeros((args.num_steps, args.num_envs) + envs.single_action_space.shape).to(device)\n logprobs = torch.zeros((args.num_steps, args.num_envs)).to(device)\n rewards = torch.zeros((args.num_steps, args.num_envs)).to(device)\n dones = torch.zeros((args.num_steps, args.num_envs)).to(device)\n values = torch.zeros((args.num_steps, args.num_envs)).to(device)\n avg_returns = deque(maxlen=20)\n\n # TRY NOT TO MODIFY: start the game\n global_step = 0\n start_time = time.time()\n next_obs = torch.Tensor(envs.reset()).to(device)\n next_done = torch.zeros(args.num_envs).to(device)\n\n for iteration in range(1, args.num_iterations + 1):\n # Annealing the rate if instructed to do so.\n if args.anneal_lr:\n frac = 1.0 - (iteration - 1.0) / args.num_iterations\n lrnow = frac * args.learning_rate\n optimizer.param_groups[0][""lr""] = lrnow\n\n for step in range(0, args.num_steps):\n global_step += args.num_envs\n obs[step] = next_obs\n dones[step] = next_done\n\n # ALGO LOGIC: action logic\n with torch.no_grad():\n action, logprob, _, value = agent.get_action_and_value(next_obs)\n values[step] = value.flatten()\n actions[step] = action\n logprobs[step] = logprob\n\n # TRY NOT TO MODIFY: execute the game and log data.\n next_obs, reward, next_done, info = envs.step(action.cpu().numpy())\n rewards[step] = torch.tensor(reward).to(device).view(-1)\n next_obs, next_done = torch.Tensor(next_obs).to(device), torch.Tensor(next_done).to(device)\n\n for idx, d in enumerate(next_done):\n if d and info[""lives""][idx] == 0:\n print(f""global_step={global_step}, episodic_return={info['r'][idx]}"")\n avg_returns.append(info[""r""][idx])\n writer.add_scalar(""charts/avg_episodic_return"", np.average(avg_returns), global_step)\n writer.add_scalar(""charts/episodic_return"", info[""r""][idx], global_step)\n writer.add_scalar(""charts/episodic_length"", info[""l""][idx], global_step)\n\n # bootstrap value if not done\n with torch.no_grad():\n next_value = agent.get_value(next_obs).reshape(1, -1)\n advantages = torch.zeros_like(rewards).to(device)\n lastgaelam = 0\n for t in reversed(range(args.num_steps)):\n if t == args.num_steps - 1:\n nextnonterminal = 1.0 - next_done\n nextvalues = next_value\n else:\n nextnonterminal = 1.0 - dones[t + 1]\n nextvalues = values[t + 1]\n delta = rewards[t] + args.gamma * nextvalues * nextnonterminal - values[t]\n advantages[t] = lastgaelam = delta + args.gamma * args.gae_lambda * nextnonterminal * lastgaelam\n returns = advantages + values\n\n # flatten the batch\n b_obs = obs.reshape((-1,) + envs.single_observation_space.shape)\n b_logprobs = logprobs.reshape(-1)\n b_actions = actions.reshape((-1,) + envs.single_action_space.shape)\n b_advantages = advantages.reshape(-1)\n b_returns = returns.reshape(-1)\n b_values = values.reshape(-1)\n\n # Optimizing the policy and value network\n b_inds = np.arange(args.batch_size)\n clipfracs = []\n for epoch in range(args.update_epochs):\n np.random.shuffle(b_inds)\n for start in range(0, args.batch_size, args.minibatch_size):\n end = start + args.minibatch_size\n mb_inds = b_inds[start:end]\n\n _, newlogprob, entropy, newvalue = agent.get_action_and_value(b_obs[mb_inds], b_actions.long()[mb_inds])\n logratio = newlogprob - b_logprobs[mb_inds]\n ratio = logratio.exp()\n\n with torch.no_grad():\n # calculate approx_kl http://joschu.net/blog/kl-approx.html\n old_approx_kl = (-logratio).mean()\n approx_kl = ((ratio - 1) - logratio).mean()\n clipfracs += [((ratio - 1.0).abs() > args.clip_coef).float().mean().item()]\n\n mb_advantages = b_advantages[mb_inds]\n if args.norm_adv:\n mb_advantages = (mb_advantages - mb_advantages.mean()) / (mb_advantages.std() + 1e-8)\n\n # Policy loss\n pg_loss1 = -mb_advantages * ratio\n pg_loss2 = -mb_advantages * torch.clamp(ratio, 1 - args.clip_coef, 1 + args.clip_coef)\n pg_loss = torch.max(pg_loss1, pg_loss2).mean()\n\n # Value loss\n newvalue = newvalue.view(-1)\n if args.clip_vloss:\n v_loss_unclipped = (newvalue - b_returns[mb_inds]) ** 2\n v_clipped = b_values[mb_inds] + torch.clamp(\n newvalue - b_values[mb_inds],\n -args.clip_coef,\n args.clip_coef,\n )\n v_loss_clipped = (v_clipped - b_returns[mb_inds]) ** 2\n v_loss_max = torch.max(v_loss_unclipped, v_loss_clipped)\n v_loss = 0.5 * v_loss_max.mean()\n else:\n v_loss = 0.5 * ((newvalue - b_returns[mb_inds]) ** 2).mean()\n\n entropy_loss = entropy.mean()\n loss = pg_loss - args.ent_coef * entropy_loss + v_loss * args.vf_coef\n\n optimizer.zero_grad()\n loss.backward()\n nn.utils.clip_grad_norm_(agent.parameters(), args.max_grad_norm)\n optimizer.step()\n\n if args.target_kl is not None and approx_kl > args.target_kl:\n break\n\n y_pred, y_true = b_values.cpu().numpy(), b_returns.cpu().numpy()\n var_y = np.var(y_true)\n explained_var = np.nan if var_y == 0 else 1 - np.var(y_true - y_pred) / var_y\n\n # TRY NOT TO MODIFY: record rewards for plotting purposes\n writer.add_scalar(""charts/learning_rate"", optimizer.param_groups[0][""lr""], global_step)\n writer.add_scalar(""losses/value_loss"", v_loss.item(), global_step)\n writer.add_scalar(""losses/policy_loss"", pg_loss.item(), global_step)\n writer.add_scalar(""losses/entropy"", entropy_loss.item(), global_step)\n writer.add_scalar(""losses/old_approx_kl"", old_approx_kl.item(), global_step)\n writer.add_scalar(""losses/approx_kl"", approx_kl.item(), global_step)\n writer.add_scalar(""losses/clipfrac"", np.mean(clipfracs), global_step)\n writer.add_scalar(""losses/explained_variance"", explained_var, global_step)\n print(""SPS:"", int(global_step / (time.time() - start_time)))\n writer.add_scalar(""charts/SPS"", int(global_step / (time.time() - start_time)), global_step)\n\n envs.close()\n writer.close()\n",python,tab
3
+ 2,142,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"9:25:04 AM [info] Activating crowd-code\n9:25:04 AM [info] Recording started\n9:25:04 AM [info] Initializing git provider using file system watchers...\n",Log,tab
4
+ 3,155,"extension-output-pdoom-org.crowd-code-#1-crowd-code",40,0,"",Log,selection_command
5
+ 4,197,"extension-output-pdoom-org.crowd-code-#1-crowd-code",150,0,"9:25:04 AM [info] Git repository found\n9:25:04 AM [info] Git provider initialized successfully\n9:25:04 AM [info] Initial git state: [object Object]\n",Log,content
6
+ 5,2189,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
7
+ 6,2191,"TERMINAL",0,0,"",,terminal_focus
8
+ 7,2213,"TERMINAL",0,0,"",,terminal_command
9
+ 8,9325,"TERMINAL",0,0,"",,terminal_command
10
+ 9,18113,"TERMINAL",0,0,"source .venv/bin/activate",,terminal_command
11
+ 10,24112,"TERMINAL",0,0,"python cleanrl/ppo_atari_envpool.py",,terminal_command
12
+ 11,24163,"TERMINAL",0,0,"]633;C",,terminal_output
13
+ 12,26014,"TERMINAL",0,0,"/fast/home/franz.srambical/cleanrl/.venv/lib/python3.10/site-packages/treevalue/tree/integration/torch.py:23: FutureWarning: `torch.utils._pytree._register_pytree_node` is deprecated. Please use `torch.utils._pytree.register_pytree_node` instead.\r\n register_for_torch(TreeValue)\r\n/fast/home/franz.srambical/cleanrl/.venv/lib/python3.10/site-packages/treevalue/tree/integration/torch.py:24: FutureWarning: `torch.utils._pytree._register_pytree_node` is deprecated. Please use `torch.utils._pytree.register_pytree_node` instead.\r\n register_for_torch(FastTreeValue)\r\n",,terminal_output
14
+ 13,26264,"TERMINAL",0,0,"Gym has been unmaintained since 2022 and does not support NumPy 2.0 amongst other critical functionality.\r\nPlease upgrade to Gymnasium, the maintained drop-in replacement of Gym, or contact the authors of your software and request that they upgrade.\r\nSee the migration guide at https://gymnasium.farama.org/introduction/migration_guide/ for additional information.\r\n",,terminal_output
15
+ 14,26759,"TERMINAL",0,0,"/fast/home/franz.srambical/cleanrl/.venv/lib/python3.10/site-packages/tyro/_parsers.py:379: UserWarning: The field `wandb-entity` is annotated with type `<class 'str'>`, but the default value `None` has type `<class 'NoneType'>`. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/fast/home/franz.srambical/cleanrl/.venv/lib/python3.10/site-packages/tyro/_parsers.py:379: UserWarning: The field `target-kl` is annotated with type `<class 'float'>`, but the default value `None` has type `<class 'NoneType'>`. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output
16
+ 15,28466,"TERMINAL",0,0,"global_step=992, episodic_return=0.0\r\nglobal_step=992, episodic_return=0.0\r\nglobal_step=992, episodic_return=0.0\r\nglobal_step=992, episodic_return=0.0\r\nglobal_step=992, episodic_return=0.0\r\nglobal_step=992, episodic_return=0.0\r\nglobal_step=992, episodic_return=0.0\r\nglobal_step=992, episodic_return=0.0\r\n",,terminal_output
17
+ 16,28939,"TERMINAL",0,0,"SPS: 1339\r\n",,terminal_output
18
+ 17,29746,"TERMINAL",0,0,"SPS: 1302\r\n",,terminal_output
19
+ 18,29996,"TERMINAL",0,0,"global_step=2776, episodic_return=2.0\r\nglobal_step=2776, episodic_return=2.0\r\nglobal_step=2776, episodic_return=2.0\r\nglobal_step=2776, episodic_return=2.0\r\nglobal_step=2776, episodic_return=2.0\r\nglobal_step=2776, episodic_return=2.0\r\nglobal_step=2776, episodic_return=2.0\r\nglobal_step=2776, episodic_return=2.0\r\n",,terminal_output
20
+ 19,30574,"TERMINAL",0,0,"SPS: 1280\r\n",,terminal_output
21
+ 20,30805,"TERMINAL",0,0,"global_step=3776, episodic_return=0.0\r\nglobal_step=3776, episodic_return=0.0\r\nglobal_step=3776, episodic_return=0.0\r\nglobal_step=3776, episodic_return=0.0\r\nglobal_step=3776, episodic_return=0.0\r\nglobal_step=3776, episodic_return=0.0\r\nglobal_step=3776, episodic_return=0.0\r\nglobal_step=3776, episodic_return=0.0\r\n",,terminal_output
22
+ 21,31406,"TERMINAL",0,0,"SPS: 1266\r\n",,terminal_output
23
+ 22,31746,"TERMINAL",0,0,"global_step=5104, episodic_return=1.0\r\nglobal_step=5104, episodic_return=1.0\r\nglobal_step=5104, episodic_return=1.0\r\nglobal_step=5104, episodic_return=1.0\r\nglobal_step=5104, episodic_return=1.0\r\nglobal_step=5104, episodic_return=1.0\r\nglobal_step=5104, episodic_return=1.0\r\nglobal_step=5104, episodic_return=1.0\r\n",,terminal_output
24
+ 23,32262,"TERMINAL",0,0,"SPS: 1252\r\n",,terminal_output
25
+ 24,33098,"TERMINAL",0,0,"SPS: 1247\r\n",,terminal_output
26
+ 25,33217,"TERMINAL",0,0,"global_step=6496, episodic_return=1.0\r\nglobal_step=6496, episodic_return=1.0\r\nglobal_step=6496, episodic_return=1.0\r\nglobal_step=6496, episodic_return=1.0\r\nglobal_step=6496, episodic_return=1.0\r\nglobal_step=6496, episodic_return=1.0\r\nglobal_step=6496, episodic_return=1.0\r\nglobal_step=6496, episodic_return=1.0\r\n",,terminal_output
27
+ 26,33964,"TERMINAL",0,0,"SPS: 1237\r\n",,terminal_output
28
+ 27,34109,"TERMINAL",0,0,"^CTraceback (most recent call last):\r\n File ""/fast/home/franz.srambical/cleanrl/cleanrl/ppo_atari_envpool.py"", line 237, in <module>\r\n next_obs, reward, next_done, info = envs.step(action.cpu().numpy())\r\n File ""/fast/home/franz.srambical/cleanrl/cleanrl/ppo_atari_envpool.py"", line 100, in step\r\n observations, rewards, dones, infos = super().step(action)\r\n File ""/fast/home/franz.srambical/cleanrl/.venv/lib/python3.10/site-packages/gym/core.py"", line 280, in step\r\n return self.env.step(action)\r\n File ""/fast/home/franz.srambical/cleanrl/.venv/lib/python3.10/site-packages/envpool/python/envpool.py"", line 144, in step\r\n return self.recv(reset=False, return_info=True)\r\n File ""/fast/home/franz.srambical/cleanrl/.venv/lib/python3.10/site-packages/envpool/python/envpool.py"", line 130, in recv\r\n state_list = self._recv()\r\nKeyboardInterrupt\r\n",,terminal_output
29
+ 28,34254,"TERMINAL",0,0,"^C",,terminal_output
30
+ 29,34258,"TERMINAL",0,0,"\r\n]0;franz.srambical@hai-login2:~/cleanrl",,terminal_output
31
+ 30,116742,"TERMINAL",0,0,"",,terminal_command
32
+ 31,278446,"cleanrl/ppo_atari_envpool.py",236,0,"",python,selection_mouse
33
+ 32,278448,"cleanrl/ppo_atari_envpool.py",235,0,"",python,selection_command
34
+ 33,290061,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"""""""\nGenerates a dataset of random-action CoinRun episodes.\nEpisodes are saved individually as memory-mapped files for efficient loading.\n""""""\n\nfrom dataclasses import dataclass\n\nfrom gym3 import types_np\nimport numpy as np\nfrom procgen import ProcgenGym3Env\nimport tyro\nimport json\nimport os\nfrom utils import save_chunks\n\n\n@dataclass\nclass Args:\n num_episodes_train: int = 10000\n num_episodes_val: int = 500\n num_episodes_test: int = 500\n output_dir: str = ""data/coinrun_episodes""\n min_episode_length: int = 1000\n max_episode_length: int = 1000\n chunk_size: int = 100\n chunks_per_file: int = 100\n seed: int = 0\n\n\nargs = tyro.cli(Args)\nassert (\n args.max_episode_length >= args.min_episode_length\n), ""Maximum episode length must be greater than or equal to minimum episode length.""\n\nif args.min_episode_length < args.chunk_size:\n print(\n ""Warning: Minimum episode length is smaller than chunk size. Note that episodes shorter than the chunk size will be discarded.""\n )\n\n\n# --- Generate episodes ---\ndef generate_episodes(num_episodes, split):\n episode_idx = 0\n episode_metadata = []\n obs_chunks = []\n act_chunks = []\n file_idx = 0\n output_dir_split = os.path.join(args.output_dir, split)\n while episode_idx < num_episodes:\n seed = np.random.randint(0, 10000)\n env = ProcgenGym3Env(num=1, env_name=""coinrun"", start_level=seed)\n\n observations_seq = []\n actions_seq = []\n episode_obs_chunks = []\n episode_act_chunks = []\n\n # --- Run episode ---\n step_t = 0\n for step_t in range(args.max_episode_length):\n action = types_np.sample(env.ac_space, bshape=(env.num,))\n env.act(action)\n _, obs, first = env.observe()\n observations_seq.append(obs[""rgb""])\n actions_seq.append(action)\n if len(observations_seq) == args.chunk_size:\n episode_obs_chunks.append(observations_seq)\n episode_act_chunks.append(actions_seq)\n observations_seq = []\n actions_seq = []\n if first:\n break\n\n # --- Save episode ---\n if step_t + 1 >= args.min_episode_length:\n if observations_seq:\n if len(observations_seq) < args.chunk_size:\n print(\n f""Warning: Inconsistent chunk_sizes. Episode has {len(observations_seq)} frames, ""\n f""which is smaller than the requested chunk_size: {args.chunk_size}. ""\n ""This might lead to performance degradation during training.""\n )\n episode_obs_chunks.append(observations_seq)\n episode_act_chunks.append(actions_seq)\n\n obs_chunks_data = [\n np.concatenate(seq, axis=0).astype(np.uint8)\n for seq in episode_obs_chunks\n ]\n act_chunks_data = [\n np.concatenate(act, axis=0) for act in episode_act_chunks\n ]\n obs_chunks.extend(obs_chunks_data)\n act_chunks.extend(act_chunks_data)\n\n ep_metadata, obs_chunks, file_idx, act_chunks = save_chunks(\n obs_chunks, file_idx, args.chunks_per_file, output_dir_split, act_chunks\n )\n episode_metadata.extend(ep_metadata)\n\n print(f""Episode {episode_idx} completed, length: {step_t + 1}."")\n episode_idx += 1\n else:\n print(f""Episode too short ({step_t + 1}), resampling..."")\n\n if len(obs_chunks) > 0:\n print(\n f""Warning: Dropping {len(obs_chunks)} chunks for consistent number of chunks per file."",\n ""Consider changing the chunk_size and chunks_per_file parameters to prevent data-loss."",\n )\n\n print(f""Done generating {split} split"")\n return episode_metadata\n\n\ndef get_action_space():\n env = ProcgenGym3Env(num=1, env_name=""coinrun"", start_level=0)\n return env.ac_space.eltype.n\n\n\ndef main():\n # Set random seed and create dataset directories\n np.random.seed(args.seed)\n # --- Generate episodes ---\n train_episode_metadata = generate_episodes(args.num_episodes_train, ""train"")\n val_episode_metadata = generate_episodes(args.num_episodes_val, ""val"")\n test_episode_metadata = generate_episodes(args.num_episodes_test, ""test"")\n\n # --- Save metadata ---\n metadata = {\n ""env"": ""coinrun"",\n ""num_actions"": get_action_space(),\n ""num_episodes_train"": args.num_episodes_train,\n ""num_episodes_val"": args.num_episodes_val,\n ""num_episodes_test"": args.num_episodes_test,\n ""avg_episode_len_train"": np.mean(\n [ep[""avg_seq_len""] for ep in train_episode_metadata]\n ),\n ""avg_episode_len_val"": np.mean(\n [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")\n\n\nif __name__ == ""__main__"":\n main()\n",python,tab
35
+ 34,299089,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
36
+ 35,300383,"cleanrl/ppo_atari_envpool.py",318,0,"",python,selection_mouse
37
+ 36,300384,"cleanrl/ppo_atari_envpool.py",317,0,"",python,selection_command
38
+ 37,314316,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"",python,tab
39
+ 38,315490,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
40
+ 39,316028,"cleanrl/ppo_atari_envpool.py",330,0,"",python,selection_mouse
41
+ 40,316031,"cleanrl/ppo_atari_envpool.py",329,0,"",python,selection_command
42
+ 41,316695,"cleanrl/ppo_atari_envpool.py",436,0,"",python,selection_mouse
43
+ 42,316698,"cleanrl/ppo_atari_envpool.py",435,0,"",python,selection_command
44
+ 43,317177,"cleanrl/ppo_atari_envpool.py",438,0,"",python,selection_mouse
45
+ 44,397100,"cleanrl/ppo_atari_envpool.py",2542,0,"",python,selection_mouse
46
+ 45,402425,"cleanrl/ppo_atari_envpool.py",4160,0,"",python,selection_mouse
47
+ 46,541375,"TERMINAL",0,0,"",,terminal_focus
48
+ 47,541842,"TERMINAL",0,0,"source /home/franz.srambical/cleanrl/.venv/bin/activate",,terminal_command
49
+ 48,544327,"TERMINAL",0,0,"cd ../jafar/",,terminal_command
50
+ 49,554362,"TERMINAL",0,0,"git checkout gt-actions",,terminal_command
51
+ 50,554410,"TERMINAL",0,0,"]633;C",,terminal_output
52
+ 51,554473,"TERMINAL",0,0,"M\t.gitignore\r\nAlready on 'gt-actions'\r\nYour branch is behind 'origin/gt-actions' by 1 commit, and can be fast-forwarded.\r\n (use ""git pull"" to update your local branch)\r\n]0;franz.srambical@hai-login2:~/jafar",,terminal_output
53
+ 52,555777,"TERMINAL",0,0,"git pull",,terminal_command
54
+ 53,555825,"TERMINAL",0,0,"]633;C",,terminal_output
55
+ 54,557974,"TERMINAL",0,0,"Updating 96d560e..1b6b878\r\nFast-forward\r\n input_pipeline/generate_coinrun_dataset.py | 2 +-\r\n 1 file changed, 1 insertion(+), 1 deletion(-)\r\n]0;franz.srambical@hai-login2:~/jafar",,terminal_output
56
+ 55,560140,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"",python,tab
57
+ 56,560211,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",563,26," chunk_size: int = 160\n",python,content
58
+ 57,561853,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
59
+ 58,570185,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"",python,tab
60
+ 59,571693,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
61
+ 60,576013,"cleanrl/ppo_atari_envpool.py",4335,0,"",python,selection_mouse
62
+ 61,725152,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"",python,tab
63
+ 62,725643,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",280,0,"",python,selection_mouse
64
+ 63,725644,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",279,0,"",python,selection_command
65
+ 64,727140,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",1442,0,"",python,selection_command
66
+ 65,728476,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
67
+ 66,782305,"cleanrl/trajectory_saver.py",0,0,"import os\nimport json\nfrom typing import List, Dict, Any\n\nimport numpy as np\n\n\nclass TrajectorySaver:\n """"""\n Collects per-environment observation frames and actions, chunks them,\n and writes chunked files to disk. Designed to be lightweight and only\n enabled when trajectory capture is requested.\n\n Behavior mirrors the high level flow in `generate_coinrun_dataset.py`:\n - Build sequences of length `chunk_size` from steps\n - At episode end, if a partial sequence exists (< chunk_size), include it\n with a warning about inconsistent chunk sizes\n - Group `chunks_per_file` chunks per output file. Any remainder at the\n very end is dropped with a warning\n - Store both observation chunks and action chunks side-by-side\n """"""\n\n def __init__(\n self,\n output_dir: str,\n num_envs: int,\n chunk_size: int = 160,\n chunks_per_file: int = 100,\n ) -> None:\n self.output_dir = output_dir\n self.chunks_dir = os.path.join(output_dir, ""chunks"")\n os.makedirs(self.chunks_dir, exist_ok=True)\n self.num_envs = num_envs\n self.chunk_size = int(chunk_size)\n self.chunks_per_file = int(chunks_per_file)\n\n # Per-env rolling buffers for the current episode\n self.env_buffers: List[Dict[str, Any]] = [\n {""obs_seq"": [], ""act_seq"": [], ""episode_obs_chunks"": [], ""episode_act_chunks"": []}\n for _ in range(self.num_envs)\n ]\n\n # Global chunk buffers waiting to be flushed to disk\n self.obs_chunks: List[np.ndarray] = []\n self.act_chunks: List[np.ndarray] = []\n self.file_idx: int = 0\n\n # Metadata (optional). We keep a minimal JSONL with per-file info\n self.metadata_path = os.path.join(self.output_dir, ""metadata.jsonl"")\n\n def _finalize_episode_for_env(self, env_idx: int) -> None:\n buf = self.env_buffers[env_idx]\n obs_seq: List[np.ndarray] = buf[""obs_seq""]\n act_seq: List[np.ndarray] = buf[""act_seq""]\n\n if len(obs_seq) > 0:\n # Partial chunk at episode end\n if len(obs_seq) < self.chunk_size:\n print(\n f""Warning: Inconsistent chunk sizes. Env {env_idx} episode ended with ""\n f""{len(obs_seq)} frames (< chunk_size={self.chunk_size}). ""\n ""Including the partial chunk may impact training I/O performance.""\n )\n self._append_chunk_from_sequences(env_idx)\n\n # Move episode chunks to global buffers\n if len(buf[""episode_obs_chunks""]) > 0:\n self.obs_chunks.extend(buf[""episode_obs_chunks""])\n self.act_chunks.extend(buf[""episode_act_chunks""])\n buf[""episode_obs_chunks""].clear()\n buf[""episode_act_chunks""].clear()\n\n self._save_if_ready()\n\n def _append_chunk_from_sequences(self, env_idx: int) -> None:\n buf = self.env_buffers[env_idx]\n obs_seq: List[np.ndarray] = buf[""obs_seq""]\n act_seq: List[np.ndarray] = buf[""act_seq""]\n\n if len(obs_seq) == 0:\n return\n\n # Stack along time dimension. Shapes become:\n # obs_chunk: (T, C, H, W) dtype=uint8\n # act_chunk: (T,) dtype=int64 (or int)\n obs_chunk = np.stack(obs_seq, axis=0).astype(np.uint8)\n act_chunk = np.asarray(act_seq, dtype=np.int64)\n\n buf[""episode_obs_chunks""].append(obs_chunk)\n buf[""episode_act_chunks""].append(act_chunk)\n\n # Reset rolling sequences\n buf[""obs_seq""] = []\n buf[""act_seq""] = []\n\n def _save_if_ready(self) -> None:\n while len(self.obs_chunks) >= self.chunks_per_file:\n obs_to_write = self.obs_chunks[: self.chunks_per_file]\n act_to_write = self.act_chunks[: self.chunks_per_file]\n\n # Remove from buffers\n self.obs_chunks = self.obs_chunks[self.chunks_per_file :]\n self.act_chunks = self.act_chunks[self.chunks_per_file :]\n\n file_stem = f""chunks_{self.file_idx:06d}""\n out_path = os.path.join(self.chunks_dir, f""{file_stem}.npz"")\n\n # Write as a .npz with numbered arrays to preserve variable chunk lengths\n save_dict: Dict[str, Any] = {}\n for idx, (o, a) in enumerate(zip(obs_to_write, act_to_write)):\n save_dict[f""obs_{idx:03d}""] = o\n save_dict[f""acts_{idx:03d}""] = a\n\n np.savez_compressed(out_path, **save_dict)\n\n # Minimal metadata per file for bookkeeping\n meta = {\n ""file"": os.path.basename(out_path),\n ""num_chunks"": len(obs_to_write),\n ""chunk_size"": self.chunk_size,\n ""chunk_indices"": [i for i in range(len(obs_to_write))],\n }\n with open(self.metadata_path, ""a"") as f:\n f.write(json.dumps(meta) + ""\n"")\n\n self.file_idx += 1\n\n def add_step(\n self,\n obs_batch: np.ndarray,\n actions_batch: np.ndarray,\n episode_done_batch: np.ndarray,\n ) -> None:\n """"""\n Add a single environment step across all envs.\n\n Args:\n obs_batch: np.ndarray with shape (num_envs, C, H, W). Values should be [0, 255].\n actions_batch: np.ndarray with shape (num_envs,).\n episode_done_batch: np.ndarray with shape (num_envs,), boolean flags indicating end-of-episode.\n """"""\n assert obs_batch.shape[0] == self.num_envs\n assert actions_batch.shape[0] == self.num_envs\n assert episode_done_batch.shape[0] == self.num_envs\n\n for env_idx in range(self.num_envs):\n buf = self.env_buffers[env_idx]\n buf[""obs_seq""].append(obs_batch[env_idx])\n buf[""act_seq""].append(int(actions_batch[env_idx]))\n\n if len(buf[""obs_seq""]) == self.chunk_size:\n self._append_chunk_from_sequences(env_idx)\n\n # Episode ended: finalize the episode\n if bool(episode_done_batch[env_idx]):\n self._finalize_episode_for_env(env_idx)\n\n def close(self) -> None:\n # Drop remainder to keep consistent chunks-per-file contract\n remainder = len(self.obs_chunks)\n if remainder > 0:\n print(\n f""Warning: Dropping {remainder} chunks to keep a consistent number of chunks per file. ""\n ""Consider adjusting chunk_size and chunks_per_file to reduce data loss.""\n )\n # Nothing else to do; per-episode buffers do not carry data unless an episode just ended\n\n\n",python,tab
68
+ 67,859278,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
69
+ 68,877159,"cleanrl/ppo_atari_envpool.py",14147,0," if traj_saver is not None:\n traj_saver.close()\n",python,content
70
+ 69,877159,"cleanrl/ppo_atari_envpool.py",8882,0," if traj_saver is not None:\n # Use true termination flags when available\n terminated = np.asarray(info.get(""terminated"", next_done.detach().cpu().numpy()), dtype=bool)\n traj_saver.add_step(obs_to_save, actions_to_save, terminated)\n\n",python,content
71
+ 70,877159,"cleanrl/ppo_atari_envpool.py",8628,0," # Capture current observation and action prior to stepping the envs\n if traj_saver is not None:\n obs_to_save = next_obs.detach().cpu().numpy().astype(np.uint8)\n actions_to_save = action.detach().cpu().numpy()\n",python,content
72
+ 71,877159,"cleanrl/ppo_atari_envpool.py",6967,0," # Optional: trajectory saver\n traj_saver = None\n if args.capture_trajectories:\n traj_saver = TrajectorySaver(\n output_dir=args.trajectories_output_dir,\n num_envs=args.num_envs,\n chunk_size=args.trajectories_chunk_size,\n chunks_per_file=args.trajectories_chunks_per_file,\n )\n\n",python,content
73
+ 72,877159,"cleanrl/ppo_atari_envpool.py",1167,0," # Trajectory capture\n capture_trajectories: bool = False\n """"""if toggled, save observation frames and actions during training""""""\n trajectories_output_dir: str = ""data/atari_trajectories""\n """"""directory to store trajectory chunks (.npz)""""""\n trajectories_chunk_size: int = 160\n """"""number of time steps per chunk before writing""""""\n trajectories_chunks_per_file: int = 100\n """"""number of chunks grouped per npz file""""""\n\n",python,content
74
+ 73,877160,"cleanrl/ppo_atari_envpool.py",437,0,"from .trajectory_saver import TrajectorySaver\n",python,content
75
+ 74,913280,"cleanrl/ppo_atari_envpool.py",15527,58,"",python,content
76
+ 75,913280,"cleanrl/ppo_atari_envpool.py",9974,288,"",python,content
77
+ 76,913280,"cleanrl/ppo_atari_envpool.py",9458,262,"",python,content
78
+ 77,913280,"cleanrl/ppo_atari_envpool.py",7454,343,"",python,content
79
+ 78,913280,"cleanrl/ppo_atari_envpool.py",1213,441,"",python,content
80
+ 79,913280,"cleanrl/ppo_atari_envpool.py",437,46,"",python,content
81
+ 80,913354,"cleanrl/ppo_atari_envpool.py",14147,0," if traj_saver is not None:\n traj_saver.close()\n",python,content
82
+ 81,913354,"cleanrl/ppo_atari_envpool.py",10671,0," # Initialize for type-checkers\n approx_kl = torch.tensor(0.0)\n old_approx_kl = torch.tensor(0.0)\n v_loss = torch.tensor(0.0)\n pg_loss = torch.tensor(0.0)\n entropy_loss = torch.tensor(0.0)\n",python,content
83
+ 82,913354,"cleanrl/ppo_atari_envpool.py",8882,0," if traj_saver is not None and obs_to_save is not None and actions_to_save is not None:\n # Use true termination flags when available\n terminated = np.asarray(info.get(""terminated"", next_done.detach().cpu().numpy()), dtype=bool)\n traj_saver.add_step(obs_to_save, actions_to_save, terminated)\n\n",python,content
84
+ 83,913354,"cleanrl/ppo_atari_envpool.py",8628,0," # Capture current observation and action prior to stepping the envs\n obs_to_save = None\n actions_to_save = None\n if traj_saver is not None:\n obs_to_save = next_obs.detach().cpu().numpy().astype(np.uint8)\n actions_to_save = action.detach().cpu().numpy()\n",python,content
85
+ 84,913354,"cleanrl/ppo_atari_envpool.py",6861,104," assert isinstance(envs.action_space, spaces.Discrete), ""only discrete action space is supported""\n\n # Optional: trajectory saver\n traj_saver = None\n if args.capture_trajectories:\n traj_saver = TrajectorySaver(\n output_dir=args.trajectories_output_dir,\n num_envs=args.num_envs,\n chunk_size=args.trajectories_chunk_size,\n chunks_per_file=args.trajectories_chunks_per_file,\n )",python,content
86
+ 85,913354,"cleanrl/ppo_atari_envpool.py",2469,27," target_kl: Optional[float] = None",python,content
87
+ 86,913355,"cleanrl/ppo_atari_envpool.py",1167,0," # Trajectory capture\n capture_trajectories: bool = False\n """"""if toggled, save observation frames and actions during training""""""\n trajectories_output_dir: str = ""data/atari_trajectories""\n """"""directory to store trajectory chunks (.npz)""""""\n trajectories_chunk_size: int = 160\n """"""number of time steps per chunk before writing""""""\n trajectories_chunks_per_file: int = 100\n """"""number of chunks grouped per npz file""""""\n\n",python,content
88
+ 87,913355,"cleanrl/ppo_atari_envpool.py",968,28," wandb_entity: Optional[str] = None",python,content
89
+ 88,913355,"cleanrl/ppo_atari_envpool.py",387,49,"from torch.utils.tensorboard.writer import SummaryWriter\nfrom typing import Optional\nfrom gym import spaces\nfrom .trajectory_saver import TrajectorySaver",python,content
90
+ 89,961375,"cleanrl/ppo_atari_envpool.py",15958,58,"",python,content
91
+ 90,961375,"cleanrl/ppo_atari_envpool.py",12251,231,"",python,content
92
+ 91,961375,"cleanrl/ppo_atari_envpool.py",10114,348,"",python,content
93
+ 92,961375,"cleanrl/ppo_atari_envpool.py",9532,328,"",python,content
94
+ 93,961375,"cleanrl/ppo_atari_envpool.py",7426,443," assert isinstance(envs.action_space, gym.spaces.Discrete), ""only discrete action space is supported""",python,content
95
+ 94,961375,"cleanrl/ppo_atari_envpool.py",3024,37," target_kl: float = None",python,content
96
+ 95,961375,"cleanrl/ppo_atari_envpool.py",1281,441,"",python,content
97
+ 96,961375,"cleanrl/ppo_atari_envpool.py",1072,38," wandb_entity: str = None",python,content
98
+ 97,961375,"cleanrl/ppo_atari_envpool.py",387,153,"from torch.utils.tensorboard import SummaryWriter",python,content
99
+ 98,961438,"cleanrl/ppo_atari_envpool.py",14147,0," if traj_saver is not None:\n traj_saver.close()\n",python,content
100
+ 99,961438,"cleanrl/ppo_atari_envpool.py",10671,0," # Initialize for type-checkers\n approx_kl = torch.tensor(0.0)\n old_approx_kl = torch.tensor(0.0)\n v_loss = torch.tensor(0.0)\n pg_loss = torch.tensor(0.0)\n entropy_loss = torch.tensor(0.0)\n",python,content
101
+ 100,961438,"cleanrl/ppo_atari_envpool.py",8882,0," if traj_saver is not None and obs_to_save is not None and actions_to_save is not None:\n # Use true termination flags when available\n terminated = np.asarray(info.get(""terminated"", next_done.detach().cpu().numpy()), dtype=bool)\n traj_saver.add_step(obs_to_save, actions_to_save, terminated)\n\n",python,content
102
+ 101,961438,"cleanrl/ppo_atari_envpool.py",8628,0," # Capture current observation and action prior to stepping the envs\n obs_to_save = None\n actions_to_save = None\n if traj_saver is not None:\n obs_to_save = next_obs.detach().cpu().numpy().astype(np.uint8)\n actions_to_save = action.detach().cpu().numpy()\n",python,content
103
+ 102,961438,"cleanrl/ppo_atari_envpool.py",7002,79," optimizer = Adam(agent.parameters(), lr=args.learning_rate, eps=1e-5)",python,content
104
+ 103,961438,"cleanrl/ppo_atari_envpool.py",6861,104," assert isinstance(envs.action_space, spaces.Discrete), ""only discrete action space is supported""\n\n # Optional: trajectory saver\n traj_saver = None\n if args.capture_trajectories:\n traj_saver = TrajectorySaver(\n output_dir=args.trajectories_output_dir,\n num_envs=args.num_envs,\n chunk_size=args.trajectories_chunk_size,\n chunks_per_file=args.trajectories_chunks_per_file,\n )",python,content
105
+ 104,961438,"cleanrl/ppo_atari_envpool.py",2982,71," self.episode_returns = np.zeros(self.num_envs, dtype=np.float32)\n self.episode_lengths = np.zeros(self.num_envs, dtype=np.int32)",python,content
106
+ 105,961438,"cleanrl/ppo_atari_envpool.py",2469,27," target_kl: Optional[float] = None",python,content
107
+ 106,961438,"cleanrl/ppo_atari_envpool.py",1167,0," # Trajectory capture\n capture_trajectories: bool = False\n """"""if toggled, save observation frames and actions during training""""""\n trajectories_output_dir: str = ""data/atari_trajectories""\n """"""directory to store trajectory chunks (.npz)""""""\n trajectories_chunk_size: int = 160\n """"""number of time steps per chunk before writing""""""\n trajectories_chunks_per_file: int = 100\n """"""number of chunks grouped per npz file""""""\n\n",python,content
108
+ 107,961438,"cleanrl/ppo_atari_envpool.py",968,28," wandb_entity: Optional[str] = None",python,content
109
+ 108,961438,"cleanrl/ppo_atari_envpool.py",387,49,"from torch.utils.tensorboard.writer import SummaryWriter\nfrom typing import Optional\nfrom gym import spaces\nfrom .trajectory_saver import TrajectorySaver",python,content
110
+ 109,961438,"cleanrl/ppo_atari_envpool.py",291,27,"from torch.optim import Adam",python,content
111
+ 110,968028,"cleanrl/ppo_atari_envpool.py",268,0,"",python,selection_mouse
112
+ 111,968030,"cleanrl/ppo_atari_envpool.py",267,0,"",python,selection_command
113
+ 112,969822,"cleanrl/ppo_atari_envpool.py",280,0,"",python,selection_command
114
+ 113,974152,"cleanrl/ppo_atari_envpool.py",291,28,"import torch.optim as optim",python,content
115
+ 114,977518,"cleanrl/ppo_atari_envpool.py",7994,0,"",python,selection_command
116
+ 115,978956,"cleanrl/ppo_atari_envpool.py",7978,73," optimizer = optim.Adam(agent.parameters(), lr=args.learning_rate, eps=1e-5)",python,content
117
+ 116,979756,"cleanrl/ppo_atari_envpool.py",8056,0,"",python,selection_command
118
+ 117,980869,"cleanrl/ppo_atari_envpool.py",0,0,"",python,selection_command
119
+ 118,1015212,"cleanrl/ppo_atari_envpool.py",12299,0,"",python,selection_mouse
120
+ 119,1015215,"cleanrl/ppo_atari_envpool.py",12298,0,"",python,selection_command
121
+ 120,1016996,"cleanrl/ppo_atari_envpool.py",12323,231,"",python,content
122
+ 121,1021028,"cleanrl/ppo_atari_envpool.py",12205,0,"",python,selection_mouse
123
+ 122,1027883,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"",python,tab
124
+ 123,1037912,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",3201,0,"",python,selection_mouse
125
+ 124,1043895,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"",python,selection_command
126
+ 125,1063321,"/home/franz.srambical/jafar/input_pipeline/utils.py",0,0,"import os\nimport pickle\nimport numpy as np\nfrom array_record.python.array_record_module import ArrayRecordWriter\n\n\ndef save_chunks(obs_chunks, file_idx, chunks_per_file, output_dir, act_chunks=None):\n os.makedirs(output_dir, exist_ok=True)\n\n metadata = []\n while len(obs_chunks) >= chunks_per_file:\n chunk_batch = obs_chunks[:chunks_per_file]\n obs_chunks = obs_chunks[chunks_per_file:]\n act_chunk_batch = None\n if act_chunks:\n act_chunk_batch = act_chunks[:chunks_per_file]\n act_chunks = act_chunks[chunks_per_file:]\n episode_path = os.path.join(output_dir, f""data_{file_idx:04d}.array_record"")\n writer = ArrayRecordWriter(str(episode_path), ""group_size:1"")\n seq_lens = []\n for idx, chunk in enumerate(chunk_batch):\n seq_len = chunk.shape[0]\n seq_lens.append(seq_len)\n chunk_record = {\n ""raw_video"": chunk.tobytes(),\n ""sequence_length"": seq_len,\n }\n if act_chunk_batch:\n assert len(chunk) == len(\n act_chunk_batch[idx]\n ), f""Observation data length and action sequence length do not match: {len(chunk)} != {len(act_chunk_batch[idx])}""\n chunk_record[""actions""] = act_chunk_batch[idx]\n writer.write(pickle.dumps(chunk_record))\n writer.close()\n file_idx += 1\n metadata.append(\n {\n ""path"": episode_path,\n ""num_chunks"": len(chunk_batch),\n ""avg_seq_len"": np.mean(seq_lens),\n }\n )\n print(f""Created {episode_path} with {len(chunk_batch)} video chunks"")\n\n return metadata, obs_chunks, file_idx, act_chunks\n",python,tab
127
+ 126,1077626,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"",python,tab
128
+ 127,1078649,"/home/franz.srambical/jafar/input_pipeline/utils.py",0,0,"",python,tab
129
+ 128,1084681,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"",python,tab
130
+ 129,1098108,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
131
+ 130,1107176,"cleanrl/ppo_atari_envpool.py",716,0,"",python,selection_mouse
132
+ 131,1107177,"cleanrl/ppo_atari_envpool.py",715,0,"",python,selection_command
133
+ 132,1163039,"cleanrl/ppo_atari_envpool.py",553,0,"",python,selection_mouse
134
+ 133,1163040,"cleanrl/ppo_atari_envpool.py",552,0,"",python,selection_command
135
+ 134,1164422,"cleanrl/ppo_atari_envpool.py",542,0,"",python,selection_mouse
136
+ 135,1167929,"cleanrl/ppo_atari_envpool.py",565,0,"",python,selection_mouse
137
+ 136,1167931,"cleanrl/ppo_atari_envpool.py",564,0,"",python,selection_command
138
+ 137,1169182,"cleanrl/ppo_atari_envpool.py",387,0,"",python,selection_command
139
+ 138,1170321,"cleanrl/ppo_atari_envpool.py",1072,0,"",python,selection_command
140
+ 139,1170873,"cleanrl/ppo_atari_envpool.py",387,0,"",python,selection_command
141
+ 140,1172070,"cleanrl/ppo_atari_envpool.py",1072,0,"",python,selection_command
142
+ 141,1172368,"cleanrl/ppo_atari_envpool.py",387,0,"",python,selection_command
143
+ 142,1172644,"cleanrl/ppo_atari_envpool.py",1072,0,"",python,selection_command
144
+ 143,4767412,"cleanrl/ppo_atari_envpool.py",840,0,"",python,selection_mouse
145
+ 144,4767415,"cleanrl/ppo_atari_envpool.py",839,0,"",python,selection_command
146
+ 145,4769432,"cleanrl/trajectory_saver.py",0,0,"",python,tab
147
+ 146,4770255,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
148
+ 147,4770257,"cleanrl/ppo_atari_envpool.py",387,0,"",python,selection_command
149
+ 148,4770719,"cleanrl/trajectory_saver.py",0,0,"",python,tab
150
+ 149,4771118,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
151
+ 150,5039789,"cleanrl/ppo_atari_envpool.py",210,0,"",python,selection_mouse
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-d9cdf338-0ddd-4679-853a-6d7bdf2b18581751046137722-2025_06_27-10.42.19.354/source.csv ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,1,"utils/dataloader.py",0,0,"import functools\nimport jax\n\nimport tensorflow as tf\n\n# reserve GPU memory for JAX only if tensorflow is built with GPU support\ntf.config.experimental.set_visible_devices([], ""GPU"")\n\n\n# --- TensorFlow function for processing: slicing, normalization ---\ndef _tf_process_episode(episode_tensor, seq_len, image_h, image_w, image_c):\n """"""\n Processes a raw episode tensor in TensorFlow.\n Takes a full episode, extracts a random sequence, and normalizes it.\n Args:\n episode_tensor: A TensorFlow tensor representing a full video episode.\n Expected shape: (dynamic_length, image_h, image_w, image_c)\n Expected dtype: e.g., tf.uint8 (raw pixel values)\n seq_len: The desired length of the sub-sequence to extract.\n image_h: The height of each frame.\n image_w: The width of each frame.\n image_c: The number of channels in each frame.\n Returns:\n A TensorFlow tensor representing the processed video sequence.\n Shape: (seq_len, image_h, image_w, image_c)\n Dtype: tf.float32 (normalized pixel values)\n """"""\n current_episode_len = tf.shape(episode_tensor)[0]\n\n max_start_idx = current_episode_len - seq_len\n\n start_idx = tf.random.uniform(\n shape=(), minval=0, maxval=max_start_idx + 1, dtype=tf.int32\n )\n\n seq = episode_tensor[start_idx : start_idx + seq_len]\n\n seq = tf.cast(seq, tf.float32) / 255.0\n\n # Ensure the final shape is statically known for batching.\n # tf.reshape is robust, but tf.ensure_shape or set_shape can also be used if confident.\n processed_sequence = tf.reshape(seq, [seq_len, image_h, image_w, image_c])\n\n return processed_sequence\n\n\ndef _parse_tfrecord_fn(example_proto, image_h, image_w, image_c):\n feature_description = {\n ""height"": tf.io.FixedLenFeature([], tf.int64),\n ""width"": tf.io.FixedLenFeature([], tf.int64),\n ""channels"": tf.io.FixedLenFeature([], tf.int64),\n ""sequence_length"": tf.io.FixedLenFeature([], tf.int64),\n ""raw_video"": tf.io.FixedLenFeature([], tf.string),\n }\n example = tf.io.parse_single_example(example_proto, feature_description)\n\n video_shape = (example[""sequence_length""], image_h, image_w, image_c)\n\n episode_tensor = tf.io.decode_raw(example[""raw_video""], out_type=tf.uint8)\n episode_tensor = tf.reshape(episode_tensor, video_shape)\n\n episode_tensor = tf.ensure_shape(episode_tensor, [None, image_h, image_w, image_c])\n return episode_tensor\n\n\ndef get_dataloader(\n tfrecord_paths: list[str], # List of TFRecord file paths\n seq_len: int,\n global_batch_size: int,\n image_h: int,\n image_w: int,\n image_c: int,\n shuffle_buffer_size: int = 1000,\n num_parallel_calls: int = tf.data.AUTOTUNE,\n seed: int = 42,\n):\n """"""\n Creates a tf.data.Dataset pipeline from TFRecord files.\n """"""\n if not tfrecord_paths:\n raise ValueError(""tfrecord_paths list cannot be empty."")\n\n process_id = jax.process_index()\n num_processes = jax.process_count()\n\n assert (\n global_batch_size % num_processes == 0\n ), ""Global batch size {global_batch_size} \\n must be divisible by the number of JAX processes {num_processes} for proper sharding.""\n per_process_batch_size = global_batch_size // num_processes\n\n dataset = tf.data.TFRecordDataset(\n tfrecord_paths, num_parallel_reads=tf.data.AUTOTUNE\n )\n\n dataset = dataset.shard(num_shards=num_processes, index=process_id)\n\n # (f.srambical) NOTE: For TFRecords, it's often good to have a large shuffle buffer.\n if shuffle_buffer_size > 0:\n dataset = dataset.shuffle(\n buffer_size=shuffle_buffer_size, seed=seed, reshuffle_each_iteration=True\n )\n parse_fn = functools.partial(\n _parse_tfrecord_fn, image_h=image_h, image_w=image_w, image_c=image_c\n )\n dataset = dataset.map(parse_fn, num_parallel_calls=num_parallel_calls)\n\n tf_process_fn = functools.partial(\n _tf_process_episode,\n seq_len=seq_len,\n image_h=image_h,\n image_w=image_w,\n image_c=image_c,\n )\n dataset = dataset.map(tf_process_fn, num_parallel_calls=num_parallel_calls)\n\n dataset = dataset.repeat(None)\n dataset = dataset.batch(per_process_batch_size, drop_remainder=True)\n dataset = dataset.prefetch(tf.data.AUTOTUNE)\n\n return dataset.as_numpy_iterator()\n",python,tab
3
+ 2,38,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
4
+ 3,67,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"10:42:19 AM [info] Activating crowd-code\n10:42:19 AM [info] Recording started\n10:42:19 AM [info] Initializing git provider using file system watchers...\n10:42:19 AM [info] Git repository found\n10:42:19 AM [info] Git provider initialized successfully\n10:42:19 AM [info] Initial git state: [object Object]\n",Log,content
5
+ 4,905,"utils/dataloader.py",0,0,"",python,tab
6
+ 5,19039,"utils/dataloader.py",2752,0,"",python,selection_command
7
+ 6,19130,"utils/dataloader.py",2715,0,"",python,selection_command
8
+ 7,30009,"models/dynamics.py",0,0,"from typing import Dict, Any\n\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\n\nfrom utils.nn import STTransformer\n\n\nclass DynamicsMaskGIT(nn.Module):\n """"""MaskGIT dynamics model""""""\n\n model_dim: int\n num_latents: int\n num_blocks: int\n num_heads: int\n dropout: float\n mask_limit: float\n\n def setup(self):\n self.dynamics = STTransformer(\n self.model_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n )\n self.patch_embed = nn.Embed(self.num_latents, self.model_dim)\n self.mask_token = self.param(\n ""mask_token"",\n nn.initializers.lecun_uniform(),\n (1, 1, 1, self.model_dim),\n )\n self.action_up = nn.Dense(self.model_dim)\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n # --- Mask videos ---\n vid_embed = self.patch_embed(batch[""video_tokens""])\n if training:\n rng1, rng2 = jax.random.split(batch[""mask_rng""])\n mask_prob = jax.random.uniform(rng1, minval=self.mask_limit)\n mask = jax.random.bernoulli(rng2, mask_prob, vid_embed.shape[:-1])\n mask = mask.at[:, 0].set(False)\n vid_embed = jnp.where(jnp.expand_dims(mask, -1), self.mask_token, vid_embed)\n else:\n mask = None\n\n # --- Predict transition ---\n act_embed = self.action_up(batch[""latent_actions""])\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n logits = self.dynamics(vid_embed)\n return dict(token_logits=logits, mask=mask)\n",python,tab
9
+ 8,30937,"models/dynamics.py",416,0,"",python,selection_command
10
+ 9,31509,"models/dynamics.py",1359,0,"",python,selection_command
11
+ 10,31790,"models/dynamics.py",1655,0,"",python,selection_command
12
+ 11,32552,"models/dynamics.py",753,0,"",python,selection_command
13
+ 12,32710,"models/dynamics.py",165,0,"",python,selection_command
14
+ 13,32898,"models/dynamics.py",0,0,"",python,selection_command
15
+ 14,35425,"train_dynamics.py",0,0,"from dataclasses import dataclass, field\nimport os\nimport time\n\nimport einops\nfrom flax.training import orbax_utils\nfrom flax.training.train_state import TrainState\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax\nfrom orbax.checkpoint import PyTreeCheckpointer\nimport numpy as np\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\n\nfrom genie import Genie, restore_genie_components\nfrom models.tokenizer import TokenizerVQVAE\nfrom models.lam import LatentActionModel\nfrom utils.dataloader import get_dataloader\n\nts = int(time.time())\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = ""data_tfrecords/coinrun""\n # Optimization\n batch_size: int = 36\n min_lr: float = 3e-6\n max_lr: float = 3e-5\n warmup_steps: int = 5000\n # Tokenizer\n tokenizer_dim: int = 512\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 8\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 8\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_dim: int = 512\n dyna_num_blocks: int = 12\n dyna_num_heads: int = 8\n dropout: float = 0.0\n mask_limit: float = 0.5\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_dynamics""\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_gradients: bool = False\n\n\nargs = tyro.cli(Args)\n\n\ndef dynamics_loss_fn(params, state, inputs):\n """"""Compute masked dynamics loss""""""\n outputs = state.apply_fn(\n params,\n inputs,\n training=True,\n rngs={""params"": inputs[""rng""], ""dropout"": inputs[""dropout_rng""]},\n )\n mask = outputs[""mask""]\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask * ce_loss).sum() / mask.sum()\n acc = outputs[""token_logits""].argmax(-1) == outputs[""video_tokens""]\n acc = (mask * acc).sum() / mask.sum()\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_accuracy=acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n )\n return ce_loss, (outputs[""recon""], metrics)\n\n\n@jax.jit\ndef train_step(state, inputs):\n """"""Update state and compute metrics""""""\n grad_fn = jax.value_and_grad(dynamics_loss_fn, has_aux=True, allow_int=True)\n (loss, (recon, metrics)), grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return state, loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.PRNGKey(args.seed)\n if args.log and jax.process_index() == 0:\n wandb.init(\n entity=args.entity,\n project=args.project,\n name=args.name,\n tags=args.tags,\n group=""debug"",\n config=args\n )\n\n # --- Initialize model ---\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n # Dynamics\n dyna_dim=args.dyna_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n mask_limit=args.mask_limit,\n )\n rng, _rng = jax.random.split(rng)\n image_shape = (args.image_height, args.image_width, args.image_channels)\n dummy_inputs = dict(\n videos=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len, *image_shape),\n dtype=jnp.float32,\n ),\n action=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len), dtype=jnp.float32\n ),\n mask_rng=_rng,\n )\n rng, _rng = jax.random.split(rng)\n init_params = genie.init(_rng, dummy_inputs)\n\n # --- Initialize optimizer ---\n lr_schedule = optax.warmup_cosine_decay_schedule(\n args.min_lr, args.max_lr, args.warmup_steps, args.num_steps\n )\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4)\n train_state = TrainState.create(apply_fn=genie.apply, params=init_params, tx=tx)\n\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n train_state = jax.device_put(train_state, replicated_sharding)\n\n # --- Restore checkpoint ---\n train_state = restore_genie_components(\n train_state, replicated_sharding, dummy_inputs, rng, args\n )\n\n # --- TRAIN LOOP ---\n tfrecord_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".tfrecord"")\n ]\n dataloader = get_dataloader(\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n tfrecord_files,\n args.seq_len,\n args.batch_size,\n *image_shape,\n )\n step = 0\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng, _rng_dropout, _rng_mask = jax.random.split(rng, 4)\n\n videos_sharding = NamedSharding(\n mesh, PartitionSpec(""data"", None, None, None, None)\n )\n videos = jax.make_array_from_process_local_data(videos_sharding, videos)\n\n inputs = dict(\n videos=videos,\n rng=_rng,\n dropout_rng=_rng_dropout,\n mask_rng=_rng_mask,\n )\n start_time = time.time()\n train_state, loss, recon, metrics = train_step(train_state, inputs)\n elapsed_time = (time.time() - start_time) * 1000\n print(f""Step {step}, loss: {loss}, step time: {elapsed_time}ms"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n ""step_time_ms"": elapsed_time,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0]\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[args.seq_len - 1])),\n recon=wandb.Image(np.asarray(recon_seq[args.seq_len - 1])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n if step % args.log_checkpoint_interval == 0:\n ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""genie_{ts}_{step}""),\n ckpt,\n save_args=save_args,\n )\n if step >= args.num_steps:\n break\n",python,tab
16
+ 15,39797,"train_dynamics.py",5890,0,"",python,selection_command
17
+ 16,40207,"train_dynamics.py",5044,0,"",python,selection_command
18
+ 17,40597,"train_dynamics.py",4217,0,"",python,selection_command
19
+ 18,40765,"train_dynamics.py",3598,0,"",python,selection_command
20
+ 19,40946,"train_dynamics.py",2859,0,"",python,selection_command
21
+ 20,41080,"train_dynamics.py",2006,0,"",python,selection_command
22
+ 21,41229,"train_dynamics.py",1466,0,"",python,selection_command
23
+ 22,41446,"train_dynamics.py",882,0,"",python,selection_command
24
+ 23,41604,"train_dynamics.py",388,0,"",python,selection_command
25
+ 24,41744,"train_dynamics.py",0,0,"",python,selection_command
26
+ 25,43268,"train_dynamics.py",617,0,"",python,selection_command
27
+ 26,46121,"train_dynamics.py",0,0,"",python,selection_command
28
+ 27,46786,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
29
+ 28,49091,"train_dynamics.py",0,0,"",python,tab
30
+ 29,49113,"TERMINAL",0,0,"",,terminal_focus
31
+ 30,104571,"requirements.txt",0,0,"dm_pix>=0.4.3\neinops>=0.8.0\nflax>=0.8.5\njax[cuda12]>=0.4.30\noptax>=0.2.3\nprocgen>=0.10.7\ntyro>=0.8.5\nwandb>=0.17.4\ntensorflow>=2.1\npre-commit>=4.2.0",pip-requirements,tab
32
+ 31,106305,"requirements.txt",73,0,"",pip-requirements,selection_command
33
+ 32,106838,"requirements.txt",73,0,"#",pip-requirements,content
34
+ 33,106844,"requirements.txt",74,0,"",pip-requirements,selection_keyboard
35
+ 34,106906,"requirements.txt",74,0," ",pip-requirements,content
36
+ 35,106910,"requirements.txt",75,0,"",pip-requirements,selection_keyboard
37
+ 36,107095,"requirements.txt",74,0,"",pip-requirements,selection_command
38
+ 37,114295,"requirements.txt",61,0,"",pip-requirements,selection_command
39
+ 38,114428,"requirements.txt",41,0,"",pip-requirements,selection_command
40
+ 39,114530,"requirements.txt",43,0,"",pip-requirements,selection_command
41
+ 40,114983,"requirements.txt",43,1,"[",pip-requirements,selection_command
42
+ 41,115082,"requirements.txt",43,7,"[cuda12",pip-requirements,selection_command
43
+ 42,115265,"requirements.txt",43,10,"[cuda12]>=",pip-requirements,selection_command
44
+ 43,115682,"requirements.txt",43,9,"[cuda12]>",pip-requirements,selection_command
45
+ 44,115832,"requirements.txt",43,8,"[cuda12]",pip-requirements,selection_command
46
+ 45,116046,"requirements.txt",43,8,"",pip-requirements,content
47
+ 46,134920,"train_dynamics.py",0,0,"",python,tab
48
+ 47,141347,"train_dynamics.py",41,0,"",python,selection_command
49
+ 48,141596,"train_dynamics.py",51,0,"",python,selection_command
50
+ 49,141626,"train_dynamics.py",63,0,"",python,selection_command
51
+ 50,141657,"train_dynamics.py",64,0,"",python,selection_command
52
+ 51,141691,"train_dynamics.py",78,0,"",python,selection_command
53
+ 52,141894,"train_dynamics.py",116,0,"",python,selection_command
54
+ 53,141964,"train_dynamics.py",121,0,"",python,selection_command
55
+ 54,142163,"train_dynamics.py",125,0,"",python,selection_command
56
+ 55,142229,"train_dynamics.py",87,0,"",python,selection_command
57
+ 56,145246,"train_dynamics.py",653,0,"",python,selection_command
58
+ 57,145597,"train_dynamics.py",1241,0,"",python,selection_command
59
+ 58,148942,"train_dynamics.py",1839,0,"",python,selection_command
60
+ 59,150414,"train_dynamics.py",2485,0,"",python,selection_command
61
+ 60,150985,"train_dynamics.py",3341,0,"",python,selection_command
62
+ 61,153046,"train_dynamics.py",2485,0,"",python,selection_command
63
+ 62,153663,"train_dynamics.py",2527,0,"",python,selection_command
64
+ 63,153914,"train_dynamics.py",2586,0,"",python,selection_command
65
+ 64,153946,"train_dynamics.py",2606,0,"",python,selection_command
66
+ 65,153981,"train_dynamics.py",2642,0,"",python,selection_command
67
+ 66,154015,"train_dynamics.py",2677,0,"",python,selection_command
68
+ 67,154046,"train_dynamics.py",2738,0,"",python,selection_command
69
+ 68,154081,"train_dynamics.py",2784,0,"",python,selection_command
70
+ 69,154113,"train_dynamics.py",2853,0,"",python,selection_command
71
+ 70,154146,"train_dynamics.py",2859,0,"",python,selection_command
72
+ 71,154180,"train_dynamics.py",2903,0,"",python,selection_command
73
+ 72,154213,"train_dynamics.py",2904,0,"",python,selection_command
74
+ 73,154246,"train_dynamics.py",2909,0,"",python,selection_command
75
+ 74,154280,"train_dynamics.py",2918,0,"",python,selection_command
76
+ 75,154508,"train_dynamics.py",3753,0,"",python,selection_command
77
+ 76,155160,"train_dynamics.py",4363,0,"",python,selection_command
78
+ 77,155948,"train_dynamics.py",5218,0,"",python,selection_command
79
+ 78,158479,"train_dynamics.py",6074,0,"",python,selection_command
80
+ 79,161795,"train_dynamics.py",6732,0,"",python,selection_command
81
+ 80,166966,"train_dynamics.py",7582,0,"",python,selection_command
82
+ 81,170485,"train_dynamics.py",8668,0,"",python,selection_command
83
+ 82,175934,"train_dynamics.py",7582,0,"",python,selection_command
84
+ 83,177106,"train_dynamics.py",7569,0,"",python,selection_command
85
+ 84,177361,"train_dynamics.py",7559,0,"",python,selection_command
86
+ 85,177390,"train_dynamics.py",7482,0,"",python,selection_command
87
+ 86,177422,"train_dynamics.py",7421,0,"",python,selection_command
88
+ 87,177457,"train_dynamics.py",7341,0,"",python,selection_command
89
+ 88,177488,"train_dynamics.py",7304,0,"",python,selection_command
90
+ 89,178494,"train_dynamics.py",7341,0,"",python,selection_command
91
+ 90,178693,"train_dynamics.py",7421,0,"",python,selection_command
92
+ 91,178849,"train_dynamics.py",7482,0,"",python,selection_command
93
+ 92,179027,"train_dynamics.py",7559,0,"",python,selection_command
94
+ 93,179210,"train_dynamics.py",7569,0,"",python,selection_command
95
+ 94,179506,"train_dynamics.py",7559,0,"",python,selection_command
96
+ 95,179674,"train_dynamics.py",7482,0,"",python,selection_command
97
+ 96,179833,"train_dynamics.py",7421,0,"",python,selection_command
98
+ 97,179926,"train_dynamics.py",7482,0,"",python,selection_command
99
+ 98,180126,"train_dynamics.py",7421,0,"",python,selection_command
100
+ 99,180497,"train_dynamics.py",7482,0,"",python,selection_command
101
+ 100,181016,"train_dynamics.py",7559,0,"",python,selection_command
102
+ 101,181259,"train_dynamics.py",7482,0,"",python,selection_command
103
+ 102,183280,"train_dynamics.py",7421,0,"",python,selection_command
104
+ 103,183526,"train_dynamics.py",7341,0,"",python,selection_command
105
+ 104,185378,"train_dynamics.py",7407,0,"",python,selection_command
106
+ 105,185754,"train_dynamics.py",7401,0,"",python,selection_command
107
+ 106,185894,"train_dynamics.py",7399,0,"",python,selection_command
108
+ 107,186075,"train_dynamics.py",7388,0,"",python,selection_command
109
+ 108,186198,"train_dynamics.py",7387,0,"",python,selection_command
110
+ 109,186477,"train_dynamics.py",7377,0,"",python,selection_command
111
+ 110,212701,"train_dynamics.py",2949,0,"",python,selection_command
112
+ 111,212882,"train_dynamics.py",2992,0,"",python,selection_command
113
+ 112,212978,"train_dynamics.py",3000,0,"",python,selection_command
114
+ 113,213228,"train_dynamics.py",3002,0,"",python,selection_command
115
+ 114,213261,"train_dynamics.py",3005,0,"",python,selection_command
116
+ 115,213294,"train_dynamics.py",3006,0,"",python,selection_command
117
+ 116,213327,"train_dynamics.py",3020,0,"",python,selection_command
118
+ 117,213590,"train_dynamics.py",3021,0,"",python,selection_command
119
+ 118,214182,"train_dynamics.py",2988,0,"",python,selection_command
120
+ 119,215003,"train_dynamics.py",2992,0,"",python,selection_command
121
+ 120,215250,"train_dynamics.py",3000,0,"",python,selection_command
122
+ 121,215283,"train_dynamics.py",3002,0,"",python,selection_command
123
+ 122,215321,"train_dynamics.py",3005,0,"",python,selection_command
124
+ 123,215353,"train_dynamics.py",3006,0,"",python,selection_command
125
+ 124,215386,"train_dynamics.py",3020,0,"",python,selection_command
126
+ 125,215420,"train_dynamics.py",3021,0,"",python,selection_command
127
+ 126,221808,"train_dynamics.py",2988,0,"",python,selection_command
128
+ 127,221880,"train_dynamics.py",2992,0,"",python,selection_command
129
+ 128,222128,"train_dynamics.py",3000,0,"",python,selection_command
130
+ 129,222161,"train_dynamics.py",3002,0,"",python,selection_command
131
+ 130,222190,"train_dynamics.py",3005,0,"",python,selection_command
132
+ 131,222224,"train_dynamics.py",3006,0,"",python,selection_command
133
+ 132,222259,"train_dynamics.py",3020,0,"",python,selection_command
134
+ 133,222293,"train_dynamics.py",3021,0,"",python,selection_command
135
+ 134,223174,"train_dynamics.py",2988,0,"",python,selection_command
136
+ 135,223299,"train_dynamics.py",2992,0,"",python,selection_command
137
+ 136,223551,"train_dynamics.py",3000,0,"",python,selection_command
138
+ 137,223581,"train_dynamics.py",3002,0,"",python,selection_command
139
+ 138,223614,"train_dynamics.py",3005,0,"",python,selection_command
140
+ 139,223648,"train_dynamics.py",3006,0,"",python,selection_command
141
+ 140,223681,"train_dynamics.py",3020,0,"",python,selection_command
142
+ 141,223715,"train_dynamics.py",3021,0,"",python,selection_command
143
+ 142,223749,"train_dynamics.py",3037,0,"",python,selection_command
144
+ 143,224380,"train_dynamics.py",3021,0,"",python,selection_command
145
+ 144,225595,"train_dynamics.py",2988,0,"",python,selection_command
146
+ 145,225681,"train_dynamics.py",2992,0,"",python,selection_command
147
+ 146,225938,"train_dynamics.py",3000,0,"",python,selection_command
148
+ 147,225966,"train_dynamics.py",3002,0,"",python,selection_command
149
+ 148,225999,"train_dynamics.py",3005,0,"",python,selection_command
150
+ 149,226037,"train_dynamics.py",3006,0,"",python,selection_command
151
+ 150,226070,"train_dynamics.py",3020,0,"",python,selection_command
152
+ 151,226103,"train_dynamics.py",3021,0,"",python,selection_command
153
+ 152,251190,"train_dynamics.py",2006,0,"",python,selection_command
154
+ 153,251439,"train_dynamics.py",2045,0,"",python,selection_command
155
+ 154,251470,"train_dynamics.py",2075,0,"",python,selection_command
156
+ 155,251506,"train_dynamics.py",2091,0,"",python,selection_command
157
+ 156,251539,"train_dynamics.py",2107,0,"",python,selection_command
158
+ 157,251574,"train_dynamics.py",2130,0,"",python,selection_command
159
+ 158,251760,"train_dynamics.py",2204,0,"",python,selection_command
160
+ 159,252115,"train_dynamics.py",2210,0,"",python,selection_command
161
+ 160,256408,"train_dynamics.py",2215,0,"",python,selection_command
162
+ 161,256583,"train_dynamics.py",2217,0,"",python,selection_command
163
+ 162,262598,"models/dynamics.py",0,0,"",python,tab
164
+ 163,263846,"models/dynamics.py",416,0,"",python,selection_keyboard
165
+ 164,271113,"models/dynamics.py",1359,0,"",python,selection_command
166
+ 165,273515,"models/dynamics.py",1373,0,"",python,selection_command
167
+ 166,273673,"models/dynamics.py",1389,0,"",python,selection_command
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-ef5ea013-ac2b-459c-8783-a7b025d58a391754900011518-2025_08_11-10.13.34.249/source.csv ADDED
The diff for this file is too large to render. See raw diff
 
507ab0ec0dfe0c18ad7778dd15e072f92367194c94623114de802c8ed9c52e20/crowd-code-b9559366-0d71-4ceb-9b37-1d3a0cf03cd61750867779082-2025_06_25-18.09.57.465/source.csv ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
2
+ 1,6,"scripts/rename_mp4_files.py",0,0,"#!/usr/bin/env python3\n""""""\nScript to rename all MP4 files in a directory with a custom prefix.\n""""""\n\nimport os\nimport argparse\nimport re\nfrom pathlib import Path\n\n\ndef rename_mp4_files(directory, prefix, dry_run=False, pattern=None, start_number=1):\n """"""\n Rename all MP4 files in the specified directory with a custom prefix.\n \n Args:\n directory (str): Path to the directory containing MP4 files\n prefix (str): Custom prefix to add to filenames\n dry_run (bool): If True, only show what would be renamed without actually renaming\n pattern (str): Optional regex pattern to filter files\n start_number (int): Starting number for sequential naming\n """"""\n directory_path = Path(directory)\n \n if not directory_path.exists():\n print(f""Error: Directory '{directory}' does not exist."")\n return\n \n if not directory_path.is_dir():\n print(f""Error: '{directory}' is not a directory."")\n return\n \n # Find all MP4 files\n mp4_files = list(directory_path.glob(""*.mp4""))\n \n if not mp4_files:\n print(f""No MP4 files found in '{directory}'"")\n return\n \n # Filter by pattern if provided\n if pattern:\n regex = re.compile(pattern)\n mp4_files = [f for f in mp4_files if regex.search(f.name)]\n \n if not mp4_files:\n print(f""No MP4 files match the pattern '{pattern}' in '{directory}'"")\n return\n \n print(f""Found {len(mp4_files)} MP4 files to rename:"")\n \n # Sort files for consistent ordering\n mp4_files.sort()\n \n for i, file_path in enumerate(mp4_files, start=start_number):\n # Get the original filename without extension\n original_name = file_path.stem\n extension = file_path.suffix\n \n # Create new filename with prefix\n new_name = f""{prefix}_{i:03d}_{original_name}{extension}""\n new_path = file_path.parent / new_name\n \n # Check if new filename already exists\n if new_path.exists() and not dry_run:\n print(f""Warning: '{new_name}' already exists, skipping '{file_path.name}'"")\n continue\n \n if dry_run:\n print(f""Would rename: '{file_path.name}' -> '{new_name}'"")\n else:\n try:\n file_path.rename(new_path)\n print(f""Renamed: '{file_path.name}' -> '{new_name}'"")\n except OSError as e:\n print(f""Error renaming '{file_path.name}': {e}"")\n \n if dry_run:\n print(f""\nDry run completed. {len(mp4_files)} files would be renamed."")\n else:\n print(f""\nRenaming completed. {len(mp4_files)} files renamed."")\n\n\ndef main():\n parser = argparse.ArgumentParser(\n description=""Rename all MP4 files in a directory with a custom prefix"",\n formatter_class=argparse.RawDescriptionHelpFormatter,\n epilog=""""""\nExamples:\n %(prog)s /path/to/videos minecraft\n %(prog)s /path/to/videos player --dry-run\n %(prog)s /path/to/videos episode --pattern ""joost"" --start-number 10\n """"""\n )\n \n parser.add_argument(\n ""directory"",\n help=""Directory containing MP4 files to rename""\n )\n \n parser.add_argument(\n ""prefix"",\n help=""Custom prefix to add to filenames""\n )\n \n parser.add_argument(\n ""--dry-run"",\n action=""store_true"",\n help=""Show what would be renamed without actually renaming files""\n )\n \n parser.add_argument(\n ""--pattern"",\n help=""Regex pattern to filter files (only rename files matching this pattern)""\n )\n \n parser.add_argument(\n ""--start-number"",\n type=int,\n default=1,\n help=""Starting number for sequential naming (default: 1)""\n )\n \n args = parser.parse_args()\n \n # Validate start number\n if args.start_number < 1:\n print(""Error: start-number must be at least 1"")\n return\n \n rename_mp4_files(\n directory=args.directory,\n prefix=args.prefix,\n dry_run=args.dry_run,\n pattern=args.pattern,\n start_number=args.start_number\n )\n\n\nif __name__ == ""__main__"":\n main() ",python,tab
3
+ 2,319,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"6:09:57 PM [info] Activating crowd-code\n6:09:57 PM [info] Welcome back tum_ind3695. Your user-id is '507ab0ec0dfe0c18ad7778dd15e072f92367194c94623114de802c8ed9c52e20'. Happy coding!\n6:09:57 PM [info] Recording started\n",Log,tab
4
+ 3,2485,"scripts/rename_mp4_files.py",0,0,"",python,tab
5
+ 4,2488,"scripts/rename_mp4_files.py",1333,0,"",python,selection_mouse
6
+ 5,2493,"scripts/rename_mp4_files.py",1332,0,"",python,selection_command
7
+ 6,4622,"TERMINAL",0,0,"",,terminal_focus
8
+ 7,301525,"TERMINAL",0,0,"",,terminal_focus
9
+ 8,419551,"TERMINAL",0,0,"cd /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/dyn/3289577",,terminal_command
10
+ 9,419570,"TERMINAL",0,0,"]633;E;2025-06-25 18:16:56 cd /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/dyn/3289577;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577]633;D;0",,terminal_output
11
+ 10,420075,"TERMINAL",0,0,"ls",,terminal_command
12
+ 11,420099,"TERMINAL",0,0,"]633;E;2025-06-25 18:16:57 ls;ef772368-fdbb-4ff2-ad99-2db92190992c]633;Cgenie_1750786631_1000 genie_1750786631_11000 genie_1750786631_12500 genie_1750786631_2500 genie_1750786631_4000 genie_1750786631_5000 genie_1750786631_6500 genie_1750786631_8000 genie_1750786631_9500\r\ngenie_1750786631_10000 genie_1750786631_11500 genie_1750786631_1500 genie_1750786631_3000 genie_1750786631_4500 genie_1750786631_5500 genie_1750786631_7000 genie_1750786631_8500\r\ngenie_1750786631_10500 genie_1750786631_12000 genie_1750786631_2000 genie_1750786631_3500 genie_1750786631_500 genie_1750786631_6000 genie_1750786631_7500 genie_1750786631_9000\r\n]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577]633;D;0",,terminal_output
13
+ 12,439480,"TERMINAL",0,0,"cd genie_1750786631_12500",,terminal_command
14
+ 13,440497,"TERMINAL",0,0,"pwd",,terminal_command
15
+ 14,440512,"TERMINAL",0,0,"]633;E;2025-06-25 18:17:17 pwd;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577/genie_1750786631_12500\r\n]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577/genie_1750786631_12500]633;D;0",,terminal_output
16
+ 15,615482,"TERMINAL",0,0,"cd ../../..",,terminal_command
17
+ 16,615494,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:12 cd ../../..;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints]633;D;0",,terminal_output
18
+ 17,616076,"TERMINAL",0,0,"ls",,terminal_command
19
+ 18,616111,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:13 ls;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C0000 3290283 3290284 3290295 3290296 3290366 3290367 3290391 3290392 3290439 3290440 3291405 dyn lam tokenizer\r\n]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints]633;D;0",,terminal_output
20
+ 19,619493,"TERMINAL",0,0,"cd tokenizer/",,terminal_command
21
+ 20,619506,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:16 cd tokenizer/;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer]633;D;0",,terminal_output
22
+ 21,619896,"TERMINAL",0,0,"ls",,terminal_command
23
+ 22,619948,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:17 ls;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C",,terminal_output
24
+ 23,619997,"TERMINAL",0,0,"3255052 3255483 3256924 3257467 3258278 3259422 3261720 3273042 3273290 3273480 3273494 3273627 3273742 3273795 3273824 3273830 3273846 3276030 3276048 3276053 3285798\r\n3255466 3256873 3256926 3257633 3258283 3260527 3261722 3273174 3273348 3273488 3273496 3273681 3273743 3273816 3273828 3273831 3275950 3276039 3276051 3276058 3285811\r\n3255482 3256921 3256929 3257812 3259405 3260932 3273026 3273229 3273476 3273489 3273503 3273687 3273746 3273820 3273829 3273841 3275991 3276043 3276052 3285784 3286114\r\n]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer]633;D;0",,terminal_output
25
+ 24,649663,"TERMINAL",0,0,"cd 3255052/",,terminal_command
26
+ 25,649679,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:47 cd 3255052/;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/3255052]633;D;0",,terminal_output
27
+ 26,650176,"TERMINAL",0,0,"ls",,terminal_command
28
+ 27,650194,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:47 ls;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/3255052]633;D;0",,terminal_output
29
+ 28,652601,"TERMINAL",0,0,"cd ..",,terminal_command
30
+ 29,652609,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:50 cd ..;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer]633;D;0",,terminal_output
31
+ 30,654392,"TERMINAL",0,0,"cd ..",,terminal_command
32
+ 31,654398,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:51 cd ..;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints]633;D;0",,terminal_output
33
+ 32,655918,"TERMINAL",0,0,"cd ..",,terminal_command
34
+ 33,655926,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:53 cd ..;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared]633;D;0",,terminal_output
35
+ 34,1148074,"TERMINAL",0,0,"queue",,terminal_command
36
+ 35,1148124,"TERMINAL",0,0,"]633;E;2025-06-25 18:29:05 queue;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C",,terminal_output
37
+ 36,1148182,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Wed Jun 25 18:29:05 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3292119 accelerat train_to tum_ind3 PD\t0:00\t 1 (Priority)3289577 accelerat train_dy tum_ind3 R 22:52:29\t 1 hkn0712",,terminal_output
38
+ 37,1149242,"TERMINAL",0,0,"630",,terminal_output
39
+ 38,1149475,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared]633;D;0",,terminal_output
40
+ 39,1150350,"TERMINAL",0,0,"queue",,terminal_command
41
+ 40,1150402,"TERMINAL",0,0,"]633;E;2025-06-25 18:29:07 queue;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C",,terminal_output
42
+ 41,1150481,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Wed Jun 25 18:29:07 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3292119 accelerat train_to tum_ind3 PD\t0:00\t 1 (Priority)3289577 accelerat train_dy tum_ind3 R 22:52:31\t 1 hkn0712",,terminal_output
43
+ 42,1151212,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared]633;D;0",,terminal_output
44
+ 43,1152587,"TERMINAL",0,0,"squeue",,terminal_command
45
+ 44,1152616,"TERMINAL",0,0,"]633;E;2025-06-25 18:29:10 squeue;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C JOBID PARTITION NAME USER ST TIME NODES NODELIST(REASON)\r\n 3292119 accelerat train_to tum_ind3 PD 0:00 1 (Priority)\r\n 3289577 accelerat train_dy tum_ind3 R 22:52:34 1 hkn0712\r\n]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared]633;D;0]633;P;Cwd=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared",,terminal_output
46
+ 45,2263248,"TERMINAL",0,0,"cd /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/dyn/3289577",,terminal_command
47
+ 46,2263255,"TERMINAL",0,0,"]633;E;2025-06-25 18:47:40 cd /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/dyn/3289577;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577]633;D;0",,terminal_output
48
+ 47,2264288,"TERMINAL",0,0,"ls",,terminal_command
49
+ 48,2264341,"TERMINAL",0,0,"]633;E;2025-06-25 18:47:41 ls;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C",,terminal_output
50
+ 49,2264401,"TERMINAL",0,0,"genie_1750786631_1000 genie_1750786631_11000 genie_1750786631_12500 genie_1750786631_2000 genie_1750786631_3500 genie_1750786631_500 genie_1750786631_6000 genie_1750786631_7500 genie_1750786631_9000\r\ngenie_1750786631_10000 genie_1750786631_11500 genie_1750786631_13000 genie_1750786631_2500 genie_1750786631_4000 genie_1750786631_5000 genie_1750786631_6500 genie_1750786631_8000 genie_1750786631_9500\r\ngenie_1750786631_10500 genie_1750786631_12000 genie_1750786631_1500 genie_1750786631_3000 genie_1750786631_4500 genie_1750786631_5500 genie_1750786631_7000 genie_1750786631_8500\r\n]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577]633;D;0",,terminal_output
51
+ 50,2269289,"TERMINAL",0,0,"cd genie_1750786631_13000",,terminal_command
52
+ 51,2269296,"TERMINAL",0,0,"]633;E;2025-06-25 18:47:46 cd genie_1750786631_13000;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577/genie_1750786631_13000]633;D;0",,terminal_output
53
+ 52,2270468,"TERMINAL",0,0,"pwd",,terminal_command
54
+ 53,2270477,"TERMINAL",0,0,"]633;E;2025-06-25 18:47:47 pwd;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577/genie_1750786631_13000\r\n]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577/genie_1750786631_13000]633;D;0",,terminal_output
55
+ 54,2272125,"TERMINAL",0,0,"os",,terminal_command
56
+ 55,2272199,"TERMINAL",0,0,"]633;E;2025-06-25 18:47:49 os;ef772368-fdbb-4ff2-ad99-2db92190992c]633;Cbash: os: command not found...\r\n",,terminal_output
57
+ 56,2274189,"TERMINAL",0,0,"ls",,terminal_command
58
+ 57,2274240,"TERMINAL",0,0,"]633;E;2025-06-25 18:47:51 ls;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C",,terminal_output
59
+ 58,2274298,"TERMINAL",0,0,"array_metadatas _CHECKPOINT_METADATA d manifest.ocdbt _METADATA ocdbt.process_0 ocdbt.process_1 ocdbt.process_2 ocdbt.process_3 _sharding\r\n]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577/genie_1750786631_13000]633;D;0",,terminal_output
60
+ 59,2708521,"scripts/file_duplicate_checker.py",0,0,"import os\nfrom collections import defaultdict\nfrom tqdm import tqdm\n\ndef find_duplicate_filenames(root_dir):\n filenames = defaultdict(list)\n file_count = 0\n\n # Use tqdm with manual update and no percentage/ETA bar\n pbar = tqdm(desc=""Files scanned"", unit=""file"", dynamic_ncols=True, bar_format=""{desc}: {n_fmt}"")\n\n # Walk the directory recursively\n for dirpath, _, files in os.walk(root_dir):\n for file in files:\n full_path = os.path.join(dirpath, file)\n if os.path.isfile(full_path):\n filenames[file].append(full_path)\n file_count += 1\n pbar.update(1)\n\n pbar.close()\n\n # Print duplicates\n duplicates = {name: paths for name, paths in filenames.items() if len(paths) > 1}\n if duplicates:\n print(""\nDuplicate filenames found:\n"")\n for name, paths in duplicates.items():\n print(f""Filename: {name}"")\n for path in paths:\n print(f"" - {path}"")\n print()\n else:\n print(""\nNo duplicate filenames found."")\n\nif __name__ == ""__main__"":\n import sys\n if len(sys.argv) < 2:\n print(""Usage: python find_duplicates.py <directory_path>"")\n else:\n find_duplicate_filenames(sys.argv[1])\n\n",python,tab
61
+ 60,2729693,"scripts/file_duplicate_checker.py",991,0,"",python,selection_mouse
62
+ 61,2729697,"scripts/file_duplicate_checker.py",990,0,"",python,selection_command
63
+ 62,2729708,"scripts/file_duplicate_checker.py",990,1,")",python,selection_mouse
64
+ 63,2729710,"scripts/file_duplicate_checker.py",991,0,"",python,selection_command