Add files using upload-large-folder tool
Browse files- 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-0e6e3fa0-317b-4a46-bdb4-752f8e86cd181758538014008-2025_09_22-12.47.12.63/source.csv +0 -0
- 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-1505f3d0-0cb4-4cc0-84bf-678810d0ac8f1757148592235-2025_09_06-10.49.56.658/source.csv +17 -0
- 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-1f52a67d-2cca-4352-b39c-6cd9f3effae01765648091845-2025_12_13-18.48.27.228/source.csv +66 -0
- 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-2790c5c3-9c96-49ac-ab9e-481a8033e7cd1757945282635-2025_09_15-16.08.09.625/source.csv +0 -0
- 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-9543d8e2-2376-4957-873e-df7016d502961763465687199-2025_11_18-12.34.49.442/source.csv +245 -0
- 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-ac4f665d-1bc1-467a-98b3-5da2178968731760857710663-2025_10_19-09.08.50.386/source.csv +6 -0
- 1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-c877f8c1-c8e0-4a7b-8720-40bb4df915221754138206219-2025_08_02-14.36.55.608/source.csv +0 -0
- 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-14f0662f-0032-43e8-be9f-8e53d6f150ad1758635869882-2025_09_23-15.57.52.616/source.csv +0 -0
- 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-1fa9b85d-3794-4f3b-b7a0-5170b7d2faaa1762362332596-2025_11_05-18.05.39.648/source.csv +19 -0
- 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-4b7a193b-6fd0-48b6-a605-a1ce6ba179221764439942282-2025_11_29-19.12.25.371/source.csv +10 -0
- 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-549b7320-9591-428d-ba6a-d3d8fb65a0001764416891089-2025_11_29-12.48.19.667/source.csv +1006 -0
- 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-8eaf046e-99c4-4091-a85d-91e359564aa51756825908437-2025_09_02-17.11.50.753/source.csv +0 -0
- 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-9543d8e2-2376-4957-873e-df7016d502961763465687199-2025_11_18-12.34.49.442/source.csv +226 -0
- 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-a7b808c2-b1d0-43a0-a38c-8b82cd2886711764488770794-2025_11_30-08.46.18.897/source.csv +32 -0
- 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-aeed47b9-f6ef-4272-b0ca-0c15ab4c25021758266694991-2025_09_19-09.25.04.660/source.csv +151 -0
- 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-d9cdf338-0ddd-4679-853a-6d7bdf2b18581751046137722-2025_06_27-10.42.19.354/source.csv +167 -0
- 4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-ef5ea013-ac2b-459c-8783-a7b025d58a391754900011518-2025_08_11-10.13.34.249/source.csv +0 -0
- 507ab0ec0dfe0c18ad7778dd15e072f92367194c94623114de802c8ed9c52e20/crowd-code-b9559366-0d71-4ceb-9b37-1d3a0cf03cd61750867779082-2025_06_25-18.09.57.465/source.csv +64 -0
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-0e6e3fa0-317b-4a46-bdb4-752f8e86cd181758538014008-2025_09_22-12.47.12.63/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-1505f3d0-0cb4-4cc0-84bf-678810d0ac8f1757148592235-2025_09_06-10.49.56.658/source.csv
ADDED
|
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,1,"train_tokenizer.py",0,0,"import os\n\nos.environ.setdefault(""XLA_PYTHON_CLIENT_MEM_FRACTION"", ""0.98"")\n\nfrom dataclasses import dataclass, field\nfrom typing import cast, Optional\n\nimport einops\nimport itertools\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom models.tokenizer import TokenizerVQVAE\nfrom utils.dataloader import get_dataloader\nfrom utils.train_utils import (\n get_lr_schedule,\n count_parameters_by_component,\n print_mem_stats,\n print_compiled_memory_stats,\n print_compiled_cost_analysis,\n)\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 300_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n vq_beta: float = 0.25\n batch_size: int = 48\n init_lr: float = 0.0\n max_lr: float = 3e-4\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 20000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n warmup_steps: int = 10000\n # Tokenizer\n model_dim: int = 512\n ffn_dim: int = 2048\n latent_dim: int = 32\n num_latents: int = 1024\n patch_size: int = 4\n num_blocks: int = 4\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.01\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_tokenizer""\n tags: list[str] = field(default_factory=lambda: [""tokenizer""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n wandb_id: str = """"\n use_flash_attention: bool = True\n\n\ndef build_model(args: Args, rng: jax.Array) -> tuple[TokenizerVQVAE, jax.Array]:\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n return (\n TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n ffn_dim=args.ffn_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n ),\n rng,\n )\n\n\ndef build_optimizer(\n model: TokenizerVQVAE, args: Args\n) -> tuple[nnx.Optimizer, optax.Schedule]:\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.param_dtype, # moments in full precision\n )\n optimizer = nnx.Optimizer(model, tx)\n return optimizer, lr_schedule\n\n\ndef build_mesh_and_sharding(\n num_devices: int,\n) -> tuple[Mesh, NamedSharding, NamedSharding]:\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n return mesh, replicated_sharding, videos_sharding\n\n\ndef shard_optimizer_states(\n optimizer: nnx.Optimizer, replicated_sharding: NamedSharding\n) -> None:\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n\ndef build_dataloader(args: Args) -> grain.DataLoaderIterator:\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n return grain_iterator\n\n\ndef build_checkpoint_manager(args: Args) -> ocp.CheckpointManager:\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n return checkpoint_manager\n\n\ndef restore_checkpoint_if_needed(\n args: Args,\n checkpoint_manager: ocp.CheckpointManager,\n optimizer: nnx.Optimizer,\n grain_iterator: grain.DataLoaderIterator,\n restore_step: Optional[int] = None,\n) -> tuple[int, nnx.Optimizer, grain.DataLoaderIterator]:\n step = 0\n if restore_step is None:\n restore_step = checkpoint_manager.latest_step()\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n restore_step,\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n grain_iterator = restored[""dataloader_state""]\n step = restore_step or 0\n print(f""Restored dataloader and model state from step {step}"")\n return step, optimizer, grain_iterator\n\n\ndef main(args: Args) -> None:\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n tokenizer, rng = build_model(args, rng)\n\n _, params, _ = nnx.split(tokenizer, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n optimizer, lr_schedule = build_optimizer(tokenizer, args)\n del tokenizer\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n mesh, replicated_sharding, videos_sharding = build_mesh_and_sharding(num_devices)\n\n shard_optimizer_states(optimizer, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n checkpoint_manager = build_checkpoint_manager(args)\n\n # --- Create DataLoaderIterator from dataloader ---\n grain_iterator = build_dataloader(args)\n\n # --- Restore checkpoint ---\n step, optimizer, grain_iterator = restore_checkpoint_if_needed(\n args, checkpoint_manager, optimizer, grain_iterator\n )\n\n # --- Define loss and train step (close over args) ---\n def tokenizer_loss_fn(\n model: TokenizerVQVAE, inputs: dict\n ) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n model.train()\n outputs = model(inputs, training=True)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n mse = jnp.square(gt - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n gt_clipped = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt_clipped, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt_clipped, recon)).mean()\n _, index_counts = jnp.unique_counts(\n jnp.ravel(outputs[""indices""]), size=args.num_latents, fill_value=0\n )\n codebook_usage = (index_counts != 0).mean()\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=codebook_usage,\n )\n return loss, (outputs[""recon""], metrics)\n\n @nnx.jit(donate_argnums=0)\n def train_step(\n optimizer: nnx.Optimizer, inputs: dict\n ) -> tuple[jax.Array, jax.Array, dict]:\n def loss_fn(model: TokenizerVQVAE) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n return tokenizer_loss_fn(model, inputs)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\n optimizer.model\n )\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""encoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""encoder""]\n )\n metrics[""vq_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""vq""]\n )\n metrics[""decoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""decoder""]\n )\n return loss, recon, metrics\n\n # --- TRAIN LOOP ---\n dataloader = (\n jax.make_array_from_process_local_data(videos_sharding, elem)\n for elem in grain_iterator\n )\n if jax.process_index() == 0:\n first_videos = next(dataloader)\n sample_inputs = dict(videos=first_videos)\n compiled = train_step.lower(optimizer, sample_inputs).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training\n dataloader = itertools.chain([first_videos], dataloader)\n print(f""Starting training from step {step}..."")\n first_step = step\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n inputs = dict(videos=videos)\n loss, recon, metrics = train_step(optimizer, inputs)\n if step == first_step:\n print_mem_stats(""After params initialized"")\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n grain_iterator # type: ignore\n ),\n ),\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n\n\nif __name__ == ""__main__"":\n args = tyro.cli(Args)\n main(args)\n",python,tab
|
| 3 |
+
2,113,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"10:49:56 AM [info] Activating crowd-code\n10:49:56 AM [info] Recording started\n10:49:56 AM [info] Initializing git provider using file system watchers...\n10:49:56 AM [info] Git repository found\n10:49:56 AM [info] Git provider initialized successfully\n",Log,tab
|
| 4 |
+
3,224,"extension-output-pdoom-org.crowd-code-#1-crowd-code",250,0,"10:49:56 AM [info] Initial git state: [object Object]\n",Log,content
|
| 5 |
+
4,188000455,"train_tokenizer.py",0,0,"",python,tab
|
| 6 |
+
5,188012251,"train_tokenizer.py",0,0,"Switched from branch 'main' to 'demo-notebook'",python,git_branch_checkout
|
| 7 |
+
6,188015548,"README.md",0,0,"<h1 align=""center"">🧞♀️ Jasmine: A simple, performant and scalable JAX-based world modeling codebase 🧞♀️</h1>\n\n<p align=""center"">\n <a href= ""https://github.com/FLAIROx/jafar/blob/main/LICENSE"">\n <img src=""https://img.shields.io/badge/license-Apache2.0-blue.svg"" /></a>\n <a href= ""https://github.com/psf/black"">\n <img src=""https://img.shields.io/badge/code%20style-black-000000.svg"" /></a>\n</p>\n\nJasmine is a production-ready JAX-based world modeling codebase. It currently implements the high-level architecture of [Genie: Generative Interactive Environments](https://arxiv.org/abs/2402.15391) (Bruce et al., 2024) with [MaskGIT](https://arxiv.org/abs/2202.04200) (Chang et al., 2022), as well as an autoregressive (causal) baseline. A diffusion baseline is coming soon.\n\nJasmine scales from single hosts to hundreds of xPUs thanks to XLA and strives to be an easily hackable, batteries-included foundation for world modeling research.\n\n<h2 name=""overview"" id=""overview"">Overview</h2>\n\n- Asynchronous & distributed checkpointing thanks to [orbax.checkpoint](https://github.com/google/orbax)\n - Jasmine also supports mixing and matching hardware topologies (e.g. train on four nodes, load the checkpoint on a single node)\n- Optimized dataloading thanks to [Grain](https://github.com/google/grain)\n - Dataloading scales with the number of processes (i.e. nodes/xPUs)\n- Checkpointing of model weights, optimizer and dataloader states\n- Full reproducibility with **identical** training curves (thanks to seeded dataloading and training, and [JAX' approach to pseudo random numbers](https://docs.jax.dev/en/latest/random-numbers.html))\n- Automatic checkpoint deletion/retention according to specified retention policy thanks to `orbax.checkpoint.CheckpointManager`\n- Mixed precision training using `bfloat16`\n - `int8` training is on the roadmap via [aqt](https://github.com/google/aqt)\n- FlashAttention thanks to [cuDNN SDPA](https://github.com/jax-ml/jax/blob/a155c5a9997924170e0067d552351a9833c12c11/jax/_src/cudnn/fused_attention_stablehlo.py#L842)\n- Frame-level KV cache resets for accelerated spatiotemporal attention in causal baseline (still in PR)\n- Activation checkpointing (even onto host memory if desired)\n- DDP (changing to FSDP requires changing **a single line of code**)\n- WSD learning rate schedule\n - No need to retrain from scratch if you want to train for longer\n- Index-shuffling during dataloading\n- Google-native stack\n - https://github.com/google/orbax for checkpointing\n - https://github.com/google/grain for dataloading\n - https://github.com/google-deepmind/dm_pix for image manipulation\n - https://github.com/google/array_record as the data format\n- Easy model inspection thanks to [treescope](https://github.com/google-deepmind/treescope)\n- Easy model surgery thanks to the new [flax.nnx](https://flax.readthedocs.io/en/latest/migrating/linen_to_nnx.html) API\n- [Shape suffixes](https://medium.com/@NoamShazeer/shape-suffixes-good-coding-style-f836e72e24fd) throughout the repository\n\n<h2 name=""start"" id=""start"">Setup 🧗</h2>\n\nJasmine requires `python 3.10`, `jax 0.6.2`, and `flax 0.10.7`. To install the requirements, run:\n\n```bash\npip install -r requirements.txt\npre-commit install\n```\n\n---\n\n<h2 name=""dataset"" id=""dataset"">Dataset 📂</h2>\n\nYou can either download our preprocessed dataset from [Hugging Face](https://huggingface.co/datasets/p-doom/open_ai_minecraft_arrayrecords_chunked) or preprocess [OpenAI's VPT dataset](https://github.com/openai/Video-Pre-Training) manually.\n\n### Option 1: Use Preprocessed Dataset (Recommended)\n\nThe easiest way to get started is to download our preprocessed dataset from Hugging Face. This script will handle downloading and extracting it:\n\n```bash\nbash input_pipeline/download/download_array_records.sh\n```\n\n---\n\n### Option 2: Manual Download & Preprocessing of OpenAI's VPT Dataset\n\nIf you prefer to use the raw VPT dataset from OpenAI and preprocess it yourself, follow these steps:\n\n1. **Download index files:**\n This will download the initial index file:\n\n ```bash\n bash input_pipeline/download/openai/download_index_files.sh\n ```\n\n2. **Download from all index files:**\n This may take a long time depending on your bandwidth:\n\n ```bash\n python input_pipeline/download/openai/download_videos.py --index_file_path data/open_ai_index_files/all_7xx_Apr_6.json\n python input_pipeline/download/openai/download_videos.py --index_file_path data/open_ai_index_files/all_8xx_Jun_29.json\n python input_pipeline/download/openai/download_videos.py --index_file_path data/open_ai_index_files/all_9xx_Jun_29.json\n python input_pipeline/download/openai/download_videos.py --index_file_path data/open_ai_index_files/all_10xx_Jun_29.json\n ```\n\n3. **Preprocess videos into ArrayRecords:**\n For efficient distributed training, convert the raw videos into the arrayrecord format (make sure to have [ffmpeg](https://github.com/FFmpeg/FFmpeg) installed on your machine):\n\n ```bash\n python input_pipeline/preprocess/video_to_array_records.py\n ```\n\n> **Note:** This is a large dataset and may take considerable time and storage to download and process.\n\n\n<h2 name=""train"" id=""train"">Quick Start 🚀 </h2>\n\nGenie has three components: a [video tokenizer](models/tokenizer.py), a [latent action model](models/lam.py), and a [dynamics model](models/dynamics.py). Each of these components are trained separately, however, the dynamics model requires a pre-trained video tokenizer (and latent action model).\n\nTo train the video tokenizer, run:\n\n```bash\npython train_tokenizer.py --ckpt_dir <path>\n```\n\nTo train the latent action model, run:\n\n```bash\npython train_lam.py --ckpt_dir <path>\n```\n\nOnce the tokenizer and LAM are trained, the dynamics model can be trained with:\n\n```bash\npython train_dynamics.py --tokenizer_checkpoint <path> --lam_checkpoint <path>\n```\n\nLogging with `wandb` is supported. To enable logging, set the `WANDB_API_KEY` environment variable or run:\n\n```bash\nwandb login\n```\n\nTraining can then be logged by setting the `--log` flag:\n\n```bash\npython train_tokenizer.py --log --entity <wandb-entity> --project <wandb-project>\n```\n\n<h2 name=""cite"" id=""cite"">Citing 📜 </h2>\n\nJasmine was built by [Mihir Mahajan](https://maharajamihir.github.io/), [Alfred Nguyen](https://avocadoali.github.io/) and [Franz Srambical](https://srambical.fr/), but started as a fork of [Jafar](https://github.com/flairox/jafar), built by [Matthew Jackson](https://matthewtjackson.com) and [Timon Willi](https://www.timonwilli.com).\n\nIf you use Jasmine in your work, please cite us, Jafar, and the original Genie paper as follows:\n\n```\n@article{\n mahajan2025jasmine,\n title={Jasmine: A simple, performant and scalable JAX-based world modeling codebase},\n author={Mihir Mahajan and Alfred Nguyen and Franz Srambical and Stefan Bauer},\n journal = {p(doom) blog},\n year={2025},\n url={https://pdoom.org/jasmine.html},\n note = {https://pdoom.org/blog.html}\n}\n```\n```\n@inproceedings{\n willi2024jafar,\n title={Jafar: An Open-Source Genie Reimplemention in Jax},\n author={Timon Willi and Matthew Thomas Jackson and Jakob Nicolaus Foerster},\n booktitle={First Workshop on Controllable Video Generation @ ICML 2024},\n year={2024},\n url={https://openreview.net/forum?id=ZZGaQHs9Jb}\n}\n```\n```\n@inproceedings{\n bruce2024genie,\n title={Genie: Generative Interactive Environments},\n author={Jake Bruce and Michael D Dennis and Ashley Edwards and Jack Parker-Holder and Yuge Shi and Edward Hughes and Matthew Lai and Aditi Mavalankar and Richie Steigerwald and Chris Apps and Yusuf Aytar and Sarah Maria Elisabeth Bechtle and Feryal Behbahani and Stephanie C.Y. Chan and Nicolas Heess and Lucy Gonzalez and Simon Osindero and Sherjil Ozair and Scott Reed and Jingwei Zhang and Konrad Zolna and Jeff Clune and Nando de Freitas and Satinder Singh and Tim Rockt{\""a}schel},\n booktitle={Forty-first International Conference on Machine Learning},\n year={2024},\n url={https://openreview.net/forum?id=bJbSbJskOS}\n}\n```\n",markdown,tab
|
| 8 |
+
7,188019111,"README.md",2726,0,"",markdown,selection_command
|
| 9 |
+
8,188019403,"README.md",2810,0,"\n",markdown,content
|
| 10 |
+
9,188019614,"README.md",2811,0," - Modularized training script for easy inspection using notebooks ([demo notebook](https://colab.research.google.com/drive/1zHkciFIZxXloJgue9F5LtFlA0m00rJIf?usp=sharing))",markdown,content
|
| 11 |
+
10,188019618,"README.md",2985,0,"",markdown,selection_keyboard
|
| 12 |
+
11,188020348,"README.md",2984,0,"",markdown,selection_command
|
| 13 |
+
12,188020659,"README.md",2815,0,"",markdown,selection_command
|
| 14 |
+
13,188020891,"README.md",2811,4,"",markdown,content
|
| 15 |
+
14,188033002,"README.md",0,0,"<h1 align=""center"">🧞♀️ Jasmine: A simple, performant and scalable JAX-based world modeling codebase 🧞♀️</h1>\n\n<p align=""center"">\n <a href= ""https://github.com/FLAIROx/jafar/blob/main/LICENSE"">\n <img src=""https://img.shields.io/badge/license-Apache2.0-blue.svg"" /></a>\n <a href= ""https://github.com/psf/black"">\n <img src=""https://img.shields.io/badge/code%20style-black-000000.svg"" /></a>\n</p>\n\nJasmine is a production-ready JAX-based world modeling codebase. It currently implements the high-level architecture of [Genie: Generative Interactive Environments](https://arxiv.org/abs/2402.15391) (Bruce et al., 2024) with [MaskGIT](https://arxiv.org/abs/2202.04200) (Chang et al., 2022), as well as an autoregressive (causal) baseline. A diffusion baseline is coming soon.\n\nJasmine scales from single hosts to hundreds of xPUs thanks to XLA and strives to be an easily hackable, batteries-included foundation for world modeling research.\n\n<h2 name=""overview"" id=""overview"">Overview</h2>\n\n- Asynchronous & distributed checkpointing thanks to [orbax.checkpoint](https://github.com/google/orbax)\n - Jasmine also supports mixing and matching hardware topologies (e.g. train on four nodes, load the checkpoint on a single node)\n- Optimized dataloading thanks to [Grain](https://github.com/google/grain)\n - Dataloading scales with the number of processes (i.e. nodes/xPUs)\n- Checkpointing of model weights, optimizer and dataloader states\n- Full reproducibility with **identical** training curves (thanks to seeded dataloading and training, and [JAX' approach to pseudo random numbers](https://docs.jax.dev/en/latest/random-numbers.html))\n- Automatic checkpoint deletion/retention according to specified retention policy thanks to `orbax.checkpoint.CheckpointManager`\n- Mixed precision training using `bfloat16`\n - `int8` training is on the roadmap via [aqt](https://github.com/google/aqt)\n- FlashAttention thanks to [cuDNN SDPA](https://github.com/jax-ml/jax/blob/a155c5a9997924170e0067d552351a9833c12c11/jax/_src/cudnn/fused_attention_stablehlo.py#L842)\n- Frame-level KV cache resets for accelerated spatiotemporal attention in causal baseline (still in PR)\n- Activation checkpointing (even onto host memory if desired)\n- DDP (changing to FSDP requires changing **a single line of code**)\n- WSD learning rate schedule\n - No need to retrain from scratch if you want to train for longer\n- Index-shuffling during dataloading\n- Google-native stack\n - https://github.com/google/orbax for checkpointing\n - https://github.com/google/grain for dataloading\n - https://github.com/google-deepmind/dm_pix for image manipulation\n - https://github.com/google/array_record as the data format\n- Easy model inspection thanks to [treescope](https://github.com/google-deepmind/treescope)\n- Modularized training script for easy inspection using notebooks ([demo notebook](https://colab.research.google.com/drive/1zHkciFIZxXloJgue9F5LtFlA0m00rJIf?usp=sharing))\n- Easy model surgery thanks to the new [flax.nnx](https://flax.readthedocs.io/en/latest/migrating/linen_to_nnx.html) API\n- [Shape suffixes](https://medium.com/@NoamShazeer/shape-suffixes-good-coding-style-f836e72e24fd) throughout the repository\n\n<h2 name=""start"" id=""start"">Setup 🧗</h2>\n\nJasmine requires `python 3.10`, `jax 0.6.2`, and `flax 0.10.7`. To install the requirements, run:\n\n```bash\npip install -r requirements.txt\npre-commit install\n```\n\n---\n\n<h2 name=""dataset"" id=""dataset"">Dataset 📂</h2>\n\nYou can either download our preprocessed dataset from [Hugging Face](https://huggingface.co/datasets/p-doom/open_ai_minecraft_arrayrecords_chunked) or preprocess [OpenAI's VPT dataset](https://github.com/openai/Video-Pre-Training) manually.\n\n### Option 1: Use Preprocessed Dataset (Recommended)\n\nThe easiest way to get started is to download our preprocessed dataset from Hugging Face. This script will handle downloading and extracting it:\n\n```bash\nbash input_pipeline/download/download_array_records.sh\n```\n\n---\n\n### Option 2: Manual Download & Preprocessing of OpenAI's VPT Dataset\n\nIf you prefer to use the raw VPT dataset from OpenAI and preprocess it yourself, follow these steps:\n\n1. **Download index files:**\n This will download the initial index file:\n\n ```bash\n bash input_pipeline/download/openai/download_index_files.sh\n ```\n\n2. **Download from all index files:**\n This may take a long time depending on your bandwidth:\n\n ```bash\n python input_pipeline/download/openai/download_videos.py --index_file_path data/open_ai_index_files/all_7xx_Apr_6.json\n python input_pipeline/download/openai/download_videos.py --index_file_path data/open_ai_index_files/all_8xx_Jun_29.json\n python input_pipeline/download/openai/download_videos.py --index_file_path data/open_ai_index_files/all_9xx_Jun_29.json\n python input_pipeline/download/openai/download_videos.py --index_file_path data/open_ai_index_files/all_10xx_Jun_29.json\n ```\n\n3. **Preprocess videos into ArrayRecords:**\n For efficient distributed training, convert the raw videos into the arrayrecord format (make sure to have [ffmpeg](https://github.com/FFmpeg/FFmpeg) installed on your machine):\n\n ```bash\n python input_pipeline/preprocess/video_to_array_records.py\n ```\n\n> **Note:** This is a large dataset and may take considerable time and storage to download and process.\n\n\n<h2 name=""train"" id=""train"">Quick Start 🚀 </h2>\n\nGenie has three components: a [video tokenizer](models/tokenizer.py), a [latent action model](models/lam.py), and a [dynamics model](models/dynamics.py). Each of these components are trained separately, however, the dynamics model requires a pre-trained video tokenizer (and latent action model).\n\nTo train the video tokenizer, run:\n\n```bash\npython train_tokenizer.py --ckpt_dir <path>\n```\n\nTo train the latent action model, run:\n\n```bash\npython train_lam.py --ckpt_dir <path>\n```\n\nOnce the tokenizer and LAM are trained, the dynamics model can be trained with:\n\n```bash\npython train_dynamics.py --tokenizer_checkpoint <path> --lam_checkpoint <path>\n```\n\nLogging with `wandb` is supported. To enable logging, set the `WANDB_API_KEY` environment variable or run:\n\n```bash\nwandb login\n```\n\nTraining can then be logged by setting the `--log` flag:\n\n```bash\npython train_tokenizer.py --log --entity <wandb-entity> --project <wandb-project>\n```\n\n<h2 name=""cite"" id=""cite"">Citing 📜 </h2>\n\nJasmine was built by [Mihir Mahajan](https://maharajamihir.github.io/), [Alfred Nguyen](https://avocadoali.github.io/) and [Franz Srambical](https://srambical.fr/), but started as a fork of [Jafar](https://github.com/flairox/jafar), built by [Matthew Jackson](https://matthewtjackson.com) and [Timon Willi](https://www.timonwilli.com).\n\nIf you use Jasmine in your work, please cite us, Jafar, and the original Genie paper as follows:\n\n```\n@article{\n mahajan2025jasmine,\n title={Jasmine: A simple, performant and scalable JAX-based world modeling codebase},\n author={Mihir Mahajan and Alfred Nguyen and Franz Srambical and Stefan Bauer},\n journal = {p(doom) blog},\n year={2025},\n url={https://pdoom.org/jasmine.html},\n note = {https://pdoom.org/blog.html}\n}\n```\n```\n@inproceedings{\n willi2024jafar,\n title={Jafar: An Open-Source Genie Reimplemention in Jax},\n author={Timon Willi and Matthew Thomas Jackson and Jakob Nicolaus Foerster},\n booktitle={First Workshop on Controllable Video Generation @ ICML 2024},\n year={2024},\n url={https://openreview.net/forum?id=ZZGaQHs9Jb}\n}\n```\n```\n@inproceedings{\n bruce2024genie,\n title={Genie: Generative Interactive Environments},\n author={Jake Bruce and Michael D Dennis and Ashley Edwards and Jack Parker-Holder and Yuge Shi and Edward Hughes and Matthew Lai and Aditi Mavalankar and Richie Steigerwald and Chris Apps and Yusuf Aytar and Sarah Maria Elisabeth Bechtle and Feryal Behbahani and Stephanie C.Y. Chan and Nicolas Heess and Lucy Gonzalez and Simon Osindero and Sherjil Ozair and Scott Reed and Jingwei Zhang and Konrad Zolna and Jeff Clune and Nando de Freitas and Satinder Singh and Tim Rockt{\""a}schel},\n booktitle={Forty-first International Conference on Machine Learning},\n year={2024},\n url={https://openreview.net/forum?id=bJbSbJskOS}\n}\n```\n",markdown,tab
|
| 16 |
+
15,188033071,"README.md",2811,0,"",markdown,selection_command
|
| 17 |
+
16,189015678,"README.md",0,0,"",markdown,tab
|
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-1f52a67d-2cca-4352-b39c-6cd9f3effae01765648091845-2025_12_13-18.48.27.228/source.csv
ADDED
|
@@ -0,0 +1,66 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,3,"crates/cli/src/main.rs",0,0,"//! CLI tool for serializing crowd-pilot IDE interaction data.\n//!\n//! This tool processes CSV session files and outputs JSONL format suitable for\n//! NeMo SFT training. It uses an embedded Python interpreter to load HuggingFace\n//! tokenizers for accurate token counting.\n\nuse std::path::PathBuf;\n\nuse clap::Parser;\nuse pyo3::prelude::*;\nuse pyo3::types::PyModule;\n\nuse crowd_pilot_serializer_core::{\n pipeline::{PipelineConfig, PipelineResult},\n process_all_sessions, write_jsonl_output, Tokenizer,\n};\n\n/// Serialize crowd-pilot CSV sessions to NeMo JSONL format.\n#[derive(Parser, Debug)]\n#[command(name = ""crowd-pilot-serialize"")]\n#[command(author, version, about, long_about = None)]\nstruct Args {\n /// Root directory containing CSV session files\n #[arg(long)]\n csv_root: PathBuf,\n\n /// Output directory for JSONL files\n #[arg(long)]\n output_dir: PathBuf,\n\n /// HuggingFace tokenizer model name or path\n #[arg(long)]\n tokenizer: String,\n\n /// Maximum tokens per conversation chunk\n #[arg(long, default_value = ""8192"")]\n max_tokens_per_conversation: usize,\n\n /// Maximum tokens per message\n #[arg(long, default_value = ""2048"")]\n max_tokens_per_message: usize,\n\n /// Minimum messages required to keep a conversation\n #[arg(long, default_value = ""5"")]\n min_conversation_messages: usize,\n\n /// Viewport radius (lines above/below cursor)\n #[arg(long, default_value = ""10"")]\n viewport_radius: usize,\n\n /// Coalesce radius for grouping nearby edits\n #[arg(long, default_value = ""5"")]\n coalesce_radius: usize,\n\n /// Fraction of sessions for validation (0.0-1.0)\n #[arg(long, default_value = ""0.1"")]\n val_ratio: f64,\n\n /// Custom system prompt (optional)\n #[arg(long)]\n system_prompt: Option<String>,\n}\n\nconst DEFAULT_SYSTEM_PROMPT: &str = r#""You are a helpful assistant that can interact multiple times with a computer shell to solve programming tasks.\nYour response must contain exactly ONE bash code block with ONE command (or commands connected with && or ||).\n\nFormat your response as shown in <format_example>.\n\n<format_example>\n```bash\nyour_command_here\n```\n</format_example>\n\nFailure to follow these rules will cause your response to be rejected.""#;\n\n/// Wrapper around Python tokenizer for exact token counting and truncation.\nstruct PythonTokenizer {\n tokenizer: Py<PyAny>,\n}\n\nimpl PythonTokenizer {\n /// Load a HuggingFace tokenizer.\n fn load(model_name: &str) -> PyResult<Self> {\n Python::with_gil(|py| {\n let transformers = PyModule::import(py, ""transformers"")?;\n let auto_tokenizer = transformers.getattr(""AutoTokenizer"")?;\n let tokenizer = auto_tokenizer.call_method1(""from_pretrained"", (model_name,))?;\n Ok(Self {\n tokenizer: tokenizer.into(),\n })\n })\n }\n}\n\nimpl Tokenizer for PythonTokenizer {\n fn count_tokens(&self, text: &str) -> usize {\n Python::with_gil(|py| {\n let tokenizer = self.tokenizer.as_ref(py);\n let tokens = tokenizer\n .call_method1(""encode"", (text,))\n .expect(""Failed to encode text with tokenizer"");\n tokens.len().unwrap()\n })\n }\n\n fn truncate_to_max_tokens(&self, text: &str, max_tokens: usize) -> String {\n Python::with_gil(|py| {\n let tokenizer = self.tokenizer.as_ref(py);\n let kwargs = pyo3::types::PyDict::new(py);\n kwargs.set_item(""max_length"", max_tokens).unwrap();\n kwargs.set_item(""truncation"", true).unwrap();\n \n let tokens = tokenizer\n .call_method(""encode"", (text,), Some(kwargs))\n .expect(""Failed to encode text with tokenizer"");\n \n tokenizer\n .call_method1(""decode"", (tokens,))\n .expect(""Failed to decode tokens"")\n .extract()\n .unwrap()\n })\n }\n}\n\nfn main() -> Result<(), Box<dyn std::error::Error>> {\n let args = Args::parse();\n\n println!(""Loading tokenizer from {}..."", args.tokenizer);\n let tokenizer = PythonTokenizer::load(&args.tokenizer)?;\n\n let config = PipelineConfig {\n max_tokens_per_conversation: args.max_tokens_per_conversation,\n max_tokens_per_message: args.max_tokens_per_message,\n min_conversation_messages: args.min_conversation_messages,\n viewport_radius: args.viewport_radius,\n coalesce_radius: args.coalesce_radius,\n val_ratio: args.val_ratio,\n };\n\n println!(""Processing CSV files from {:?}..."", args.csv_root);\n let session_results = process_all_sessions(\n &args.csv_root,\n &tokenizer,\n &config,\n )?;\n\n let total_sessions = session_results.len();\n println!(""Processed {} sessions"", total_sessions);\n\n let system_prompt = args.system_prompt.as_deref().unwrap_or(DEFAULT_SYSTEM_PROMPT);\n\n println!(""Writing output to {:?}..."", args.output_dir);\n let result: PipelineResult = write_jsonl_output(\n session_results,\n &args.output_dir,\n args.val_ratio,\n system_prompt,\n )?;\n\n let metadata_path = args.output_dir.join(""metadata.json"");\n let metadata = serde_json::json!({\n ""config"": {\n ""csv_root"": args.csv_root.to_string_lossy(),\n ""output_dir"": args.output_dir.to_string_lossy(),\n ""tokenizer"": args.tokenizer,\n ""max_tokens_per_conversation"": args.max_tokens_per_conversation,\n ""max_tokens_per_message"": args.max_tokens_per_message,\n ""min_conversation_messages"": args.min_conversation_messages,\n ""viewport_radius"": args.viewport_radius,\n ""coalesce_radius"": args.coalesce_radius,\n ""val_ratio"": args.val_ratio,\n },\n ""counts"": {\n ""total_sessions"": result.total_sessions,\n ""total_conversations"": result.total_conversations,\n ""train_conversations"": result.train_conversations,\n ""val_conversations"": result.val_conversations,\n },\n ""stats"": {\n ""total_messages"": result.total_messages,\n ""total_tokens"": result.total_tokens,\n ""avg_messages_per_conversation"": if result.total_conversations > 0 {\n result.total_messages as f64 / result.total_conversations as f64\n } else {\n 0.0\n },\n ""avg_tokens_per_conversation"": if result.total_conversations > 0 {\n result.total_tokens as f64 / result.total_conversations as f64\n } else {\n 0.0\n },\n },\n ""files"": {\n ""train_path"": args.output_dir.join(""training.jsonl"").to_string_lossy(),\n ""val_path"": args.output_dir.join(""validation.jsonl"").to_string_lossy(),\n },\n });\n std::fs::write(&metadata_path, serde_json::to_string_pretty(&metadata)?)?;\n\n println!(""\n[summary]"");\n println!("" Total sessions processed: {}"", result.total_sessions);\n println!("" Train conversations: {}"", result.train_conversations);\n println!("" Val conversations: {}"", result.val_conversations);\n println!("" Total messages: {}"", result.total_messages);\n println!("" Total tokens: {}"", result.total_tokens);\n println!("" Output: {:?}/{{training,validation}}.jsonl"", args.output_dir);\n println!("" Metadata: {:?}"", metadata_path);\n\n Ok(())\n}\n\n",rust,tab
|
| 3 |
+
2,248,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"6:48:27 PM [info] Activating crowd-code\n6:48:27 PM [info] Recording started\n6:48:27 PM [info] Initializing git provider using file system watchers...\n6:48:27 PM [info] Git repository found\n",Log,tab
|
| 4 |
+
3,345,"extension-output-pdoom-org.crowd-code-#1-crowd-code",189,0,"6:48:27 PM [info] Git provider initialized successfully\n6:48:27 PM [info] Initial git state: [object Object]\n",Log,content
|
| 5 |
+
4,3836,"TERMINAL",0,0,"",,terminal_focus
|
| 6 |
+
5,3838,"crates/cli/src/main.rs",0,0,"",rust,tab
|
| 7 |
+
6,5667,"TERMINAL",0,0,"squeue",,terminal_command
|
| 8 |
+
7,5680,"TERMINAL",0,0,"]633;C JOBID USER PARTITION NODES CPUS ST SUBMIT_TIME START_TIME TIME TIME_LIMIT NODELIST(REASON)\r\n 36313 xiao.liu interacti 1 128 R 2025-12-13T12:56:48 2025-12-13T12:56:48 5:51:44 23:59:00 hai005\r\n 36303 xiao.liu interacti 1 128 R 2025-12-12T23:25:52 2025-12-12T23:25:52 19:22:40 23:59:00 hai006\r\n 36317 mihir.maha standard 1 10 R 2025-12-13T15:27:10 2025-12-13T15:27:10 3:21:22 1-00:00:00 hai004\r\n 36314 xiao.liu standard 1 128 R 2025-12-13T13:27:57 2025-12-13T15:13:56 3:34:36 23:59:00 hai007\r\n 36304 nishant.ku standard 3 624 R 2025-12-13T07:56:43 2025-12-13T07:56:43 10:51:49 1-00:00:00 hai[001-003]\r\n]0;franz.srambical@hai-login1:~/crowd-pilot-serializer",,terminal_output
|
| 9 |
+
8,8439,"TERMINAL",0,0,"bash",,terminal_focus
|
| 10 |
+
9,9983,"TERMINAL",0,0,"bash",,terminal_focus
|
| 11 |
+
10,27972,"TERMINAL",0,0,"source .venv/bin/activate",,terminal_command
|
| 12 |
+
11,27988,"TERMINAL",0,0,"]633;C]0;franz.srambical@hai-login1:~/crowd-pilot-serializer",,terminal_output
|
| 13 |
+
12,30366,"TERMINAL",0,0,"cargo clean",,terminal_command
|
| 14 |
+
13,30367,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 15 |
+
14,30559,"TERMINAL",0,0,"[1m[96m Cleaning[0m [ ] 0.00% \r",,terminal_output
|
| 16 |
+
15,30651,"TERMINAL",0,0,"[1m[96m Cleaning[0m [ ] 0.63% \r",,terminal_output
|
| 17 |
+
16,30739,"TERMINAL",0,0,"[1m[96m Cleaning[0m [ ] 1.28% \r",,terminal_output
|
| 18 |
+
17,30856,"TERMINAL",0,0,"[1m[96m Cleaning[0m [ ] 1.91% \r",,terminal_output
|
| 19 |
+
18,31022,"TERMINAL",0,0,"[1m[96m Cleaning[0m [ ] 2.58% \r",,terminal_output
|
| 20 |
+
19,31127,"TERMINAL",0,0,"[1m[96m Cleaning[0m [ ] 3.21% \r",,terminal_output
|
| 21 |
+
20,31249,"TERMINAL",0,0,"[1m[96m Cleaning[0m [ ] 3.89% \r[1m[96m Cleaning[0m [> ] 4.58% \r",,terminal_output
|
| 22 |
+
21,31341,"TERMINAL",0,0,"[1m[96m Cleaning[0m [> ] 5.30% \r",,terminal_output
|
| 23 |
+
22,31473,"TERMINAL",0,0,"[1m[96m Cleaning[0m [> ] 6.09% \r",,terminal_output
|
| 24 |
+
23,31558,"TERMINAL",0,0,"[1m[96m Cleaning[0m [> ] 6.97% \r",,terminal_output
|
| 25 |
+
24,31713,"TERMINAL",0,0,"[1m[96m Cleaning[0m [> ] 7.82% \r",,terminal_output
|
| 26 |
+
25,31775,"TERMINAL",0,0,"[1m[96m Cleaning[0m [=> ] 8.70% \r",,terminal_output
|
| 27 |
+
26,31883,"TERMINAL",0,0,"[1m[96m Cleaning[0m [=> ] 9.57% \r",,terminal_output
|
| 28 |
+
27,32024,"TERMINAL",0,0,"[1m[96m Cleaning[0m [=> ] 10.43% \r",,terminal_output
|
| 29 |
+
28,32160,"TERMINAL",0,0,"[1m[96m Cleaning[0m [=> ] 11.28% \r",,terminal_output
|
| 30 |
+
29,32237,"TERMINAL",0,0,"[1m[96m Cleaning[0m [==> ] 12.13% \r",,terminal_output
|
| 31 |
+
30,32307,"TERMINAL",0,0,"[1m[96m Cleaning[0m [==> ] 12.99% \r",,terminal_output
|
| 32 |
+
31,32403,"TERMINAL",0,0,"[1m[96m Cleaning[0m [==> ] 13.82% \r",,terminal_output
|
| 33 |
+
32,32476,"TERMINAL",0,0,"[1m[96m Cleaning[0m [==> ] 14.65% \r",,terminal_output
|
| 34 |
+
33,32598,"TERMINAL",0,0,"[1m[96m Cleaning[0m [==> ] 15.51% \r",,terminal_output
|
| 35 |
+
34,32705,"TERMINAL",0,0,"[1m[96m Cleaning[0m [===> ] 16.36% \r",,terminal_output
|
| 36 |
+
35,32785,"TERMINAL",0,0,"[1m[96m Cleaning[0m [===> ] 17.21% \r",,terminal_output
|
| 37 |
+
36,32907,"TERMINAL",0,0,"[1m[96m Cleaning[0m [===> ] 18.07% \r",,terminal_output
|
| 38 |
+
37,32989,"TERMINAL",0,0,"[1m[96m Cleaning[0m [===> ] 18.92% \r",,terminal_output
|
| 39 |
+
38,33122,"TERMINAL",0,0,"[1m[96m Cleaning[0m [===> ] 19.78% \r",,terminal_output
|
| 40 |
+
39,33221,"TERMINAL",0,0,"[1m[96m Cleaning[0m [====> ] 20.63% \r",,terminal_output
|
| 41 |
+
40,33352,"TERMINAL",0,0,"[1m[96m Cleaning[0m [====> ] 21.48% \r",,terminal_output
|
| 42 |
+
41,33419,"TERMINAL",0,0,"[1m[96m Cleaning[0m [====> ] 22.16% \r",,terminal_output
|
| 43 |
+
42,33518,"TERMINAL",0,0,"[1m[96m Cleaning[0m [====> ] 22.79% \r",,terminal_output
|
| 44 |
+
43,33637,"TERMINAL",0,0,"[1m[96m Cleaning[0m [====> ] 23.44% \r",,terminal_output
|
| 45 |
+
44,33812,"TERMINAL",0,0,"[1m[96m Cleaning[0m [=====> ] 24.07% \r",,terminal_output
|
| 46 |
+
45,33894,"TERMINAL",0,0,"[1m[96m Cleaning[0m [=====> ] 24.67% \r[1m[96m Cleaning[0m [=====> ] 25.37% \r",,terminal_output
|
| 47 |
+
46,34019,"TERMINAL",0,0,"[1m[96m Cleaning[0m [=====> ] 26.02% \r",,terminal_output
|
| 48 |
+
47,34217,"TERMINAL",0,0,"[1m[96m Cleaning[0m [=====> ] 26.72% \r[1m[96m Cleaning[0m [=====> ] 27.33% \r",,terminal_output
|
| 49 |
+
48,34297,"TERMINAL",0,0,"[1m[96m Cleaning[0m [======> ] 28.02% \r",,terminal_output
|
| 50 |
+
49,34407,"TERMINAL",0,0,"[1m[96m Cleaning[0m [======> ] 28.74% \r",,terminal_output
|
| 51 |
+
50,34526,"TERMINAL",0,0,"[1m[96m Cleaning[0m [======> ] 29.42% \r",,terminal_output
|
| 52 |
+
51,34743,"TERMINAL",0,0,"[1m[96m Cleaning[0m [======> ] 30.09% \r",,terminal_output
|
| 53 |
+
52,34746,"TERMINAL",0,0,"[1m[96m Cleaning[0m [======> ] 30.76% \r",,terminal_output
|
| 54 |
+
53,34813,"TERMINAL",0,0,"[1m[96m Cleaning[0m [======> ] 31.42% \r",,terminal_output
|
| 55 |
+
54,34910,"TERMINAL",0,0,"[1m[96m Cleaning[0m [=======> ] 32.07% \r",,terminal_output
|
| 56 |
+
55,35053,"TERMINAL",0,0,"[1m[96m Cleaning[0m [=======> ] 32.72% \r",,terminal_output
|
| 57 |
+
56,35073,"TERMINAL",0,0,"[1m[96m Cleaning[0m [=======> ] 33.33% \r[K[1m[91merror[0m: failed to remove file `/fast/home/franz.srambical/crowd-pilot-serializer/target/debug/deps/.nfs000000012d584f9d00000407`\r\n\r\nCaused by:\r\n Device or resource busy (os error 16)\r\n\r\nError: failed to remove directory `/fast/home/franz.srambical/crowd-pilot-serializer/target/debug/deps`\r\n\r\nCaused by:\r\n Device or resource busy (os error 16)\r\n]0;franz.srambical@hai-login1:~/crowd-pilot-serializer",,terminal_output
|
| 58 |
+
57,41510,"TERMINAL",0,0,"bash",,terminal_focus
|
| 59 |
+
58,43518,"TERMINAL",0,0,"exit",,terminal_command
|
| 60 |
+
59,44223,"TERMINAL",0,0,"]633;Cexit\r\n",,terminal_output
|
| 61 |
+
60,44226,"TERMINAL",0,0,"bash",,terminal_focus
|
| 62 |
+
61,52525,"TERMINAL",0,0,"rm -rf target/",,terminal_command
|
| 63 |
+
62,52574,"TERMINAL",0,0,"]633;Crm: cannot remove 'target/debug/deps/.nfs000000012d584f9d00000407': Device or resource busy\r\nrm: cannot remove 'target/debug/deps/.nfs000000012ce413230000040b': Device or resource busy\r\nrm: cannot remove 'target/debug/deps/.nfs000000012cf90e1000000408': Device or resource busy\r\nrm: cannot remove 'target/debug/deps/.nfs000000012cb3183b0000040a': Device or resource busy\r\nrm: cannot remove 'target/debug/deps/.nfs000000012c0f154900000406': Device or resource busy\r\nrm: cannot remove 'target/debug/deps/.nfs000000012c166c5400000409': Device or resource busy\r\nrm: cannot remove 'target/debug/deps/.nfs000000012d1819c400000405': Device or resource busy\r\n",,terminal_output
|
| 64 |
+
63,56197,"TERMINAL",0,0,"^C\r\n]0;franz.srambical@hai-login1:~/crowd-pilot-serializer",,terminal_output
|
| 65 |
+
64,59615,"TERMINAL",0,0,"squeue",,terminal_command
|
| 66 |
+
65,59616,"TERMINAL",0,0,"]633;C JOBID USER PARTITION NODES CPUS ST SUBMIT_TIME START_TIME TIME TIME_LIMIT NODELIST(REASON)\r\n 36313 xiao.liu interacti 1 128 R 2025-12-13T12:56:48 2025-12-13T12:56:48 5:52:38 23:59:00 hai005\r\n 36303 xiao.liu interacti 1 128 R 2025-12-12T23:25:52 2025-12-12T23:25:52 19:23:34 23:59:00 hai006\r\n 36317 mihir.maha standard 1 10 R 2025-12-13T15:27:10 2025-12-13T15:27:10 3:22:16 1-00:00:00 hai004\r\n 36314 xiao.liu standard 1 128 R 2025-12-13T13:27:57 2025-12-13T15:13:56 3:35:30 23:59:00 hai007\r\n 36304 nishant.ku standard 3 624 R 2025-12-13T07:56:43 2025-12-13T07:56:43 10:52:43 1-00:00:00 hai[001-003]\r\n]0;franz.srambical@hai-login1:~/crowd-pilot-serializer",,terminal_output
|
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-2790c5c3-9c96-49ac-ab9e-481a8033e7cd1757945282635-2025_09_15-16.08.09.625/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-9543d8e2-2376-4957-873e-df7016d502961763465687199-2025_11_18-12.34.49.442/source.csv
ADDED
|
@@ -0,0 +1,245 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,67,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"12:34:49 PM [info] Activating crowd-code\n12:34:49 PM [info] Recording started\n12:34:49 PM [info] Initializing git provider using file system watchers...\n12:34:49 PM [info] Git repository found\n12:34:49 PM [info] Git provider initialized successfully\n",Log,tab
|
| 3 |
+
3,116,"extension-output-pdoom-org.crowd-code-#1-crowd-code",250,0,"12:34:49 PM [info] Initial git state: [object Object]\n",Log,content
|
| 4 |
+
4,2975,"extension-output-pdoom-org.crowd-code-#1-crowd-code",304,0,"",Log,selection_mouse
|
| 5 |
+
5,13030,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,tab
|
| 6 |
+
6,20469,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",64,0,"",python,selection_command
|
| 7 |
+
7,20661,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",66,0,"",python,selection_command
|
| 8 |
+
8,20694,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",132,0,"",python,selection_command
|
| 9 |
+
9,20728,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",199,0,"",python,selection_command
|
| 10 |
+
10,20761,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",241,0,"",python,selection_command
|
| 11 |
+
11,20794,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",243,0,"",python,selection_command
|
| 12 |
+
12,20828,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",292,0,"",python,selection_command
|
| 13 |
+
13,20863,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",294,0,"",python,selection_command
|
| 14 |
+
14,20904,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",364,0,"",python,selection_command
|
| 15 |
+
15,20936,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",432,0,"",python,selection_command
|
| 16 |
+
16,20962,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",507,0,"",python,selection_command
|
| 17 |
+
17,20996,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",577,0,"",python,selection_command
|
| 18 |
+
18,21030,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",610,0,"",python,selection_command
|
| 19 |
+
19,21068,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",611,0,"",python,selection_command
|
| 20 |
+
20,21097,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",639,0,"",python,selection_command
|
| 21 |
+
21,21131,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",640,0,"",python,selection_command
|
| 22 |
+
22,21166,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",671,0,"",python,selection_command
|
| 23 |
+
23,21203,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",694,0,"",python,selection_command
|
| 24 |
+
24,21233,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",707,0,"",python,selection_command
|
| 25 |
+
25,21265,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",795,0,"",python,selection_command
|
| 26 |
+
26,21297,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",796,0,"",python,selection_command
|
| 27 |
+
27,21331,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",852,0,"",python,selection_command
|
| 28 |
+
28,21364,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",914,0,"",python,selection_command
|
| 29 |
+
29,21397,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",965,0,"",python,selection_command
|
| 30 |
+
30,21431,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1047,0,"",python,selection_command
|
| 31 |
+
31,21464,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1148,0,"",python,selection_command
|
| 32 |
+
32,21497,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1245,0,"",python,selection_command
|
| 33 |
+
33,21530,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1319,0,"",python,selection_command
|
| 34 |
+
34,21565,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1369,0,"",python,selection_command
|
| 35 |
+
35,21598,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1370,0,"",python,selection_command
|
| 36 |
+
36,21631,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1393,0,"",python,selection_command
|
| 37 |
+
37,21668,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1394,0,"",python,selection_command
|
| 38 |
+
38,21701,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1395,0,"",python,selection_command
|
| 39 |
+
39,21890,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1423,0,"",python,selection_command
|
| 40 |
+
40,23050,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1427,0,"",python,selection_command
|
| 41 |
+
41,23197,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1432,0,"",python,selection_command
|
| 42 |
+
42,23489,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1435,0,"",python,selection_command
|
| 43 |
+
43,23702,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1438,0,"",python,selection_command
|
| 44 |
+
44,23912,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1441,0,"",python,selection_command
|
| 45 |
+
45,24119,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1442,0,"",python,selection_command
|
| 46 |
+
46,24327,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1448,0,"",python,selection_command
|
| 47 |
+
47,24581,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1449,0,"",python,selection_command
|
| 48 |
+
48,24945,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1451,0,"",python,selection_command
|
| 49 |
+
49,25331,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1452,0,"",python,selection_command
|
| 50 |
+
50,26280,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1423,0,"",python,selection_command
|
| 51 |
+
51,26667,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1395,0,"",python,selection_command
|
| 52 |
+
52,26920,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1394,0,"",python,selection_command
|
| 53 |
+
53,26954,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1393,0,"",python,selection_command
|
| 54 |
+
54,146021,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
|
| 55 |
+
55,146654,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,tab
|
| 56 |
+
56,203332,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2838,0,"",python,selection_keyboard
|
| 57 |
+
57,203923,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",5124,0,"",python,selection_keyboard
|
| 58 |
+
58,204398,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7126,0,"",python,selection_keyboard
|
| 59 |
+
59,204829,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9057,0,"",python,selection_keyboard
|
| 60 |
+
60,205408,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_keyboard
|
| 61 |
+
61,210105,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7721,0,"",python,selection_keyboard
|
| 62 |
+
62,210235,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",5922,0,"",python,selection_keyboard
|
| 63 |
+
63,210382,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3490,0,"",python,selection_keyboard
|
| 64 |
+
64,210771,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1906,0,"",python,selection_keyboard
|
| 65 |
+
65,210944,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",64,0,"",python,selection_keyboard
|
| 66 |
+
66,211099,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_keyboard
|
| 67 |
+
67,214839,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10,0,"",python,selection_command
|
| 68 |
+
68,2267039,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2294,0,"",python,selection_mouse
|
| 69 |
+
69,2272461,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2114,0,"",python,selection_mouse
|
| 70 |
+
70,3052561,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,10041,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,selection_command
|
| 71 |
+
71,3052668,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_command
|
| 72 |
+
72,3145153,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_command
|
| 73 |
+
73,3148126,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7487,0,"",python,selection_command
|
| 74 |
+
74,3149770,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7876,0,"",python,selection_command
|
| 75 |
+
75,3150226,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",8235,0,"",python,selection_command
|
| 76 |
+
76,3150898,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7876,0,"",python,selection_command
|
| 77 |
+
77,3151966,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",8235,0,"",python,selection_command
|
| 78 |
+
78,3161866,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9070,0,"",python,selection_command
|
| 79 |
+
79,3164224,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9131,0,"",python,selection_command
|
| 80 |
+
80,3166810,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9194,0,"",python,selection_command
|
| 81 |
+
81,3167056,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9257,0,"",python,selection_command
|
| 82 |
+
82,3167090,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9321,0,"",python,selection_command
|
| 83 |
+
83,3167123,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9386,0,"",python,selection_command
|
| 84 |
+
84,3167156,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9443,0,"",python,selection_command
|
| 85 |
+
85,3167189,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9481,0,"",python,selection_command
|
| 86 |
+
86,3167223,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9531,0,"",python,selection_command
|
| 87 |
+
87,3168286,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9532,0,"",python,selection_command
|
| 88 |
+
88,3168527,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9538,0,"",python,selection_command
|
| 89 |
+
89,3168566,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9539,0,"",python,selection_command
|
| 90 |
+
90,3168597,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9546,0,"",python,selection_command
|
| 91 |
+
91,3168633,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9547,0,"",python,selection_command
|
| 92 |
+
92,3168887,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9555,0,"",python,selection_command
|
| 93 |
+
93,3169118,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
|
| 94 |
+
94,3169292,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9583,0,"",python,selection_command
|
| 95 |
+
95,3169900,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
|
| 96 |
+
96,3170394,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
|
| 97 |
+
97,3171165,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
|
| 98 |
+
98,3171795,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
|
| 99 |
+
99,3171945,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
|
| 100 |
+
100,3172433,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
|
| 101 |
+
101,3173957,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
|
| 102 |
+
102,3174398,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
|
| 103 |
+
103,3174801,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9682,0,"",python,selection_command
|
| 104 |
+
104,3175123,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9746,0,"",python,selection_command
|
| 105 |
+
105,3175686,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9682,0,"",python,selection_command
|
| 106 |
+
106,3175821,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
|
| 107 |
+
107,3175981,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
|
| 108 |
+
108,3240534,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_command
|
| 109 |
+
109,3241210,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1898,0,"",python,selection_keyboard
|
| 110 |
+
110,3241373,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3420,0,"",python,selection_keyboard
|
| 111 |
+
111,3242843,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3370,0,"",python,selection_command
|
| 112 |
+
112,3243091,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3298,0,"",python,selection_command
|
| 113 |
+
113,3243124,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3288,0,"",python,selection_command
|
| 114 |
+
114,3243158,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3287,0,"",python,selection_command
|
| 115 |
+
115,3243191,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3211,0,"",python,selection_command
|
| 116 |
+
116,3243225,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3137,0,"",python,selection_command
|
| 117 |
+
117,3243258,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3058,0,"",python,selection_command
|
| 118 |
+
118,3243291,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3057,0,"",python,selection_command
|
| 119 |
+
119,3243324,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2999,0,"",python,selection_command
|
| 120 |
+
120,3243357,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2991,0,"",python,selection_command
|
| 121 |
+
121,3243391,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2973,0,"",python,selection_command
|
| 122 |
+
122,3243424,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2956,0,"",python,selection_command
|
| 123 |
+
123,3243458,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2932,0,"",python,selection_command
|
| 124 |
+
124,3243491,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2915,0,"",python,selection_command
|
| 125 |
+
125,3243525,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2898,0,"",python,selection_command
|
| 126 |
+
126,3243558,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2876,0,"",python,selection_command
|
| 127 |
+
127,3243591,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2854,0,"",python,selection_command
|
| 128 |
+
128,3243632,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2838,0,"",python,selection_command
|
| 129 |
+
129,3243658,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2817,0,"",python,selection_command
|
| 130 |
+
130,3243691,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2793,0,"",python,selection_command
|
| 131 |
+
131,3243734,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2767,0,"",python,selection_command
|
| 132 |
+
132,3243758,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2756,0,"",python,selection_command
|
| 133 |
+
133,3243791,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2721,0,"",python,selection_command
|
| 134 |
+
134,3243824,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2688,0,"",python,selection_command
|
| 135 |
+
135,3243857,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2655,0,"",python,selection_command
|
| 136 |
+
136,3243891,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2621,0,"",python,selection_command
|
| 137 |
+
137,3243925,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2585,0,"",python,selection_command
|
| 138 |
+
138,3243958,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2547,0,"",python,selection_command
|
| 139 |
+
139,3243991,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2512,0,"",python,selection_command
|
| 140 |
+
140,3244025,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2483,0,"",python,selection_command
|
| 141 |
+
141,3244059,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2451,0,"",python,selection_command
|
| 142 |
+
142,3244092,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2427,0,"",python,selection_command
|
| 143 |
+
143,3244125,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2388,0,"",python,selection_command
|
| 144 |
+
144,3286103,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2411,0,"",python,selection_command
|
| 145 |
+
145,3296751,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2388,23,"",python,content
|
| 146 |
+
146,3298859,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2388,0," sequence_parallelis",python,content
|
| 147 |
+
147,3298862,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2388,0,"",python,selection_command
|
| 148 |
+
148,3483131,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
|
| 149 |
+
149,3484741,"TERMINAL",0,0,"",,terminal_focus
|
| 150 |
+
150,3484752,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,tab
|
| 151 |
+
151,3788287,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2427,0,"",python,selection_command
|
| 152 |
+
152,3788457,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2428,0,"",python,selection_command
|
| 153 |
+
153,3788946,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2429,0,"",python,selection_command
|
| 154 |
+
154,3789583,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2427,0,"",python,selection_command
|
| 155 |
+
155,4549882,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,10041,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,selection_command
|
| 156 |
+
156,4550471,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_command
|
| 157 |
+
157,4559897,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,10041,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,selection_command
|
| 158 |
+
158,4560021,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_command
|
| 159 |
+
159,4560890,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,content
|
| 160 |
+
160,4583906,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18686,0,"",python,selection_mouse
|
| 161 |
+
161,4583914,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18685,0,"",python,selection_command
|
| 162 |
+
162,4584529,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18699,0,"",python,selection_command
|
| 163 |
+
163,4585394,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18743,0,"",python,selection_command
|
| 164 |
+
164,4585610,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18752,0,"",python,selection_command
|
| 165 |
+
165,4591082,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,tab
|
| 166 |
+
166,4591147,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",64,0,"",python,selection_command
|
| 167 |
+
167,4595172,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,10041,"",python,content
|
| 168 |
+
168,4597038,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,tab
|
| 169 |
+
169,5454427,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9372,0,"",python,selection_mouse
|
| 170 |
+
170,5454796,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_command
|
| 171 |
+
171,5455466,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",64,0,"",python,selection_command
|
| 172 |
+
172,5455793,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1906,0,"",python,selection_keyboard
|
| 173 |
+
173,5456269,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3490,0,"",python,selection_keyboard
|
| 174 |
+
174,5456884,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3420,0,"",python,selection_command
|
| 175 |
+
175,5457139,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3370,0,"",python,selection_command
|
| 176 |
+
176,5457172,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3298,0,"",python,selection_command
|
| 177 |
+
177,5457202,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3288,0,"",python,selection_command
|
| 178 |
+
178,5457232,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3287,0,"",python,selection_command
|
| 179 |
+
179,5457265,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3211,0,"",python,selection_command
|
| 180 |
+
180,5457298,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3137,0,"",python,selection_command
|
| 181 |
+
181,5457331,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3058,0,"",python,selection_command
|
| 182 |
+
182,5457367,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3057,0,"",python,selection_command
|
| 183 |
+
183,5457402,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2999,0,"",python,selection_command
|
| 184 |
+
184,5457431,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2991,0,"",python,selection_command
|
| 185 |
+
185,5457465,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2973,0,"",python,selection_command
|
| 186 |
+
186,5457497,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2956,0,"",python,selection_command
|
| 187 |
+
187,5457533,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2932,0,"",python,selection_command
|
| 188 |
+
188,5457565,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2915,0,"",python,selection_command
|
| 189 |
+
189,5457598,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2898,0,"",python,selection_command
|
| 190 |
+
190,5457631,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2876,0,"",python,selection_command
|
| 191 |
+
191,5457665,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2854,0,"",python,selection_command
|
| 192 |
+
192,5457699,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2838,0,"",python,selection_command
|
| 193 |
+
193,5457732,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2817,0,"",python,selection_command
|
| 194 |
+
194,5457764,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2793,0,"",python,selection_command
|
| 195 |
+
195,5457799,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2767,0,"",python,selection_command
|
| 196 |
+
196,5457833,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2756,0,"",python,selection_command
|
| 197 |
+
197,5457866,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2721,0,"",python,selection_command
|
| 198 |
+
198,5457899,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2688,0,"",python,selection_command
|
| 199 |
+
199,5457931,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2655,0,"",python,selection_command
|
| 200 |
+
200,5457964,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2621,0,"",python,selection_command
|
| 201 |
+
201,5457999,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2585,0,"",python,selection_command
|
| 202 |
+
202,5458033,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2547,0,"",python,selection_command
|
| 203 |
+
203,5458067,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2512,0,"",python,selection_command
|
| 204 |
+
204,5458102,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2483,0,"",python,selection_command
|
| 205 |
+
205,5458137,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2451,0,"",python,selection_command
|
| 206 |
+
206,5459402,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2483,0,"",python,selection_command
|
| 207 |
+
207,5459593,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2512,0,"",python,selection_command
|
| 208 |
+
208,5459748,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2547,0,"",python,selection_command
|
| 209 |
+
209,5460152,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2551,0,"",python,selection_command
|
| 210 |
+
210,5461248,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2589,0,"",python,selection_command
|
| 211 |
+
211,5461431,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2625,0,"",python,selection_command
|
| 212 |
+
212,5461588,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2659,0,"",python,selection_command
|
| 213 |
+
213,5461782,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2692,0,"",python,selection_command
|
| 214 |
+
214,5462187,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2725,0,"",python,selection_command
|
| 215 |
+
215,5462694,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2692,0,"",python,selection_command
|
| 216 |
+
216,5462868,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2659,0,"",python,selection_command
|
| 217 |
+
217,5463112,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2625,0,"",python,selection_command
|
| 218 |
+
218,5463146,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2589,0,"",python,selection_command
|
| 219 |
+
219,5463184,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2551,0,"",python,selection_command
|
| 220 |
+
220,5463214,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2516,0,"",python,selection_command
|
| 221 |
+
221,5463329,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2487,0,"",python,selection_command
|
| 222 |
+
222,8520791,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,10041,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,selection_command
|
| 223 |
+
223,8520860,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_command
|
| 224 |
+
224,8685736,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2797,0,"",python,selection_command
|
| 225 |
+
225,8686757,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2813,0,"",python,selection_command
|
| 226 |
+
226,8686946,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2814,0,"",python,selection_command
|
| 227 |
+
227,10758031,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2813,0,"",python,selection_command
|
| 228 |
+
228,10758522,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2797,0,"",python,selection_command
|
| 229 |
+
229,10759094,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2813,0,"",python,selection_command
|
| 230 |
+
230,10759461,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2797,0,"",python,selection_command
|
| 231 |
+
231,10762705,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2771,0,"",python,selection_command
|
| 232 |
+
232,10880161,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2797,0,"",python,selection_command
|
| 233 |
+
233,14746153,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9275,0,"",python,selection_mouse
|
| 234 |
+
234,14750022,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9465,0,"",python,selection_mouse
|
| 235 |
+
235,14757418,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9932,0,"",python,selection_mouse
|
| 236 |
+
236,14758360,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9464,0,"",python,selection_mouse
|
| 237 |
+
237,15922202,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_command
|
| 238 |
+
238,15924565,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1008,0,"",python,selection_command
|
| 239 |
+
239,15925176,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1016,0,"",python,selection_command
|
| 240 |
+
240,15925334,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1023,0,"",python,selection_command
|
| 241 |
+
241,15925510,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1047,0,"",python,selection_command
|
| 242 |
+
242,15925934,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1023,0,"",python,selection_command
|
| 243 |
+
243,15926409,"nemo/collections/llm/recipes/finetune_default.py",0,0,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import TYPE_CHECKING, Any, Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\n\nimport nemo.lightning as nl\nfrom nemo.collections import llm\nfrom nemo.collections.llm.gpt.data.packed_sequence import PackedSequenceSpecs\nfrom nemo.collections.llm.peft import DoRA, LoRA\nfrom nemo.collections.llm.recipes.log.default import tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.precision.mixed_precision import bf16_mixed\nfrom nemo.lightning.pytorch.callbacks import PEFT\nfrom nemo.utils.exp_manager import TimingCallback\n\nif TYPE_CHECKING:\n from lightning.pytorch.loggers import TensorBoardLogger, WandbLogger\n\nTokenizerType = Any\n\n\ndef default_finetune_recipe(\n model: run.Config[pl.LightningModule],\n resume_path: str,\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n packed_sequence: bool = False, # once packing recipe is well tested, change this default to true\n tokenizer: Optional[TokenizerType] = ""model"",\n) -> run.Partial:\n """"""\n Create a default fine-tuning recipe for any model.\n\n This function sets up a template for a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n\n Args:\n model (run.Config[pl.LightningModule]): Configuration for a NeMo model.\n resume_path (str): Path to the Huggingface model or pretrained distributed checkpoint for resume\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n packed_sequence (bool): Whether to use packed sequence.\n tokenizer (Optional[TokenizerType]): Tokenizer setting to be applied. Can be 'data' or 'model'\n or an instance of TokenizerSpec.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n See usages of this recipe for further details.\n """"""\n if packed_sequence:\n datamodule = run.Config(\n llm.SquadDataModule,\n seq_length=2048,\n global_batch_size=8,\n micro_batch_size=1,\n packed_sequence_specs=PackedSequenceSpecs(packed_sequence_size=2048),\n )\n else:\n datamodule = run.Config(llm.SquadDataModule, seq_length=2048, global_batch_size=128, micro_batch_size=1)\n recipe = run.Partial(\n llm.finetune,\n model=model,\n trainer=default_finetune_trainer(\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n ),\n data=datamodule,\n log=default_finetune_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(max_lr=1e-4, min_lr=0, warmup_steps=50, adam_beta2=0.98),\n resume=nemo_resume(resume_path),\n tokenizer=tokenizer,\n )\n\n return recipe\n\n\ndef default_finetune_trainer(\n tensor_parallelism=1,\n pipeline_parallelism=1,\n pipeline_parallelism_type=torch.bfloat16,\n virtual_pipeline_parallelism=None,\n context_parallelism=1,\n sequence_parallelism=False,\n num_nodes=1,\n num_gpus_per_node=8,\n max_steps=1000,\n limit_test_batches=None,\n limit_val_batches=None,\n val_check_interval=30,\n):\n """"""\n Create a default fine-tuning trainer for any model.\n\n This function sets up a template for strategy and trainer.\n\n Args:\n See docstrings of MegatronStrategy and Trainer.\n\n Returns:\n run.Config: Config for a finetuning trainer.\n\n See usages of this in recipes for further details.\n """"""\n strategy = run.Config(\n nl.MegatronStrategy,\n tensor_model_parallel_size=tensor_parallelism,\n pipeline_model_parallel_size=pipeline_parallelism,\n pipeline_dtype=pipeline_parallelism_type,\n virtual_pipeline_model_parallel_size=virtual_pipeline_parallelism,\n context_parallel_size=context_parallelism,\n sequence_parallel=sequence_parallelism,\n gradient_as_bucket_view=True,\n ckpt_load_strictness=""log_all"",\n )\n\n trainer = run.Config(\n nl.Trainer,\n accelerator=""gpu"",\n accumulate_grad_batches=1,\n devices=num_gpus_per_node,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=1,\n max_steps=max_steps,\n num_nodes=num_nodes,\n plugins=bf16_mixed(),\n strategy=strategy,\n use_distributed_sampler=False,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n )\n\n return trainer\n\n\ndef default_finetune_log(\n dir: Optional[str] = None,\n name: str = ""default"",\n tensorboard_logger: Optional[run.Config['TensorBoardLogger']] = None,\n wandb_logger: Optional[run.Config['WandbLogger']] = None,\n) -> run.Config[nl.NeMoLogger]:\n """"""\n Create a default fine-tuning logger for any model.\n\n This function sets up a template for ModelCheckpoint and NeMoLogger.\n\n Args:\n See docstrings of ModelCheckpoint and NeMoLogger.\n\n Returns:\n run.Config: Config for a finetuning NeMoLogger.\n\n See usages of this in recipes for further details.\n """"""\n\n ckpt = run.Config(\n nl.ModelCheckpoint,\n save_last=""link"",\n save_top_k=2,\n every_n_train_steps=50,\n filename=""{model_name}--{val_loss:.2f}-{step}-{consumed_samples}"",\n )\n\n return run.Config(\n nl.NeMoLogger,\n ckpt=ckpt,\n name=name,\n tensorboard=tensorboard_logger,\n wandb=wandb_logger,\n log_dir=dir,\n )\n\n\ndef nemo_resume(model_id: str) -> run.Config[nl.AutoResume]:\n """"""\n Configure automatic resumption from a NeMo checkpoint converted from Huggingface for\n https://huggingface.co/{model_id}.\n\n This NeMo checkpoint should be converted from Huggingface beforehand, using nemo.collections.llm.import_ckpt.\n When converting the checkpoint, the NeMo checkpoint will be saved in NEMO_HOME (set to ~/.cache/nemo by default).\n\n This function sets up the configuration to resume training from path nemo://{model_id}.\n This translates to the full path {NEMO_HOME}/models/{model_id}.\n\n Args:\n model_id (str): Path to the Huggingface model or pretrained distributed checkpoint for resume\n\n Returns:\n run.Config[nl.AutoResume]: Configuration for resuming from NeMo checkpoint.\n """"""\n return run.Config(\n nl.AutoResume,\n restore_config=run.Config(nl.RestoreConfig, path=f""nemo://{model_id}""),\n )\n\n\n@run.cli.factory(name='lora')\ndef lora() -> run.Config[PEFT]:\n """"""\n Factory function to create a LoRA configuration.\n\n Returns:\n run.Config[PEFT]: Configuration for the LoRA class.\n\n Examples:\n CLI usage:\n $ nemo llm finetune -f llama3_8b peft=lora\n\n Python API usage:\n >>> lora_config = lora()\n >>> print(lora_config)\n """"""\n return run.Config(LoRA)\n\n\n@run.cli.factory(name='dora')\ndef dora() -> run.Config[PEFT]:\n """"""\n Factory function to create a DoRA configuration.\n\n Returns:\n run.Config[PEFT]: Configuration for the DoRA class.\n\n Examples:\n CLI usage:\n $ nemo llm finetune -f llama3_8b peft=dora\n\n Python API usage:\n >>> dora_config = dora()\n >>> print(dora_config)\n """"""\n return run.Config(DoRA)\n",python,tab
|
| 244 |
+
244,15926417,"nemo/collections/llm/recipes/finetune_default.py",1382,0,"",python,selection_command
|
| 245 |
+
245,15929448,"nemo/collections/llm/recipes/finetune_default.py",3372,0,"",python,selection_keyboard
|
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-ac4f665d-1bc1-467a-98b3-5da2178968731760857710663-2025_10_19-09.08.50.386/source.csv
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,3,"slurm/dev/franz/berlin/coinrun/sample/maskgit/sample_mila_submission_case_study_vanilla.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=24:00:00\n#SBATCH --cpus-per-task=8\n#SBATCH --gres=gpu:1\n#SBATCH --output=/fast/project/HFMI_SynergyUnit/jafar_ws/logs/franz/coinrun/dynamics_sample/%x_%j.log\n#SBATCH --error=/fast/project/HFMI_SynergyUnit/jafar_ws/logs/franz/coinrun/dynamics_sample/%x_%j.log\n#SBATCH --job-name=coinrun_sample_maskgit_mila_submission_case_study_vanilla\n\n# Activate virtual environment\nsource .venv/bin/activate\n\narray_records_dir=""/fast/project/HFMI_SynergyUnit/jafar_ws/data/coinrun/array_records_10M_npy_arr_rec/array_record/test""\nCHECKPOINT_PATH=""/fast/project/HFMI_SynergyUnit/jafar_ws/checkpoints/coinrun/dynamics/dynamics_case_study_dataset_10M_30031""\n\ncurrent_branch=$(git rev-parse --abbrev-ref HEAD)\nif [ ""$current_branch"" != ""main"" ]; then\n echo ""This script must be run from the main branch. Current branch is $current_branch. Exiting.""\n exit 1\nfi\n\necho ""Sampling from checkpoint: $CHECKPOINT_PATH""\n\nsrun python jasmine/sample.py \\n --seed=1 \\n --maskgit_steps=1 \\n --tokenizer_ffn_dim=512 \\n --tokenizer_num_blocks=8 \\n --dyna_ffn_dim=512 \\n --dyna_num_blocks=12 \\n --output_dir=gifs/dynamics_case_study_dataset_10M_vanilla \\n --checkpoint $CHECKPOINT_PATH \\n --data_dir=$array_records_dir \\n --seq_len=16 \\n --batch_size=32 \\n --patch_size=4 \\n --start_frame=4 \\n --image_height=64 \\n --image_width=64 \\n --dyna_type=maskgit\n",shellscript,tab
|
| 3 |
+
2,262,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"9:08:50 AM [info] Activating crowd-code\n9:08:50 AM [info] Recording started\n9:08:50 AM [info] Initializing git provider using file system watchers...\n9:08:50 AM [info] Git repository found\n9:08:50 AM [info] Git provider initialized successfully\n9:08:50 AM [info] Initial git state: [object Object]\n",Log,tab
|
| 4 |
+
3,4665,"TERMINAL",0,0,"",,terminal_command
|
| 5 |
+
4,11496,"TERMINAL",0,0,"",,terminal_command
|
| 6 |
+
5,106853,"TERMINAL",0,0,"",,terminal_command
|
1f15334ab7e6820c9fda17c961659882ef9853cc80f7356b9a9b22f286fd7389/crowd-code-c877f8c1-c8e0-4a7b-8720-40bb4df915221754138206219-2025_08_02-14.36.55.608/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-14f0662f-0032-43e8-be9f-8e53d6f150ad1758635869882-2025_09_23-15.57.52.616/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-1fa9b85d-3794-4f3b-b7a0-5170b7d2faaa1762362332596-2025_11_05-18.05.39.648/source.csv
ADDED
|
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,3,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",0,0,"/*---------------------------------------------------------------------------------------------\n * Copyright (c) Microsoft Corporation. All rights reserved.\n * Licensed under the MIT License. See License.txt in the project root for license information.\n *--------------------------------------------------------------------------------------------*/\n\n/** @jsxRuntime automatic */\n/** @jsxImportSource ../../../../prompt/jsx-runtime/ */\nimport { CopilotContentExclusionManager, StatusBarEvent } from '../../contentExclusion/contentExclusionManager';\nimport { ICompletionsContextService } from '../../context';\nimport { logger, LogTarget } from '../../logger';\n\nimport { IInstantiationService, ServicesAccessor } from '../../../../../../../util/vs/platform/instantiation/common/instantiation';\nimport { ICompletionsTelemetryService } from '../../../../bridge/src/completionsTelemetryServiceBridge';\nimport { DataPipe, VirtualPrompt } from '../../../../prompt/src/components/virtualPrompt';\nimport { TokenizerName } from '../../../../prompt/src/tokenization';\nimport { CancellationToken, Position } from '../../../../types/src';\nimport { CompletionState } from '../../completionState';\nimport { telemetryException, TelemetryWithExp } from '../../telemetry';\nimport { TextDocumentContents } from '../../textDocument';\nimport { CodeSnippets } from '../components/codeSnippets';\nimport { CompletionsContext } from '../components/completionsContext';\nimport { CompletionsPromptOk, CompletionsPromptRenderer } from '../components/completionsPromptRenderer';\nimport { ContextProviderBridge } from '../components/contextProviderBridge';\nimport { CurrentFile } from '../components/currentFile';\nimport { DocumentMarker } from '../components/marker';\nimport { RecentEdits } from '../components/recentEdits';\nimport { SimilarFiles } from '../components/similarFiles';\nimport { splitContextCompletionsPrompt } from '../components/splitContextPrompt';\nimport { SplitContextPromptRenderer } from '../components/splitContextPromptRenderer';\nimport { Traits } from '../components/traits';\nimport {\n\tContextProviderTelemetry,\n\tmatchContextItems,\n\tResolvedContextItem,\n\ttelemetrizeContextItems,\n\tuseContextProviderAPI,\n} from '../contextProviderRegistry';\nimport { getCodeSnippetsFromContextItems } from '../contextProviders/codeSnippets';\nimport {\n\tCodeSnippetWithId,\n\tSupportedContextItemWithId,\n\tTraitWithId,\n} from '../contextProviders/contextItemSchemas';\nimport { getTraitsFromContextItems, ReportTraitsTelemetry } from '../contextProviders/traits';\nimport { componentStatisticsToPromptMatcher, ContextProviderStatistics } from '../contextProviderStatistics';\nimport {\n\t_contextTooShort,\n\t_copilotContentExclusion,\n\t_promptCancelled,\n\t_promptError,\n\tgetPromptOptions,\n\tMIN_PROMPT_CHARS,\n\tPromptResponse,\n\ttrimLastLine,\n} from '../prompt';\nimport { isIncludeNeighborFilesActive } from '../similarFiles/neighborFiles';\nimport {\n\tCompletionsPromptFactory,\n\tCompletionsPromptOptions,\n\tPromptOpts,\n} from './completionsPromptFactory';\n\nexport type CompletionRequestDocument = TextDocumentContents;\n\nexport type CompletionRequestData = {\n\tdocument: CompletionRequestDocument;\n\tposition: Position;\n\ttelemetryData: TelemetryWithExp;\n\tcancellationToken?: CancellationToken;\n\t// see inlineCompletions data param\n\tdata?: unknown;\n\t// Context provider items\n\ttraits?: TraitWithId[];\n\tcodeSnippets?: CodeSnippetWithId[];\n\tturnOffSimilarFiles?: boolean;\n\tsuffixMatchThreshold?: number;\n\tmaxPromptTokens: number;\n\ttokenizer?: TokenizerName;\n};\n\nexport function isCompletionRequestData(data: unknown): data is CompletionRequestData {\n\tif (!data || typeof data !== 'object') { return false; }\n\n\tconst req = data as Partial<CompletionRequestData>;\n\n\t// Check document\n\tif (!req.document) { return false; }\n\n\t// Check position\n\tif (!req.position) { return false; }\n\tif (req.position.line === undefined) { return false; }\n\tif (req.position.character === undefined) { return false; }\n\n\t// Check telemetryData\n\tif (!req.telemetryData) { return false; }\n\n\treturn true;\n}\n\nexport enum PromptOrdering {\n\tDefault = 'default',\n\tSplitContext = 'splitContext',\n}\n\ntype DeclarativePromptFunction = typeof defaultCompletionsPrompt;\ntype AvailableDeclarativePrompts = {\n\t[K in PromptOrdering]: {\n\t\tpromptFunction: DeclarativePromptFunction;\n\t\trenderer: typeof CompletionsPromptRenderer;\n\t};\n};\n\nconst availableDeclarativePrompts: AvailableDeclarativePrompts = {\n\t[PromptOrdering.Default]: {\n\t\tpromptFunction: defaultCompletionsPrompt,\n\t\trenderer: CompletionsPromptRenderer,\n\t},\n\t[PromptOrdering.SplitContext]: {\n\t\tpromptFunction: splitContextCompletionsPrompt,\n\t\trenderer: SplitContextPromptRenderer,\n\t},\n};\n\n// The weights mimic the PromptPriorityList from prompt/src/wishlist.ts\nfunction defaultCompletionsPrompt(accessor: ServicesAccessor) {\n\tconst ctx = accessor.get(ICompletionsContextService);\n\treturn (\n\t\t<>\n\t\t\t<CompletionsContext>\n\t\t\t\t<DocumentMarker ctx={ctx} weight={0.7} />\n\t\t\t\t<Traits weight={0.6} />\n\t\t\t\t<CodeSnippets ctx={ctx} weight={0.9} />\n\t\t\t\t<SimilarFiles ctx={ctx} weight={0.8} />\n\t\t\t\t<RecentEdits ctx={ctx} weight={0.99} />\n\t\t\t</CompletionsContext>\n\t\t\t<CurrentFile weight={1} />\n\t\t</>\n\t);\n}\n\n// Exported for testing\nexport class ComponentsCompletionsPromptFactory implements CompletionsPromptFactory {\n\tprivate virtualPrompt: VirtualPrompt;\n\tprivate pipe: DataPipe;\n\tprivate renderer: CompletionsPromptRenderer;\n\tprivate promptOrdering: PromptOrdering;\n\tprivate logTarget;\n\n\tconstructor(\n\t\tvirtualPrompt: VirtualPrompt | undefined = undefined,\n\t\tordering: PromptOrdering | undefined = undefined,\n\t\t@ICompletionsContextService private readonly ctx: ICompletionsContextService,\n\t\t@IInstantiationService private readonly instantiationService: IInstantiationService,\n\t\t@ICompletionsTelemetryService private readonly completionsTelemetryService: ICompletionsTelemetryService,\n\t) {\n\t\tthis.logTarget = this.ctx.get(LogTarget);\n\t\tthis.promptOrdering = ordering ?? PromptOrdering.Default;\n\t\tthis.virtualPrompt = virtualPrompt ?? new VirtualPrompt(this.completionsPrompt());\n\t\tthis.pipe = this.virtualPrompt.createPipe();\n\t\tthis.renderer = this.getRenderer();\n\t}\n\n\tasync prompt(opts: CompletionsPromptOptions, cancellationToken?: CancellationToken): Promise<PromptResponse> {\n\t\ttry {\n\t\t\treturn await this.createPromptUnsafe(opts, cancellationToken);\n\t\t} catch (e) {\n\t\t\treturn this.errorPrompt(e as Error);\n\t\t}\n\t}\n\n\tasync createPromptUnsafe(\n\t\t{ completionId, completionState, telemetryData, promptOpts }: CompletionsPromptOptions,\n\t\tcancellationToken?: CancellationToken\n\t): Promise<PromptResponse> {\n\t\tconst { maxPromptLength, suffixPercent, suffixMatchThreshold } = this.instantiationService.invokeFunction(getPromptOptions,\n\t\t\ttelemetryData,\n\t\t\tcompletionState.textDocument.detectedLanguageId\n\t\t);\n\n\t\tconst failFastPrompt = await this.failFastPrompt(\n\t\t\tcompletionState.textDocument,\n\t\t\tcompletionState.position,\n\t\t\tsuffixPercent,\n\t\t\tcancellationToken\n\t\t);\n\t\tif (failFastPrompt) {\n\t\t\treturn failFastPrompt;\n\t\t}\n\n\t\t// TODO: Prompt ordering changes are triggered by ExP changes.\n\t\t// TODO@benibenj remove this as its always true (except in tests)\n\t\tconst promptOrdering = promptOpts?.separateContext ? PromptOrdering.SplitContext : PromptOrdering.Default;\n\t\tthis.setPromptOrdering(promptOrdering);\n\n\t\tconst start = performance.now();\n\n\t\tconst { traits, codeSnippets, turnOffSimilarFiles, resolvedContextItems } = await this.resolveContext(\n\t\t\tcompletionId,\n\t\t\tcompletionState,\n\t\t\ttelemetryData,\n\t\t\tcancellationToken,\n\t\t\tpromptOpts\n\t\t);\n\n\t\tawait this.updateComponentData(\n\t\t\tcompletionState.textDocument,\n\t\t\tcompletionState.position,\n\t\t\ttraits,\n\t\t\tcodeSnippets,\n\t\t\ttelemetryData,\n\t\t\tturnOffSimilarFiles,\n\t\t\tmaxPromptLength,\n\t\t\tcancellationToken,\n\t\t\tpromptOpts,\n\t\t\tsuffixMatchThreshold,\n\t\t\tpromptOpts?.tokenizer\n\t\t);\n\n\t\tif (cancellationToken?.isCancellationRequested) {\n\t\t\treturn _promptCancelled;\n\t\t}\n\n\t\tconst snapshot = this.virtualPrompt.snapshot(cancellationToken);\n\t\tconst snapshotStatus = snapshot.status;\n\t\tif (snapshotStatus === 'cancelled') {\n\t\t\treturn _promptCancelled;\n\t\t} else if (snapshotStatus === 'error') {\n\t\t\treturn this.errorPrompt(snapshot.error);\n\t\t}\n\n\t\tconst rendered = this.renderer.render(\n\t\t\tsnapshot.snapshot!,\n\t\t\t{\n\t\t\t\tdelimiter: '\n',\n\t\t\t\ttokenizer: promptOpts?.tokenizer,\n\t\t\t\tpromptTokenLimit: maxPromptLength,\n\t\t\t\tsuffixPercent: suffixPercent,\n\t\t\t\tlanguageId: completionState.textDocument.detectedLanguageId,\n\t\t\t},\n\t\t\tcancellationToken\n\t\t);\n\t\tif (rendered.status === 'cancelled') {\n\t\t\treturn _promptCancelled;\n\t\t} else if (rendered.status === 'error') {\n\t\t\treturn this.errorPrompt(rendered.error);\n\t\t}\n\n\t\tconst [prefix, trailingWs] = trimLastLine(rendered.prefix);\n\t\tconst renderedTrimmed = { ...rendered, prefix };\n\n\t\tlet contextProvidersTelemetry: ContextProviderTelemetry[] | undefined = undefined;\n\t\tconst languageId = completionState.textDocument.detectedLanguageId;\n\t\tif (this.instantiationService.invokeFunction(useContextProviderAPI, languageId, telemetryData)) {\n\t\t\tconst promptMatcher = componentStatisticsToPromptMatcher(rendered.metadata.componentStatistics);\n\t\t\tthis.ctx\n\t\t\t\t.get(ContextProviderStatistics)\n\t\t\t\t.getStatisticsForCompletion(completionId)\n\t\t\t\t.computeMatch(promptMatcher);\n\t\t\tcontextProvidersTelemetry = telemetrizeContextItems(this.ctx, completionId, resolvedContextItems);\n\t\t\t// To support generating context provider metrics of completion in COffE.\n\t\t\tlogger.debug(this.logTarget, `Context providers telemetry: '${JSON.stringify(contextProvidersTelemetry)}'`);\n\t\t}\n\t\tconst end = performance.now();\n\t\tthis.resetIfEmpty(rendered);\n\t\treturn this.successPrompt(renderedTrimmed, end, start, trailingWs, contextProvidersTelemetry);\n\t}\n\n\tprivate async updateComponentData(\n\t\ttextDocument: CompletionRequestDocument,\n\t\tposition: Position,\n\t\ttraits: TraitWithId[] | undefined,\n\t\tcodeSnippets: CodeSnippetWithId[] | undefined,\n\t\ttelemetryData: TelemetryWithExp,\n\t\tturnOffSimilarFiles: boolean,\n\t\tmaxPromptLength: number,\n\t\tcancellationToken?: CancellationToken,\n\t\topts: PromptOpts = {},\n\t\tsuffixMatchThreshold?: number,\n\t\ttokenizer?: TokenizerName\n\t) {\n\t\tconst completionRequestData = this.createRequestData(\n\t\t\ttextDocument,\n\t\t\tposition,\n\t\t\ttelemetryData,\n\t\t\tcancellationToken,\n\t\t\topts,\n\t\t\tmaxPromptLength,\n\t\t\ttraits,\n\t\t\tcodeSnippets,\n\t\t\tturnOffSimilarFiles,\n\t\t\tsuffixMatchThreshold,\n\t\t\ttokenizer\n\t\t);\n\t\tawait this.pipe.pump(completionRequestData);\n\t}\n\n\tprivate async resolveContext(\n\t\tcompletionId: string,\n\t\tcompletionState: CompletionState,\n\t\ttelemetryData: TelemetryWithExp,\n\t\tcancellationToken?: CancellationToken,\n\t\topts: PromptOpts = {}\n\t): Promise<{\n\t\ttraits: TraitWithId[] | undefined;\n\t\tcodeSnippets: CodeSnippetWithId[] | undefined;\n\t\tturnOffSimilarFiles: boolean;\n\t\tresolvedContextItems: ResolvedContextItem[];\n\t}> {\n\t\tlet resolvedContextItems: ResolvedContextItem[] = [];\n\t\tlet traits: TraitWithId[] | undefined;\n\t\tlet codeSnippets: CodeSnippetWithId[] | undefined;\n\t\tlet turnOffSimilarFiles = false;\n\t\tif (this.instantiationService.invokeFunction(useContextProviderAPI, completionState.textDocument.detectedLanguageId, telemetryData)) {\n\t\t\tresolvedContextItems = await this.ctx.get(ContextProviderBridge).resolution(completionId);\n\t\t\tconst { textDocument } = completionState;\n\t\t\t// Turn off neighboring files if:\n\t\t\t// - it's not explicitly enabled via EXP flag\n\t\t\t// - there are matched context providers\n\t\t\tconst matchedContextItems = resolvedContextItems.filter(matchContextItems);\n\t\t\tif (!this.instantiationService.invokeFunction(similarFilesEnabled, textDocument.detectedLanguageId, matchedContextItems, telemetryData)) {\n\t\t\t\tturnOffSimilarFiles = true;\n\t\t\t}\n\n\t\t\ttraits = await this.instantiationService.invokeFunction(getTraitsFromContextItems, completionId, matchedContextItems);\n\t\t\tvoid this.instantiationService.invokeFunction(ReportTraitsTelemetry,\n\t\t\t\t`contextProvider.traits`,\n\t\t\t\ttraits,\n\t\t\t\ttextDocument.detectedLanguageId,\n\t\t\t\ttextDocument.detectedLanguageId, // TextDocumentContext does not have clientLanguageId\n\t\t\t\ttelemetryData\n\t\t\t);\n\n\t\t\tcodeSnippets = await this.instantiationService.invokeFunction(getCodeSnippetsFromContextItems,\n\t\t\t\tcompletionId,\n\t\t\t\tmatchedContextItems,\n\t\t\t\ttextDocument.detectedLanguageId\n\t\t\t);\n\t\t}\n\t\treturn { traits, codeSnippets, turnOffSimilarFiles, resolvedContextItems };\n\t}\n\n\tprivate async failFastPrompt(\n\t\ttextDocument: TextDocumentContents,\n\t\tposition: Position,\n\t\tsuffixPercent: number,\n\t\tcancellationToken: CancellationToken | undefined\n\t) {\n\t\tif (cancellationToken?.isCancellationRequested) {\n\t\t\treturn _promptCancelled;\n\t\t}\n\t\tif (\n\t\t\t(\n\t\t\t\tawait this.ctx\n\t\t\t\t\t.get(CopilotContentExclusionManager)\n\t\t\t\t\t.evaluate(textDocument.uri, textDocument.getText(), StatusBarEvent.UPDATE)\n\t\t\t).isBlocked\n\t\t) {\n\t\t\treturn _copilotContentExclusion;\n\t\t}\n\n\t\tconst eligibleChars = suffixPercent > 0 ? textDocument.getText().length : textDocument.offsetAt(position);\n\t\tif (eligibleChars < MIN_PROMPT_CHARS) {\n\t\t\t// Too short context\n\t\t\treturn _contextTooShort;\n\t\t}\n\t}\n\n\tprivate createRequestData(\n\t\ttextDocument: CompletionRequestDocument,\n\t\tposition: Position,\n\t\ttelemetryData: TelemetryWithExp,\n\t\tcancellationToken: CancellationToken | undefined,\n\t\topts: PromptOpts,\n\t\tmaxPromptLength: number,\n\t\ttraits?: TraitWithId[],\n\t\tcodeSnippets?: CodeSnippetWithId[],\n\t\tturnOffSimilarFiles?: boolean,\n\t\tsuffixMatchThreshold?: number,\n\t\ttokenizer?: TokenizerName\n\t): CompletionRequestData {\n\t\treturn {\n\t\t\tdocument: textDocument,\n\t\t\tposition,\n\t\t\ttelemetryData,\n\t\t\tcancellationToken,\n\t\t\tdata: opts.data,\n\t\t\ttraits,\n\t\t\tcodeSnippets,\n\t\t\tturnOffSimilarFiles,\n\t\t\tsuffixMatchThreshold,\n\t\t\tmaxPromptTokens: maxPromptLength,\n\t\t\ttokenizer,\n\t\t};\n\t}\n\n\tprivate resetIfEmpty(rendered: CompletionsPromptOk) {\n\t\tif (rendered.prefix.length === 0 && rendered.suffix.length === 0) {\n\t\t\tthis.reset();\n\t\t}\n\t}\n\n\tprivate successPrompt(\n\t\trendered: CompletionsPromptOk,\n\t\tend: number,\n\t\tstart: number,\n\t\ttrailingWs: string,\n\t\tcontextProvidersTelemetry?: ContextProviderTelemetry[]\n\t): PromptResponse {\n\t\treturn {\n\t\t\ttype: 'prompt',\n\t\t\tprompt: {\n\t\t\t\tprefix: rendered.prefix,\n\t\t\t\tprefixTokens: rendered.prefixTokens,\n\t\t\t\tsuffix: rendered.suffix,\n\t\t\t\tsuffixTokens: rendered.suffixTokens,\n\t\t\t\tcontext: rendered.context,\n\t\t\t\tisFimEnabled: rendered.suffix.length > 0,\n\t\t\t},\n\t\t\tcomputeTimeMs: end - start,\n\t\t\ttrailingWs,\n\t\t\tneighborSource: new Map(),\n\t\t\tmetadata: rendered.metadata,\n\t\t\tcontextProvidersTelemetry,\n\t\t};\n\t}\n\n\tprivate errorPrompt(error: Error): PromptResponse {\n\t\ttelemetryException(this.completionsTelemetryService, error, 'PromptComponents.CompletionsPromptFactory');\n\t\tthis.reset();\n\t\treturn _promptError;\n\t}\n\n\tprivate reset() {\n\t\tthis.renderer = this.getRenderer();\n\t\tthis.virtualPrompt = new VirtualPrompt(this.completionsPrompt());\n\t\tthis.pipe = this.virtualPrompt.createPipe();\n\t}\n\n\tprivate setPromptOrdering(ordering: PromptOrdering) {\n\t\tif (this.promptOrdering !== ordering) {\n\t\t\tthis.promptOrdering = ordering;\n\t\t\tthis.reset();\n\t\t}\n\t}\n\n\tprivate completionsPrompt() {\n\t\tconst promptFunction =\n\t\t\tavailableDeclarativePrompts[this.promptOrdering]?.promptFunction ?? defaultCompletionsPrompt;\n\t\treturn this.instantiationService.invokeFunction(promptFunction);\n\t}\n\n\tprivate getRenderer() {\n\t\tconst promptInfo =\n\t\t\tavailableDeclarativePrompts[this.promptOrdering] ?? availableDeclarativePrompts[PromptOrdering.Default];\n\t\treturn new promptInfo.renderer();\n\t}\n}\n\n// Similar files is enabled if:\n// - the languageId is C/C++.\n// - it's explicitly enabled via EXP flag or config.\n// - no code snippets are provided (which includes the case when all providers error).\nfunction similarFilesEnabled(\n\taccessor: ServicesAccessor,\n\tdetectedLanguageId: string,\n\tmatchedContextItems: ResolvedContextItem<SupportedContextItemWithId>[],\n\ttelemetryData: TelemetryWithExp\n) {\n\tconst cppLanguageIds = ['cpp', 'c'];\n\tconst includeNeighboringFiles =\n\t\tisIncludeNeighborFilesActive(accessor, detectedLanguageId, telemetryData) || cppLanguageIds.includes(detectedLanguageId);\n\treturn (\n\t\tincludeNeighboringFiles || !matchedContextItems.some(ci => ci.data.some(item => item.type === 'CodeSnippet'))\n\t);\n}\n",typescriptreact,tab
|
| 3 |
+
2,289,"tasks",0,0,"",Log,tab
|
| 4 |
+
3,290,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",0,0,"",typescriptreact,tab
|
| 5 |
+
4,343,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"6:05:39 PM [info] Activating crowd-code\n6:05:39 PM [info] Recording started\n6:05:39 PM [info] Initializing git provider using file system watchers...\n6:05:39 PM [info] Git repository found\n6:05:39 PM [info] Git provider initialized successfully\n6:05:39 PM [info] Initial git state: [object Object]\n",Log,tab
|
| 6 |
+
5,12297,"extension-output-pdoom-org.crowd-code-#1-crowd-code",298,0,"",Log,selection_mouse
|
| 7 |
+
6,13219,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",0,0,"",typescriptreact,tab
|
| 8 |
+
7,347453,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",5114,0,"",typescriptreact,selection_command
|
| 9 |
+
8,347571,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",5070,0,"",typescriptreact,selection_command
|
| 10 |
+
9,1004602,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",5026,0,"",typescriptreact,selection_command
|
| 11 |
+
10,1004729,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",4982,0,"",typescriptreact,selection_command
|
| 12 |
+
11,1004853,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",4954,0,"",typescriptreact,selection_command
|
| 13 |
+
12,1005293,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",4908,0,"",typescriptreact,selection_command
|
| 14 |
+
13,2090202,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
|
| 15 |
+
14,2092221,"TERMINAL",0,0,"",,terminal_focus
|
| 16 |
+
15,2092222,"src/extension/completions-core/vscode-node/lib/src/prompt/completionsPromptFactory/componentsCompletionsPromptFactory.tsx",0,0,"",typescriptreact,tab
|
| 17 |
+
16,2093995,"TERMINAL",0,0,"cd ..",,terminal_command
|
| 18 |
+
17,2094451,"TERMINAL",0,0,"ls",,terminal_command
|
| 19 |
+
18,2094466,"TERMINAL",0,0,"]633;C[0m[01;34mcleanrl[0m [01;34mcrowd-code-player[0m [01;34mcrowd-pilot[0m [01;34mjafar[0m [01;34mjax_cache[0m [01;34mmaxtext[0m [01;34mnpm-global[0m [01;34moai-compatible-copilot[0m [01;34msbatch-runner[0m [01;34mStoix[0m [01;34mvscode-crowd-pilot-chat[0m\r\n]0;franz.srambical@hai-login2:~",,terminal_output
|
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-4b7a193b-6fd0-48b6-a605-a1ce6ba179221764439942282-2025_11_29-19.12.25.371/source.csv
ADDED
|
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,106,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"7:12:25 PM [info] Activating crowd-code\n7:12:25 PM [info] Recording started\n7:12:25 PM [info] Initializing git provider using file system watchers...\n7:12:25 PM [info] No workspace folder found\n",Log,tab
|
| 3 |
+
3,2027,"extension-output-pdoom-org.crowd-code-#1-crowd-code",194,0,"7:12:27 PM [info] Retrying git provider initialization...\n7:12:27 PM [info] No workspace folder found\n",Log,content
|
| 4 |
+
4,9498,"Untitled-1",0,0,"",plaintext,tab
|
| 5 |
+
5,11149,"TERMINAL",0,0,"Test",,terminal_focus
|
| 6 |
+
6,11153,"Untitled-1",0,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
|
| 7 |
+
7,18176,"Untitled-1",46,0,"",plaintext,selection_command
|
| 8 |
+
8,18275,"Untitled-1",39,0,"",plaintext,selection_command
|
| 9 |
+
9,18422,"Untitled-1",32,0,"",plaintext,selection_command
|
| 10 |
+
10,18611,"Untitled-1",0,0,"",plaintext,selection_command
|
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-549b7320-9591-428d-ba6a-d3d8fb65a0001764416891089-2025_11_29-12.48.19.667/source.csv
ADDED
|
@@ -0,0 +1,1006 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,103,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"12:48:19 PM [info] Activating crowd-code\n12:48:19 PM [info] Recording started\n12:48:19 PM [info] Initializing git provider using file system watchers...\n12:48:19 PM [info] Git repository found\n12:48:19 PM [info] Git provider initialized successfully\n12:48:19 PM [info] Initial git state: [object Object]\n",Log,tab
|
| 3 |
+
3,1736,"TERMINAL",0,0,"",,terminal_focus
|
| 4 |
+
4,4333,"TERMINAL",0,0,"zsh",,terminal_focus
|
| 5 |
+
5,20864,"TERMINAL",0,0,"cd webview-ui/sto",,terminal_command
|
| 6 |
+
6,20866,"TERMINAL",0,0,"]633;Ccd: no such file or directory: webview-ui/sto\r\n[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 7 |
+
7,22604,"TERMINAL",0,0,"cd webview-ui/sto",,terminal_command
|
| 8 |
+
8,22605,"TERMINAL",0,0,"]633;Ccd: no such file or directory: webview-ui/sto\r\n[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 9 |
+
9,28427,"TERMINAL",0,0,"cd webview-ui/.storybook",,terminal_command
|
| 10 |
+
10,28428,"TERMINAL",0,0,"]633;C[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 11 |
+
11,31719,"TERMINAL",0,0,"npm run storybook",,terminal_command
|
| 12 |
+
12,31770,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 13 |
+
13,32548,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\n[1G[0Ksh: storybook: command not found\r\n[1G[0K⠙[1G[0K[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 14 |
+
14,201178,"TERMINAL",0,0,"cd ..",,terminal_command
|
| 15 |
+
15,201179,"TERMINAL",0,0,"]633;C[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 16 |
+
16,202166,"TERMINAL",0,0,"cd ..",,terminal_command
|
| 17 |
+
17,202171,"TERMINAL",0,0,"]633;C[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 18 |
+
18,208467,"TERMINAL",0,0,"npm install",,terminal_command
|
| 19 |
+
19,208518,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 20 |
+
20,211999,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 21 |
+
21,212258,"TERMINAL",0,0,"[1G[0K[1mnpm[22m [33mwarn[39m [94mEBADENGINE[39m Unsupported engine {\r\n[1mnpm[22m [33mwarn[39m [94mEBADENGINE[39m package: [32m'@sap/xsenv@6.0.0'[39m,\r\n[1mnpm[22m [33mwarn[39m [94mEBADENGINE[39m required: { node: [32m'^20.0.0 || ^22.0.0 || ^24.0.0'[39m },\r\n[1mnpm[22m [33mwarn[39m [94mEBADENGINE[39m current: { node: [32m'v23.11.0'[39m, npm: [32m'10.9.2'[39m }\r\n[1mnpm[22m [33mwarn[39m [94mEBADENGINE[39m }\r\n[1G[0K⠹[1G[0K[1mnpm[22m [33mwarn[39m [94mEBADENGINE[39m Unsupported engine {\r\n[1mnpm[22m [33mwarn[39m [94mEBADENGINE[39m package: [32m'opossum@9.0.0'[39m,\r\n[1mnpm[22m [33mwarn[39m [94mEBADENGINE[39m required: { node: [32m'^24 || ^22 || ^20'[39m },\r\n[1mnpm[22m [33mwarn[39m [94mEBADENGINE[39m current: { node: [32m'v23.11.0'[39m, npm: [32m'10.9.2'[39m }\r\n[1mnpm[22m [33mwarn[39m [94mEBADENGINE[39m }\r\n[1G[0K⠹",,terminal_output
|
| 22 |
+
22,212340,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 23 |
+
23,212476,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 24 |
+
24,212566,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 25 |
+
25,212680,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 26 |
+
26,212761,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 27 |
+
27,212842,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 28 |
+
28,212922,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 29 |
+
29,213003,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 30 |
+
30,213084,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 31 |
+
31,213163,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 32 |
+
32,213244,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 33 |
+
33,213325,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 34 |
+
34,213403,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 35 |
+
35,213484,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 36 |
+
36,213784,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 37 |
+
37,213862,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 38 |
+
38,213949,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 39 |
+
39,214040,"TERMINAL",0,0,"[1G[0K⠋[1G[0K[1mnpm[22m [33mwarn[39m [94mdeprecated[39m rimraf@3.0.2: Rimraf versions prior to v4 are no longer supported\r\n[1G[0K⠋[1G[0K[1mnpm[22m [33mwarn[39m [94mdeprecated[39m rimraf@3.0.2: Rimraf versions prior to v4 are no longer supported\r\n[1G[0K⠋[1G[0K[1mnpm[22m [33mwarn[39m [94mdeprecated[39m rimraf@3.0.2: Rimraf versions prior to v4 are no longer supported\r\n[1G[0K⠋",,terminal_output
|
| 40 |
+
40,214240,"TERMINAL",0,0,"[1G[0K⠙[1G[0K[1mnpm[22m [33mwarn[39m [94mdeprecated[39m glob@7.2.3: Glob versions prior to v9 are no longer supported\r\n[1G[0K⠙[1G[0K[1mnpm[22m [33mwarn[39m [94mdeprecated[39m glob@7.2.3: Glob versions prior to v9 are no longer supported\r\n[1G[0K⠙[1G[0K[1mnpm[22m [33mwarn[39m [94mdeprecated[39m glob@7.2.3: Glob versions prior to v9 are no longer supported\r\n[1G[0K⠙[1G[0K[1mnpm[22m [33mwarn[39m [94mdeprecated[39m glob@7.2.3: Glob versions prior to v9 are no longer supported\r\n[1G[0K⠙",,terminal_output
|
| 41 |
+
41,214321,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 42 |
+
42,214401,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 43 |
+
43,214483,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 44 |
+
44,214563,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 45 |
+
45,214643,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 46 |
+
46,214722,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 47 |
+
47,214803,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 48 |
+
48,214882,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 49 |
+
49,214963,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 50 |
+
50,215043,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 51 |
+
51,215122,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 52 |
+
52,215200,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 53 |
+
53,215288,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 54 |
+
54,215364,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 55 |
+
55,215443,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 56 |
+
56,215524,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 57 |
+
57,215604,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 58 |
+
58,215684,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 59 |
+
59,215765,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 60 |
+
60,215845,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 61 |
+
61,215925,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 62 |
+
62,216004,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 63 |
+
63,216085,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 64 |
+
64,216164,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 65 |
+
65,216254,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 66 |
+
66,216326,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 67 |
+
67,216408,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 68 |
+
68,216487,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 69 |
+
69,216567,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 70 |
+
70,216647,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 71 |
+
71,216728,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 72 |
+
72,216809,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 73 |
+
73,216888,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 74 |
+
74,216971,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 75 |
+
75,217050,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 76 |
+
76,217132,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 77 |
+
77,217211,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 78 |
+
78,217294,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 79 |
+
79,217371,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 80 |
+
80,217452,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 81 |
+
81,217532,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 82 |
+
82,217613,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 83 |
+
83,217694,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 84 |
+
84,217774,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 85 |
+
85,217856,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 86 |
+
86,217936,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 87 |
+
87,218018,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 88 |
+
88,218098,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 89 |
+
89,218176,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 90 |
+
90,218258,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 91 |
+
91,218339,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 92 |
+
92,218427,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 93 |
+
93,218501,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 94 |
+
94,218582,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 95 |
+
95,218662,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 96 |
+
96,218742,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 97 |
+
97,218821,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 98 |
+
98,218900,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 99 |
+
99,218982,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 100 |
+
100,219063,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 101 |
+
101,219142,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 102 |
+
102,219223,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 103 |
+
103,219304,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 104 |
+
104,219385,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 105 |
+
105,219464,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 106 |
+
106,219544,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 107 |
+
107,219625,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 108 |
+
108,219704,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 109 |
+
109,219785,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 110 |
+
110,219866,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 111 |
+
111,219946,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 112 |
+
112,220024,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 113 |
+
113,220107,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 114 |
+
114,220188,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 115 |
+
115,220268,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 116 |
+
116,220349,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 117 |
+
117,220431,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 118 |
+
118,220509,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 119 |
+
119,220591,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 120 |
+
120,220668,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 121 |
+
121,220751,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 122 |
+
122,220831,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 123 |
+
123,220911,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 124 |
+
124,220991,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 125 |
+
125,221072,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 126 |
+
126,221151,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 127 |
+
127,221232,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 128 |
+
128,221314,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 129 |
+
129,221396,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 130 |
+
130,221474,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 131 |
+
131,221556,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 132 |
+
132,221634,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 133 |
+
133,221715,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 134 |
+
134,221794,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 135 |
+
135,221881,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 136 |
+
136,221958,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 137 |
+
137,222038,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 138 |
+
138,222119,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 139 |
+
139,222202,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 140 |
+
140,222281,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 141 |
+
141,222361,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 142 |
+
142,222441,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 143 |
+
143,222522,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 144 |
+
144,222603,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 145 |
+
145,222683,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 146 |
+
146,222763,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 147 |
+
147,222847,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 148 |
+
148,222944,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 149 |
+
149,223014,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 150 |
+
150,223095,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 151 |
+
151,223175,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 152 |
+
152,223255,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 153 |
+
153,223335,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 154 |
+
154,223417,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 155 |
+
155,223497,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 156 |
+
156,223577,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 157 |
+
157,223664,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 158 |
+
158,223739,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 159 |
+
159,223820,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 160 |
+
160,223900,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 161 |
+
161,223982,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 162 |
+
162,224062,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 163 |
+
163,224142,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 164 |
+
164,224221,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 165 |
+
165,224304,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 166 |
+
166,224385,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 167 |
+
167,224465,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 168 |
+
168,224545,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 169 |
+
169,224624,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 170 |
+
170,224703,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 171 |
+
171,224796,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 172 |
+
172,224874,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 173 |
+
173,224954,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 174 |
+
174,225033,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 175 |
+
175,225113,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 176 |
+
176,225193,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 177 |
+
177,225275,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 178 |
+
178,225354,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 179 |
+
179,225434,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 180 |
+
180,225514,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 181 |
+
181,225594,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 182 |
+
182,225677,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 183 |
+
183,225757,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 184 |
+
184,225837,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 185 |
+
185,225918,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 186 |
+
186,225996,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 187 |
+
187,226078,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 188 |
+
188,226158,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 189 |
+
189,226237,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 190 |
+
190,226319,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 191 |
+
191,226396,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 192 |
+
192,226478,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 193 |
+
193,226559,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 194 |
+
194,226637,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 195 |
+
195,226721,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 196 |
+
196,226800,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 197 |
+
197,226880,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 198 |
+
198,226960,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 199 |
+
199,227039,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 200 |
+
200,227137,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 201 |
+
201,227200,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 202 |
+
202,227279,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 203 |
+
203,227362,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 204 |
+
204,227442,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 205 |
+
205,227532,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 206 |
+
206,227612,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 207 |
+
207,227696,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 208 |
+
208,227774,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 209 |
+
209,227859,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 210 |
+
210,227946,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 211 |
+
211,228025,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 212 |
+
212,228107,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 213 |
+
213,228186,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 214 |
+
214,228267,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 215 |
+
215,228347,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 216 |
+
216,228427,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 217 |
+
217,228508,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 218 |
+
218,228586,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 219 |
+
219,228667,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 220 |
+
220,228747,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 221 |
+
221,228827,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 222 |
+
222,228906,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 223 |
+
223,228987,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 224 |
+
224,229067,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 225 |
+
225,229147,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 226 |
+
226,229226,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 227 |
+
227,229306,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 228 |
+
228,229452,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 229 |
+
229,229562,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 230 |
+
230,229645,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 231 |
+
231,229725,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 232 |
+
232,229802,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 233 |
+
233,229888,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 234 |
+
234,229968,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 235 |
+
235,230048,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 236 |
+
236,230129,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 237 |
+
237,230211,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 238 |
+
238,230294,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 239 |
+
239,230373,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 240 |
+
240,230452,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 241 |
+
241,230532,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 242 |
+
242,230615,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 243 |
+
243,230693,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 244 |
+
244,230774,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 245 |
+
245,230854,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 246 |
+
246,230933,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 247 |
+
247,231014,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 248 |
+
248,231093,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 249 |
+
249,231174,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 250 |
+
250,231253,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 251 |
+
251,231333,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 252 |
+
252,231413,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 253 |
+
253,231492,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 254 |
+
254,231573,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 255 |
+
255,231653,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 256 |
+
256,231732,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 257 |
+
257,231813,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 258 |
+
258,231894,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 259 |
+
259,231973,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 260 |
+
260,232054,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 261 |
+
261,232134,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 262 |
+
262,232213,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 263 |
+
263,232293,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 264 |
+
264,232373,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 265 |
+
265,232453,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 266 |
+
266,232534,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 267 |
+
267,232613,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 268 |
+
268,232694,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 269 |
+
269,232773,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 270 |
+
270,232854,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 271 |
+
271,232939,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 272 |
+
272,233013,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 273 |
+
273,233094,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 274 |
+
274,233173,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 275 |
+
275,233253,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 276 |
+
276,233334,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 277 |
+
277,233415,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 278 |
+
278,233494,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 279 |
+
279,233574,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 280 |
+
280,233655,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 281 |
+
281,233735,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 282 |
+
282,233816,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 283 |
+
283,233895,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 284 |
+
284,233974,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 285 |
+
285,234055,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 286 |
+
286,234134,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 287 |
+
287,234219,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 288 |
+
288,234299,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 289 |
+
289,234378,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 290 |
+
290,234458,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 291 |
+
291,234539,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 292 |
+
292,234618,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 293 |
+
293,234716,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 294 |
+
294,234793,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 295 |
+
295,234872,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 296 |
+
296,234952,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 297 |
+
297,235033,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 298 |
+
298,235112,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 299 |
+
299,235194,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 300 |
+
300,235274,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 301 |
+
301,235352,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 302 |
+
302,235434,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 303 |
+
303,235513,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 304 |
+
304,235594,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 305 |
+
305,235673,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 306 |
+
306,235752,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 307 |
+
307,235833,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 308 |
+
308,235914,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 309 |
+
309,235994,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 310 |
+
310,236073,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 311 |
+
311,236154,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 312 |
+
312,236260,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 313 |
+
313,236339,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 314 |
+
314,236419,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 315 |
+
315,236500,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 316 |
+
316,236581,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 317 |
+
317,236668,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 318 |
+
318,236741,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 319 |
+
319,236824,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 320 |
+
320,236902,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 321 |
+
321,236984,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 322 |
+
322,237064,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 323 |
+
323,237145,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 324 |
+
324,237225,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 325 |
+
325,237305,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 326 |
+
326,237386,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 327 |
+
327,237467,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 328 |
+
328,237549,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 329 |
+
329,237628,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 330 |
+
330,237709,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 331 |
+
331,237788,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 332 |
+
332,237871,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 333 |
+
333,237953,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 334 |
+
334,238063,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 335 |
+
335,238116,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 336 |
+
336,238198,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 337 |
+
337,238280,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 338 |
+
338,238359,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 339 |
+
339,238440,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 340 |
+
340,238522,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 341 |
+
341,238604,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 342 |
+
342,238685,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 343 |
+
343,238765,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 344 |
+
344,238845,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 345 |
+
345,238928,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 346 |
+
346,239008,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 347 |
+
347,239088,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 348 |
+
348,239171,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 349 |
+
349,239252,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 350 |
+
350,239333,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 351 |
+
351,239413,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 352 |
+
352,239494,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 353 |
+
353,239573,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 354 |
+
354,239655,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 355 |
+
355,239736,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 356 |
+
356,239817,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 357 |
+
357,239898,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 358 |
+
358,239978,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 359 |
+
359,240059,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 360 |
+
360,240140,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 361 |
+
361,240222,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 362 |
+
362,240303,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 363 |
+
363,240384,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 364 |
+
364,240464,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 365 |
+
365,240544,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 366 |
+
366,240627,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 367 |
+
367,240707,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 368 |
+
368,240789,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 369 |
+
369,240867,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 370 |
+
370,240950,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 371 |
+
371,241029,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 372 |
+
372,241110,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 373 |
+
373,241192,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 374 |
+
374,241273,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 375 |
+
375,241356,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 376 |
+
376,241434,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 377 |
+
377,241515,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 378 |
+
378,241595,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 379 |
+
379,241676,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 380 |
+
380,241759,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 381 |
+
381,241841,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 382 |
+
382,241921,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 383 |
+
383,242003,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 384 |
+
384,242082,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 385 |
+
385,242163,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 386 |
+
386,242244,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 387 |
+
387,242325,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 388 |
+
388,242406,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 389 |
+
389,242487,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 390 |
+
390,242568,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 391 |
+
391,242648,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 392 |
+
392,242732,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 393 |
+
393,242813,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 394 |
+
394,242892,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 395 |
+
395,242971,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 396 |
+
396,243051,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 397 |
+
397,243135,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 398 |
+
398,243215,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 399 |
+
399,243295,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 400 |
+
400,243377,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 401 |
+
401,243457,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 402 |
+
402,243540,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 403 |
+
403,243621,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 404 |
+
404,243700,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 405 |
+
405,243781,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 406 |
+
406,243864,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 407 |
+
407,243946,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 408 |
+
408,244025,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 409 |
+
409,244106,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 410 |
+
410,244186,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 411 |
+
411,244267,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 412 |
+
412,244348,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 413 |
+
413,244436,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 414 |
+
414,244515,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 415 |
+
415,244597,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 416 |
+
416,244678,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 417 |
+
417,244768,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 418 |
+
418,244846,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 419 |
+
419,244927,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 420 |
+
420,245007,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 421 |
+
421,245089,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 422 |
+
422,245170,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 423 |
+
423,245251,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 424 |
+
424,245332,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 425 |
+
425,245411,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 426 |
+
426,245495,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 427 |
+
427,245573,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 428 |
+
428,245656,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 429 |
+
429,245735,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 430 |
+
430,245818,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 431 |
+
431,245898,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 432 |
+
432,245979,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 433 |
+
433,246061,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 434 |
+
434,246142,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 435 |
+
435,246222,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 436 |
+
436,246312,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 437 |
+
437,246384,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 438 |
+
438,246467,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 439 |
+
439,246549,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 440 |
+
440,246635,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 441 |
+
441,246716,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 442 |
+
442,246798,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 443 |
+
443,246879,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 444 |
+
444,246961,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 445 |
+
445,247041,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 446 |
+
446,247122,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 447 |
+
447,247201,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 448 |
+
448,247282,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 449 |
+
449,247362,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 450 |
+
450,247444,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 451 |
+
451,247525,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 452 |
+
452,247607,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 453 |
+
453,247686,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 454 |
+
454,247768,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 455 |
+
455,247846,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 456 |
+
456,247929,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 457 |
+
457,248011,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 458 |
+
458,248096,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 459 |
+
459,248177,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 460 |
+
460,248257,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 461 |
+
461,248340,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 462 |
+
462,248420,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 463 |
+
463,248499,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 464 |
+
464,248580,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 465 |
+
465,248663,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 466 |
+
466,248743,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 467 |
+
467,248824,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 468 |
+
468,248906,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 469 |
+
469,248986,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 470 |
+
470,249068,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 471 |
+
471,249149,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 472 |
+
472,249244,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 473 |
+
473,249326,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 474 |
+
474,249407,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 475 |
+
475,249488,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 476 |
+
476,249568,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 477 |
+
477,249655,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 478 |
+
478,249767,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 479 |
+
479,249840,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 480 |
+
480,249912,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 481 |
+
481,249993,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 482 |
+
482,250073,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 483 |
+
483,250155,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 484 |
+
484,250235,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 485 |
+
485,250316,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 486 |
+
486,250398,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 487 |
+
487,250479,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 488 |
+
488,250558,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 489 |
+
489,250642,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 490 |
+
490,250720,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 491 |
+
491,250800,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 492 |
+
492,250881,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 493 |
+
493,250963,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 494 |
+
494,251048,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 495 |
+
495,251123,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 496 |
+
496,251205,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 497 |
+
497,251286,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 498 |
+
498,251367,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 499 |
+
499,251447,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 500 |
+
500,251528,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 501 |
+
501,251609,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 502 |
+
502,251689,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 503 |
+
503,251771,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 504 |
+
504,251850,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 505 |
+
505,251932,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 506 |
+
506,252013,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 507 |
+
507,252091,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 508 |
+
508,252173,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 509 |
+
509,252255,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 510 |
+
510,252335,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 511 |
+
511,252415,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 512 |
+
512,252497,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 513 |
+
513,252579,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 514 |
+
514,252659,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 515 |
+
515,252740,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 516 |
+
516,252819,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 517 |
+
517,252899,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 518 |
+
518,252981,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 519 |
+
519,253062,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 520 |
+
520,253144,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 521 |
+
521,253224,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 522 |
+
522,253305,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 523 |
+
523,253386,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 524 |
+
524,253469,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 525 |
+
525,253549,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 526 |
+
526,253628,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 527 |
+
527,253710,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 528 |
+
528,253791,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 529 |
+
529,253870,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 530 |
+
530,253953,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 531 |
+
531,254033,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 532 |
+
532,254113,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 533 |
+
533,254194,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 534 |
+
534,254276,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 535 |
+
535,254357,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 536 |
+
536,254435,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 537 |
+
537,254516,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 538 |
+
538,254598,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 539 |
+
539,254678,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 540 |
+
540,254759,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 541 |
+
541,254840,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 542 |
+
542,254920,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 543 |
+
543,255002,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 544 |
+
544,255083,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 545 |
+
545,255162,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 546 |
+
546,255248,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 547 |
+
547,255326,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 548 |
+
548,255408,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 549 |
+
549,255487,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 550 |
+
550,255572,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 551 |
+
551,255648,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 552 |
+
552,255729,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 553 |
+
553,255811,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 554 |
+
554,255894,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 555 |
+
555,255976,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 556 |
+
556,256056,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 557 |
+
557,256141,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 558 |
+
558,256218,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 559 |
+
559,256297,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 560 |
+
560,256380,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 561 |
+
561,256475,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 562 |
+
562,256559,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 563 |
+
563,256632,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 564 |
+
564,256712,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 565 |
+
565,256795,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 566 |
+
566,256874,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 567 |
+
567,256956,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 568 |
+
568,257036,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 569 |
+
569,257117,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 570 |
+
570,257205,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 571 |
+
571,257281,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 572 |
+
572,257360,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 573 |
+
573,257441,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 574 |
+
574,257521,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 575 |
+
575,257601,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 576 |
+
576,257682,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 577 |
+
577,257764,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 578 |
+
578,257845,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 579 |
+
579,257927,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 580 |
+
580,258008,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 581 |
+
581,258091,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 582 |
+
582,258171,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 583 |
+
583,258249,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 584 |
+
584,258334,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 585 |
+
585,258431,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 586 |
+
586,258506,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 587 |
+
587,258584,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 588 |
+
588,258666,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 589 |
+
589,258758,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 590 |
+
590,258826,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 591 |
+
591,258906,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 592 |
+
592,258991,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 593 |
+
593,259069,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 594 |
+
594,259150,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 595 |
+
595,259231,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 596 |
+
596,259296,"TERMINAL",0,0,"",,terminal_focus
|
| 597 |
+
597,259404,"TERMINAL",0,0,"[1G[0K⠧[1G[0K⠇",,terminal_output
|
| 598 |
+
598,259475,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 599 |
+
599,259553,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 600 |
+
600,259636,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 601 |
+
601,259715,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 602 |
+
602,259796,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 603 |
+
603,259878,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 604 |
+
604,259959,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 605 |
+
605,260043,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 606 |
+
606,260122,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 607 |
+
607,260203,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 608 |
+
608,260283,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 609 |
+
609,260364,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 610 |
+
610,260445,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 611 |
+
611,260525,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 612 |
+
612,260608,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 613 |
+
613,260688,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 614 |
+
614,260772,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 615 |
+
615,260852,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 616 |
+
616,260943,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 617 |
+
617,261012,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 618 |
+
618,261091,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 619 |
+
619,261171,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 620 |
+
620,261253,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 621 |
+
621,261331,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 622 |
+
622,261413,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 623 |
+
623,261492,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 624 |
+
624,261572,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 625 |
+
625,261652,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 626 |
+
626,261733,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 627 |
+
627,261813,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 628 |
+
628,261894,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 629 |
+
629,261973,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 630 |
+
630,262055,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 631 |
+
631,262085,"TERMINAL",0,0,"cd webview-ui",,terminal_command
|
| 632 |
+
632,262088,"TERMINAL",0,0,"]633;C[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 633 |
+
633,262136,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 634 |
+
634,262216,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 635 |
+
635,262297,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 636 |
+
636,262377,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 637 |
+
637,262460,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 638 |
+
638,262540,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 639 |
+
639,262625,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 640 |
+
640,262701,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 641 |
+
641,262783,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 642 |
+
642,262862,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 643 |
+
643,262946,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 644 |
+
644,263027,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 645 |
+
645,263108,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 646 |
+
646,263188,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 647 |
+
647,263270,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 648 |
+
648,263350,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 649 |
+
649,263431,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 650 |
+
650,263513,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 651 |
+
651,263593,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 652 |
+
652,263674,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 653 |
+
653,263755,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 654 |
+
654,263836,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 655 |
+
655,263917,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 656 |
+
656,263999,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 657 |
+
657,264076,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 658 |
+
658,264158,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 659 |
+
659,264238,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 660 |
+
660,264320,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 661 |
+
661,264400,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 662 |
+
662,264481,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 663 |
+
663,264563,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 664 |
+
664,264644,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 665 |
+
665,264725,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 666 |
+
666,264806,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 667 |
+
667,264886,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 668 |
+
668,264968,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 669 |
+
669,265048,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 670 |
+
670,265132,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 671 |
+
671,265212,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 672 |
+
672,265293,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 673 |
+
673,265340,"TERMINAL",0,0,"npm run storybook",,terminal_command
|
| 674 |
+
674,265373,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 675 |
+
675,265391,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 676 |
+
676,265456,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 677 |
+
677,265535,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 678 |
+
678,265582,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\n[1G[0Ksh: storybook: command not found\r\n[1G[0K⠙[1G[0K[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 679 |
+
679,265617,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 680 |
+
680,265697,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 681 |
+
681,265780,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 682 |
+
682,265860,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 683 |
+
683,265943,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 684 |
+
684,266022,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 685 |
+
685,266104,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 686 |
+
686,266184,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 687 |
+
687,266264,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 688 |
+
688,266344,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 689 |
+
689,266427,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 690 |
+
690,266508,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 691 |
+
691,266589,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 692 |
+
692,266670,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 693 |
+
693,266751,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 694 |
+
694,266833,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 695 |
+
695,266915,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 696 |
+
696,266995,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 697 |
+
697,267076,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 698 |
+
698,267154,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 699 |
+
699,267238,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 700 |
+
700,267319,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 701 |
+
701,267400,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 702 |
+
702,267480,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 703 |
+
703,267559,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 704 |
+
704,267774,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 705 |
+
705,267845,"TERMINAL",0,0,"[1G[0K\r\n> claude-dev@3.38.3 prepare\r\n> husky\r\n\r\n[1G[0K⠋[1G[0K",,terminal_output
|
| 706 |
+
706,268035,"TERMINAL",0,0,"[1G[0K⠙[1G[0K\r\nadded 1505 packages, and audited 1506 packages in 59s\r\n[1G[0K⠙[1G[0K\r\n[1G[0K⠙[1G[0K277 packages are looking for funding\r\n[1G[0K⠙[1G[0K run `npm fund` for details\r\n[1G[0K⠙[1G[0K\r\nfound [32m[1m0[22m[39m vulnerabilities\r\n[1G[0K⠙[1G[0K[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 707 |
+
707,270587,"TERMINAL",0,0,"zsh",,terminal_focus
|
| 708 |
+
708,275890,"TERMINAL",0,0,"cd webview-ui",,terminal_command
|
| 709 |
+
709,275896,"TERMINAL",0,0,"]633;C[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 710 |
+
710,278185,"TERMINAL",0,0,"npm run storybook",,terminal_command
|
| 711 |
+
711,278237,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 712 |
+
712,278317,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\n[1G[0Ksh: storybook: command not found\r\n[1G[0K⠙[1G[0K[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 713 |
+
713,282871,"TERMINAL",0,0,"npm install",,terminal_command
|
| 714 |
+
714,282923,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 715 |
+
715,283584,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 716 |
+
716,283850,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 717 |
+
717,283971,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 718 |
+
718,284051,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 719 |
+
719,284140,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 720 |
+
720,284240,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 721 |
+
721,284501,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 722 |
+
722,284569,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 723 |
+
723,284653,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 724 |
+
724,284737,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 725 |
+
725,284817,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 726 |
+
726,284900,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 727 |
+
727,284984,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 728 |
+
728,285063,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 729 |
+
729,285144,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 730 |
+
730,285227,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 731 |
+
731,285306,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 732 |
+
732,285385,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 733 |
+
733,285465,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 734 |
+
734,285545,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 735 |
+
735,285626,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 736 |
+
736,285705,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 737 |
+
737,285785,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 738 |
+
738,285866,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 739 |
+
739,285946,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 740 |
+
740,286027,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 741 |
+
741,286107,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 742 |
+
742,286187,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 743 |
+
743,286300,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 744 |
+
744,286374,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 745 |
+
745,286452,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 746 |
+
746,286534,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 747 |
+
747,286613,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 748 |
+
748,286694,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 749 |
+
749,286774,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 750 |
+
750,286854,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 751 |
+
751,286934,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 752 |
+
752,287014,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 753 |
+
753,287095,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 754 |
+
754,287175,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 755 |
+
755,287256,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 756 |
+
756,287336,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 757 |
+
757,287415,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 758 |
+
758,287496,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 759 |
+
759,287576,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 760 |
+
760,287656,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 761 |
+
761,287734,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 762 |
+
762,287817,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 763 |
+
763,287928,"TERMINAL",0,0,"[1G[0K⠏[1G[0K[1mnpm[22m [33mwarn[39m [94mdeprecated[39m @vscode/webview-ui-toolkit@1.4.0: This package has been deprecated, https://github.com/microsoft/vscode-webview-ui-toolkit/issues/561\r\n[1G[0K⠏",,terminal_output
|
| 764 |
+
764,288009,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 765 |
+
765,288091,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 766 |
+
766,288171,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 767 |
+
767,288249,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 768 |
+
768,288329,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 769 |
+
769,288411,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 770 |
+
770,288491,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 771 |
+
771,288572,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 772 |
+
772,288653,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 773 |
+
773,288734,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 774 |
+
774,288814,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 775 |
+
775,288894,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 776 |
+
776,288976,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 777 |
+
777,289057,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 778 |
+
778,289135,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 779 |
+
779,289218,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 780 |
+
780,289298,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 781 |
+
781,289378,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 782 |
+
782,289456,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 783 |
+
783,289537,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 784 |
+
784,289617,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 785 |
+
785,289697,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 786 |
+
786,289776,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 787 |
+
787,289858,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 788 |
+
788,289937,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 789 |
+
789,290016,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 790 |
+
790,290098,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 791 |
+
791,290176,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 792 |
+
792,290258,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 793 |
+
793,290339,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 794 |
+
794,290418,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 795 |
+
795,290496,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 796 |
+
796,290578,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 797 |
+
797,290659,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 798 |
+
798,290738,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 799 |
+
799,290820,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 800 |
+
800,290900,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 801 |
+
801,290982,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 802 |
+
802,291061,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 803 |
+
803,291142,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 804 |
+
804,291222,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 805 |
+
805,291303,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 806 |
+
806,291381,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 807 |
+
807,291463,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 808 |
+
808,291542,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 809 |
+
809,291623,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 810 |
+
810,291702,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 811 |
+
811,291781,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 812 |
+
812,291865,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 813 |
+
813,291944,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 814 |
+
814,292025,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 815 |
+
815,292106,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 816 |
+
816,292186,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 817 |
+
817,292266,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 818 |
+
818,292349,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 819 |
+
819,292426,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 820 |
+
820,292506,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 821 |
+
821,292585,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 822 |
+
822,292667,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 823 |
+
823,292747,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 824 |
+
824,292828,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 825 |
+
825,292908,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 826 |
+
826,292999,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 827 |
+
827,293078,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 828 |
+
828,293160,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 829 |
+
829,293239,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 830 |
+
830,293320,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 831 |
+
831,293399,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 832 |
+
832,293481,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 833 |
+
833,293558,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 834 |
+
834,293639,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 835 |
+
835,293720,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 836 |
+
836,293802,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 837 |
+
837,293880,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 838 |
+
838,293960,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 839 |
+
839,294040,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 840 |
+
840,294121,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 841 |
+
841,294202,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 842 |
+
842,294283,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 843 |
+
843,294363,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 844 |
+
844,294441,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 845 |
+
845,294523,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 846 |
+
846,294621,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 847 |
+
847,294701,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 848 |
+
848,294780,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 849 |
+
849,294861,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 850 |
+
850,294943,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 851 |
+
851,295021,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 852 |
+
852,295102,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 853 |
+
853,295186,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 854 |
+
854,295286,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 855 |
+
855,295359,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 856 |
+
856,295440,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 857 |
+
857,295521,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 858 |
+
858,295601,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 859 |
+
859,295682,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 860 |
+
860,295763,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 861 |
+
861,295844,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 862 |
+
862,295926,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 863 |
+
863,296013,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 864 |
+
864,296086,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 865 |
+
865,296170,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 866 |
+
866,296251,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 867 |
+
867,296330,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 868 |
+
868,296411,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 869 |
+
869,296492,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 870 |
+
870,296572,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 871 |
+
871,296651,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 872 |
+
872,296732,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 873 |
+
873,296815,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 874 |
+
874,296891,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 875 |
+
875,296970,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 876 |
+
876,297050,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 877 |
+
877,297130,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 878 |
+
878,297211,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 879 |
+
879,297291,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 880 |
+
880,297374,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 881 |
+
881,297453,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 882 |
+
882,297533,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 883 |
+
883,297613,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 884 |
+
884,297694,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 885 |
+
885,297775,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 886 |
+
886,297853,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 887 |
+
887,297934,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 888 |
+
888,298014,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 889 |
+
889,298094,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 890 |
+
890,298173,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 891 |
+
891,298254,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 892 |
+
892,298334,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 893 |
+
893,298415,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 894 |
+
894,298494,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 895 |
+
895,298573,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 896 |
+
896,298656,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 897 |
+
897,298737,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 898 |
+
898,298817,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 899 |
+
899,298899,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 900 |
+
900,298979,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 901 |
+
901,299058,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 902 |
+
902,299139,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 903 |
+
903,299220,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 904 |
+
904,299313,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 905 |
+
905,299394,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 906 |
+
906,299475,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 907 |
+
907,299555,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 908 |
+
908,299638,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 909 |
+
909,299717,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 910 |
+
910,299800,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 911 |
+
911,299879,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 912 |
+
912,299961,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 913 |
+
913,300041,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 914 |
+
914,300122,"TERMINAL",0,0,"[1G[0K⠋",,terminal_output
|
| 915 |
+
915,300204,"TERMINAL",0,0,"[1G[0K⠙",,terminal_output
|
| 916 |
+
916,300284,"TERMINAL",0,0,"[1G[0K⠹",,terminal_output
|
| 917 |
+
917,300366,"TERMINAL",0,0,"[1G[0K⠸",,terminal_output
|
| 918 |
+
918,300447,"TERMINAL",0,0,"[1G[0K⠼",,terminal_output
|
| 919 |
+
919,300528,"TERMINAL",0,0,"[1G[0K⠴",,terminal_output
|
| 920 |
+
920,300610,"TERMINAL",0,0,"[1G[0K⠦",,terminal_output
|
| 921 |
+
921,300691,"TERMINAL",0,0,"[1G[0K⠧",,terminal_output
|
| 922 |
+
922,300773,"TERMINAL",0,0,"[1G[0K⠇",,terminal_output
|
| 923 |
+
923,300989,"TERMINAL",0,0,"[1G[0K⠏",,terminal_output
|
| 924 |
+
924,301052,"TERMINAL",0,0,"[1G[0K⠋[1G[0K\r\nadded 978 packages, and audited 984 packages in 18s\r\n[1G[0K⠋[1G[0K\r\n[1G[0K⠋[1G[0K187 packages are looking for funding\r\n[1G[0K⠋[1G[0K run `npm fund` for details\r\n[1G[0K⠋[1G[0K\r\nfound [32m[1m0[22m[39m vulnerabilities\r\n[1G[0K⠋[1G[0K[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 925 |
+
925,374417,"TERMINAL",0,0,"npm run storybook",,terminal_command
|
| 926 |
+
926,374468,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 927 |
+
927,374749,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\n[1G[0K",,terminal_output
|
| 928 |
+
928,375925,"TERMINAL",0,0,"[1mstorybook v9.1.7[22m[0m\r\n[0m\r\n",,terminal_output
|
| 929 |
+
929,376672,"TERMINAL",0,0,"(node:36013) ExperimentalWarning: Type Stripping is an experimental feature and might change at any time\r\n(Use `node --trace-warnings ...` to show where the warning was created)\r\n",,terminal_output
|
| 930 |
+
930,376730,"TERMINAL",0,0,"[34mAttention:[39m Storybook now collects completely anonymous telemetry regarding usage. This information is used to shape Storybook's roadmap and prioritize features.\r\nYou can learn more, including how to opt-out if you'd not like to participate in this anonymous program, by visiting the following URL:\r\n[36mhttps://storybook.js.org/telemetry[39m\r\n\r\n",,terminal_output
|
| 931 |
+
931,377210,"TERMINAL",0,0,"[32minfo[0m[35m[0m => Starting manager..\r\n[0m",,terminal_output
|
| 932 |
+
932,377331,"TERMINAL",0,0,"No story files found for the specified pattern: [34msrc/**/*.mdx[39m\r\n",,terminal_output
|
| 933 |
+
933,377521,"TERMINAL",0,0,"[32minfo[0m[35m[0m => Starting preview..\r\n[0m",,terminal_output
|
| 934 |
+
934,379138,"TERMINAL",0,0,"Building webview for vscode\r\n",,terminal_output
|
| 935 |
+
935,380938,"TERMINAL",0,0,"[32minfo[0m[35m[0m Using tsconfig paths for react-docgen\r\n[0m[33mThe `define` option contains an object with ""PATH"" for ""process.env"" key. It looks like you may have passed the entire `process.env` object to `define`, which can unintentionally expose all environment variables. This poses a security risk and is discouraged.[39m\r\n",,terminal_output
|
| 936 |
+
936,387648,"TERMINAL",0,0,"[38;2;241;97;140m╭──────────────────────────────────────────────────────────────────────────────────────────╮[39m\r\n[38;2;241;97;140m│[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [38;2;162;224;94mStorybook [1m9.1.7[22m for [1mreact-vite[22m started[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [90m[4m312 ms[24m for manager and [4m10 s[24m for preview[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m Local: [36mhttp://localhost:6006/[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m On your network: [36mhttp://192.168.178.68:6006/[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [38;2;243;173;56mA new version ([1m10.1.0[22m) is available![39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [90mUpgrade now:[39m [38;2;162;224;94mnpx storybook@latest upgrade[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [90mRead full changelog:[39m [90m[4mhttps://github.com/storybookjs/storybook/blob/main/CHANGELOG.md[24m[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m╰──────────────────────────────────────────────────────────────────────────────────────────╯[39m\r\n",,terminal_output
|
| 937 |
+
937,390868,"TERMINAL",0,0,"[baseline-browser-mapping] The data in this module is over two months old. To ensure accurate Baseline data, please update: `npm i baseline-browser-mapping@latest -D`\r\n",,terminal_output
|
| 938 |
+
938,395179,"TERMINAL",0,0,"[2m12:54:54 PM[22m [31m[1m[vite][22m[39m [31m[2m(client)[22m[39m Pre-transform error: Failed to resolve import ""@shared/proto/cline/common"" from ""src/context/ClineAuthContext.tsx"". Does the file exist?\r\n Plugin: [35mvite:import-analysis[39m\r\n File: [36m/Users/franzsrambical/Documents/pdoom/cline/webview-ui/src/context/ClineAuthContext.tsx[39m:2:29\r\n[33m 20 | import { jsxDEV as _jsxDEV } from ""react/jsx-dev-runtime"";\r\n 21 | var _s = $RefreshSig$(), _s1 = $RefreshSig$();\r\n 22 | import { EmptyRequest } from ""@shared/proto/cline/common"";\r\n | ^\r\n 23 | import deepEqual from ""fast-deep-equal"";\r\n 24 | import { createContext, useCallback, useContext, useEffect, useMemo, useState } from ""react"";[39m\r\n[2m12:54:54 PM[22m [31m[1m[vite][22m[39m [31m[2m(client)[22m[39m Pre-transform error: Failed to resolve import ""../services/grpc-client"" from ""src/context/ExtensionStateContext.tsx"". Does the file exist?\r\n Plugin: [35mvite:import-analysis[39m\r\n File: [36m/Users/franzsrambical/Documents/pdoom/cline/webview-ui/src/context/ExtensionStateContext.tsx[39m:31:91\r\n[33m 34 | import { Environment } from ""../../../src/config"";\r\n 35 | import { basetenDefaultModelId, basetenModels, groqDefaultModelId, groqModels, openRouterDefaultModelId, openRouterDe...\r\n 36 | import { McpServiceClient, ModelsServiceClient, StateServiceClient, UiServiceClient } from ""../services/grpc-client"";\r\n | ^\r\n 37 | export const ExtensionStateContext = /*#__PURE__*/ createContext(undefined);\r\n 38 | export const ExtensionStateContextProvider = ({ children })=>{[39m\r\n",,terminal_output
|
| 939 |
+
939,395343,"TERMINAL",0,0,"[2m12:54:54 PM[22m [31m[1m[vite][22m[39m [31m[2m(client)[22m[39m Pre-transform error: Failed to resolve import ""@shared/proto/cline/ui"" from ""../src/shared/proto-conversions/cline-message.ts"". Does the file exist?\r\n Plugin: [35mvite:import-analysis[39m\r\n File: [36m/Users/franzsrambical/Documents/pdoom/cline/src/shared/proto-conversions/cline-message.ts[39m:3:88\r\n[33m 1 | import { ClineAsk, ClineMessageType, ClineSay } from ""@shared/proto/cline/ui"";\r\n | ^\r\n 2 | // Helper function to convert ClineAsk string to enum\r\n 3 | function convertClineAskToProtoEnum(ask) {[39m\r\n[2m12:54:54 PM[22m [31m[1m[vite][22m[39m [31m[2m(client)[22m[39m Pre-transform error: Failed to resolve import ""@shared/proto/cline/mcp"" from ""../src/shared/proto-conversions/mcp/mcp-server-conversion.ts"". Does the file exist?\r\n Plugin: [35mvite:import-analysis[39m\r\n File: [36m/Users/franzsrambical/Documents/pdoom/cline/src/shared/proto-conversions/mcp/mcp-server-conversion.ts[39m:7:7\r\n[33m 1 | import { McpServerStatus } from ""@shared/proto/cline/mcp"";\r\n | ^\r\n 2 | // Helper to convert TS status to Proto enum\r\n 3 | function convertMcpStatusToProto(status) {[39m\r\n[2m12:54:54 PM[22m [31m[1m[vite][22m[39m [31m[2m(client)[22m[39m Pre-transform error: Failed to resolve import ""@shared/proto/cline/models"" from ""../src/shared/proto-conversions/models/typeConversion.ts"". Does the file exist?\r\n Plugin: [35mvite:import-analysis[39m\r\n File: [36m/Users/franzsrambical/Documents/pdoom/cline/src/shared/proto-conversions/models/typeConversion.ts[39m:8:7\r\n[33m 1 | import { OpenRouterModelInfo, ThinkingConfig } from ""@shared/proto/cline/models"";\r\n | ^\r\n 2 | /**\r\n 3 | * Convert protobuf ThinkingConfig to application ThinkingConfig[39m\r\n",,terminal_output
|
| 940 |
+
940,398882,"TERMINAL",0,0,"\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[1;1H[0J[2m12:54:58 PM[22m [31m[1m[vite][22m[39m [31mInternal server error: Failed to resolve import ""@shared/proto/cline/common"" from ""src/context/ClineAuthContext.tsx"". Does the file exist?[39m\r\n Plugin: [35mvite:import-analysis[39m\r\n File: [36m/Users/franzsrambical/Documents/pdoom/cline/webview-ui/src/context/ClineAuthContext.tsx[39m:2:29\r\n[33m 20 | import { jsxDEV as _jsxDEV } from ""react/jsx-dev-runtime"";\r\n 21 | var _s = $RefreshSig$(), _s1 = $RefreshSig$();\r\n 22 | import { EmptyRequest } from ""@shared/proto/cline/common"";\r\n | ^\r\n 23 | import deepEqual from ""fast-deep-equal"";\r\n 24 | import { createContext, useCallback, useContext, useEffect, useMemo, useState } from ""react"";[39m\r\n at TransformPluginContext._formatLog (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:29618:43)\r\n at TransformPluginContext.error (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:29615:14)\r\n at normalizeUrl (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:27738:18)\r\n at async file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:27796:32\r\n at async Promise.all (index 3)\r\n at async TransformPluginContext.transform (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:27764:4)\r\n at async EnvironmentPluginContainer.transform (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:29416:14)\r\n at async loadAndTransform (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:23287:26)\r\n at async viteTransformMiddleware (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:25159:20)\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[1;1H[0J[2m12:54:58 PM[22m [31m[1m[vite][22m[39m [31mInternal server error: Failed to resolve import ""../services/grpc-client"" from ""src/context/ExtensionStateContext.tsx"". Does the file exist?[39m\r\n Plugin: [35mvite:import-analysis[39m\r\n File: [36m/Users/franzsrambical/Documents/pdoom/cline/webview-ui/src/context/ExtensionStateContext.tsx[39m:31:91\r\n[33m 34 | import { Environment } from ""../../../src/config"";\r\n 35 | import { basetenDefaultModelId, basetenModels, groqDefaultModelId, groqModels, openRouterDefaultModelId, openRouterDe...\r\n 36 | import { McpServiceClient, ModelsServiceClient, StateServiceClient, UiServiceClient } from ""../services/grpc-client"";\r\n | ^\r\n 37 | export const ExtensionStateContext = /*#__PURE__*/ createContext(undefined);\r\n 38 | export const ExtensionStateContextProvider = ({ children })=>{[39m\r\n at TransformPluginContext._formatLog (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:29618:43)\r\n at TransformPluginContext.error (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:29615:14)\r\n at normalizeUrl (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:27738:18)\r\n at async file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:27796:32\r\n at async Promise.all (index 17)\r\n at async TransformPluginContext.transform (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:27764:4)\r\n at async EnvironmentPluginContainer.transform (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:29416:14)\r\n at async loadAndTransform (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:23287:26)\r\n at async viteTransformMiddleware (file:///Users/franzsrambical/Documents/pdoom/cline/webview-ui/node_modules/vite/dist/node/chunks/config.js:25159:20)\r\n[2m12:54:58 PM[22m [31m[1m[vite][22m[39m [31m[2m(client)[22m[39m Pre-transform error: Failed to resolve import ""@shared/proto/cline/mcp"" from ""../src/shared/proto-conversions/mcp/mcp-server-conversion.ts"". Does the file exist?\r\n Plugin: [35mvite:import-analysis[39m\r\n File: [36m/Users/franzsrambical/Documents/pdoom/cline/src/shared/proto-conversions/mcp/mcp-server-conversion.ts[39m:7:7\r\n[33m 1 | import { McpServerStatus } from ""@shared/proto/cline/mcp"";\r\n | ^\r\n 2 | // Helper to convert TS status to Proto enum\r\n 3 | function convertMcpStatusToProto(status) {[39m\r\n[2m12:54:58 PM[22m [31m[1m[vite][22m[39m [31m[2m(client)[22m[39m Pre-transform error: Failed to resolve import ""@shared/proto/cline/models"" from ""../src/shared/proto-conversions/models/typeConversion.ts"". Does the file exist?\r\n Plugin: [35mvite:import-analysis[39m\r\n File: [36m/Users/franzsrambical/Documents/pdoom/cline/src/shared/proto-conversions/models/typeConversion.ts[39m:8:7\r\n[33m 1 | import { OpenRouterModelInfo, ThinkingConfig } from ""@shared/proto/cline/models"";\r\n | ^\r\n 2 | /**\r\n 3 | * Convert protobuf ThinkingConfig to application ThinkingConfig[39m\r\n[2m12:54:58 PM[22m [31m[1m[vite][22m[39m [31m[2m(client)[22m[39m Pre-transform error: Failed to resolve import ""@shared/proto/cline/ui"" from ""../src/shared/proto-conversions/cline-message.ts"". Does the file exist?\r\n Plugin: [35mvite:import-analysis[39m\r\n File: [36m/Users/franzsrambical/Documents/pdoom/cline/src/shared/proto-conversions/cline-message.ts[39m:3:88\r\n[33m 1 | import { ClineAsk, ClineMessageType, ClineSay } from ""@shared/proto/cline/ui"";\r\n | ^\r\n 2 | // Helper function to convert ClineAsk string to enum\r\n 3 | function convertClineAskToProtoEnum(ask) {[39m\r\n",,terminal_output
|
| 941 |
+
941,2981449,"TERMINAL",0,0,"\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[1;1H[0J[2m1:38:01 PM[22m [36m[1m[vite][22m[39m [90m[2m(client)[22m[39m [32mpage reload [39m[2msrc/services/grpc-client.ts[22m\r\n",,terminal_output
|
| 942 |
+
942,2987755,"TERMINAL",0,0,"[2m1:38:07 PM[22m [36m[1m[vite][22m[39m [90m[2m(client)[22m[39m [32m✨ new dependencies optimized: [33m@bufbuild/protobuf/wire, uuid[32m[39m\r\n[2m1:38:07 PM[22m [36m[1m[vite][22m[39m [90m[2m(client)[22m[39m [32m✨ optimized dependencies changed. reloading[39m\r\n",,terminal_output
|
| 943 |
+
943,2992777,"TERMINAL",0,0,"\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[1;1H[0J[2m1:38:12 PM[22m [36m[1m[vite][22m[39m [90m[2m(client)[22m[39m [32mhmr update [39m[2m/src/index.css[22m\r\n",,terminal_output
|
| 944 |
+
944,3008542,"TERMINAL",0,0,"\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[1;1H[0J[2m1:38:28 PM[22m [36m[1m[vite][22m[39m [90m[2m(client)[22m[39m [32mhmr update [39m[2m/src/index.css[22m [33m(x2)[39m\r\n",,terminal_output
|
| 945 |
+
945,3026313,"TERMINAL",0,0,"\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[1;1H[0J[2m1:38:45 PM[22m [36m[1m[vite][22m[39m [90m[2m(client)[22m[39m [32mhmr update [39m[2m/src/index.css[22m [33m(x3)[39m\r\n",,terminal_output
|
| 946 |
+
946,3034948,"TERMINAL",0,0,"^C",,terminal_output
|
| 947 |
+
947,3035530,"TERMINAL",0,0,"[1G[0K⠙[1G[0K[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 948 |
+
948,3036908,"TERMINAL",0,0,"npm run storybook",,terminal_command
|
| 949 |
+
949,3036960,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 950 |
+
950,3037124,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\n[1G[0K",,terminal_output
|
| 951 |
+
951,3038143,"TERMINAL",0,0,"[1mstorybook v9.1.7[22m[0m\r\n[0m\r\n",,terminal_output
|
| 952 |
+
952,3038311,"TERMINAL",0,0,"(node:43008) ExperimentalWarning: Type Stripping is an experimental feature and might change at any time\r\n(Use `node --trace-warnings ...` to show where the warning was created)\r\n",,terminal_output
|
| 953 |
+
953,3038770,"TERMINAL",0,0,"[32minfo[0m[35m[0m => Starting manager..\r\n[0m",,terminal_output
|
| 954 |
+
954,3038852,"TERMINAL",0,0,"No story files found for the specified pattern: [34msrc/**/*.mdx[39m\r\n",,terminal_output
|
| 955 |
+
955,3039062,"TERMINAL",0,0,"[32minfo[0m[35m[0m => Starting preview..\r\n[0m",,terminal_output
|
| 956 |
+
956,3040101,"TERMINAL",0,0,"Building webview for vscode\r\n",,terminal_output
|
| 957 |
+
957,3042195,"TERMINAL",0,0,"[32minfo[0m[35m[0m Using tsconfig paths for react-docgen\r\n[0m[38;2;241;97;97m=> Failed to build the preview[39m\r\n[38;2;241;97;97mReferenceError: __dirname is not defined[39m\r\n at Object.viteFinal (./.storybook/main.ts:13:17)\r\n at ./node_modules/storybook/dist/common/index.cjs:23650:18\r\n at async createViteServer (./node_modules/@storybook/builder-vite/dist/index.js:92:25)\r\n at async Module.start (./node_modules/@storybook/builder-vite/dist/index.js:92:658)\r\n at async storybookDevServer (./node_modules/storybook/dist/core-server/index.cjs:42917:79)\r\n at async buildOrThrow (./node_modules/storybook/dist/core-server/index.cjs:39285:12)\r\n at async buildDevStandalone (./node_modules/storybook/dist/core-server/index.cjs:44170:78)\r\n at async withTelemetry (./node_modules/storybook/dist/core-server/index.cjs:42279:12)\r\n at async dev (./node_modules/storybook/dist/cli/bin/index.cjs:5905:3)\r\n at async r.<anonymous> (./node_modules/storybook/dist/cli/bin/index.cjs:6015:74)\r\n\r\nBroken build, fix the error above.\r\nYou may need to refresh the browser.\r\n\r\n[?25l[2K[1G[36m?[39m [1mWould you like to help improve Storybook by sending anonymous crash reports?[22m [90m›[39m [90m(Y/n)[39m",,terminal_output
|
| 958 |
+
958,3056937,"TERMINAL",0,0,"[2K[G[2K[1G[32m✔[39m [1mWould you like to help improve Storybook by sending anonymous crash reports?[22m [90m…[39m no\r\n[?25h",,terminal_output
|
| 959 |
+
959,3057351,"TERMINAL",0,0,"[1G[0K⠙[1G[0K[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 960 |
+
960,3058995,"TERMINAL",0,0,"npm run storybook",,terminal_command
|
| 961 |
+
961,3059046,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 962 |
+
962,3059206,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\n[1G[0K",,terminal_output
|
| 963 |
+
963,3059534,"TERMINAL",0,0,"[1mstorybook v9.1.7[22m[0m\r\n[0m\r\n",,terminal_output
|
| 964 |
+
964,3059688,"TERMINAL",0,0,"(node:43148) ExperimentalWarning: Type Stripping is an experimental feature and might change at any time\r\n(Use `node --trace-warnings ...` to show where the warning was created)\r\n",,terminal_output
|
| 965 |
+
965,3060001,"TERMINAL",0,0,"[32minfo[0m[35m[0m => Starting manager..\r\n[0mNo story files found for the specified pattern: [34msrc/**/*.mdx[39m\r\n",,terminal_output
|
| 966 |
+
966,3060120,"TERMINAL",0,0,"[32minfo[0m[35m[0m => Starting preview..\r\n[0m",,terminal_output
|
| 967 |
+
967,3060243,"TERMINAL",0,0,"Building webview for vscode\r\n",,terminal_output
|
| 968 |
+
968,3060516,"TERMINAL",0,0,"[32minfo[0m[35m[0m Using tsconfig paths for react-docgen\r\n[0m[38;2;241;97;97m=> Failed to build the preview[39m\r\n[38;2;241;97;97mReferenceError: __dirname is not defined[39m\r\n at Object.viteFinal (./.storybook/main.ts:13:17)\r\n at ./node_modules/storybook/dist/common/index.cjs:23650:18\r\n at async createViteServer (./node_modules/@storybook/builder-vite/dist/index.js:92:25)\r\n at async Module.start (./node_modules/@storybook/builder-vite/dist/index.js:92:658)\r\n at async storybookDevServer (./node_modules/storybook/dist/core-server/index.cjs:42917:79)\r\n at async buildOrThrow (./node_modules/storybook/dist/core-server/index.cjs:39285:12)\r\n at async buildDevStandalone (./node_modules/storybook/dist/core-server/index.cjs:44170:78)\r\n at async withTelemetry (./node_modules/storybook/dist/core-server/index.cjs:42279:12)\r\n at async dev (./node_modules/storybook/dist/cli/bin/index.cjs:5905:3)\r\n at async r.<anonymous> (./node_modules/storybook/dist/cli/bin/index.cjs:6015:74)\r\n\r\nBroken build, fix the error above.\r\nYou may need to refresh the browser.\r\n\r\n",,terminal_output
|
| 969 |
+
969,3060956,"TERMINAL",0,0,"[1G[0K⠙[1G[0K[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 970 |
+
970,3078567,"TERMINAL",0,0,"npm run storybook",,terminal_command
|
| 971 |
+
971,3078619,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 972 |
+
972,3078759,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\n[1G[0K",,terminal_output
|
| 973 |
+
973,3079085,"TERMINAL",0,0,"[1mstorybook v9.1.7[22m[0m\r\n[0m\r\n",,terminal_output
|
| 974 |
+
974,3079242,"TERMINAL",0,0,"(node:43265) ExperimentalWarning: Type Stripping is an experimental feature and might change at any time\r\n(Use `node --trace-warnings ...` to show where the warning was created)\r\n",,terminal_output
|
| 975 |
+
975,3079508,"TERMINAL",0,0,"[32minfo[0m[35m[0m => Starting manager..\r\n[0mNo story files found for the specified pattern: [34msrc/**/*.mdx[39m\r\n",,terminal_output
|
| 976 |
+
976,3079609,"TERMINAL",0,0,"[32minfo[0m[35m[0m => Starting preview..\r\n[0m",,terminal_output
|
| 977 |
+
977,3079857,"TERMINAL",0,0,"Building webview for vscode\r\n",,terminal_output
|
| 978 |
+
978,3080137,"TERMINAL",0,0,"[32minfo[0m[35m[0m Using tsconfig paths for react-docgen\r\n[0m[38;2;241;97;97m=> Failed to build the preview[39m\r\n[38;2;241;97;97mReferenceError: __dirname is not defined[39m\r\n at Object.viteFinal (./.storybook/main.ts:13:17)\r\n at ./node_modules/storybook/dist/common/index.cjs:23650:18\r\n at async createViteServer (./node_modules/@storybook/builder-vite/dist/index.js:92:25)\r\n at async Module.start (./node_modules/@storybook/builder-vite/dist/index.js:92:658)\r\n at async storybookDevServer (./node_modules/storybook/dist/core-server/index.cjs:42917:79)\r\n at async buildOrThrow (./node_modules/storybook/dist/core-server/index.cjs:39285:12)\r\n at async buildDevStandalone (./node_modules/storybook/dist/core-server/index.cjs:44170:78)\r\n at async withTelemetry (./node_modules/storybook/dist/core-server/index.cjs:42279:12)\r\n at async dev (./node_modules/storybook/dist/cli/bin/index.cjs:5905:3)\r\n at async r.<anonymous> (./node_modules/storybook/dist/cli/bin/index.cjs:6015:74)\r\n\r\nBroken build, fix the error above.\r\nYou may need to refresh the browser.\r\n\r\n",,terminal_output
|
| 979 |
+
979,3080559,"TERMINAL",0,0,"[1G[0K⠙[1G[0K[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 980 |
+
980,3082050,"TERMINAL",0,0,"cd ..",,terminal_command
|
| 981 |
+
981,3082051,"TERMINAL",0,0,"]633;C[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 982 |
+
982,3087158,"TERMINAL",0,0,"npm run storybook",,terminal_command
|
| 983 |
+
983,3087209,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 984 |
+
984,3087345,"TERMINAL",0,0,"\r\n> claude-dev@3.38.3 storybook\r\n> cd webview-ui && npm run storybook\r\n\r\n[1G[0K",,terminal_output
|
| 985 |
+
985,3087427,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\n[1G[0K",,terminal_output
|
| 986 |
+
986,3087717,"TERMINAL",0,0,"[1mstorybook v9.1.7[22m[0m\r\n[0m\r\n",,terminal_output
|
| 987 |
+
987,3087864,"TERMINAL",0,0,"(node:43453) ExperimentalWarning: Type Stripping is an experimental feature and might change at any time\r\n(Use `node --trace-warnings ...` to show where the warning was created)\r\n",,terminal_output
|
| 988 |
+
988,3088122,"TERMINAL",0,0,"[32minfo[0m[35m[0m => Starting manager..\r\n[0mNo story files found for the specified pattern: [34msrc/**/*.mdx[39m\r\n",,terminal_output
|
| 989 |
+
989,3088223,"TERMINAL",0,0,"[32minfo[0m[35m[0m => Starting preview..\r\n[0m",,terminal_output
|
| 990 |
+
990,3088362,"TERMINAL",0,0,"Building webview for vscode\r\n",,terminal_output
|
| 991 |
+
991,3088601,"TERMINAL",0,0,"[32minfo[0m[35m[0m Using tsconfig paths for react-docgen\r\n[0m[38;2;241;97;97m=> Failed to build the preview[39m\r\n[38;2;241;97;97mReferenceError: __dirname is not defined[39m\r\n at Object.viteFinal (./.storybook/main.ts:13:17)\r\n at ./node_modules/storybook/dist/common/index.cjs:23650:18\r\n at async createViteServer (./node_modules/@storybook/builder-vite/dist/index.js:92:25)\r\n at async Module.start (./node_modules/@storybook/builder-vite/dist/index.js:92:658)\r\n at async storybookDevServer (./node_modules/storybook/dist/core-server/index.cjs:42917:79)\r\n at async buildOrThrow (./node_modules/storybook/dist/core-server/index.cjs:39285:12)\r\n at async buildDevStandalone (./node_modules/storybook/dist/core-server/index.cjs:44170:78)\r\n at async withTelemetry (./node_modules/storybook/dist/core-server/index.cjs:42279:12)\r\n at async dev (./node_modules/storybook/dist/cli/bin/index.cjs:5905:3)\r\n at async r.<anonymous> (./node_modules/storybook/dist/cli/bin/index.cjs:6015:74)\r\n\r\nBroken build, fix the error above.\r\nYou may need to refresh the browser.\r\n\r\n",,terminal_output
|
| 992 |
+
992,3088992,"TERMINAL",0,0,"[1G[0K⠙[1G[0K[1G[0K⠙[1G[0K[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 993 |
+
993,3161357,"TERMINAL",0,0,"cd webview-ui",,terminal_command
|
| 994 |
+
994,3161357,"TERMINAL",0,0,"]633;C[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 995 |
+
995,3162625,"TERMINAL",0,0,"npm run storybook",,terminal_command
|
| 996 |
+
996,3162677,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 997 |
+
997,3162941,"TERMINAL",0,0,"\r\n> webview-ui@0.3.0 storybook\r\n> storybook dev -p 6006\r\n\r\n[1G[0K",,terminal_output
|
| 998 |
+
998,3163284,"TERMINAL",0,0,"[1mstorybook v9.1.7[22m[0m\r\n[0m\r\n",,terminal_output
|
| 999 |
+
999,3163477,"TERMINAL",0,0,"(node:43708) ExperimentalWarning: Type Stripping is an experimental feature and might change at any time\r\n(Use `node --trace-warnings ...` to show where the warning was created)\r\n",,terminal_output
|
| 1000 |
+
1000,3163764,"TERMINAL",0,0,"[32minfo[0m[35m[0m => Starting manager..\r\n[0mNo story files found for the specified pattern: [34msrc/**/*.mdx[39m\r\n",,terminal_output
|
| 1001 |
+
1001,3163883,"TERMINAL",0,0,"[32minfo[0m[35m[0m => Starting preview..\r\n[0m",,terminal_output
|
| 1002 |
+
1002,3164047,"TERMINAL",0,0,"Building webview for vscode\r\n",,terminal_output
|
| 1003 |
+
1003,3164319,"TERMINAL",0,0,"[32minfo[0m[35m[0m Using tsconfig paths for react-docgen\r\n[0m[33mThe `define` option contains an object with ""PATH"" for ""process.env"" key. It looks like you may have passed the entire `process.env` object to `define`, which can unintentionally expose all environment variables. This poses a security risk and is discouraged.[39m\r\n",,terminal_output
|
| 1004 |
+
1004,3169884,"TERMINAL",0,0,"[38;2;241;97;140m╭──────────────────────────────────────────────────────────────────────────────────────────╮[39m\r\n[38;2;241;97;140m│[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [38;2;162;224;94mStorybook [1m9.1.7[22m for [1mreact-vite[22m started[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [90m[4m157 ms[24m for manager and [4m5.95 s[24m for preview[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m Local: [36mhttp://localhost:6006/[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m On your network: [36mhttp://192.168.178.68:6006/[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [38;2;243;173;56mA new version ([1m10.1.0[22m) is available![39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [90mUpgrade now:[39m [38;2;162;224;94mnpx storybook@latest upgrade[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [90mRead full changelog:[39m [90m[4mhttps://github.com/storybookjs/storybook/blob/main/CHANGELOG.md[24m[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m│[39m [38;2;241;97;140m│[39m\r\n[38;2;241;97;140m╰──────────────────────────────────────────────────────────────────────────────────────────╯[39m\r\n",,terminal_output
|
| 1005 |
+
1005,3170751,"TERMINAL",0,0,"[baseline-browser-mapping] The data in this module is over two months old. To ensure accurate Baseline data, please update: `npm i baseline-browser-mapping@latest -D`\r\n",,terminal_output
|
| 1006 |
+
1006,3210633,"TERMINAL",0,0,"The request id ""/Users/franzsrambical/Documents/pdoom/cline/node_modules/@vscode/codicons/dist/codicon.ttf"" is outside of Vite serving allow list.\r\n\r\n- /Users/franzsrambical/Documents/pdoom/cline/webview-ui\r\n\r\nRefer to docs https://vite.dev/config/server-options.html#server-fs-allow for configurations and more details.\r\n\r\n",,terminal_output
|
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-8eaf046e-99c4-4091-a85d-91e359564aa51756825908437-2025_09_02-17.11.50.753/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-9543d8e2-2376-4957-873e-df7016d502961763465687199-2025_11_18-12.34.49.442/source.csv
ADDED
|
@@ -0,0 +1,226 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
2,67,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"12:34:49 PM [info] Activating crowd-code\n12:34:49 PM [info] Recording started\n12:34:49 PM [info] Initializing git provider using file system watchers...\n12:34:49 PM [info] Git repository found\n12:34:49 PM [info] Git provider initialized successfully\n",Log,tab
|
| 3 |
+
3,116,"extension-output-pdoom-org.crowd-code-#1-crowd-code",250,0,"12:34:49 PM [info] Initial git state: [object Object]\n",Log,content
|
| 4 |
+
4,2975,"extension-output-pdoom-org.crowd-code-#1-crowd-code",304,0,"",Log,selection_mouse
|
| 5 |
+
5,13030,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,tab
|
| 6 |
+
6,20469,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",64,0,"",python,selection_command
|
| 7 |
+
7,20661,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",66,0,"",python,selection_command
|
| 8 |
+
8,20694,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",132,0,"",python,selection_command
|
| 9 |
+
9,20728,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",199,0,"",python,selection_command
|
| 10 |
+
10,20761,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",241,0,"",python,selection_command
|
| 11 |
+
11,20794,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",243,0,"",python,selection_command
|
| 12 |
+
12,20828,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",292,0,"",python,selection_command
|
| 13 |
+
13,20863,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",294,0,"",python,selection_command
|
| 14 |
+
14,20904,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",364,0,"",python,selection_command
|
| 15 |
+
15,20936,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",432,0,"",python,selection_command
|
| 16 |
+
16,20962,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",507,0,"",python,selection_command
|
| 17 |
+
17,20996,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",577,0,"",python,selection_command
|
| 18 |
+
18,21030,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",610,0,"",python,selection_command
|
| 19 |
+
19,21068,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",611,0,"",python,selection_command
|
| 20 |
+
20,21097,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",639,0,"",python,selection_command
|
| 21 |
+
21,21131,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",640,0,"",python,selection_command
|
| 22 |
+
22,21166,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",671,0,"",python,selection_command
|
| 23 |
+
23,21203,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",694,0,"",python,selection_command
|
| 24 |
+
24,21233,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",707,0,"",python,selection_command
|
| 25 |
+
25,21265,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",795,0,"",python,selection_command
|
| 26 |
+
26,21297,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",796,0,"",python,selection_command
|
| 27 |
+
27,21331,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",852,0,"",python,selection_command
|
| 28 |
+
28,21364,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",914,0,"",python,selection_command
|
| 29 |
+
29,21397,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",965,0,"",python,selection_command
|
| 30 |
+
30,21431,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1047,0,"",python,selection_command
|
| 31 |
+
31,21464,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1148,0,"",python,selection_command
|
| 32 |
+
32,21497,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1245,0,"",python,selection_command
|
| 33 |
+
33,21530,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1319,0,"",python,selection_command
|
| 34 |
+
34,21565,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1369,0,"",python,selection_command
|
| 35 |
+
35,21598,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1370,0,"",python,selection_command
|
| 36 |
+
36,21631,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1393,0,"",python,selection_command
|
| 37 |
+
37,21668,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1394,0,"",python,selection_command
|
| 38 |
+
38,21701,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1395,0,"",python,selection_command
|
| 39 |
+
39,21890,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1423,0,"",python,selection_command
|
| 40 |
+
40,23050,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1427,0,"",python,selection_command
|
| 41 |
+
41,23197,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1432,0,"",python,selection_command
|
| 42 |
+
42,23489,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1435,0,"",python,selection_command
|
| 43 |
+
43,23702,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1438,0,"",python,selection_command
|
| 44 |
+
44,23912,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1441,0,"",python,selection_command
|
| 45 |
+
45,24119,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1442,0,"",python,selection_command
|
| 46 |
+
46,24327,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1448,0,"",python,selection_command
|
| 47 |
+
47,24581,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1449,0,"",python,selection_command
|
| 48 |
+
48,24945,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1451,0,"",python,selection_command
|
| 49 |
+
49,25331,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1452,0,"",python,selection_command
|
| 50 |
+
50,26280,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1423,0,"",python,selection_command
|
| 51 |
+
51,26667,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1395,0,"",python,selection_command
|
| 52 |
+
52,26920,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1394,0,"",python,selection_command
|
| 53 |
+
53,26954,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1393,0,"",python,selection_command
|
| 54 |
+
54,146021,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
|
| 55 |
+
55,146654,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,tab
|
| 56 |
+
56,203332,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2838,0,"",python,selection_keyboard
|
| 57 |
+
57,203923,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",5124,0,"",python,selection_keyboard
|
| 58 |
+
58,204398,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7126,0,"",python,selection_keyboard
|
| 59 |
+
59,204829,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9057,0,"",python,selection_keyboard
|
| 60 |
+
60,205408,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_keyboard
|
| 61 |
+
61,210105,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7721,0,"",python,selection_keyboard
|
| 62 |
+
62,210235,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",5922,0,"",python,selection_keyboard
|
| 63 |
+
63,210382,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3490,0,"",python,selection_keyboard
|
| 64 |
+
64,210771,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1906,0,"",python,selection_keyboard
|
| 65 |
+
65,210944,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",64,0,"",python,selection_keyboard
|
| 66 |
+
66,211099,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_keyboard
|
| 67 |
+
67,214839,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10,0,"",python,selection_command
|
| 68 |
+
68,2267039,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2294,0,"",python,selection_mouse
|
| 69 |
+
69,2272461,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2114,0,"",python,selection_mouse
|
| 70 |
+
70,3052561,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,10041,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,selection_command
|
| 71 |
+
71,3052668,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_command
|
| 72 |
+
72,3145153,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_command
|
| 73 |
+
73,3148126,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7487,0,"",python,selection_command
|
| 74 |
+
74,3149770,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7876,0,"",python,selection_command
|
| 75 |
+
75,3150226,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",8235,0,"",python,selection_command
|
| 76 |
+
76,3150898,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",7876,0,"",python,selection_command
|
| 77 |
+
77,3151966,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",8235,0,"",python,selection_command
|
| 78 |
+
78,3161866,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9070,0,"",python,selection_command
|
| 79 |
+
79,3164224,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9131,0,"",python,selection_command
|
| 80 |
+
80,3166810,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9194,0,"",python,selection_command
|
| 81 |
+
81,3167056,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9257,0,"",python,selection_command
|
| 82 |
+
82,3167090,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9321,0,"",python,selection_command
|
| 83 |
+
83,3167123,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9386,0,"",python,selection_command
|
| 84 |
+
84,3167156,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9443,0,"",python,selection_command
|
| 85 |
+
85,3167189,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9481,0,"",python,selection_command
|
| 86 |
+
86,3167223,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9531,0,"",python,selection_command
|
| 87 |
+
87,3168286,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9532,0,"",python,selection_command
|
| 88 |
+
88,3168527,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9538,0,"",python,selection_command
|
| 89 |
+
89,3168566,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9539,0,"",python,selection_command
|
| 90 |
+
90,3168597,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9546,0,"",python,selection_command
|
| 91 |
+
91,3168633,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9547,0,"",python,selection_command
|
| 92 |
+
92,3168887,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9555,0,"",python,selection_command
|
| 93 |
+
93,3169118,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
|
| 94 |
+
94,3169292,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9583,0,"",python,selection_command
|
| 95 |
+
95,3169900,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
|
| 96 |
+
96,3170394,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
|
| 97 |
+
97,3171165,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
|
| 98 |
+
98,3171795,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
|
| 99 |
+
99,3171945,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
|
| 100 |
+
100,3172433,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
|
| 101 |
+
101,3173957,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
|
| 102 |
+
102,3174398,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
|
| 103 |
+
103,3174801,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9682,0,"",python,selection_command
|
| 104 |
+
104,3175123,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9746,0,"",python,selection_command
|
| 105 |
+
105,3175686,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9682,0,"",python,selection_command
|
| 106 |
+
106,3175821,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9619,0,"",python,selection_command
|
| 107 |
+
107,3175981,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9556,0,"",python,selection_command
|
| 108 |
+
108,3240534,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_command
|
| 109 |
+
109,3241210,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1898,0,"",python,selection_keyboard
|
| 110 |
+
110,3241373,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3420,0,"",python,selection_keyboard
|
| 111 |
+
111,3242843,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3370,0,"",python,selection_command
|
| 112 |
+
112,3243091,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3298,0,"",python,selection_command
|
| 113 |
+
113,3243124,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3288,0,"",python,selection_command
|
| 114 |
+
114,3243158,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3287,0,"",python,selection_command
|
| 115 |
+
115,3243191,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3211,0,"",python,selection_command
|
| 116 |
+
116,3243225,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3137,0,"",python,selection_command
|
| 117 |
+
117,3243258,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3058,0,"",python,selection_command
|
| 118 |
+
118,3243291,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3057,0,"",python,selection_command
|
| 119 |
+
119,3243324,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2999,0,"",python,selection_command
|
| 120 |
+
120,3243357,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2991,0,"",python,selection_command
|
| 121 |
+
121,3243391,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2973,0,"",python,selection_command
|
| 122 |
+
122,3243424,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2956,0,"",python,selection_command
|
| 123 |
+
123,3243458,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2932,0,"",python,selection_command
|
| 124 |
+
124,3243491,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2915,0,"",python,selection_command
|
| 125 |
+
125,3243525,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2898,0,"",python,selection_command
|
| 126 |
+
126,3243558,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2876,0,"",python,selection_command
|
| 127 |
+
127,3243591,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2854,0,"",python,selection_command
|
| 128 |
+
128,3243632,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2838,0,"",python,selection_command
|
| 129 |
+
129,3243658,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2817,0,"",python,selection_command
|
| 130 |
+
130,3243691,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2793,0,"",python,selection_command
|
| 131 |
+
131,3243734,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2767,0,"",python,selection_command
|
| 132 |
+
132,3243758,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2756,0,"",python,selection_command
|
| 133 |
+
133,3243791,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2721,0,"",python,selection_command
|
| 134 |
+
134,3243824,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2688,0,"",python,selection_command
|
| 135 |
+
135,3243857,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2655,0,"",python,selection_command
|
| 136 |
+
136,3243891,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2621,0,"",python,selection_command
|
| 137 |
+
137,3243925,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2585,0,"",python,selection_command
|
| 138 |
+
138,3243958,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2547,0,"",python,selection_command
|
| 139 |
+
139,3243991,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2512,0,"",python,selection_command
|
| 140 |
+
140,3244025,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2483,0,"",python,selection_command
|
| 141 |
+
141,3244059,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2451,0,"",python,selection_command
|
| 142 |
+
142,3244092,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2427,0,"",python,selection_command
|
| 143 |
+
143,3244125,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2388,0,"",python,selection_command
|
| 144 |
+
144,3286103,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2411,0,"",python,selection_command
|
| 145 |
+
145,3296751,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2388,23,"",python,content
|
| 146 |
+
146,3298859,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2388,0," sequence_parallelis",python,content
|
| 147 |
+
147,3298862,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2388,0,"",python,selection_command
|
| 148 |
+
148,3483131,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
|
| 149 |
+
149,3484741,"TERMINAL",0,0,"",,terminal_focus
|
| 150 |
+
150,3484752,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,tab
|
| 151 |
+
151,3788287,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2427,0,"",python,selection_command
|
| 152 |
+
152,3788457,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2428,0,"",python,selection_command
|
| 153 |
+
153,3788946,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2429,0,"",python,selection_command
|
| 154 |
+
154,3789583,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2427,0,"",python,selection_command
|
| 155 |
+
155,4549882,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,10041,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,selection_command
|
| 156 |
+
156,4550471,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_command
|
| 157 |
+
157,4559897,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,10041,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,selection_command
|
| 158 |
+
158,4560021,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_command
|
| 159 |
+
159,4560890,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,content
|
| 160 |
+
160,4583906,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18686,0,"",python,selection_mouse
|
| 161 |
+
161,4583914,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18685,0,"",python,selection_command
|
| 162 |
+
162,4584529,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18699,0,"",python,selection_command
|
| 163 |
+
163,4585394,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18743,0,"",python,selection_command
|
| 164 |
+
164,4585610,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",18752,0,"",python,selection_command
|
| 165 |
+
165,4591082,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,tab
|
| 166 |
+
166,4591147,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",64,0,"",python,selection_command
|
| 167 |
+
167,4595172,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,10041,"",python,content
|
| 168 |
+
168,4597038,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,tab
|
| 169 |
+
169,5454427,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",9372,0,"",python,selection_mouse
|
| 170 |
+
170,5454796,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,0,"",python,selection_command
|
| 171 |
+
171,5455466,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",64,0,"",python,selection_command
|
| 172 |
+
172,5455793,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",1906,0,"",python,selection_keyboard
|
| 173 |
+
173,5456269,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3490,0,"",python,selection_keyboard
|
| 174 |
+
174,5456884,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3420,0,"",python,selection_command
|
| 175 |
+
175,5457139,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3370,0,"",python,selection_command
|
| 176 |
+
176,5457172,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3298,0,"",python,selection_command
|
| 177 |
+
177,5457202,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3288,0,"",python,selection_command
|
| 178 |
+
178,5457232,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3287,0,"",python,selection_command
|
| 179 |
+
179,5457265,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3211,0,"",python,selection_command
|
| 180 |
+
180,5457298,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3137,0,"",python,selection_command
|
| 181 |
+
181,5457331,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3058,0,"",python,selection_command
|
| 182 |
+
182,5457367,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",3057,0,"",python,selection_command
|
| 183 |
+
183,5457402,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2999,0,"",python,selection_command
|
| 184 |
+
184,5457431,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2991,0,"",python,selection_command
|
| 185 |
+
185,5457465,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2973,0,"",python,selection_command
|
| 186 |
+
186,5457497,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2956,0,"",python,selection_command
|
| 187 |
+
187,5457533,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2932,0,"",python,selection_command
|
| 188 |
+
188,5457565,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2915,0,"",python,selection_command
|
| 189 |
+
189,5457598,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2898,0,"",python,selection_command
|
| 190 |
+
190,5457631,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2876,0,"",python,selection_command
|
| 191 |
+
191,5457665,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2854,0,"",python,selection_command
|
| 192 |
+
192,5457699,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2838,0,"",python,selection_command
|
| 193 |
+
193,5457732,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2817,0,"",python,selection_command
|
| 194 |
+
194,5457764,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2793,0,"",python,selection_command
|
| 195 |
+
195,5457799,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2767,0,"",python,selection_command
|
| 196 |
+
196,5457833,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2756,0,"",python,selection_command
|
| 197 |
+
197,5457866,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2721,0,"",python,selection_command
|
| 198 |
+
198,5457899,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2688,0,"",python,selection_command
|
| 199 |
+
199,5457931,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2655,0,"",python,selection_command
|
| 200 |
+
200,5457964,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2621,0,"",python,selection_command
|
| 201 |
+
201,5457999,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2585,0,"",python,selection_command
|
| 202 |
+
202,5458033,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2547,0,"",python,selection_command
|
| 203 |
+
203,5458067,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2512,0,"",python,selection_command
|
| 204 |
+
204,5458102,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2483,0,"",python,selection_command
|
| 205 |
+
205,5458137,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2451,0,"",python,selection_command
|
| 206 |
+
206,5459402,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2483,0,"",python,selection_command
|
| 207 |
+
207,5459593,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2512,0,"",python,selection_command
|
| 208 |
+
208,5459748,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2547,0,"",python,selection_command
|
| 209 |
+
209,5460152,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2551,0,"",python,selection_command
|
| 210 |
+
210,5461248,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2589,0,"",python,selection_command
|
| 211 |
+
211,5461431,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2625,0,"",python,selection_command
|
| 212 |
+
212,5461588,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2659,0,"",python,selection_command
|
| 213 |
+
213,5461782,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2692,0,"",python,selection_command
|
| 214 |
+
214,5462187,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2725,0,"",python,selection_command
|
| 215 |
+
215,5462694,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2692,0,"",python,selection_command
|
| 216 |
+
216,5462868,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2659,0,"",python,selection_command
|
| 217 |
+
217,5463112,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2625,0,"",python,selection_command
|
| 218 |
+
218,5463146,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2589,0,"",python,selection_command
|
| 219 |
+
219,5463184,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2551,0,"",python,selection_command
|
| 220 |
+
220,5463214,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2516,0,"",python,selection_command
|
| 221 |
+
221,5463329,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2487,0,"",python,selection_command
|
| 222 |
+
222,8520791,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",0,10041,"# Copyright (c) 2025, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Optional\n\nimport lightning.pytorch as pl\nimport nemo_run as run\nimport torch\nfrom nemo.collections.common.tokenizers.huggingface.auto_tokenizer import AutoTokenizer\n\nfrom nemo.collections.llm.api import finetune, pretrain\nfrom nemo.collections.llm.gpt.data.mock import MockDataModule\nfrom nemo.collections.llm.peft import PEFT_STR2CLS\nfrom nemo.collections.llm.recipes.finetune_default import default_finetune_recipe\nfrom nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger\nfrom nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing\nfrom nemo.collections.llm.recipes.qwen3 import qwen3_model, qwen3_trainer\nfrom nemo.utils.exp_manager import TimingCallback\n\nNAME = ""qwen3_30b_a3b""\n\n\n@run.cli.factory(name=NAME)\ndef model() -> run.Config[pl.LightningModule]:\n """"""\n Factory function to create a Qwen3 30B-A3B model configuration.\n This is a MoE (Mixture of Experts) model with 128 experts.\n\n Returns:\n run.Config[pl.LightningModule]: Configuration for the Qwen3 30B-A3B model.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain model=qwen3_30b_a3b ...\n\n Python API usage:\n >>> model_config = model()\n >>> print(model_config)\n """"""\n return qwen3_model(version=NAME)\n\n\n@run.cli.factory(target=pretrain, name=NAME)\ndef pretrain_recipe(\n # General\n dir: Optional[str] = None,\n name: str = ""default"",\n # Trainer\n tensor_parallelism: int = 4, # Default for 30B-A3B model\n pipeline_parallelism: int = 2,\n pipeline_parallelism_type: Optional[torch.dtype] = None,\n virtual_pipeline_parallelism: Optional[int] = None,\n context_parallelism: int = 1,\n expert_parallelism: Optional[int] = 4,\n sequence_parallelism: bool = True,\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n max_steps: int = 300000,\n precision: str = ""bf16-mixed"",\n accumulate_grad_batches: int = 1,\n gradient_clip_val: float = 1.0,\n limit_test_batches: int = 32,\n limit_val_batches: int = 32,\n log_every_n_steps: int = 10,\n val_check_interval: int = 500,\n # Data\n global_batch_size=32,\n micro_batch_size=2,\n seq_length=4096,\n # Optimizer\n warmup_steps=500,\n constant_steps=0,\n min_lr=3e-5,\n max_lr=3e-4,\n # Training function\n fn=pretrain,\n) -> run.Partial:\n """"""\n Create a pre-training recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for pre-training, including\n model, trainer, data, logging, optimization, and resumption settings.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the pre-training run.\n tensor_parallelism (int): Degree of tensor model parallelism.\n pipeline_parallelism (int): Degree of pipeline model parallelism.\n pipeline_parallelism_type (Optional[torch.dtype]): Data type for pipeline parallelism.\n virtual_pipeline_parallelism (Optional[int]): Size of virtual pipeline parallelism.\n context_parallelism (int): Degree of context parallelism.\n sequence_parallelism (bool): Whether to use sequence parallelism.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n max_steps (int): Maximum number of training steps.\n precision (str): Precision configuration, one of fp32, 16-mixed or bf16-mixed.\n accumulate_grad_batches (int): Number of steps per gradient accumulation.\n gradient_clip_val (float): Value for gradient clipping.\n limit_test_batches (int): Limit the number of test batches.\n limit_val_batches (int): Limit the number of validation batches.\n log_every_n_steps (int): Log every n steps.\n val_check_interval (int): Run validation every N steps.\n global_batch_size (int): Global batch size.\n micro_batch_size (int): Micro batch size.\n seq_length (int): Sequence length.\n warmup_steps (int): Number of warmup steps.\n constant_steps (int): Number of constant steps.\n min_lr (float): Minimum learning rate.\n max_lr (float): Maximum learning rate.\n fn (Callable): The pre-training function to use.\n\n Returns:\n run.Partial: Partial configuration for pre-training.\n\n Examples:\n CLI usage:\n $ nemo llm pretrain --factory qwen3_30b_a3b\n $ nemo llm pretrain --factory ""qwen3_30b_a3b(num_nodes=1, name='my_qwen3_pretrain')""\n\n Python API usage:\n >>> recipe = pretrain_recipe(name=""qwen3_pretrain"", num_nodes=1)\n >>> print(recipe)\n\n Note:\n This recipe uses a mock dataset, look for the finetune examples to see how to change the dataset.\n """"""\n recipe = run.Partial(\n fn,\n model=model(),\n trainer=qwen3_trainer(\n tensor_parallelism=tensor_parallelism,\n pipeline_parallelism=pipeline_parallelism,\n pipeline_parallelism_type=pipeline_parallelism_type,\n virtual_pipeline_parallelism=virtual_pipeline_parallelism,\n context_parallelism=context_parallelism,\n sequence_parallelism=sequence_parallelism,\n expert_parallelism=expert_parallelism,\n num_nodes=num_nodes,\n num_gpus_per_node=num_gpus_per_node,\n max_steps=max_steps,\n precision=precision,\n accumulate_grad_batches=accumulate_grad_batches,\n limit_test_batches=limit_test_batches,\n limit_val_batches=limit_val_batches,\n log_every_n_steps=log_every_n_steps,\n val_check_interval=val_check_interval,\n callbacks=[run.Config(TimingCallback)],\n ),\n data=run.Config(\n MockDataModule,\n seq_length=seq_length,\n global_batch_size=global_batch_size,\n micro_batch_size=micro_batch_size,\n tokenizer=run.Config(AutoTokenizer, ""Qwen/Qwen3-30B-A3B""),\n ),\n log=default_log(dir=dir, name=name, tensorboard_logger=tensorboard_logger(name=name)),\n optim=distributed_fused_adam_with_cosine_annealing(\n precision=precision,\n warmup_steps=warmup_steps,\n constant_steps=constant_steps,\n min_lr=min_lr,\n max_lr=max_lr,\n clip_grad=gradient_clip_val,\n ),\n resume=default_resume(),\n )\n recipe.model.config.recompute_granularity = ""full""\n recipe.model.config.recompute_method = ""uniform""\n recipe.model.config.recompute_num_layers = 1\n return recipe\n\n\n@run.cli.factory(target=finetune, name=NAME)\ndef finetune_recipe(\n dir: Optional[str] = None,\n name: str = ""default"",\n num_nodes: int = 1,\n num_gpus_per_node: int = 8,\n peft_scheme: Optional[str] = 'lora',\n packed_sequence: bool = False,\n) -> run.Partial:\n """"""\n Create a fine-tuning recipe for Qwen3 30B-A3B model.\n\n This function sets up a complete configuration for fine-tuning, including\n model, trainer, data, logging, optimization, and resumption settings.\n The recipe uses LoRA (Low-Rank Adaptation) for efficient fine-tuning, unless peft_scheme is set to None.\n This model uses Mixture of Experts (MoE) architecture with 128 experts.\n\n Args:\n dir (Optional[str]): Directory for saving logs and checkpoints.\n name (str): Name of the fine-tuning run.\n num_nodes (int): Number of compute nodes to use.\n num_gpus_per_node (int): Number of GPUs per node.\n peft_scheme (Optional[str]): Name of the peft scheme to use for fine-tuning.\n Allowed values: 'lora'/'dora'/'none'/None.\n packed_sequence (Optional[bool]): Packing multiple training sequences into one long sequence for training\n efficiency. Default sequence length is 2048.\n\n Returns:\n run.Partial: Partial configuration for fine-tuning.\n\n Examples:\n CLI usage:\n $ nemo llm finetune --factory qwen3_30b_a3b\n\n Python API usage:\n >>> recipe = finetune_recipe(name=""qwen3_30b_a3b_finetune"", num_nodes=2)\n >>> print(recipe)\n\n Note:\n This recipe uses the SQuAD dataset for fine-tuning.\n """"""\n recipe = default_finetune_recipe(\n model(), ""Qwen/Qwen3-30B-A3B"", dir, name, num_nodes, num_gpus_per_node, packed_sequence\n )\n if peft_scheme is None or peft_scheme.lower() == 'none':\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.pipeline_model_parallel_size = 2\n recipe.trainer.strategy.sequence_parallel = True\n recipe.optim.config.lr = 5e-6\n elif peft_scheme.lower() in ['lora', 'dora']:\n recipe.trainer.strategy.tensor_model_parallel_size = 4\n recipe.trainer.strategy.expert_model_parallel_size = 4\n recipe.trainer.strategy.expert_tensor_parallel_size = 1\n recipe.trainer.strategy.sequence_parallel = True\n recipe.peft = run.Config(PEFT_STR2CLS[peft_scheme.lower()])\n recipe.peft.target_modules = ['linear_qkv', 'linear_proj']\n recipe.optim.config.lr = 1e-4\n else:\n raise ValueError(f""Unrecognized peft scheme: {peft_scheme}"")\n return recipe\n",python,selection_command
|
| 223 |
+
223,8520860,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",10041,0,"",python,selection_command
|
| 224 |
+
224,8685736,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2797,0,"",python,selection_command
|
| 225 |
+
225,8686757,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2813,0,"",python,selection_command
|
| 226 |
+
226,8686946,"nemo/collections/llm/recipes/qwen3_30b_a3b.py",2814,0,"",python,selection_command
|
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-a7b808c2-b1d0-43a0-a38c-8b82cd2886711764488770794-2025_11_30-08.46.18.897/source.csv
ADDED
|
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,13,"Untitled-1",0,0,"",plaintext,tab
|
| 3 |
+
2,77,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"8:46:18 AM [info] Activating crowd-code\n8:46:18 AM [info] Recording started\n8:46:18 AM [info] Initializing git provider using file system watchers...\n8:46:18 AM [info] No workspace folder found\n",Log,tab
|
| 4 |
+
3,2028,"extension-output-pdoom-org.crowd-code-#1-crowd-code",194,0,"8:46:20 AM [info] Retrying git provider initialization...\n8:46:20 AM [info] No workspace folder found\n",Log,content
|
| 5 |
+
4,2227,"Untitled-1",0,0,"",plaintext,tab
|
| 6 |
+
5,3588,"TERMINAL",0,0,"Test",,terminal_focus
|
| 7 |
+
6,3593,"Untitled-1",0,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
|
| 8 |
+
7,4572,"Untitled-1",76,0,"/* crowd-pilot: replacement */\nREPLACED LINE 1\nREPLACED LINE 2",plaintext,content
|
| 9 |
+
8,6374,"Untitled-1",122,0,"",plaintext,selection_command
|
| 10 |
+
9,6522,"Untitled-1",91,0,"",plaintext,selection_command
|
| 11 |
+
10,6765,"Untitled-1",122,0,"",plaintext,selection_command
|
| 12 |
+
11,6932,"Untitled-1",138,0,"",plaintext,selection_command
|
| 13 |
+
12,7497,"Untitled-1",123,15,"",plaintext,content
|
| 14 |
+
13,8248,"TERMINAL",0,0,"echo ""Hello World""",,terminal_command
|
| 15 |
+
14,8249,"TERMINAL",0,0,"]633;CHello World\r\n[1m[7m%[27m[1m[0m \r \r",,terminal_output
|
| 16 |
+
15,9755,"Untitled-1",107,0,"",plaintext,selection_command
|
| 17 |
+
16,9857,"Untitled-1",76,0,"",plaintext,selection_command
|
| 18 |
+
17,10024,"Untitled-1",46,0,"",plaintext,selection_command
|
| 19 |
+
18,10659,"Untitled-1",39,0,"",plaintext,selection_command
|
| 20 |
+
19,10826,"Untitled-1",32,0,"",plaintext,selection_command
|
| 21 |
+
20,10955,"Untitled-1",0,0,"",plaintext,selection_command
|
| 22 |
+
21,14633,"Untitled-1",0,31,"/* crowd-pilot: insert start */",plaintext,selection_command
|
| 23 |
+
22,15287,"Untitled-1",0,38,"/* crowd-pilot: insert start */\nline A",plaintext,selection_command
|
| 24 |
+
23,15426,"Untitled-1",0,45,"/* crowd-pilot: insert start */\nline A\nline B",plaintext,selection_command
|
| 25 |
+
24,15551,"Untitled-1",0,75,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */",plaintext,selection_command
|
| 26 |
+
25,15693,"Untitled-1",0,106,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n/* crowd-pilot: replacement */",plaintext,selection_command
|
| 27 |
+
26,15820,"Untitled-1",0,122,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n/* crowd-pilot: replacement */\nREPLACED LINE 1",plaintext,selection_command
|
| 28 |
+
27,15909,"Untitled-1",0,122,"",plaintext,content
|
| 29 |
+
28,16827,"Untitled-1",1,0,"/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,content
|
| 30 |
+
29,19025,"Untitled-1",0,1,"\n",plaintext,selection_command
|
| 31 |
+
30,19300,"Untitled-1",0,77,"\n/* crowd-pilot: insert start */\nline A\nline B\n/* crowd-pilot: insert end */\n",plaintext,selection_command
|
| 32 |
+
31,19520,"Untitled-1",0,77,"",plaintext,content
|
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-aeed47b9-f6ef-4272-b0ca-0c15ab4c25021758266694991-2025_09_19-09.25.04.660/source.csv
ADDED
|
@@ -0,0 +1,151 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,3,"cleanrl/ppo_atari_envpool.py",0,0,"# docs and experiment results can be found at https://docs.cleanrl.dev/rl-algorithms/ppo/#ppo_atari_envpoolpy\nimport os\nimport random\nimport time\nfrom collections import deque\nfrom dataclasses import dataclass\n\nimport envpool\nimport gym\nimport numpy as np\nimport torch\nimport torch.nn as nn\nimport torch.optim as optim\nimport tyro\nfrom torch.distributions.categorical import Categorical\nfrom torch.utils.tensorboard import SummaryWriter\n\n\n@dataclass\nclass Args:\n exp_name: str = os.path.basename(__file__)[: -len("".py"")]\n """"""the name of this experiment""""""\n seed: int = 1\n """"""seed of the experiment""""""\n torch_deterministic: bool = True\n """"""if toggled, `torch.backends.cudnn.deterministic=False`""""""\n cuda: bool = True\n """"""if toggled, cuda will be enabled by default""""""\n track: bool = False\n """"""if toggled, this experiment will be tracked with Weights and Biases""""""\n wandb_project_name: str = ""cleanRL""\n """"""the wandb's project name""""""\n wandb_entity: str = None\n """"""the entity (team) of wandb's project""""""\n capture_video: bool = False\n """"""whether to capture videos of the agent performances (check out `videos` folder)""""""\n\n # Algorithm specific arguments\n env_id: str = ""Breakout-v5""\n """"""the id of the environment""""""\n total_timesteps: int = 10000000\n """"""total timesteps of the experiments""""""\n learning_rate: float = 2.5e-4\n """"""the learning rate of the optimizer""""""\n num_envs: int = 8\n """"""the number of parallel game environments""""""\n num_steps: int = 128\n """"""the number of steps to run in each environment per policy rollout""""""\n anneal_lr: bool = True\n """"""Toggle learning rate annealing for policy and value networks""""""\n gamma: float = 0.99\n """"""the discount factor gamma""""""\n gae_lambda: float = 0.95\n """"""the lambda for the general advantage estimation""""""\n num_minibatches: int = 4\n """"""the number of mini-batches""""""\n update_epochs: int = 4\n """"""the K epochs to update the policy""""""\n norm_adv: bool = True\n """"""Toggles advantages normalization""""""\n clip_coef: float = 0.1\n """"""the surrogate clipping coefficient""""""\n clip_vloss: bool = True\n """"""Toggles whether or not to use a clipped loss for the value function, as per the paper.""""""\n ent_coef: float = 0.01\n """"""coefficient of the entropy""""""\n vf_coef: float = 0.5\n """"""coefficient of the value function""""""\n max_grad_norm: float = 0.5\n """"""the maximum norm for the gradient clipping""""""\n target_kl: float = None\n """"""the target KL divergence threshold""""""\n\n # to be filled in runtime\n batch_size: int = 0\n """"""the batch size (computed in runtime)""""""\n minibatch_size: int = 0\n """"""the mini-batch size (computed in runtime)""""""\n num_iterations: int = 0\n """"""the number of iterations (computed in runtime)""""""\n\n\nclass RecordEpisodeStatistics(gym.Wrapper):\n def __init__(self, env, deque_size=100):\n super().__init__(env)\n self.num_envs = getattr(env, ""num_envs"", 1)\n self.episode_returns = None\n self.episode_lengths = None\n\n def reset(self, **kwargs):\n observations = super().reset(**kwargs)\n self.episode_returns = np.zeros(self.num_envs, dtype=np.float32)\n self.episode_lengths = np.zeros(self.num_envs, dtype=np.int32)\n self.lives = np.zeros(self.num_envs, dtype=np.int32)\n self.returned_episode_returns = np.zeros(self.num_envs, dtype=np.float32)\n self.returned_episode_lengths = np.zeros(self.num_envs, dtype=np.int32)\n return observations\n\n def step(self, action):\n observations, rewards, dones, infos = super().step(action)\n self.episode_returns += infos[""reward""]\n self.episode_lengths += 1\n self.returned_episode_returns[:] = self.episode_returns\n self.returned_episode_lengths[:] = self.episode_lengths\n self.episode_returns *= 1 - infos[""terminated""]\n self.episode_lengths *= 1 - infos[""terminated""]\n infos[""r""] = self.returned_episode_returns\n infos[""l""] = self.returned_episode_lengths\n return (\n observations,\n rewards,\n dones,\n infos,\n )\n\n\ndef layer_init(layer, std=np.sqrt(2), bias_const=0.0):\n torch.nn.init.orthogonal_(layer.weight, std)\n torch.nn.init.constant_(layer.bias, bias_const)\n return layer\n\n\nclass Agent(nn.Module):\n def __init__(self, envs):\n super().__init__()\n self.network = nn.Sequential(\n layer_init(nn.Conv2d(4, 32, 8, stride=4)),\n nn.ReLU(),\n layer_init(nn.Conv2d(32, 64, 4, stride=2)),\n nn.ReLU(),\n layer_init(nn.Conv2d(64, 64, 3, stride=1)),\n nn.ReLU(),\n nn.Flatten(),\n layer_init(nn.Linear(64 * 7 * 7, 512)),\n nn.ReLU(),\n )\n self.actor = layer_init(nn.Linear(512, envs.single_action_space.n), std=0.01)\n self.critic = layer_init(nn.Linear(512, 1), std=1)\n\n def get_value(self, x):\n return self.critic(self.network(x / 255.0))\n\n def get_action_and_value(self, x, action=None):\n hidden = self.network(x / 255.0)\n logits = self.actor(hidden)\n probs = Categorical(logits=logits)\n if action is None:\n action = probs.sample()\n return action, probs.log_prob(action), probs.entropy(), self.critic(hidden)\n\n\nif __name__ == ""__main__"":\n args = tyro.cli(Args)\n args.batch_size = int(args.num_envs * args.num_steps)\n args.minibatch_size = int(args.batch_size // args.num_minibatches)\n args.num_iterations = args.total_timesteps // args.batch_size\n run_name = f""{args.env_id}__{args.exp_name}__{args.seed}__{int(time.time())}""\n if args.track:\n import wandb\n\n wandb.init(\n project=args.wandb_project_name,\n entity=args.wandb_entity,\n sync_tensorboard=True,\n config=vars(args),\n name=run_name,\n monitor_gym=True,\n save_code=True,\n )\n writer = SummaryWriter(f""runs/{run_name}"")\n writer.add_text(\n ""hyperparameters"",\n ""|param|value|\n|-|-|\n%s"" % (""\n"".join([f""|{key}|{value}|"" for key, value in vars(args).items()])),\n )\n\n # TRY NOT TO MODIFY: seeding\n random.seed(args.seed)\n np.random.seed(args.seed)\n torch.manual_seed(args.seed)\n torch.backends.cudnn.deterministic = args.torch_deterministic\n\n device = torch.device(""cuda"" if torch.cuda.is_available() and args.cuda else ""cpu"")\n\n # env setup\n envs = envpool.make(\n args.env_id,\n env_type=""gym"",\n num_envs=args.num_envs,\n episodic_life=True,\n reward_clip=True,\n seed=args.seed,\n )\n envs.num_envs = args.num_envs\n envs.single_action_space = envs.action_space\n envs.single_observation_space = envs.observation_space\n envs = RecordEpisodeStatistics(envs)\n assert isinstance(envs.action_space, gym.spaces.Discrete), ""only discrete action space is supported""\n\n agent = Agent(envs).to(device)\n optimizer = optim.Adam(agent.parameters(), lr=args.learning_rate, eps=1e-5)\n\n # ALGO Logic: Storage setup\n obs = torch.zeros((args.num_steps, args.num_envs) + envs.single_observation_space.shape).to(device)\n actions = torch.zeros((args.num_steps, args.num_envs) + envs.single_action_space.shape).to(device)\n logprobs = torch.zeros((args.num_steps, args.num_envs)).to(device)\n rewards = torch.zeros((args.num_steps, args.num_envs)).to(device)\n dones = torch.zeros((args.num_steps, args.num_envs)).to(device)\n values = torch.zeros((args.num_steps, args.num_envs)).to(device)\n avg_returns = deque(maxlen=20)\n\n # TRY NOT TO MODIFY: start the game\n global_step = 0\n start_time = time.time()\n next_obs = torch.Tensor(envs.reset()).to(device)\n next_done = torch.zeros(args.num_envs).to(device)\n\n for iteration in range(1, args.num_iterations + 1):\n # Annealing the rate if instructed to do so.\n if args.anneal_lr:\n frac = 1.0 - (iteration - 1.0) / args.num_iterations\n lrnow = frac * args.learning_rate\n optimizer.param_groups[0][""lr""] = lrnow\n\n for step in range(0, args.num_steps):\n global_step += args.num_envs\n obs[step] = next_obs\n dones[step] = next_done\n\n # ALGO LOGIC: action logic\n with torch.no_grad():\n action, logprob, _, value = agent.get_action_and_value(next_obs)\n values[step] = value.flatten()\n actions[step] = action\n logprobs[step] = logprob\n\n # TRY NOT TO MODIFY: execute the game and log data.\n next_obs, reward, next_done, info = envs.step(action.cpu().numpy())\n rewards[step] = torch.tensor(reward).to(device).view(-1)\n next_obs, next_done = torch.Tensor(next_obs).to(device), torch.Tensor(next_done).to(device)\n\n for idx, d in enumerate(next_done):\n if d and info[""lives""][idx] == 0:\n print(f""global_step={global_step}, episodic_return={info['r'][idx]}"")\n avg_returns.append(info[""r""][idx])\n writer.add_scalar(""charts/avg_episodic_return"", np.average(avg_returns), global_step)\n writer.add_scalar(""charts/episodic_return"", info[""r""][idx], global_step)\n writer.add_scalar(""charts/episodic_length"", info[""l""][idx], global_step)\n\n # bootstrap value if not done\n with torch.no_grad():\n next_value = agent.get_value(next_obs).reshape(1, -1)\n advantages = torch.zeros_like(rewards).to(device)\n lastgaelam = 0\n for t in reversed(range(args.num_steps)):\n if t == args.num_steps - 1:\n nextnonterminal = 1.0 - next_done\n nextvalues = next_value\n else:\n nextnonterminal = 1.0 - dones[t + 1]\n nextvalues = values[t + 1]\n delta = rewards[t] + args.gamma * nextvalues * nextnonterminal - values[t]\n advantages[t] = lastgaelam = delta + args.gamma * args.gae_lambda * nextnonterminal * lastgaelam\n returns = advantages + values\n\n # flatten the batch\n b_obs = obs.reshape((-1,) + envs.single_observation_space.shape)\n b_logprobs = logprobs.reshape(-1)\n b_actions = actions.reshape((-1,) + envs.single_action_space.shape)\n b_advantages = advantages.reshape(-1)\n b_returns = returns.reshape(-1)\n b_values = values.reshape(-1)\n\n # Optimizing the policy and value network\n b_inds = np.arange(args.batch_size)\n clipfracs = []\n for epoch in range(args.update_epochs):\n np.random.shuffle(b_inds)\n for start in range(0, args.batch_size, args.minibatch_size):\n end = start + args.minibatch_size\n mb_inds = b_inds[start:end]\n\n _, newlogprob, entropy, newvalue = agent.get_action_and_value(b_obs[mb_inds], b_actions.long()[mb_inds])\n logratio = newlogprob - b_logprobs[mb_inds]\n ratio = logratio.exp()\n\n with torch.no_grad():\n # calculate approx_kl http://joschu.net/blog/kl-approx.html\n old_approx_kl = (-logratio).mean()\n approx_kl = ((ratio - 1) - logratio).mean()\n clipfracs += [((ratio - 1.0).abs() > args.clip_coef).float().mean().item()]\n\n mb_advantages = b_advantages[mb_inds]\n if args.norm_adv:\n mb_advantages = (mb_advantages - mb_advantages.mean()) / (mb_advantages.std() + 1e-8)\n\n # Policy loss\n pg_loss1 = -mb_advantages * ratio\n pg_loss2 = -mb_advantages * torch.clamp(ratio, 1 - args.clip_coef, 1 + args.clip_coef)\n pg_loss = torch.max(pg_loss1, pg_loss2).mean()\n\n # Value loss\n newvalue = newvalue.view(-1)\n if args.clip_vloss:\n v_loss_unclipped = (newvalue - b_returns[mb_inds]) ** 2\n v_clipped = b_values[mb_inds] + torch.clamp(\n newvalue - b_values[mb_inds],\n -args.clip_coef,\n args.clip_coef,\n )\n v_loss_clipped = (v_clipped - b_returns[mb_inds]) ** 2\n v_loss_max = torch.max(v_loss_unclipped, v_loss_clipped)\n v_loss = 0.5 * v_loss_max.mean()\n else:\n v_loss = 0.5 * ((newvalue - b_returns[mb_inds]) ** 2).mean()\n\n entropy_loss = entropy.mean()\n loss = pg_loss - args.ent_coef * entropy_loss + v_loss * args.vf_coef\n\n optimizer.zero_grad()\n loss.backward()\n nn.utils.clip_grad_norm_(agent.parameters(), args.max_grad_norm)\n optimizer.step()\n\n if args.target_kl is not None and approx_kl > args.target_kl:\n break\n\n y_pred, y_true = b_values.cpu().numpy(), b_returns.cpu().numpy()\n var_y = np.var(y_true)\n explained_var = np.nan if var_y == 0 else 1 - np.var(y_true - y_pred) / var_y\n\n # TRY NOT TO MODIFY: record rewards for plotting purposes\n writer.add_scalar(""charts/learning_rate"", optimizer.param_groups[0][""lr""], global_step)\n writer.add_scalar(""losses/value_loss"", v_loss.item(), global_step)\n writer.add_scalar(""losses/policy_loss"", pg_loss.item(), global_step)\n writer.add_scalar(""losses/entropy"", entropy_loss.item(), global_step)\n writer.add_scalar(""losses/old_approx_kl"", old_approx_kl.item(), global_step)\n writer.add_scalar(""losses/approx_kl"", approx_kl.item(), global_step)\n writer.add_scalar(""losses/clipfrac"", np.mean(clipfracs), global_step)\n writer.add_scalar(""losses/explained_variance"", explained_var, global_step)\n print(""SPS:"", int(global_step / (time.time() - start_time)))\n writer.add_scalar(""charts/SPS"", int(global_step / (time.time() - start_time)), global_step)\n\n envs.close()\n writer.close()\n",python,tab
|
| 3 |
+
2,142,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"9:25:04 AM [info] Activating crowd-code\n9:25:04 AM [info] Recording started\n9:25:04 AM [info] Initializing git provider using file system watchers...\n",Log,tab
|
| 4 |
+
3,155,"extension-output-pdoom-org.crowd-code-#1-crowd-code",40,0,"",Log,selection_command
|
| 5 |
+
4,197,"extension-output-pdoom-org.crowd-code-#1-crowd-code",150,0,"9:25:04 AM [info] Git repository found\n9:25:04 AM [info] Git provider initialized successfully\n9:25:04 AM [info] Initial git state: [object Object]\n",Log,content
|
| 6 |
+
5,2189,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
|
| 7 |
+
6,2191,"TERMINAL",0,0,"",,terminal_focus
|
| 8 |
+
7,2213,"TERMINAL",0,0,"",,terminal_command
|
| 9 |
+
8,9325,"TERMINAL",0,0,"",,terminal_command
|
| 10 |
+
9,18113,"TERMINAL",0,0,"source .venv/bin/activate",,terminal_command
|
| 11 |
+
10,24112,"TERMINAL",0,0,"python cleanrl/ppo_atari_envpool.py",,terminal_command
|
| 12 |
+
11,24163,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 13 |
+
12,26014,"TERMINAL",0,0,"/fast/home/franz.srambical/cleanrl/.venv/lib/python3.10/site-packages/treevalue/tree/integration/torch.py:23: FutureWarning: `torch.utils._pytree._register_pytree_node` is deprecated. Please use `torch.utils._pytree.register_pytree_node` instead.\r\n register_for_torch(TreeValue)\r\n/fast/home/franz.srambical/cleanrl/.venv/lib/python3.10/site-packages/treevalue/tree/integration/torch.py:24: FutureWarning: `torch.utils._pytree._register_pytree_node` is deprecated. Please use `torch.utils._pytree.register_pytree_node` instead.\r\n register_for_torch(FastTreeValue)\r\n",,terminal_output
|
| 14 |
+
13,26264,"TERMINAL",0,0,"Gym has been unmaintained since 2022 and does not support NumPy 2.0 amongst other critical functionality.\r\nPlease upgrade to Gymnasium, the maintained drop-in replacement of Gym, or contact the authors of your software and request that they upgrade.\r\nSee the migration guide at https://gymnasium.farama.org/introduction/migration_guide/ for additional information.\r\n",,terminal_output
|
| 15 |
+
14,26759,"TERMINAL",0,0,"/fast/home/franz.srambical/cleanrl/.venv/lib/python3.10/site-packages/tyro/_parsers.py:379: UserWarning: The field `wandb-entity` is annotated with type `<class 'str'>`, but the default value `None` has type `<class 'NoneType'>`. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/fast/home/franz.srambical/cleanrl/.venv/lib/python3.10/site-packages/tyro/_parsers.py:379: UserWarning: The field `target-kl` is annotated with type `<class 'float'>`, but the default value `None` has type `<class 'NoneType'>`. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output
|
| 16 |
+
15,28466,"TERMINAL",0,0,"global_step=992, episodic_return=0.0\r\nglobal_step=992, episodic_return=0.0\r\nglobal_step=992, episodic_return=0.0\r\nglobal_step=992, episodic_return=0.0\r\nglobal_step=992, episodic_return=0.0\r\nglobal_step=992, episodic_return=0.0\r\nglobal_step=992, episodic_return=0.0\r\nglobal_step=992, episodic_return=0.0\r\n",,terminal_output
|
| 17 |
+
16,28939,"TERMINAL",0,0,"SPS: 1339\r\n",,terminal_output
|
| 18 |
+
17,29746,"TERMINAL",0,0,"SPS: 1302\r\n",,terminal_output
|
| 19 |
+
18,29996,"TERMINAL",0,0,"global_step=2776, episodic_return=2.0\r\nglobal_step=2776, episodic_return=2.0\r\nglobal_step=2776, episodic_return=2.0\r\nglobal_step=2776, episodic_return=2.0\r\nglobal_step=2776, episodic_return=2.0\r\nglobal_step=2776, episodic_return=2.0\r\nglobal_step=2776, episodic_return=2.0\r\nglobal_step=2776, episodic_return=2.0\r\n",,terminal_output
|
| 20 |
+
19,30574,"TERMINAL",0,0,"SPS: 1280\r\n",,terminal_output
|
| 21 |
+
20,30805,"TERMINAL",0,0,"global_step=3776, episodic_return=0.0\r\nglobal_step=3776, episodic_return=0.0\r\nglobal_step=3776, episodic_return=0.0\r\nglobal_step=3776, episodic_return=0.0\r\nglobal_step=3776, episodic_return=0.0\r\nglobal_step=3776, episodic_return=0.0\r\nglobal_step=3776, episodic_return=0.0\r\nglobal_step=3776, episodic_return=0.0\r\n",,terminal_output
|
| 22 |
+
21,31406,"TERMINAL",0,0,"SPS: 1266\r\n",,terminal_output
|
| 23 |
+
22,31746,"TERMINAL",0,0,"global_step=5104, episodic_return=1.0\r\nglobal_step=5104, episodic_return=1.0\r\nglobal_step=5104, episodic_return=1.0\r\nglobal_step=5104, episodic_return=1.0\r\nglobal_step=5104, episodic_return=1.0\r\nglobal_step=5104, episodic_return=1.0\r\nglobal_step=5104, episodic_return=1.0\r\nglobal_step=5104, episodic_return=1.0\r\n",,terminal_output
|
| 24 |
+
23,32262,"TERMINAL",0,0,"SPS: 1252\r\n",,terminal_output
|
| 25 |
+
24,33098,"TERMINAL",0,0,"SPS: 1247\r\n",,terminal_output
|
| 26 |
+
25,33217,"TERMINAL",0,0,"global_step=6496, episodic_return=1.0\r\nglobal_step=6496, episodic_return=1.0\r\nglobal_step=6496, episodic_return=1.0\r\nglobal_step=6496, episodic_return=1.0\r\nglobal_step=6496, episodic_return=1.0\r\nglobal_step=6496, episodic_return=1.0\r\nglobal_step=6496, episodic_return=1.0\r\nglobal_step=6496, episodic_return=1.0\r\n",,terminal_output
|
| 27 |
+
26,33964,"TERMINAL",0,0,"SPS: 1237\r\n",,terminal_output
|
| 28 |
+
27,34109,"TERMINAL",0,0,"^CTraceback (most recent call last):\r\n File ""/fast/home/franz.srambical/cleanrl/cleanrl/ppo_atari_envpool.py"", line 237, in <module>\r\n next_obs, reward, next_done, info = envs.step(action.cpu().numpy())\r\n File ""/fast/home/franz.srambical/cleanrl/cleanrl/ppo_atari_envpool.py"", line 100, in step\r\n observations, rewards, dones, infos = super().step(action)\r\n File ""/fast/home/franz.srambical/cleanrl/.venv/lib/python3.10/site-packages/gym/core.py"", line 280, in step\r\n return self.env.step(action)\r\n File ""/fast/home/franz.srambical/cleanrl/.venv/lib/python3.10/site-packages/envpool/python/envpool.py"", line 144, in step\r\n return self.recv(reset=False, return_info=True)\r\n File ""/fast/home/franz.srambical/cleanrl/.venv/lib/python3.10/site-packages/envpool/python/envpool.py"", line 130, in recv\r\n state_list = self._recv()\r\nKeyboardInterrupt\r\n[0m",,terminal_output
|
| 29 |
+
28,34254,"TERMINAL",0,0,"^C",,terminal_output
|
| 30 |
+
29,34258,"TERMINAL",0,0,"\r\n]0;franz.srambical@hai-login2:~/cleanrl",,terminal_output
|
| 31 |
+
30,116742,"TERMINAL",0,0,"",,terminal_command
|
| 32 |
+
31,278446,"cleanrl/ppo_atari_envpool.py",236,0,"",python,selection_mouse
|
| 33 |
+
32,278448,"cleanrl/ppo_atari_envpool.py",235,0,"",python,selection_command
|
| 34 |
+
33,290061,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"""""""\nGenerates a dataset of random-action CoinRun episodes.\nEpisodes are saved individually as memory-mapped files for efficient loading.\n""""""\n\nfrom dataclasses import dataclass\n\nfrom gym3 import types_np\nimport numpy as np\nfrom procgen import ProcgenGym3Env\nimport tyro\nimport json\nimport os\nfrom utils import save_chunks\n\n\n@dataclass\nclass Args:\n num_episodes_train: int = 10000\n num_episodes_val: int = 500\n num_episodes_test: int = 500\n output_dir: str = ""data/coinrun_episodes""\n min_episode_length: int = 1000\n max_episode_length: int = 1000\n chunk_size: int = 100\n chunks_per_file: int = 100\n seed: int = 0\n\n\nargs = tyro.cli(Args)\nassert (\n args.max_episode_length >= args.min_episode_length\n), ""Maximum episode length must be greater than or equal to minimum episode length.""\n\nif args.min_episode_length < args.chunk_size:\n print(\n ""Warning: Minimum episode length is smaller than chunk size. Note that episodes shorter than the chunk size will be discarded.""\n )\n\n\n# --- Generate episodes ---\ndef generate_episodes(num_episodes, split):\n episode_idx = 0\n episode_metadata = []\n obs_chunks = []\n act_chunks = []\n file_idx = 0\n output_dir_split = os.path.join(args.output_dir, split)\n while episode_idx < num_episodes:\n seed = np.random.randint(0, 10000)\n env = ProcgenGym3Env(num=1, env_name=""coinrun"", start_level=seed)\n\n observations_seq = []\n actions_seq = []\n episode_obs_chunks = []\n episode_act_chunks = []\n\n # --- Run episode ---\n step_t = 0\n for step_t in range(args.max_episode_length):\n action = types_np.sample(env.ac_space, bshape=(env.num,))\n env.act(action)\n _, obs, first = env.observe()\n observations_seq.append(obs[""rgb""])\n actions_seq.append(action)\n if len(observations_seq) == args.chunk_size:\n episode_obs_chunks.append(observations_seq)\n episode_act_chunks.append(actions_seq)\n observations_seq = []\n actions_seq = []\n if first:\n break\n\n # --- Save episode ---\n if step_t + 1 >= args.min_episode_length:\n if observations_seq:\n if len(observations_seq) < args.chunk_size:\n print(\n f""Warning: Inconsistent chunk_sizes. Episode has {len(observations_seq)} frames, ""\n f""which is smaller than the requested chunk_size: {args.chunk_size}. ""\n ""This might lead to performance degradation during training.""\n )\n episode_obs_chunks.append(observations_seq)\n episode_act_chunks.append(actions_seq)\n\n obs_chunks_data = [\n np.concatenate(seq, axis=0).astype(np.uint8)\n for seq in episode_obs_chunks\n ]\n act_chunks_data = [\n np.concatenate(act, axis=0) for act in episode_act_chunks\n ]\n obs_chunks.extend(obs_chunks_data)\n act_chunks.extend(act_chunks_data)\n\n ep_metadata, obs_chunks, file_idx, act_chunks = save_chunks(\n obs_chunks, file_idx, args.chunks_per_file, output_dir_split, act_chunks\n )\n episode_metadata.extend(ep_metadata)\n\n print(f""Episode {episode_idx} completed, length: {step_t + 1}."")\n episode_idx += 1\n else:\n print(f""Episode too short ({step_t + 1}), resampling..."")\n\n if len(obs_chunks) > 0:\n print(\n f""Warning: Dropping {len(obs_chunks)} chunks for consistent number of chunks per file."",\n ""Consider changing the chunk_size and chunks_per_file parameters to prevent data-loss."",\n )\n\n print(f""Done generating {split} split"")\n return episode_metadata\n\n\ndef get_action_space():\n env = ProcgenGym3Env(num=1, env_name=""coinrun"", start_level=0)\n return env.ac_space.eltype.n\n\n\ndef main():\n # Set random seed and create dataset directories\n np.random.seed(args.seed)\n # --- Generate episodes ---\n train_episode_metadata = generate_episodes(args.num_episodes_train, ""train"")\n val_episode_metadata = generate_episodes(args.num_episodes_val, ""val"")\n test_episode_metadata = generate_episodes(args.num_episodes_test, ""test"")\n\n # --- Save metadata ---\n metadata = {\n ""env"": ""coinrun"",\n ""num_actions"": get_action_space(),\n ""num_episodes_train"": args.num_episodes_train,\n ""num_episodes_val"": args.num_episodes_val,\n ""num_episodes_test"": args.num_episodes_test,\n ""avg_episode_len_train"": np.mean(\n [ep[""avg_seq_len""] for ep in train_episode_metadata]\n ),\n ""avg_episode_len_val"": np.mean(\n [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")\n\n\nif __name__ == ""__main__"":\n main()\n",python,tab
|
| 35 |
+
34,299089,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
|
| 36 |
+
35,300383,"cleanrl/ppo_atari_envpool.py",318,0,"",python,selection_mouse
|
| 37 |
+
36,300384,"cleanrl/ppo_atari_envpool.py",317,0,"",python,selection_command
|
| 38 |
+
37,314316,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"",python,tab
|
| 39 |
+
38,315490,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
|
| 40 |
+
39,316028,"cleanrl/ppo_atari_envpool.py",330,0,"",python,selection_mouse
|
| 41 |
+
40,316031,"cleanrl/ppo_atari_envpool.py",329,0,"",python,selection_command
|
| 42 |
+
41,316695,"cleanrl/ppo_atari_envpool.py",436,0,"",python,selection_mouse
|
| 43 |
+
42,316698,"cleanrl/ppo_atari_envpool.py",435,0,"",python,selection_command
|
| 44 |
+
43,317177,"cleanrl/ppo_atari_envpool.py",438,0,"",python,selection_mouse
|
| 45 |
+
44,397100,"cleanrl/ppo_atari_envpool.py",2542,0,"",python,selection_mouse
|
| 46 |
+
45,402425,"cleanrl/ppo_atari_envpool.py",4160,0,"",python,selection_mouse
|
| 47 |
+
46,541375,"TERMINAL",0,0,"",,terminal_focus
|
| 48 |
+
47,541842,"TERMINAL",0,0,"source /home/franz.srambical/cleanrl/.venv/bin/activate",,terminal_command
|
| 49 |
+
48,544327,"TERMINAL",0,0,"cd ../jafar/",,terminal_command
|
| 50 |
+
49,554362,"TERMINAL",0,0,"git checkout gt-actions",,terminal_command
|
| 51 |
+
50,554410,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 52 |
+
51,554473,"TERMINAL",0,0,"M\t.gitignore\r\nAlready on 'gt-actions'\r\nYour branch is behind 'origin/gt-actions' by 1 commit, and can be fast-forwarded.\r\n (use ""git pull"" to update your local branch)\r\n]0;franz.srambical@hai-login2:~/jafar",,terminal_output
|
| 53 |
+
52,555777,"TERMINAL",0,0,"git pull",,terminal_command
|
| 54 |
+
53,555825,"TERMINAL",0,0,"]633;C",,terminal_output
|
| 55 |
+
54,557974,"TERMINAL",0,0,"Updating 96d560e..1b6b878\r\nFast-forward\r\n input_pipeline/generate_coinrun_dataset.py | 2 [32m+[m[31m-[m\r\n 1 file changed, 1 insertion(+), 1 deletion(-)\r\n]0;franz.srambical@hai-login2:~/jafar",,terminal_output
|
| 56 |
+
55,560140,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"",python,tab
|
| 57 |
+
56,560211,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",563,26," chunk_size: int = 160\n",python,content
|
| 58 |
+
57,561853,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
|
| 59 |
+
58,570185,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"",python,tab
|
| 60 |
+
59,571693,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
|
| 61 |
+
60,576013,"cleanrl/ppo_atari_envpool.py",4335,0,"",python,selection_mouse
|
| 62 |
+
61,725152,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"",python,tab
|
| 63 |
+
62,725643,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",280,0,"",python,selection_mouse
|
| 64 |
+
63,725644,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",279,0,"",python,selection_command
|
| 65 |
+
64,727140,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",1442,0,"",python,selection_command
|
| 66 |
+
65,728476,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
|
| 67 |
+
66,782305,"cleanrl/trajectory_saver.py",0,0,"import os\nimport json\nfrom typing import List, Dict, Any\n\nimport numpy as np\n\n\nclass TrajectorySaver:\n """"""\n Collects per-environment observation frames and actions, chunks them,\n and writes chunked files to disk. Designed to be lightweight and only\n enabled when trajectory capture is requested.\n\n Behavior mirrors the high level flow in `generate_coinrun_dataset.py`:\n - Build sequences of length `chunk_size` from steps\n - At episode end, if a partial sequence exists (< chunk_size), include it\n with a warning about inconsistent chunk sizes\n - Group `chunks_per_file` chunks per output file. Any remainder at the\n very end is dropped with a warning\n - Store both observation chunks and action chunks side-by-side\n """"""\n\n def __init__(\n self,\n output_dir: str,\n num_envs: int,\n chunk_size: int = 160,\n chunks_per_file: int = 100,\n ) -> None:\n self.output_dir = output_dir\n self.chunks_dir = os.path.join(output_dir, ""chunks"")\n os.makedirs(self.chunks_dir, exist_ok=True)\n self.num_envs = num_envs\n self.chunk_size = int(chunk_size)\n self.chunks_per_file = int(chunks_per_file)\n\n # Per-env rolling buffers for the current episode\n self.env_buffers: List[Dict[str, Any]] = [\n {""obs_seq"": [], ""act_seq"": [], ""episode_obs_chunks"": [], ""episode_act_chunks"": []}\n for _ in range(self.num_envs)\n ]\n\n # Global chunk buffers waiting to be flushed to disk\n self.obs_chunks: List[np.ndarray] = []\n self.act_chunks: List[np.ndarray] = []\n self.file_idx: int = 0\n\n # Metadata (optional). We keep a minimal JSONL with per-file info\n self.metadata_path = os.path.join(self.output_dir, ""metadata.jsonl"")\n\n def _finalize_episode_for_env(self, env_idx: int) -> None:\n buf = self.env_buffers[env_idx]\n obs_seq: List[np.ndarray] = buf[""obs_seq""]\n act_seq: List[np.ndarray] = buf[""act_seq""]\n\n if len(obs_seq) > 0:\n # Partial chunk at episode end\n if len(obs_seq) < self.chunk_size:\n print(\n f""Warning: Inconsistent chunk sizes. Env {env_idx} episode ended with ""\n f""{len(obs_seq)} frames (< chunk_size={self.chunk_size}). ""\n ""Including the partial chunk may impact training I/O performance.""\n )\n self._append_chunk_from_sequences(env_idx)\n\n # Move episode chunks to global buffers\n if len(buf[""episode_obs_chunks""]) > 0:\n self.obs_chunks.extend(buf[""episode_obs_chunks""])\n self.act_chunks.extend(buf[""episode_act_chunks""])\n buf[""episode_obs_chunks""].clear()\n buf[""episode_act_chunks""].clear()\n\n self._save_if_ready()\n\n def _append_chunk_from_sequences(self, env_idx: int) -> None:\n buf = self.env_buffers[env_idx]\n obs_seq: List[np.ndarray] = buf[""obs_seq""]\n act_seq: List[np.ndarray] = buf[""act_seq""]\n\n if len(obs_seq) == 0:\n return\n\n # Stack along time dimension. Shapes become:\n # obs_chunk: (T, C, H, W) dtype=uint8\n # act_chunk: (T,) dtype=int64 (or int)\n obs_chunk = np.stack(obs_seq, axis=0).astype(np.uint8)\n act_chunk = np.asarray(act_seq, dtype=np.int64)\n\n buf[""episode_obs_chunks""].append(obs_chunk)\n buf[""episode_act_chunks""].append(act_chunk)\n\n # Reset rolling sequences\n buf[""obs_seq""] = []\n buf[""act_seq""] = []\n\n def _save_if_ready(self) -> None:\n while len(self.obs_chunks) >= self.chunks_per_file:\n obs_to_write = self.obs_chunks[: self.chunks_per_file]\n act_to_write = self.act_chunks[: self.chunks_per_file]\n\n # Remove from buffers\n self.obs_chunks = self.obs_chunks[self.chunks_per_file :]\n self.act_chunks = self.act_chunks[self.chunks_per_file :]\n\n file_stem = f""chunks_{self.file_idx:06d}""\n out_path = os.path.join(self.chunks_dir, f""{file_stem}.npz"")\n\n # Write as a .npz with numbered arrays to preserve variable chunk lengths\n save_dict: Dict[str, Any] = {}\n for idx, (o, a) in enumerate(zip(obs_to_write, act_to_write)):\n save_dict[f""obs_{idx:03d}""] = o\n save_dict[f""acts_{idx:03d}""] = a\n\n np.savez_compressed(out_path, **save_dict)\n\n # Minimal metadata per file for bookkeeping\n meta = {\n ""file"": os.path.basename(out_path),\n ""num_chunks"": len(obs_to_write),\n ""chunk_size"": self.chunk_size,\n ""chunk_indices"": [i for i in range(len(obs_to_write))],\n }\n with open(self.metadata_path, ""a"") as f:\n f.write(json.dumps(meta) + ""\n"")\n\n self.file_idx += 1\n\n def add_step(\n self,\n obs_batch: np.ndarray,\n actions_batch: np.ndarray,\n episode_done_batch: np.ndarray,\n ) -> None:\n """"""\n Add a single environment step across all envs.\n\n Args:\n obs_batch: np.ndarray with shape (num_envs, C, H, W). Values should be [0, 255].\n actions_batch: np.ndarray with shape (num_envs,).\n episode_done_batch: np.ndarray with shape (num_envs,), boolean flags indicating end-of-episode.\n """"""\n assert obs_batch.shape[0] == self.num_envs\n assert actions_batch.shape[0] == self.num_envs\n assert episode_done_batch.shape[0] == self.num_envs\n\n for env_idx in range(self.num_envs):\n buf = self.env_buffers[env_idx]\n buf[""obs_seq""].append(obs_batch[env_idx])\n buf[""act_seq""].append(int(actions_batch[env_idx]))\n\n if len(buf[""obs_seq""]) == self.chunk_size:\n self._append_chunk_from_sequences(env_idx)\n\n # Episode ended: finalize the episode\n if bool(episode_done_batch[env_idx]):\n self._finalize_episode_for_env(env_idx)\n\n def close(self) -> None:\n # Drop remainder to keep consistent chunks-per-file contract\n remainder = len(self.obs_chunks)\n if remainder > 0:\n print(\n f""Warning: Dropping {remainder} chunks to keep a consistent number of chunks per file. ""\n ""Consider adjusting chunk_size and chunks_per_file to reduce data loss.""\n )\n # Nothing else to do; per-episode buffers do not carry data unless an episode just ended\n\n\n",python,tab
|
| 68 |
+
67,859278,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
|
| 69 |
+
68,877159,"cleanrl/ppo_atari_envpool.py",14147,0," if traj_saver is not None:\n traj_saver.close()\n",python,content
|
| 70 |
+
69,877159,"cleanrl/ppo_atari_envpool.py",8882,0," if traj_saver is not None:\n # Use true termination flags when available\n terminated = np.asarray(info.get(""terminated"", next_done.detach().cpu().numpy()), dtype=bool)\n traj_saver.add_step(obs_to_save, actions_to_save, terminated)\n\n",python,content
|
| 71 |
+
70,877159,"cleanrl/ppo_atari_envpool.py",8628,0," # Capture current observation and action prior to stepping the envs\n if traj_saver is not None:\n obs_to_save = next_obs.detach().cpu().numpy().astype(np.uint8)\n actions_to_save = action.detach().cpu().numpy()\n",python,content
|
| 72 |
+
71,877159,"cleanrl/ppo_atari_envpool.py",6967,0," # Optional: trajectory saver\n traj_saver = None\n if args.capture_trajectories:\n traj_saver = TrajectorySaver(\n output_dir=args.trajectories_output_dir,\n num_envs=args.num_envs,\n chunk_size=args.trajectories_chunk_size,\n chunks_per_file=args.trajectories_chunks_per_file,\n )\n\n",python,content
|
| 73 |
+
72,877159,"cleanrl/ppo_atari_envpool.py",1167,0," # Trajectory capture\n capture_trajectories: bool = False\n """"""if toggled, save observation frames and actions during training""""""\n trajectories_output_dir: str = ""data/atari_trajectories""\n """"""directory to store trajectory chunks (.npz)""""""\n trajectories_chunk_size: int = 160\n """"""number of time steps per chunk before writing""""""\n trajectories_chunks_per_file: int = 100\n """"""number of chunks grouped per npz file""""""\n\n",python,content
|
| 74 |
+
73,877160,"cleanrl/ppo_atari_envpool.py",437,0,"from .trajectory_saver import TrajectorySaver\n",python,content
|
| 75 |
+
74,913280,"cleanrl/ppo_atari_envpool.py",15527,58,"",python,content
|
| 76 |
+
75,913280,"cleanrl/ppo_atari_envpool.py",9974,288,"",python,content
|
| 77 |
+
76,913280,"cleanrl/ppo_atari_envpool.py",9458,262,"",python,content
|
| 78 |
+
77,913280,"cleanrl/ppo_atari_envpool.py",7454,343,"",python,content
|
| 79 |
+
78,913280,"cleanrl/ppo_atari_envpool.py",1213,441,"",python,content
|
| 80 |
+
79,913280,"cleanrl/ppo_atari_envpool.py",437,46,"",python,content
|
| 81 |
+
80,913354,"cleanrl/ppo_atari_envpool.py",14147,0," if traj_saver is not None:\n traj_saver.close()\n",python,content
|
| 82 |
+
81,913354,"cleanrl/ppo_atari_envpool.py",10671,0," # Initialize for type-checkers\n approx_kl = torch.tensor(0.0)\n old_approx_kl = torch.tensor(0.0)\n v_loss = torch.tensor(0.0)\n pg_loss = torch.tensor(0.0)\n entropy_loss = torch.tensor(0.0)\n",python,content
|
| 83 |
+
82,913354,"cleanrl/ppo_atari_envpool.py",8882,0," if traj_saver is not None and obs_to_save is not None and actions_to_save is not None:\n # Use true termination flags when available\n terminated = np.asarray(info.get(""terminated"", next_done.detach().cpu().numpy()), dtype=bool)\n traj_saver.add_step(obs_to_save, actions_to_save, terminated)\n\n",python,content
|
| 84 |
+
83,913354,"cleanrl/ppo_atari_envpool.py",8628,0," # Capture current observation and action prior to stepping the envs\n obs_to_save = None\n actions_to_save = None\n if traj_saver is not None:\n obs_to_save = next_obs.detach().cpu().numpy().astype(np.uint8)\n actions_to_save = action.detach().cpu().numpy()\n",python,content
|
| 85 |
+
84,913354,"cleanrl/ppo_atari_envpool.py",6861,104," assert isinstance(envs.action_space, spaces.Discrete), ""only discrete action space is supported""\n\n # Optional: trajectory saver\n traj_saver = None\n if args.capture_trajectories:\n traj_saver = TrajectorySaver(\n output_dir=args.trajectories_output_dir,\n num_envs=args.num_envs,\n chunk_size=args.trajectories_chunk_size,\n chunks_per_file=args.trajectories_chunks_per_file,\n )",python,content
|
| 86 |
+
85,913354,"cleanrl/ppo_atari_envpool.py",2469,27," target_kl: Optional[float] = None",python,content
|
| 87 |
+
86,913355,"cleanrl/ppo_atari_envpool.py",1167,0," # Trajectory capture\n capture_trajectories: bool = False\n """"""if toggled, save observation frames and actions during training""""""\n trajectories_output_dir: str = ""data/atari_trajectories""\n """"""directory to store trajectory chunks (.npz)""""""\n trajectories_chunk_size: int = 160\n """"""number of time steps per chunk before writing""""""\n trajectories_chunks_per_file: int = 100\n """"""number of chunks grouped per npz file""""""\n\n",python,content
|
| 88 |
+
87,913355,"cleanrl/ppo_atari_envpool.py",968,28," wandb_entity: Optional[str] = None",python,content
|
| 89 |
+
88,913355,"cleanrl/ppo_atari_envpool.py",387,49,"from torch.utils.tensorboard.writer import SummaryWriter\nfrom typing import Optional\nfrom gym import spaces\nfrom .trajectory_saver import TrajectorySaver",python,content
|
| 90 |
+
89,961375,"cleanrl/ppo_atari_envpool.py",15958,58,"",python,content
|
| 91 |
+
90,961375,"cleanrl/ppo_atari_envpool.py",12251,231,"",python,content
|
| 92 |
+
91,961375,"cleanrl/ppo_atari_envpool.py",10114,348,"",python,content
|
| 93 |
+
92,961375,"cleanrl/ppo_atari_envpool.py",9532,328,"",python,content
|
| 94 |
+
93,961375,"cleanrl/ppo_atari_envpool.py",7426,443," assert isinstance(envs.action_space, gym.spaces.Discrete), ""only discrete action space is supported""",python,content
|
| 95 |
+
94,961375,"cleanrl/ppo_atari_envpool.py",3024,37," target_kl: float = None",python,content
|
| 96 |
+
95,961375,"cleanrl/ppo_atari_envpool.py",1281,441,"",python,content
|
| 97 |
+
96,961375,"cleanrl/ppo_atari_envpool.py",1072,38," wandb_entity: str = None",python,content
|
| 98 |
+
97,961375,"cleanrl/ppo_atari_envpool.py",387,153,"from torch.utils.tensorboard import SummaryWriter",python,content
|
| 99 |
+
98,961438,"cleanrl/ppo_atari_envpool.py",14147,0," if traj_saver is not None:\n traj_saver.close()\n",python,content
|
| 100 |
+
99,961438,"cleanrl/ppo_atari_envpool.py",10671,0," # Initialize for type-checkers\n approx_kl = torch.tensor(0.0)\n old_approx_kl = torch.tensor(0.0)\n v_loss = torch.tensor(0.0)\n pg_loss = torch.tensor(0.0)\n entropy_loss = torch.tensor(0.0)\n",python,content
|
| 101 |
+
100,961438,"cleanrl/ppo_atari_envpool.py",8882,0," if traj_saver is not None and obs_to_save is not None and actions_to_save is not None:\n # Use true termination flags when available\n terminated = np.asarray(info.get(""terminated"", next_done.detach().cpu().numpy()), dtype=bool)\n traj_saver.add_step(obs_to_save, actions_to_save, terminated)\n\n",python,content
|
| 102 |
+
101,961438,"cleanrl/ppo_atari_envpool.py",8628,0," # Capture current observation and action prior to stepping the envs\n obs_to_save = None\n actions_to_save = None\n if traj_saver is not None:\n obs_to_save = next_obs.detach().cpu().numpy().astype(np.uint8)\n actions_to_save = action.detach().cpu().numpy()\n",python,content
|
| 103 |
+
102,961438,"cleanrl/ppo_atari_envpool.py",7002,79," optimizer = Adam(agent.parameters(), lr=args.learning_rate, eps=1e-5)",python,content
|
| 104 |
+
103,961438,"cleanrl/ppo_atari_envpool.py",6861,104," assert isinstance(envs.action_space, spaces.Discrete), ""only discrete action space is supported""\n\n # Optional: trajectory saver\n traj_saver = None\n if args.capture_trajectories:\n traj_saver = TrajectorySaver(\n output_dir=args.trajectories_output_dir,\n num_envs=args.num_envs,\n chunk_size=args.trajectories_chunk_size,\n chunks_per_file=args.trajectories_chunks_per_file,\n )",python,content
|
| 105 |
+
104,961438,"cleanrl/ppo_atari_envpool.py",2982,71," self.episode_returns = np.zeros(self.num_envs, dtype=np.float32)\n self.episode_lengths = np.zeros(self.num_envs, dtype=np.int32)",python,content
|
| 106 |
+
105,961438,"cleanrl/ppo_atari_envpool.py",2469,27," target_kl: Optional[float] = None",python,content
|
| 107 |
+
106,961438,"cleanrl/ppo_atari_envpool.py",1167,0," # Trajectory capture\n capture_trajectories: bool = False\n """"""if toggled, save observation frames and actions during training""""""\n trajectories_output_dir: str = ""data/atari_trajectories""\n """"""directory to store trajectory chunks (.npz)""""""\n trajectories_chunk_size: int = 160\n """"""number of time steps per chunk before writing""""""\n trajectories_chunks_per_file: int = 100\n """"""number of chunks grouped per npz file""""""\n\n",python,content
|
| 108 |
+
107,961438,"cleanrl/ppo_atari_envpool.py",968,28," wandb_entity: Optional[str] = None",python,content
|
| 109 |
+
108,961438,"cleanrl/ppo_atari_envpool.py",387,49,"from torch.utils.tensorboard.writer import SummaryWriter\nfrom typing import Optional\nfrom gym import spaces\nfrom .trajectory_saver import TrajectorySaver",python,content
|
| 110 |
+
109,961438,"cleanrl/ppo_atari_envpool.py",291,27,"from torch.optim import Adam",python,content
|
| 111 |
+
110,968028,"cleanrl/ppo_atari_envpool.py",268,0,"",python,selection_mouse
|
| 112 |
+
111,968030,"cleanrl/ppo_atari_envpool.py",267,0,"",python,selection_command
|
| 113 |
+
112,969822,"cleanrl/ppo_atari_envpool.py",280,0,"",python,selection_command
|
| 114 |
+
113,974152,"cleanrl/ppo_atari_envpool.py",291,28,"import torch.optim as optim",python,content
|
| 115 |
+
114,977518,"cleanrl/ppo_atari_envpool.py",7994,0,"",python,selection_command
|
| 116 |
+
115,978956,"cleanrl/ppo_atari_envpool.py",7978,73," optimizer = optim.Adam(agent.parameters(), lr=args.learning_rate, eps=1e-5)",python,content
|
| 117 |
+
116,979756,"cleanrl/ppo_atari_envpool.py",8056,0,"",python,selection_command
|
| 118 |
+
117,980869,"cleanrl/ppo_atari_envpool.py",0,0,"",python,selection_command
|
| 119 |
+
118,1015212,"cleanrl/ppo_atari_envpool.py",12299,0,"",python,selection_mouse
|
| 120 |
+
119,1015215,"cleanrl/ppo_atari_envpool.py",12298,0,"",python,selection_command
|
| 121 |
+
120,1016996,"cleanrl/ppo_atari_envpool.py",12323,231,"",python,content
|
| 122 |
+
121,1021028,"cleanrl/ppo_atari_envpool.py",12205,0,"",python,selection_mouse
|
| 123 |
+
122,1027883,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"",python,tab
|
| 124 |
+
123,1037912,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",3201,0,"",python,selection_mouse
|
| 125 |
+
124,1043895,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"",python,selection_command
|
| 126 |
+
125,1063321,"/home/franz.srambical/jafar/input_pipeline/utils.py",0,0,"import os\nimport pickle\nimport numpy as np\nfrom array_record.python.array_record_module import ArrayRecordWriter\n\n\ndef save_chunks(obs_chunks, file_idx, chunks_per_file, output_dir, act_chunks=None):\n os.makedirs(output_dir, exist_ok=True)\n\n metadata = []\n while len(obs_chunks) >= chunks_per_file:\n chunk_batch = obs_chunks[:chunks_per_file]\n obs_chunks = obs_chunks[chunks_per_file:]\n act_chunk_batch = None\n if act_chunks:\n act_chunk_batch = act_chunks[:chunks_per_file]\n act_chunks = act_chunks[chunks_per_file:]\n episode_path = os.path.join(output_dir, f""data_{file_idx:04d}.array_record"")\n writer = ArrayRecordWriter(str(episode_path), ""group_size:1"")\n seq_lens = []\n for idx, chunk in enumerate(chunk_batch):\n seq_len = chunk.shape[0]\n seq_lens.append(seq_len)\n chunk_record = {\n ""raw_video"": chunk.tobytes(),\n ""sequence_length"": seq_len,\n }\n if act_chunk_batch:\n assert len(chunk) == len(\n act_chunk_batch[idx]\n ), f""Observation data length and action sequence length do not match: {len(chunk)} != {len(act_chunk_batch[idx])}""\n chunk_record[""actions""] = act_chunk_batch[idx]\n writer.write(pickle.dumps(chunk_record))\n writer.close()\n file_idx += 1\n metadata.append(\n {\n ""path"": episode_path,\n ""num_chunks"": len(chunk_batch),\n ""avg_seq_len"": np.mean(seq_lens),\n }\n )\n print(f""Created {episode_path} with {len(chunk_batch)} video chunks"")\n\n return metadata, obs_chunks, file_idx, act_chunks\n",python,tab
|
| 127 |
+
126,1077626,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"",python,tab
|
| 128 |
+
127,1078649,"/home/franz.srambical/jafar/input_pipeline/utils.py",0,0,"",python,tab
|
| 129 |
+
128,1084681,"/home/franz.srambical/jafar/input_pipeline/generate_coinrun_dataset.py",0,0,"",python,tab
|
| 130 |
+
129,1098108,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
|
| 131 |
+
130,1107176,"cleanrl/ppo_atari_envpool.py",716,0,"",python,selection_mouse
|
| 132 |
+
131,1107177,"cleanrl/ppo_atari_envpool.py",715,0,"",python,selection_command
|
| 133 |
+
132,1163039,"cleanrl/ppo_atari_envpool.py",553,0,"",python,selection_mouse
|
| 134 |
+
133,1163040,"cleanrl/ppo_atari_envpool.py",552,0,"",python,selection_command
|
| 135 |
+
134,1164422,"cleanrl/ppo_atari_envpool.py",542,0,"",python,selection_mouse
|
| 136 |
+
135,1167929,"cleanrl/ppo_atari_envpool.py",565,0,"",python,selection_mouse
|
| 137 |
+
136,1167931,"cleanrl/ppo_atari_envpool.py",564,0,"",python,selection_command
|
| 138 |
+
137,1169182,"cleanrl/ppo_atari_envpool.py",387,0,"",python,selection_command
|
| 139 |
+
138,1170321,"cleanrl/ppo_atari_envpool.py",1072,0,"",python,selection_command
|
| 140 |
+
139,1170873,"cleanrl/ppo_atari_envpool.py",387,0,"",python,selection_command
|
| 141 |
+
140,1172070,"cleanrl/ppo_atari_envpool.py",1072,0,"",python,selection_command
|
| 142 |
+
141,1172368,"cleanrl/ppo_atari_envpool.py",387,0,"",python,selection_command
|
| 143 |
+
142,1172644,"cleanrl/ppo_atari_envpool.py",1072,0,"",python,selection_command
|
| 144 |
+
143,4767412,"cleanrl/ppo_atari_envpool.py",840,0,"",python,selection_mouse
|
| 145 |
+
144,4767415,"cleanrl/ppo_atari_envpool.py",839,0,"",python,selection_command
|
| 146 |
+
145,4769432,"cleanrl/trajectory_saver.py",0,0,"",python,tab
|
| 147 |
+
146,4770255,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
|
| 148 |
+
147,4770257,"cleanrl/ppo_atari_envpool.py",387,0,"",python,selection_command
|
| 149 |
+
148,4770719,"cleanrl/trajectory_saver.py",0,0,"",python,tab
|
| 150 |
+
149,4771118,"cleanrl/ppo_atari_envpool.py",0,0,"",python,tab
|
| 151 |
+
150,5039789,"cleanrl/ppo_atari_envpool.py",210,0,"",python,selection_mouse
|
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-d9cdf338-0ddd-4679-853a-6d7bdf2b18581751046137722-2025_06_27-10.42.19.354/source.csv
ADDED
|
@@ -0,0 +1,167 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,1,"utils/dataloader.py",0,0,"import functools\nimport jax\n\nimport tensorflow as tf\n\n# reserve GPU memory for JAX only if tensorflow is built with GPU support\ntf.config.experimental.set_visible_devices([], ""GPU"")\n\n\n# --- TensorFlow function for processing: slicing, normalization ---\ndef _tf_process_episode(episode_tensor, seq_len, image_h, image_w, image_c):\n """"""\n Processes a raw episode tensor in TensorFlow.\n Takes a full episode, extracts a random sequence, and normalizes it.\n Args:\n episode_tensor: A TensorFlow tensor representing a full video episode.\n Expected shape: (dynamic_length, image_h, image_w, image_c)\n Expected dtype: e.g., tf.uint8 (raw pixel values)\n seq_len: The desired length of the sub-sequence to extract.\n image_h: The height of each frame.\n image_w: The width of each frame.\n image_c: The number of channels in each frame.\n Returns:\n A TensorFlow tensor representing the processed video sequence.\n Shape: (seq_len, image_h, image_w, image_c)\n Dtype: tf.float32 (normalized pixel values)\n """"""\n current_episode_len = tf.shape(episode_tensor)[0]\n\n max_start_idx = current_episode_len - seq_len\n\n start_idx = tf.random.uniform(\n shape=(), minval=0, maxval=max_start_idx + 1, dtype=tf.int32\n )\n\n seq = episode_tensor[start_idx : start_idx + seq_len]\n\n seq = tf.cast(seq, tf.float32) / 255.0\n\n # Ensure the final shape is statically known for batching.\n # tf.reshape is robust, but tf.ensure_shape or set_shape can also be used if confident.\n processed_sequence = tf.reshape(seq, [seq_len, image_h, image_w, image_c])\n\n return processed_sequence\n\n\ndef _parse_tfrecord_fn(example_proto, image_h, image_w, image_c):\n feature_description = {\n ""height"": tf.io.FixedLenFeature([], tf.int64),\n ""width"": tf.io.FixedLenFeature([], tf.int64),\n ""channels"": tf.io.FixedLenFeature([], tf.int64),\n ""sequence_length"": tf.io.FixedLenFeature([], tf.int64),\n ""raw_video"": tf.io.FixedLenFeature([], tf.string),\n }\n example = tf.io.parse_single_example(example_proto, feature_description)\n\n video_shape = (example[""sequence_length""], image_h, image_w, image_c)\n\n episode_tensor = tf.io.decode_raw(example[""raw_video""], out_type=tf.uint8)\n episode_tensor = tf.reshape(episode_tensor, video_shape)\n\n episode_tensor = tf.ensure_shape(episode_tensor, [None, image_h, image_w, image_c])\n return episode_tensor\n\n\ndef get_dataloader(\n tfrecord_paths: list[str], # List of TFRecord file paths\n seq_len: int,\n global_batch_size: int,\n image_h: int,\n image_w: int,\n image_c: int,\n shuffle_buffer_size: int = 1000,\n num_parallel_calls: int = tf.data.AUTOTUNE,\n seed: int = 42,\n):\n """"""\n Creates a tf.data.Dataset pipeline from TFRecord files.\n """"""\n if not tfrecord_paths:\n raise ValueError(""tfrecord_paths list cannot be empty."")\n\n process_id = jax.process_index()\n num_processes = jax.process_count()\n\n assert (\n global_batch_size % num_processes == 0\n ), ""Global batch size {global_batch_size} \\n must be divisible by the number of JAX processes {num_processes} for proper sharding.""\n per_process_batch_size = global_batch_size // num_processes\n\n dataset = tf.data.TFRecordDataset(\n tfrecord_paths, num_parallel_reads=tf.data.AUTOTUNE\n )\n\n dataset = dataset.shard(num_shards=num_processes, index=process_id)\n\n # (f.srambical) NOTE: For TFRecords, it's often good to have a large shuffle buffer.\n if shuffle_buffer_size > 0:\n dataset = dataset.shuffle(\n buffer_size=shuffle_buffer_size, seed=seed, reshuffle_each_iteration=True\n )\n parse_fn = functools.partial(\n _parse_tfrecord_fn, image_h=image_h, image_w=image_w, image_c=image_c\n )\n dataset = dataset.map(parse_fn, num_parallel_calls=num_parallel_calls)\n\n tf_process_fn = functools.partial(\n _tf_process_episode,\n seq_len=seq_len,\n image_h=image_h,\n image_w=image_w,\n image_c=image_c,\n )\n dataset = dataset.map(tf_process_fn, num_parallel_calls=num_parallel_calls)\n\n dataset = dataset.repeat(None)\n dataset = dataset.batch(per_process_batch_size, drop_remainder=True)\n dataset = dataset.prefetch(tf.data.AUTOTUNE)\n\n return dataset.as_numpy_iterator()\n",python,tab
|
| 3 |
+
2,38,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
|
| 4 |
+
3,67,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"10:42:19 AM [info] Activating crowd-code\n10:42:19 AM [info] Recording started\n10:42:19 AM [info] Initializing git provider using file system watchers...\n10:42:19 AM [info] Git repository found\n10:42:19 AM [info] Git provider initialized successfully\n10:42:19 AM [info] Initial git state: [object Object]\n",Log,content
|
| 5 |
+
4,905,"utils/dataloader.py",0,0,"",python,tab
|
| 6 |
+
5,19039,"utils/dataloader.py",2752,0,"",python,selection_command
|
| 7 |
+
6,19130,"utils/dataloader.py",2715,0,"",python,selection_command
|
| 8 |
+
7,30009,"models/dynamics.py",0,0,"from typing import Dict, Any\n\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\n\nfrom utils.nn import STTransformer\n\n\nclass DynamicsMaskGIT(nn.Module):\n """"""MaskGIT dynamics model""""""\n\n model_dim: int\n num_latents: int\n num_blocks: int\n num_heads: int\n dropout: float\n mask_limit: float\n\n def setup(self):\n self.dynamics = STTransformer(\n self.model_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n )\n self.patch_embed = nn.Embed(self.num_latents, self.model_dim)\n self.mask_token = self.param(\n ""mask_token"",\n nn.initializers.lecun_uniform(),\n (1, 1, 1, self.model_dim),\n )\n self.action_up = nn.Dense(self.model_dim)\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n # --- Mask videos ---\n vid_embed = self.patch_embed(batch[""video_tokens""])\n if training:\n rng1, rng2 = jax.random.split(batch[""mask_rng""])\n mask_prob = jax.random.uniform(rng1, minval=self.mask_limit)\n mask = jax.random.bernoulli(rng2, mask_prob, vid_embed.shape[:-1])\n mask = mask.at[:, 0].set(False)\n vid_embed = jnp.where(jnp.expand_dims(mask, -1), self.mask_token, vid_embed)\n else:\n mask = None\n\n # --- Predict transition ---\n act_embed = self.action_up(batch[""latent_actions""])\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n logits = self.dynamics(vid_embed)\n return dict(token_logits=logits, mask=mask)\n",python,tab
|
| 9 |
+
8,30937,"models/dynamics.py",416,0,"",python,selection_command
|
| 10 |
+
9,31509,"models/dynamics.py",1359,0,"",python,selection_command
|
| 11 |
+
10,31790,"models/dynamics.py",1655,0,"",python,selection_command
|
| 12 |
+
11,32552,"models/dynamics.py",753,0,"",python,selection_command
|
| 13 |
+
12,32710,"models/dynamics.py",165,0,"",python,selection_command
|
| 14 |
+
13,32898,"models/dynamics.py",0,0,"",python,selection_command
|
| 15 |
+
14,35425,"train_dynamics.py",0,0,"from dataclasses import dataclass, field\nimport os\nimport time\n\nimport einops\nfrom flax.training import orbax_utils\nfrom flax.training.train_state import TrainState\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax\nfrom orbax.checkpoint import PyTreeCheckpointer\nimport numpy as np\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\n\nfrom genie import Genie, restore_genie_components\nfrom models.tokenizer import TokenizerVQVAE\nfrom models.lam import LatentActionModel\nfrom utils.dataloader import get_dataloader\n\nts = int(time.time())\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = ""data_tfrecords/coinrun""\n # Optimization\n batch_size: int = 36\n min_lr: float = 3e-6\n max_lr: float = 3e-5\n warmup_steps: int = 5000\n # Tokenizer\n tokenizer_dim: int = 512\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 8\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 8\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_dim: int = 512\n dyna_num_blocks: int = 12\n dyna_num_heads: int = 8\n dropout: float = 0.0\n mask_limit: float = 0.5\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_dynamics""\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_gradients: bool = False\n\n\nargs = tyro.cli(Args)\n\n\ndef dynamics_loss_fn(params, state, inputs):\n """"""Compute masked dynamics loss""""""\n outputs = state.apply_fn(\n params,\n inputs,\n training=True,\n rngs={""params"": inputs[""rng""], ""dropout"": inputs[""dropout_rng""]},\n )\n mask = outputs[""mask""]\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask * ce_loss).sum() / mask.sum()\n acc = outputs[""token_logits""].argmax(-1) == outputs[""video_tokens""]\n acc = (mask * acc).sum() / mask.sum()\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_accuracy=acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n )\n return ce_loss, (outputs[""recon""], metrics)\n\n\n@jax.jit\ndef train_step(state, inputs):\n """"""Update state and compute metrics""""""\n grad_fn = jax.value_and_grad(dynamics_loss_fn, has_aux=True, allow_int=True)\n (loss, (recon, metrics)), grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return state, loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.PRNGKey(args.seed)\n if args.log and jax.process_index() == 0:\n wandb.init(\n entity=args.entity,\n project=args.project,\n name=args.name,\n tags=args.tags,\n group=""debug"",\n config=args\n )\n\n # --- Initialize model ---\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n # Dynamics\n dyna_dim=args.dyna_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n mask_limit=args.mask_limit,\n )\n rng, _rng = jax.random.split(rng)\n image_shape = (args.image_height, args.image_width, args.image_channels)\n dummy_inputs = dict(\n videos=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len, *image_shape),\n dtype=jnp.float32,\n ),\n action=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len), dtype=jnp.float32\n ),\n mask_rng=_rng,\n )\n rng, _rng = jax.random.split(rng)\n init_params = genie.init(_rng, dummy_inputs)\n\n # --- Initialize optimizer ---\n lr_schedule = optax.warmup_cosine_decay_schedule(\n args.min_lr, args.max_lr, args.warmup_steps, args.num_steps\n )\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4)\n train_state = TrainState.create(apply_fn=genie.apply, params=init_params, tx=tx)\n\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n train_state = jax.device_put(train_state, replicated_sharding)\n\n # --- Restore checkpoint ---\n train_state = restore_genie_components(\n train_state, replicated_sharding, dummy_inputs, rng, args\n )\n\n # --- TRAIN LOOP ---\n tfrecord_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".tfrecord"")\n ]\n dataloader = get_dataloader(\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n tfrecord_files,\n args.seq_len,\n args.batch_size,\n *image_shape,\n )\n step = 0\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng, _rng_dropout, _rng_mask = jax.random.split(rng, 4)\n\n videos_sharding = NamedSharding(\n mesh, PartitionSpec(""data"", None, None, None, None)\n )\n videos = jax.make_array_from_process_local_data(videos_sharding, videos)\n\n inputs = dict(\n videos=videos,\n rng=_rng,\n dropout_rng=_rng_dropout,\n mask_rng=_rng_mask,\n )\n start_time = time.time()\n train_state, loss, recon, metrics = train_step(train_state, inputs)\n elapsed_time = (time.time() - start_time) * 1000\n print(f""Step {step}, loss: {loss}, step time: {elapsed_time}ms"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n ""step_time_ms"": elapsed_time,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0]\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[args.seq_len - 1])),\n recon=wandb.Image(np.asarray(recon_seq[args.seq_len - 1])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n if step % args.log_checkpoint_interval == 0:\n ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""genie_{ts}_{step}""),\n ckpt,\n save_args=save_args,\n )\n if step >= args.num_steps:\n break\n",python,tab
|
| 16 |
+
15,39797,"train_dynamics.py",5890,0,"",python,selection_command
|
| 17 |
+
16,40207,"train_dynamics.py",5044,0,"",python,selection_command
|
| 18 |
+
17,40597,"train_dynamics.py",4217,0,"",python,selection_command
|
| 19 |
+
18,40765,"train_dynamics.py",3598,0,"",python,selection_command
|
| 20 |
+
19,40946,"train_dynamics.py",2859,0,"",python,selection_command
|
| 21 |
+
20,41080,"train_dynamics.py",2006,0,"",python,selection_command
|
| 22 |
+
21,41229,"train_dynamics.py",1466,0,"",python,selection_command
|
| 23 |
+
22,41446,"train_dynamics.py",882,0,"",python,selection_command
|
| 24 |
+
23,41604,"train_dynamics.py",388,0,"",python,selection_command
|
| 25 |
+
24,41744,"train_dynamics.py",0,0,"",python,selection_command
|
| 26 |
+
25,43268,"train_dynamics.py",617,0,"",python,selection_command
|
| 27 |
+
26,46121,"train_dynamics.py",0,0,"",python,selection_command
|
| 28 |
+
27,46786,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab
|
| 29 |
+
28,49091,"train_dynamics.py",0,0,"",python,tab
|
| 30 |
+
29,49113,"TERMINAL",0,0,"",,terminal_focus
|
| 31 |
+
30,104571,"requirements.txt",0,0,"dm_pix>=0.4.3\neinops>=0.8.0\nflax>=0.8.5\njax[cuda12]>=0.4.30\noptax>=0.2.3\nprocgen>=0.10.7\ntyro>=0.8.5\nwandb>=0.17.4\ntensorflow>=2.1\npre-commit>=4.2.0",pip-requirements,tab
|
| 32 |
+
31,106305,"requirements.txt",73,0,"",pip-requirements,selection_command
|
| 33 |
+
32,106838,"requirements.txt",73,0,"#",pip-requirements,content
|
| 34 |
+
33,106844,"requirements.txt",74,0,"",pip-requirements,selection_keyboard
|
| 35 |
+
34,106906,"requirements.txt",74,0," ",pip-requirements,content
|
| 36 |
+
35,106910,"requirements.txt",75,0,"",pip-requirements,selection_keyboard
|
| 37 |
+
36,107095,"requirements.txt",74,0,"",pip-requirements,selection_command
|
| 38 |
+
37,114295,"requirements.txt",61,0,"",pip-requirements,selection_command
|
| 39 |
+
38,114428,"requirements.txt",41,0,"",pip-requirements,selection_command
|
| 40 |
+
39,114530,"requirements.txt",43,0,"",pip-requirements,selection_command
|
| 41 |
+
40,114983,"requirements.txt",43,1,"[",pip-requirements,selection_command
|
| 42 |
+
41,115082,"requirements.txt",43,7,"[cuda12",pip-requirements,selection_command
|
| 43 |
+
42,115265,"requirements.txt",43,10,"[cuda12]>=",pip-requirements,selection_command
|
| 44 |
+
43,115682,"requirements.txt",43,9,"[cuda12]>",pip-requirements,selection_command
|
| 45 |
+
44,115832,"requirements.txt",43,8,"[cuda12]",pip-requirements,selection_command
|
| 46 |
+
45,116046,"requirements.txt",43,8,"",pip-requirements,content
|
| 47 |
+
46,134920,"train_dynamics.py",0,0,"",python,tab
|
| 48 |
+
47,141347,"train_dynamics.py",41,0,"",python,selection_command
|
| 49 |
+
48,141596,"train_dynamics.py",51,0,"",python,selection_command
|
| 50 |
+
49,141626,"train_dynamics.py",63,0,"",python,selection_command
|
| 51 |
+
50,141657,"train_dynamics.py",64,0,"",python,selection_command
|
| 52 |
+
51,141691,"train_dynamics.py",78,0,"",python,selection_command
|
| 53 |
+
52,141894,"train_dynamics.py",116,0,"",python,selection_command
|
| 54 |
+
53,141964,"train_dynamics.py",121,0,"",python,selection_command
|
| 55 |
+
54,142163,"train_dynamics.py",125,0,"",python,selection_command
|
| 56 |
+
55,142229,"train_dynamics.py",87,0,"",python,selection_command
|
| 57 |
+
56,145246,"train_dynamics.py",653,0,"",python,selection_command
|
| 58 |
+
57,145597,"train_dynamics.py",1241,0,"",python,selection_command
|
| 59 |
+
58,148942,"train_dynamics.py",1839,0,"",python,selection_command
|
| 60 |
+
59,150414,"train_dynamics.py",2485,0,"",python,selection_command
|
| 61 |
+
60,150985,"train_dynamics.py",3341,0,"",python,selection_command
|
| 62 |
+
61,153046,"train_dynamics.py",2485,0,"",python,selection_command
|
| 63 |
+
62,153663,"train_dynamics.py",2527,0,"",python,selection_command
|
| 64 |
+
63,153914,"train_dynamics.py",2586,0,"",python,selection_command
|
| 65 |
+
64,153946,"train_dynamics.py",2606,0,"",python,selection_command
|
| 66 |
+
65,153981,"train_dynamics.py",2642,0,"",python,selection_command
|
| 67 |
+
66,154015,"train_dynamics.py",2677,0,"",python,selection_command
|
| 68 |
+
67,154046,"train_dynamics.py",2738,0,"",python,selection_command
|
| 69 |
+
68,154081,"train_dynamics.py",2784,0,"",python,selection_command
|
| 70 |
+
69,154113,"train_dynamics.py",2853,0,"",python,selection_command
|
| 71 |
+
70,154146,"train_dynamics.py",2859,0,"",python,selection_command
|
| 72 |
+
71,154180,"train_dynamics.py",2903,0,"",python,selection_command
|
| 73 |
+
72,154213,"train_dynamics.py",2904,0,"",python,selection_command
|
| 74 |
+
73,154246,"train_dynamics.py",2909,0,"",python,selection_command
|
| 75 |
+
74,154280,"train_dynamics.py",2918,0,"",python,selection_command
|
| 76 |
+
75,154508,"train_dynamics.py",3753,0,"",python,selection_command
|
| 77 |
+
76,155160,"train_dynamics.py",4363,0,"",python,selection_command
|
| 78 |
+
77,155948,"train_dynamics.py",5218,0,"",python,selection_command
|
| 79 |
+
78,158479,"train_dynamics.py",6074,0,"",python,selection_command
|
| 80 |
+
79,161795,"train_dynamics.py",6732,0,"",python,selection_command
|
| 81 |
+
80,166966,"train_dynamics.py",7582,0,"",python,selection_command
|
| 82 |
+
81,170485,"train_dynamics.py",8668,0,"",python,selection_command
|
| 83 |
+
82,175934,"train_dynamics.py",7582,0,"",python,selection_command
|
| 84 |
+
83,177106,"train_dynamics.py",7569,0,"",python,selection_command
|
| 85 |
+
84,177361,"train_dynamics.py",7559,0,"",python,selection_command
|
| 86 |
+
85,177390,"train_dynamics.py",7482,0,"",python,selection_command
|
| 87 |
+
86,177422,"train_dynamics.py",7421,0,"",python,selection_command
|
| 88 |
+
87,177457,"train_dynamics.py",7341,0,"",python,selection_command
|
| 89 |
+
88,177488,"train_dynamics.py",7304,0,"",python,selection_command
|
| 90 |
+
89,178494,"train_dynamics.py",7341,0,"",python,selection_command
|
| 91 |
+
90,178693,"train_dynamics.py",7421,0,"",python,selection_command
|
| 92 |
+
91,178849,"train_dynamics.py",7482,0,"",python,selection_command
|
| 93 |
+
92,179027,"train_dynamics.py",7559,0,"",python,selection_command
|
| 94 |
+
93,179210,"train_dynamics.py",7569,0,"",python,selection_command
|
| 95 |
+
94,179506,"train_dynamics.py",7559,0,"",python,selection_command
|
| 96 |
+
95,179674,"train_dynamics.py",7482,0,"",python,selection_command
|
| 97 |
+
96,179833,"train_dynamics.py",7421,0,"",python,selection_command
|
| 98 |
+
97,179926,"train_dynamics.py",7482,0,"",python,selection_command
|
| 99 |
+
98,180126,"train_dynamics.py",7421,0,"",python,selection_command
|
| 100 |
+
99,180497,"train_dynamics.py",7482,0,"",python,selection_command
|
| 101 |
+
100,181016,"train_dynamics.py",7559,0,"",python,selection_command
|
| 102 |
+
101,181259,"train_dynamics.py",7482,0,"",python,selection_command
|
| 103 |
+
102,183280,"train_dynamics.py",7421,0,"",python,selection_command
|
| 104 |
+
103,183526,"train_dynamics.py",7341,0,"",python,selection_command
|
| 105 |
+
104,185378,"train_dynamics.py",7407,0,"",python,selection_command
|
| 106 |
+
105,185754,"train_dynamics.py",7401,0,"",python,selection_command
|
| 107 |
+
106,185894,"train_dynamics.py",7399,0,"",python,selection_command
|
| 108 |
+
107,186075,"train_dynamics.py",7388,0,"",python,selection_command
|
| 109 |
+
108,186198,"train_dynamics.py",7387,0,"",python,selection_command
|
| 110 |
+
109,186477,"train_dynamics.py",7377,0,"",python,selection_command
|
| 111 |
+
110,212701,"train_dynamics.py",2949,0,"",python,selection_command
|
| 112 |
+
111,212882,"train_dynamics.py",2992,0,"",python,selection_command
|
| 113 |
+
112,212978,"train_dynamics.py",3000,0,"",python,selection_command
|
| 114 |
+
113,213228,"train_dynamics.py",3002,0,"",python,selection_command
|
| 115 |
+
114,213261,"train_dynamics.py",3005,0,"",python,selection_command
|
| 116 |
+
115,213294,"train_dynamics.py",3006,0,"",python,selection_command
|
| 117 |
+
116,213327,"train_dynamics.py",3020,0,"",python,selection_command
|
| 118 |
+
117,213590,"train_dynamics.py",3021,0,"",python,selection_command
|
| 119 |
+
118,214182,"train_dynamics.py",2988,0,"",python,selection_command
|
| 120 |
+
119,215003,"train_dynamics.py",2992,0,"",python,selection_command
|
| 121 |
+
120,215250,"train_dynamics.py",3000,0,"",python,selection_command
|
| 122 |
+
121,215283,"train_dynamics.py",3002,0,"",python,selection_command
|
| 123 |
+
122,215321,"train_dynamics.py",3005,0,"",python,selection_command
|
| 124 |
+
123,215353,"train_dynamics.py",3006,0,"",python,selection_command
|
| 125 |
+
124,215386,"train_dynamics.py",3020,0,"",python,selection_command
|
| 126 |
+
125,215420,"train_dynamics.py",3021,0,"",python,selection_command
|
| 127 |
+
126,221808,"train_dynamics.py",2988,0,"",python,selection_command
|
| 128 |
+
127,221880,"train_dynamics.py",2992,0,"",python,selection_command
|
| 129 |
+
128,222128,"train_dynamics.py",3000,0,"",python,selection_command
|
| 130 |
+
129,222161,"train_dynamics.py",3002,0,"",python,selection_command
|
| 131 |
+
130,222190,"train_dynamics.py",3005,0,"",python,selection_command
|
| 132 |
+
131,222224,"train_dynamics.py",3006,0,"",python,selection_command
|
| 133 |
+
132,222259,"train_dynamics.py",3020,0,"",python,selection_command
|
| 134 |
+
133,222293,"train_dynamics.py",3021,0,"",python,selection_command
|
| 135 |
+
134,223174,"train_dynamics.py",2988,0,"",python,selection_command
|
| 136 |
+
135,223299,"train_dynamics.py",2992,0,"",python,selection_command
|
| 137 |
+
136,223551,"train_dynamics.py",3000,0,"",python,selection_command
|
| 138 |
+
137,223581,"train_dynamics.py",3002,0,"",python,selection_command
|
| 139 |
+
138,223614,"train_dynamics.py",3005,0,"",python,selection_command
|
| 140 |
+
139,223648,"train_dynamics.py",3006,0,"",python,selection_command
|
| 141 |
+
140,223681,"train_dynamics.py",3020,0,"",python,selection_command
|
| 142 |
+
141,223715,"train_dynamics.py",3021,0,"",python,selection_command
|
| 143 |
+
142,223749,"train_dynamics.py",3037,0,"",python,selection_command
|
| 144 |
+
143,224380,"train_dynamics.py",3021,0,"",python,selection_command
|
| 145 |
+
144,225595,"train_dynamics.py",2988,0,"",python,selection_command
|
| 146 |
+
145,225681,"train_dynamics.py",2992,0,"",python,selection_command
|
| 147 |
+
146,225938,"train_dynamics.py",3000,0,"",python,selection_command
|
| 148 |
+
147,225966,"train_dynamics.py",3002,0,"",python,selection_command
|
| 149 |
+
148,225999,"train_dynamics.py",3005,0,"",python,selection_command
|
| 150 |
+
149,226037,"train_dynamics.py",3006,0,"",python,selection_command
|
| 151 |
+
150,226070,"train_dynamics.py",3020,0,"",python,selection_command
|
| 152 |
+
151,226103,"train_dynamics.py",3021,0,"",python,selection_command
|
| 153 |
+
152,251190,"train_dynamics.py",2006,0,"",python,selection_command
|
| 154 |
+
153,251439,"train_dynamics.py",2045,0,"",python,selection_command
|
| 155 |
+
154,251470,"train_dynamics.py",2075,0,"",python,selection_command
|
| 156 |
+
155,251506,"train_dynamics.py",2091,0,"",python,selection_command
|
| 157 |
+
156,251539,"train_dynamics.py",2107,0,"",python,selection_command
|
| 158 |
+
157,251574,"train_dynamics.py",2130,0,"",python,selection_command
|
| 159 |
+
158,251760,"train_dynamics.py",2204,0,"",python,selection_command
|
| 160 |
+
159,252115,"train_dynamics.py",2210,0,"",python,selection_command
|
| 161 |
+
160,256408,"train_dynamics.py",2215,0,"",python,selection_command
|
| 162 |
+
161,256583,"train_dynamics.py",2217,0,"",python,selection_command
|
| 163 |
+
162,262598,"models/dynamics.py",0,0,"",python,tab
|
| 164 |
+
163,263846,"models/dynamics.py",416,0,"",python,selection_keyboard
|
| 165 |
+
164,271113,"models/dynamics.py",1359,0,"",python,selection_command
|
| 166 |
+
165,273515,"models/dynamics.py",1373,0,"",python,selection_command
|
| 167 |
+
166,273673,"models/dynamics.py",1389,0,"",python,selection_command
|
4de8d861ed2563988d5f1871647ebc5fe70861b32d24a4b32f9363518653a328/crowd-code-ef5ea013-ac2b-459c-8783-a7b025d58a391754900011518-2025_08_11-10.13.34.249/source.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
507ab0ec0dfe0c18ad7778dd15e072f92367194c94623114de802c8ed9c52e20/crowd-code-b9559366-0d71-4ceb-9b37-1d3a0cf03cd61750867779082-2025_06_25-18.09.57.465/source.csv
ADDED
|
@@ -0,0 +1,64 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type
|
| 2 |
+
1,6,"scripts/rename_mp4_files.py",0,0,"#!/usr/bin/env python3\n""""""\nScript to rename all MP4 files in a directory with a custom prefix.\n""""""\n\nimport os\nimport argparse\nimport re\nfrom pathlib import Path\n\n\ndef rename_mp4_files(directory, prefix, dry_run=False, pattern=None, start_number=1):\n """"""\n Rename all MP4 files in the specified directory with a custom prefix.\n \n Args:\n directory (str): Path to the directory containing MP4 files\n prefix (str): Custom prefix to add to filenames\n dry_run (bool): If True, only show what would be renamed without actually renaming\n pattern (str): Optional regex pattern to filter files\n start_number (int): Starting number for sequential naming\n """"""\n directory_path = Path(directory)\n \n if not directory_path.exists():\n print(f""Error: Directory '{directory}' does not exist."")\n return\n \n if not directory_path.is_dir():\n print(f""Error: '{directory}' is not a directory."")\n return\n \n # Find all MP4 files\n mp4_files = list(directory_path.glob(""*.mp4""))\n \n if not mp4_files:\n print(f""No MP4 files found in '{directory}'"")\n return\n \n # Filter by pattern if provided\n if pattern:\n regex = re.compile(pattern)\n mp4_files = [f for f in mp4_files if regex.search(f.name)]\n \n if not mp4_files:\n print(f""No MP4 files match the pattern '{pattern}' in '{directory}'"")\n return\n \n print(f""Found {len(mp4_files)} MP4 files to rename:"")\n \n # Sort files for consistent ordering\n mp4_files.sort()\n \n for i, file_path in enumerate(mp4_files, start=start_number):\n # Get the original filename without extension\n original_name = file_path.stem\n extension = file_path.suffix\n \n # Create new filename with prefix\n new_name = f""{prefix}_{i:03d}_{original_name}{extension}""\n new_path = file_path.parent / new_name\n \n # Check if new filename already exists\n if new_path.exists() and not dry_run:\n print(f""Warning: '{new_name}' already exists, skipping '{file_path.name}'"")\n continue\n \n if dry_run:\n print(f""Would rename: '{file_path.name}' -> '{new_name}'"")\n else:\n try:\n file_path.rename(new_path)\n print(f""Renamed: '{file_path.name}' -> '{new_name}'"")\n except OSError as e:\n print(f""Error renaming '{file_path.name}': {e}"")\n \n if dry_run:\n print(f""\nDry run completed. {len(mp4_files)} files would be renamed."")\n else:\n print(f""\nRenaming completed. {len(mp4_files)} files renamed."")\n\n\ndef main():\n parser = argparse.ArgumentParser(\n description=""Rename all MP4 files in a directory with a custom prefix"",\n formatter_class=argparse.RawDescriptionHelpFormatter,\n epilog=""""""\nExamples:\n %(prog)s /path/to/videos minecraft\n %(prog)s /path/to/videos player --dry-run\n %(prog)s /path/to/videos episode --pattern ""joost"" --start-number 10\n """"""\n )\n \n parser.add_argument(\n ""directory"",\n help=""Directory containing MP4 files to rename""\n )\n \n parser.add_argument(\n ""prefix"",\n help=""Custom prefix to add to filenames""\n )\n \n parser.add_argument(\n ""--dry-run"",\n action=""store_true"",\n help=""Show what would be renamed without actually renaming files""\n )\n \n parser.add_argument(\n ""--pattern"",\n help=""Regex pattern to filter files (only rename files matching this pattern)""\n )\n \n parser.add_argument(\n ""--start-number"",\n type=int,\n default=1,\n help=""Starting number for sequential naming (default: 1)""\n )\n \n args = parser.parse_args()\n \n # Validate start number\n if args.start_number < 1:\n print(""Error: start-number must be at least 1"")\n return\n \n rename_mp4_files(\n directory=args.directory,\n prefix=args.prefix,\n dry_run=args.dry_run,\n pattern=args.pattern,\n start_number=args.start_number\n )\n\n\nif __name__ == ""__main__"":\n main() ",python,tab
|
| 3 |
+
2,319,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"6:09:57 PM [info] Activating crowd-code\n6:09:57 PM [info] Welcome back tum_ind3695. Your user-id is '507ab0ec0dfe0c18ad7778dd15e072f92367194c94623114de802c8ed9c52e20'. Happy coding!\n6:09:57 PM [info] Recording started\n",Log,tab
|
| 4 |
+
3,2485,"scripts/rename_mp4_files.py",0,0,"",python,tab
|
| 5 |
+
4,2488,"scripts/rename_mp4_files.py",1333,0,"",python,selection_mouse
|
| 6 |
+
5,2493,"scripts/rename_mp4_files.py",1332,0,"",python,selection_command
|
| 7 |
+
6,4622,"TERMINAL",0,0,"",,terminal_focus
|
| 8 |
+
7,301525,"TERMINAL",0,0,"",,terminal_focus
|
| 9 |
+
8,419551,"TERMINAL",0,0,"cd /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/dyn/3289577",,terminal_command
|
| 10 |
+
9,419570,"TERMINAL",0,0,"]633;E;2025-06-25 18:16:56 cd /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/dyn/3289577;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577]633;D;0",,terminal_output
|
| 11 |
+
10,420075,"TERMINAL",0,0,"ls",,terminal_command
|
| 12 |
+
11,420099,"TERMINAL",0,0,"]633;E;2025-06-25 18:16:57 ls;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C[0m[01;34mgenie_1750786631_1000[0m [01;34mgenie_1750786631_11000[0m [01;34mgenie_1750786631_12500[0m [01;34mgenie_1750786631_2500[0m [01;34mgenie_1750786631_4000[0m [01;34mgenie_1750786631_5000[0m [01;34mgenie_1750786631_6500[0m [01;34mgenie_1750786631_8000[0m [01;34mgenie_1750786631_9500[0m\r\n[01;34mgenie_1750786631_10000[0m [01;34mgenie_1750786631_11500[0m [01;34mgenie_1750786631_1500[0m [01;34mgenie_1750786631_3000[0m [01;34mgenie_1750786631_4500[0m [01;34mgenie_1750786631_5500[0m [01;34mgenie_1750786631_7000[0m [01;34mgenie_1750786631_8500[0m\r\n[01;34mgenie_1750786631_10500[0m [01;34mgenie_1750786631_12000[0m [01;34mgenie_1750786631_2000[0m [01;34mgenie_1750786631_3500[0m [01;34mgenie_1750786631_500[0m [01;34mgenie_1750786631_6000[0m [01;34mgenie_1750786631_7500[0m [01;34mgenie_1750786631_9000[0m\r\n]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577]633;D;0",,terminal_output
|
| 13 |
+
12,439480,"TERMINAL",0,0,"cd genie_1750786631_12500",,terminal_command
|
| 14 |
+
13,440497,"TERMINAL",0,0,"pwd",,terminal_command
|
| 15 |
+
14,440512,"TERMINAL",0,0,"]633;E;2025-06-25 18:17:17 pwd;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577/genie_1750786631_12500\r\n]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577/genie_1750786631_12500]633;D;0",,terminal_output
|
| 16 |
+
15,615482,"TERMINAL",0,0,"cd ../../..",,terminal_command
|
| 17 |
+
16,615494,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:12 cd ../../..;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints]633;D;0",,terminal_output
|
| 18 |
+
17,616076,"TERMINAL",0,0,"ls",,terminal_command
|
| 19 |
+
18,616111,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:13 ls;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C[0m[01;34m0000[0m [01;34m3290283[0m [01;34m3290284[0m [01;34m3290295[0m [01;34m3290296[0m [01;34m3290366[0m [01;34m3290367[0m [01;34m3290391[0m [01;34m3290392[0m [01;34m3290439[0m [01;34m3290440[0m [01;34m3291405[0m [01;34mdyn[0m [01;34mlam[0m [01;34mtokenizer[0m\r\n]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints]633;D;0",,terminal_output
|
| 20 |
+
19,619493,"TERMINAL",0,0,"cd tokenizer/",,terminal_command
|
| 21 |
+
20,619506,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:16 cd tokenizer/;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer]633;D;0",,terminal_output
|
| 22 |
+
21,619896,"TERMINAL",0,0,"ls",,terminal_command
|
| 23 |
+
22,619948,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:17 ls;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C",,terminal_output
|
| 24 |
+
23,619997,"TERMINAL",0,0,"[0m[01;34m3255052[0m [01;34m3255483[0m [01;34m3256924[0m [01;34m3257467[0m [01;34m3258278[0m [01;34m3259422[0m [01;34m3261720[0m [01;34m3273042[0m [01;34m3273290[0m [01;34m3273480[0m [01;34m3273494[0m [01;34m3273627[0m [01;34m3273742[0m [01;34m3273795[0m [01;34m3273824[0m [01;34m3273830[0m [01;34m3273846[0m [01;34m3276030[0m [01;34m3276048[0m [01;34m3276053[0m [01;34m3285798[0m\r\n[01;34m3255466[0m [01;34m3256873[0m [01;34m3256926[0m [01;34m3257633[0m [01;34m3258283[0m [01;34m3260527[0m [01;34m3261722[0m [01;34m3273174[0m [01;34m3273348[0m [01;34m3273488[0m [01;34m3273496[0m [01;34m3273681[0m [01;34m3273743[0m [01;34m3273816[0m [01;34m3273828[0m [01;34m3273831[0m [01;34m3275950[0m [01;34m3276039[0m [01;34m3276051[0m [01;34m3276058[0m [01;34m3285811[0m\r\n[01;34m3255482[0m [01;34m3256921[0m [01;34m3256929[0m [01;34m3257812[0m [01;34m3259405[0m [01;34m3260932[0m [01;34m3273026[0m [01;34m3273229[0m [01;34m3273476[0m [01;34m3273489[0m [01;34m3273503[0m [01;34m3273687[0m [01;34m3273746[0m [01;34m3273820[0m [01;34m3273829[0m [01;34m3273841[0m [01;34m3275991[0m [01;34m3276043[0m [01;34m3276052[0m [01;34m3285784[0m [01;34m3286114[0m\r\n]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer]633;D;0",,terminal_output
|
| 25 |
+
24,649663,"TERMINAL",0,0,"cd 3255052/",,terminal_command
|
| 26 |
+
25,649679,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:47 cd 3255052/;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/3255052]633;D;0",,terminal_output
|
| 27 |
+
26,650176,"TERMINAL",0,0,"ls",,terminal_command
|
| 28 |
+
27,650194,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:47 ls;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/3255052]633;D;0",,terminal_output
|
| 29 |
+
28,652601,"TERMINAL",0,0,"cd ..",,terminal_command
|
| 30 |
+
29,652609,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:50 cd ..;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer]633;D;0",,terminal_output
|
| 31 |
+
30,654392,"TERMINAL",0,0,"cd ..",,terminal_command
|
| 32 |
+
31,654398,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:51 cd ..;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints]633;D;0",,terminal_output
|
| 33 |
+
32,655918,"TERMINAL",0,0,"cd ..",,terminal_command
|
| 34 |
+
33,655926,"TERMINAL",0,0,"]633;E;2025-06-25 18:20:53 cd ..;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared]633;D;0",,terminal_output
|
| 35 |
+
34,1148074,"TERMINAL",0,0,"queue",,terminal_command
|
| 36 |
+
35,1148124,"TERMINAL",0,0,"]633;E;2025-06-25 18:29:05 queue;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C",,terminal_output
|
| 37 |
+
36,1148182,"TERMINAL",0,0,"[?1049h[22;0;0t[1;106r(B[m[4l[?7h[H[2JEvery 1.0s: squeue --me[1;173Hhkn1993.localdomain: Wed Jun 25 18:29:05 2025[3;14HJOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)[4;12H3292119 accelerat train_to tum_ind3 PD\t0:00\t 1 (Priority)[5;12H3289577 accelerat train_dy tum_ind3 R 22:52:29\t 1 hkn0712[106;217H",,terminal_output
|
| 38 |
+
37,1149242,"TERMINAL",0,0,"[1;212H6[5;59H30[106;217H",,terminal_output
|
| 39 |
+
38,1149475,"TERMINAL",0,0,"[106;1H[?1049l[23;0;0t\r[?1l>]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared]633;D;0",,terminal_output
|
| 40 |
+
39,1150350,"TERMINAL",0,0,"queue",,terminal_command
|
| 41 |
+
40,1150402,"TERMINAL",0,0,"]633;E;2025-06-25 18:29:07 queue;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C",,terminal_output
|
| 42 |
+
41,1150481,"TERMINAL",0,0,"[?1049h[22;0;0t[1;106r(B[m[4l[?7h[H[2JEvery 1.0s: squeue --me[1;173Hhkn1993.localdomain: Wed Jun 25 18:29:07 2025[3;14HJOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)[4;12H3292119 accelerat train_to tum_ind3 PD\t0:00\t 1 (Priority)[5;12H3289577 accelerat train_dy tum_ind3 R 22:52:31\t 1 hkn0712[106;217H",,terminal_output
|
| 43 |
+
42,1151212,"TERMINAL",0,0,"[106;1H[?1049l[23;0;0t\r[?1l>]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared]633;D;0",,terminal_output
|
| 44 |
+
43,1152587,"TERMINAL",0,0,"squeue",,terminal_command
|
| 45 |
+
44,1152616,"TERMINAL",0,0,"]633;E;2025-06-25 18:29:10 squeue;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C JOBID PARTITION NAME USER ST TIME NODES NODELIST(REASON)\r\n 3292119 accelerat train_to tum_ind3 PD 0:00 1 (Priority)\r\n 3289577 accelerat train_dy tum_ind3 R 22:52:34 1 hkn0712\r\n]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared]633;D;0]633;P;Cwd=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared",,terminal_output
|
| 46 |
+
45,2263248,"TERMINAL",0,0,"cd /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/dyn/3289577",,terminal_command
|
| 47 |
+
46,2263255,"TERMINAL",0,0,"]633;E;2025-06-25 18:47:40 cd /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/dyn/3289577;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577]633;D;0",,terminal_output
|
| 48 |
+
47,2264288,"TERMINAL",0,0,"ls",,terminal_command
|
| 49 |
+
48,2264341,"TERMINAL",0,0,"]633;E;2025-06-25 18:47:41 ls;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C",,terminal_output
|
| 50 |
+
49,2264401,"TERMINAL",0,0,"[0m[01;34mgenie_1750786631_1000[0m [01;34mgenie_1750786631_11000[0m [01;34mgenie_1750786631_12500[0m [01;34mgenie_1750786631_2000[0m [01;34mgenie_1750786631_3500[0m [01;34mgenie_1750786631_500[0m [01;34mgenie_1750786631_6000[0m [01;34mgenie_1750786631_7500[0m [01;34mgenie_1750786631_9000[0m\r\n[01;34mgenie_1750786631_10000[0m [01;34mgenie_1750786631_11500[0m [01;34mgenie_1750786631_13000[0m [01;34mgenie_1750786631_2500[0m [01;34mgenie_1750786631_4000[0m [01;34mgenie_1750786631_5000[0m [01;34mgenie_1750786631_6500[0m [01;34mgenie_1750786631_8000[0m [01;34mgenie_1750786631_9500[0m\r\n[01;34mgenie_1750786631_10500[0m [01;34mgenie_1750786631_12000[0m [01;34mgenie_1750786631_1500[0m [01;34mgenie_1750786631_3000[0m [01;34mgenie_1750786631_4500[0m [01;34mgenie_1750786631_5500[0m [01;34mgenie_1750786631_7000[0m [01;34mgenie_1750786631_8500[0m\r\n]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577]633;D;0",,terminal_output
|
| 51 |
+
50,2269289,"TERMINAL",0,0,"cd genie_1750786631_13000",,terminal_command
|
| 52 |
+
51,2269296,"TERMINAL",0,0,"]633;E;2025-06-25 18:47:46 cd genie_1750786631_13000;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577/genie_1750786631_13000]633;D;0",,terminal_output
|
| 53 |
+
52,2270468,"TERMINAL",0,0,"pwd",,terminal_command
|
| 54 |
+
53,2270477,"TERMINAL",0,0,"]633;E;2025-06-25 18:47:47 pwd;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577/genie_1750786631_13000\r\n]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577/genie_1750786631_13000]633;D;0",,terminal_output
|
| 55 |
+
54,2272125,"TERMINAL",0,0,"os",,terminal_command
|
| 56 |
+
55,2272199,"TERMINAL",0,0,"]633;E;2025-06-25 18:47:49 os;ef772368-fdbb-4ff2-ad99-2db92190992c]633;Cbash: os: command not found...\r\n",,terminal_output
|
| 57 |
+
56,2274189,"TERMINAL",0,0,"ls",,terminal_command
|
| 58 |
+
57,2274240,"TERMINAL",0,0,"]633;E;2025-06-25 18:47:51 ls;ef772368-fdbb-4ff2-ad99-2db92190992c]633;C",,terminal_output
|
| 59 |
+
58,2274298,"TERMINAL",0,0,"[0m[01;34marray_metadatas[0m _CHECKPOINT_METADATA [01;34md[0m manifest.ocdbt _METADATA [01;34mocdbt.process_0[0m [01;34mocdbt.process_1[0m [01;34mocdbt.process_2[0m [01;34mocdbt.process_3[0m _sharding\r\n]0;tum_ind3695@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/dyn/3289577/genie_1750786631_13000]633;D;0",,terminal_output
|
| 60 |
+
59,2708521,"scripts/file_duplicate_checker.py",0,0,"import os\nfrom collections import defaultdict\nfrom tqdm import tqdm\n\ndef find_duplicate_filenames(root_dir):\n filenames = defaultdict(list)\n file_count = 0\n\n # Use tqdm with manual update and no percentage/ETA bar\n pbar = tqdm(desc=""Files scanned"", unit=""file"", dynamic_ncols=True, bar_format=""{desc}: {n_fmt}"")\n\n # Walk the directory recursively\n for dirpath, _, files in os.walk(root_dir):\n for file in files:\n full_path = os.path.join(dirpath, file)\n if os.path.isfile(full_path):\n filenames[file].append(full_path)\n file_count += 1\n pbar.update(1)\n\n pbar.close()\n\n # Print duplicates\n duplicates = {name: paths for name, paths in filenames.items() if len(paths) > 1}\n if duplicates:\n print(""\nDuplicate filenames found:\n"")\n for name, paths in duplicates.items():\n print(f""Filename: {name}"")\n for path in paths:\n print(f"" - {path}"")\n print()\n else:\n print(""\nNo duplicate filenames found."")\n\nif __name__ == ""__main__"":\n import sys\n if len(sys.argv) < 2:\n print(""Usage: python find_duplicates.py <directory_path>"")\n else:\n find_duplicate_filenames(sys.argv[1])\n\n",python,tab
|
| 61 |
+
60,2729693,"scripts/file_duplicate_checker.py",991,0,"",python,selection_mouse
|
| 62 |
+
61,2729697,"scripts/file_duplicate_checker.py",990,0,"",python,selection_command
|
| 63 |
+
62,2729708,"scripts/file_duplicate_checker.py",990,1,")",python,selection_mouse
|
| 64 |
+
63,2729710,"scripts/file_duplicate_checker.py",991,0,"",python,selection_command
|