hydra: run: dir: ${output_dir} output_subdir: ${output_dir}/code/hydra # Store hydra's config breakdown here for debugging searchpath: # Only in these paths are discoverable - pkg://navsim.planning.script.config.common job: chdir: False defaults: - default_common - default_evaluation - default_train_val_test_log_split - agent: ego_status_mlp_agent - _self_ split: trainval cache_path: ${oc.env:NAVSIM_EXP_ROOT}/training_cache use_cache_without_dataset: false # load the training samples from the cache. scene-filter will be ignored force_cache_computation: true seed: 0 dataloader: params: batch_size: 64 # number of samples per batch num_workers: 4 # number of workers for data loading pin_memory: true # pin memory for faster GPU transfer prefetch_factor: 2 # number of samples loaded in advance by each worker trainer: params: max_epochs: 100 # maximum number of training epochs check_val_every_n_epoch: 1 # run validation set every n training epochs val_check_interval: 1.0 # [%] run validation set every X% of training set limit_train_batches: 1.0 # how much of training dataset to check (float = fraction, int = num_batches) limit_val_batches: 1.0 # how much of validation dataset to check (float = fraction, int = num_batches) accelerator: gpu # distribution method strategy: ddp precision: 16-mixed # floating point precision num_nodes: 1 # Number of nodes used for training num_sanity_val_steps: 0 # number of validation steps to run before training begins fast_dev_run: false # runs 1 batch of train/val/test for sanity accumulate_grad_batches: 1 # accumulates gradients every n batches # track_grad_norm: -1 # logs the p-norm for inspection gradient_clip_val: 0.0 # value to clip gradients gradient_clip_algorithm: norm # [value, norm] method to clip gradients default_root_dir: ${output_dir}