Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes. See raw diff
- .gitattributes +1 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/.hydra/config.yaml +115 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/.hydra/hydra.yaml +156 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/.hydra/overrides.yaml +1 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/config.json +42 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/generation_config.json +7 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/merges.txt +0 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/mlp_projector.bin +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/model.safetensors +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/optimizer.pt +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/rng_state.pth +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/scheduler.pt +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/special_tokens_map.json +34 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/tokenizer.json +0 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/tokenizer_config.json +155 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/trainer_state.json +0 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/training_args.bin +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/vocab.json +0 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/config.json +42 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/generation_config.json +7 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/merges.txt +0 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/mlp_projector.bin +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/model.safetensors +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/optimizer.pt +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/rng_state.pth +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/scheduler.pt +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/special_tokens_map.json +34 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/tokenizer.json +0 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/tokenizer_config.json +155 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/trainer_state.json +0 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/training_args.bin +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/vocab.json +0 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/config.json +42 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/generation_config.json +7 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/merges.txt +0 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/mlp_projector.bin +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/model.safetensors +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/optimizer.pt +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/rng_state.pth +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/scheduler.pt +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/special_tokens_map.json +34 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/tokenizer.json +0 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/tokenizer_config.json +155 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/trainer_state.json +0 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/training_args.bin +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/vocab.json +0 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/config.json +42 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/generation_config.json +7 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/model.safetensors +3 -0
- 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/normalizer.pt +3 -0
.gitattributes
CHANGED
|
@@ -39,3 +39,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 39 |
2025.10.01/22.05.45_train_llm_lowdim_PegInsertionSide-v1/wandb/run-20251001_220546-8mqefeab/run-8mqefeab.wandb filter=lfs diff=lfs merge=lfs -text
|
| 40 |
2025.10.01/22.07.23_train_llm_lowdim_blockpush_lowdim_seed/wandb/run-20251001_220724-uktwy92f/run-uktwy92f.wandb filter=lfs diff=lfs merge=lfs -text
|
| 41 |
2025.10.03/09.42.26_train_llm_lowdim_push-back-v2/wandb/run-20251003_094227-po9rvkh2/run-po9rvkh2.wandb filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 39 |
2025.10.01/22.05.45_train_llm_lowdim_PegInsertionSide-v1/wandb/run-20251001_220546-8mqefeab/run-8mqefeab.wandb filter=lfs diff=lfs merge=lfs -text
|
| 40 |
2025.10.01/22.07.23_train_llm_lowdim_blockpush_lowdim_seed/wandb/run-20251001_220724-uktwy92f/run-uktwy92f.wandb filter=lfs diff=lfs merge=lfs -text
|
| 41 |
2025.10.03/09.42.26_train_llm_lowdim_push-back-v2/wandb/run-20251003_094227-po9rvkh2/run-po9rvkh2.wandb filter=lfs diff=lfs merge=lfs -text
|
| 42 |
+
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/wandb/run-20251016_115323-y9v0s119/run-y9v0s119.wandb filter=lfs diff=lfs merge=lfs -text
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/.hydra/config.yaml
ADDED
|
@@ -0,0 +1,115 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
name: train_llm_lowdim
|
| 2 |
+
_target_: llmbc.workspace.train_llm_workspace.TrainLLMWorkspace
|
| 3 |
+
obs_dim: ${task.obs_dim}
|
| 4 |
+
action_dim: ${task.action_dim}
|
| 5 |
+
horizon: 1
|
| 6 |
+
n_obs_steps: 1
|
| 7 |
+
n_action_steps: 1
|
| 8 |
+
task_name: ${task.name}
|
| 9 |
+
exp_name: train llm
|
| 10 |
+
model_name: ${llm.name}
|
| 11 |
+
use_quantization: ${llm.use_quantization}
|
| 12 |
+
lora_config: ${llm.lora_config}
|
| 13 |
+
dataset:
|
| 14 |
+
test_data_ratio: 0.01
|
| 15 |
+
debug: false
|
| 16 |
+
training:
|
| 17 |
+
seed: 42
|
| 18 |
+
per_device_train_batch_size: 128
|
| 19 |
+
per_device_eval_batch_size: 128
|
| 20 |
+
gradient_accumulation_steps: 1
|
| 21 |
+
optim: paged_adamw_32bit
|
| 22 |
+
num_train_epochs: 10
|
| 23 |
+
eval_strategy: steps
|
| 24 |
+
logging_steps: 1
|
| 25 |
+
warmup_steps: 10
|
| 26 |
+
logging_strategy: steps
|
| 27 |
+
learning_rate: 1.0e-05
|
| 28 |
+
fp16: false
|
| 29 |
+
bf16: true
|
| 30 |
+
tf32: true
|
| 31 |
+
group_by_length: true
|
| 32 |
+
report_to: wandb
|
| 33 |
+
save_steps: 5000
|
| 34 |
+
eval_steps: 10
|
| 35 |
+
use_joint_mlp_projector: ${llm.use_joint_mlp_projector}
|
| 36 |
+
joint_obs_action_mlp_lr: 1.0e-05
|
| 37 |
+
trainer:
|
| 38 |
+
obs_dim: ${obs_dim}
|
| 39 |
+
action_dim: ${action_dim}
|
| 40 |
+
use_joint_mlp_projector: ${llm.use_joint_mlp_projector}
|
| 41 |
+
max_seq_length: ${llm.max_length}
|
| 42 |
+
dataset_text_field: text
|
| 43 |
+
packing: false
|
| 44 |
+
logging:
|
| 45 |
+
project: llm_module_finetuning
|
| 46 |
+
resume: true
|
| 47 |
+
mode: online
|
| 48 |
+
name: ${now:%Y.%m.%d-%H.%M.%S}_${name}_${task_name}
|
| 49 |
+
tags:
|
| 50 |
+
- ${name}
|
| 51 |
+
- ${task_name}
|
| 52 |
+
- ${exp_name}
|
| 53 |
+
id: null
|
| 54 |
+
group: null
|
| 55 |
+
multi_run:
|
| 56 |
+
run_dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
|
| 57 |
+
wandb_name_base: ${now:%Y.%m.%d-%H.%M.%S}_${name}_${task_name}
|
| 58 |
+
task:
|
| 59 |
+
name: maze2d-medium-v0
|
| 60 |
+
obs_dim: 6
|
| 61 |
+
action_dim: 2
|
| 62 |
+
env_runner:
|
| 63 |
+
_target_: llmbc.env_runner.pointmaze_lowdim_runner.PointmazeLowdimRunner
|
| 64 |
+
env_name: llf-pointmaze-maze2d-medium-v0
|
| 65 |
+
n_train: 10
|
| 66 |
+
n_test: 50
|
| 67 |
+
n_envs: 10
|
| 68 |
+
max_steps: 250
|
| 69 |
+
n_obs_steps: ${n_obs_steps}
|
| 70 |
+
n_action_steps: ${n_action_steps}
|
| 71 |
+
instruction_type: b
|
| 72 |
+
feedback_type:
|
| 73 |
+
- hp
|
| 74 |
+
- hn
|
| 75 |
+
- fp
|
| 76 |
+
visual: false
|
| 77 |
+
discount: 0.99
|
| 78 |
+
dataset:
|
| 79 |
+
_target_: llmbc.dataset.pointmaze_lowdim_dataset.PointmazeLowdimDataset
|
| 80 |
+
data_path: datasets/maze2d-medium-v0-general.pt
|
| 81 |
+
data_path2: datasets/maze2d-medium-v0-general.pt
|
| 82 |
+
horizon: ${horizon}
|
| 83 |
+
pad_before: ${eval:'${n_obs_steps}-1'}
|
| 84 |
+
pad_after: ${eval:'${n_action_steps}-1'}
|
| 85 |
+
obs_eef_target: true
|
| 86 |
+
use_manual_normalizer: false
|
| 87 |
+
val_ratio: 0.02
|
| 88 |
+
dummy_normalizer: true
|
| 89 |
+
instructor:
|
| 90 |
+
_target_: llmbc.translator.instructor.pointmaze_instructor.maze2d_medium_v0_instructor.Maze2dMediumV0Instructor
|
| 91 |
+
llm:
|
| 92 |
+
name: HuggingFaceTB/SmolLM2-135M-Instruct
|
| 93 |
+
model_name: SmolLM2-135M-Instruct
|
| 94 |
+
config_target: llmbc.model.llm.llama_lowdim_model.LowdimLlamaConfig
|
| 95 |
+
causal_lm_target: llmbc.model.llm.llama_lowdim_model.LowdimLlamaForCausalLM
|
| 96 |
+
use_quantization: false
|
| 97 |
+
use_joint_mlp_projector: true
|
| 98 |
+
llm_mode: mlp-finetuned
|
| 99 |
+
finetune_mode: orig
|
| 100 |
+
checkpoint: data/outputs/2025.10.14/13.17.23_train_mlp_projector_maze2d-medium-v0/checkpoints/latest.ckpt
|
| 101 |
+
max_length: 100
|
| 102 |
+
lora_config:
|
| 103 |
+
r: 32
|
| 104 |
+
lora_alpha: 64
|
| 105 |
+
lora_dropout: 0.05
|
| 106 |
+
bias: none
|
| 107 |
+
task_type: CAUSAL_LM
|
| 108 |
+
prompter:
|
| 109 |
+
_target_: llmbc.translator.prompter.smollm2_prompter.SmolLM2Prompter
|
| 110 |
+
use_joint_mlp_projector: true
|
| 111 |
+
hydra:
|
| 112 |
+
job:
|
| 113 |
+
override_dirname: ${model_name}
|
| 114 |
+
run:
|
| 115 |
+
dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${model_name}
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/.hydra/hydra.yaml
ADDED
|
@@ -0,0 +1,156 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
hydra:
|
| 2 |
+
run:
|
| 3 |
+
dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
|
| 4 |
+
sweep:
|
| 5 |
+
dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
|
| 6 |
+
subdir: ${hydra.job.num}
|
| 7 |
+
launcher:
|
| 8 |
+
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
|
| 9 |
+
sweeper:
|
| 10 |
+
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
|
| 11 |
+
max_batch_size: null
|
| 12 |
+
params: null
|
| 13 |
+
help:
|
| 14 |
+
app_name: ${hydra.job.name}
|
| 15 |
+
header: '${hydra.help.app_name} is powered by Hydra.
|
| 16 |
+
|
| 17 |
+
'
|
| 18 |
+
footer: 'Powered by Hydra (https://hydra.cc)
|
| 19 |
+
|
| 20 |
+
Use --hydra-help to view Hydra specific help
|
| 21 |
+
|
| 22 |
+
'
|
| 23 |
+
template: '${hydra.help.header}
|
| 24 |
+
|
| 25 |
+
== Configuration groups ==
|
| 26 |
+
|
| 27 |
+
Compose your configuration from those groups (group=option)
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
$APP_CONFIG_GROUPS
|
| 31 |
+
|
| 32 |
+
|
| 33 |
+
== Config ==
|
| 34 |
+
|
| 35 |
+
Override anything in the config (foo.bar=value)
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
$CONFIG
|
| 39 |
+
|
| 40 |
+
|
| 41 |
+
${hydra.help.footer}
|
| 42 |
+
|
| 43 |
+
'
|
| 44 |
+
hydra_help:
|
| 45 |
+
template: 'Hydra (${hydra.runtime.version})
|
| 46 |
+
|
| 47 |
+
See https://hydra.cc for more info.
|
| 48 |
+
|
| 49 |
+
|
| 50 |
+
== Flags ==
|
| 51 |
+
|
| 52 |
+
$FLAGS_HELP
|
| 53 |
+
|
| 54 |
+
|
| 55 |
+
== Configuration groups ==
|
| 56 |
+
|
| 57 |
+
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
|
| 58 |
+
to command line)
|
| 59 |
+
|
| 60 |
+
|
| 61 |
+
$HYDRA_CONFIG_GROUPS
|
| 62 |
+
|
| 63 |
+
|
| 64 |
+
Use ''--cfg hydra'' to Show the Hydra config.
|
| 65 |
+
|
| 66 |
+
'
|
| 67 |
+
hydra_help: ???
|
| 68 |
+
hydra_logging:
|
| 69 |
+
version: 1
|
| 70 |
+
formatters:
|
| 71 |
+
simple:
|
| 72 |
+
format: '[%(asctime)s][HYDRA] %(message)s'
|
| 73 |
+
handlers:
|
| 74 |
+
console:
|
| 75 |
+
class: logging.StreamHandler
|
| 76 |
+
formatter: simple
|
| 77 |
+
stream: ext://sys.stdout
|
| 78 |
+
root:
|
| 79 |
+
level: INFO
|
| 80 |
+
handlers:
|
| 81 |
+
- console
|
| 82 |
+
loggers:
|
| 83 |
+
logging_example:
|
| 84 |
+
level: DEBUG
|
| 85 |
+
disable_existing_loggers: false
|
| 86 |
+
job_logging:
|
| 87 |
+
version: 1
|
| 88 |
+
formatters:
|
| 89 |
+
simple:
|
| 90 |
+
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
|
| 91 |
+
handlers:
|
| 92 |
+
console:
|
| 93 |
+
class: logging.StreamHandler
|
| 94 |
+
formatter: simple
|
| 95 |
+
stream: ext://sys.stdout
|
| 96 |
+
file:
|
| 97 |
+
class: logging.FileHandler
|
| 98 |
+
formatter: simple
|
| 99 |
+
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
|
| 100 |
+
root:
|
| 101 |
+
level: INFO
|
| 102 |
+
handlers:
|
| 103 |
+
- console
|
| 104 |
+
- file
|
| 105 |
+
disable_existing_loggers: false
|
| 106 |
+
env: {}
|
| 107 |
+
mode: RUN
|
| 108 |
+
searchpath: []
|
| 109 |
+
callbacks: {}
|
| 110 |
+
output_subdir: .hydra
|
| 111 |
+
overrides:
|
| 112 |
+
hydra:
|
| 113 |
+
- hydra.mode=RUN
|
| 114 |
+
task: []
|
| 115 |
+
job:
|
| 116 |
+
name: train_llm_workspace
|
| 117 |
+
chdir: null
|
| 118 |
+
override_dirname: ''
|
| 119 |
+
id: ???
|
| 120 |
+
num: ???
|
| 121 |
+
config_name: train_llm_workspace
|
| 122 |
+
env_set: {}
|
| 123 |
+
env_copy: []
|
| 124 |
+
config:
|
| 125 |
+
override_dirname:
|
| 126 |
+
kv_sep: '='
|
| 127 |
+
item_sep: ','
|
| 128 |
+
exclude_keys: []
|
| 129 |
+
runtime:
|
| 130 |
+
version: 1.2.0
|
| 131 |
+
version_base: '1.2'
|
| 132 |
+
cwd: /home/chyang/workspace/LLM-BC
|
| 133 |
+
config_sources:
|
| 134 |
+
- path: hydra.conf
|
| 135 |
+
schema: pkg
|
| 136 |
+
provider: hydra
|
| 137 |
+
- path: /home/chyang/workspace/LLM-BC/llmbc/config
|
| 138 |
+
schema: file
|
| 139 |
+
provider: main
|
| 140 |
+
- path: ''
|
| 141 |
+
schema: structured
|
| 142 |
+
provider: schema
|
| 143 |
+
output_dir: /home/chyang/workspace/LLM-BC/data/outputs/2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0
|
| 144 |
+
choices:
|
| 145 |
+
llm: smollm2-135m-instruct
|
| 146 |
+
task: maze2d-medium-v0
|
| 147 |
+
hydra/env: default
|
| 148 |
+
hydra/callbacks: null
|
| 149 |
+
hydra/job_logging: default
|
| 150 |
+
hydra/hydra_logging: default
|
| 151 |
+
hydra/hydra_help: default
|
| 152 |
+
hydra/help: default
|
| 153 |
+
hydra/sweeper: basic
|
| 154 |
+
hydra/launcher: basic
|
| 155 |
+
hydra/output: default
|
| 156 |
+
verbose: false
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/.hydra/overrides.yaml
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
[]
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/config.json
ADDED
|
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "HuggingFaceTB/SmolLM2-135M-Instruct",
|
| 3 |
+
"action_dim": 2,
|
| 4 |
+
"architectures": [
|
| 5 |
+
"LowdimLlamaForCausalLM"
|
| 6 |
+
],
|
| 7 |
+
"attention_bias": false,
|
| 8 |
+
"attention_dropout": 0.0,
|
| 9 |
+
"bos_token_id": 1,
|
| 10 |
+
"eos_token_id": 2,
|
| 11 |
+
"head_dim": 64,
|
| 12 |
+
"hidden_act": "silu",
|
| 13 |
+
"hidden_size": 576,
|
| 14 |
+
"initializer_range": 0.041666666666666664,
|
| 15 |
+
"intermediate_size": 1536,
|
| 16 |
+
"is_llama_config": true,
|
| 17 |
+
"max_position_embeddings": 8192,
|
| 18 |
+
"mlp_bias": false,
|
| 19 |
+
"model_type": "llama_lowdim",
|
| 20 |
+
"num_attention_heads": 9,
|
| 21 |
+
"num_hidden_layers": 30,
|
| 22 |
+
"num_key_value_heads": 3,
|
| 23 |
+
"obs_dim": 6,
|
| 24 |
+
"pad_token_id": 2,
|
| 25 |
+
"pretraining_tp": 1,
|
| 26 |
+
"rms_norm_eps": 1e-05,
|
| 27 |
+
"rope_interleaved": false,
|
| 28 |
+
"rope_scaling": null,
|
| 29 |
+
"rope_theta": 100000,
|
| 30 |
+
"tie_word_embeddings": true,
|
| 31 |
+
"torch_dtype": "float32",
|
| 32 |
+
"transformers.js_config": {
|
| 33 |
+
"kv_cache_dtype": {
|
| 34 |
+
"fp16": "float16",
|
| 35 |
+
"q4f16": "float16"
|
| 36 |
+
}
|
| 37 |
+
},
|
| 38 |
+
"transformers_version": "4.47.1",
|
| 39 |
+
"use_cache": false,
|
| 40 |
+
"use_joint_mlp_projector": true,
|
| 41 |
+
"vocab_size": 49152
|
| 42 |
+
}
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/generation_config.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"bos_token_id": 1,
|
| 4 |
+
"eos_token_id": 2,
|
| 5 |
+
"pad_token_id": 2,
|
| 6 |
+
"transformers_version": "4.47.1"
|
| 7 |
+
}
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/merges.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/mlp_projector.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5772a53afc6bd0228788f1070597710ed5c2fd894fdb8e0c60504896bcfb4590
|
| 3 |
+
size 1352320
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f631a5ae310a208d9c7bf046cff8cc7f455840e8aca03fccee52a8470109d8ec
|
| 3 |
+
size 539441032
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:260bf78e9b154b4467f9291eb5de3d59a41dee14778b8f2287f2bff464c9a197
|
| 3 |
+
size 1078989882
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:23f694cc049a249c7913981f2ac716d798be392b5b087b08cf3ecdcb75397653
|
| 3 |
+
size 14244
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f8c9ac5a49fa3e3d4135d2414df5cdedecb4ad2b52224edc0a38352c38e84eb7
|
| 3 |
+
size 1064
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/special_tokens_map.json
ADDED
|
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"additional_special_tokens": [
|
| 3 |
+
"<|im_start|>",
|
| 4 |
+
"<|im_end|>"
|
| 5 |
+
],
|
| 6 |
+
"bos_token": {
|
| 7 |
+
"content": "<|im_start|>",
|
| 8 |
+
"lstrip": false,
|
| 9 |
+
"normalized": false,
|
| 10 |
+
"rstrip": false,
|
| 11 |
+
"single_word": false
|
| 12 |
+
},
|
| 13 |
+
"eos_token": {
|
| 14 |
+
"content": "<|im_end|>",
|
| 15 |
+
"lstrip": false,
|
| 16 |
+
"normalized": false,
|
| 17 |
+
"rstrip": false,
|
| 18 |
+
"single_word": false
|
| 19 |
+
},
|
| 20 |
+
"pad_token": {
|
| 21 |
+
"content": "<|im_end|>",
|
| 22 |
+
"lstrip": false,
|
| 23 |
+
"normalized": false,
|
| 24 |
+
"rstrip": false,
|
| 25 |
+
"single_word": false
|
| 26 |
+
},
|
| 27 |
+
"unk_token": {
|
| 28 |
+
"content": "<|endoftext|>",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false
|
| 33 |
+
}
|
| 34 |
+
}
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/tokenizer_config.json
ADDED
|
@@ -0,0 +1,155 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_prefix_space": false,
|
| 3 |
+
"added_tokens_decoder": {
|
| 4 |
+
"0": {
|
| 5 |
+
"content": "<|endoftext|>",
|
| 6 |
+
"lstrip": false,
|
| 7 |
+
"normalized": false,
|
| 8 |
+
"rstrip": false,
|
| 9 |
+
"single_word": false,
|
| 10 |
+
"special": true
|
| 11 |
+
},
|
| 12 |
+
"1": {
|
| 13 |
+
"content": "<|im_start|>",
|
| 14 |
+
"lstrip": false,
|
| 15 |
+
"normalized": false,
|
| 16 |
+
"rstrip": false,
|
| 17 |
+
"single_word": false,
|
| 18 |
+
"special": true
|
| 19 |
+
},
|
| 20 |
+
"2": {
|
| 21 |
+
"content": "<|im_end|>",
|
| 22 |
+
"lstrip": false,
|
| 23 |
+
"normalized": false,
|
| 24 |
+
"rstrip": false,
|
| 25 |
+
"single_word": false,
|
| 26 |
+
"special": true
|
| 27 |
+
},
|
| 28 |
+
"3": {
|
| 29 |
+
"content": "<repo_name>",
|
| 30 |
+
"lstrip": false,
|
| 31 |
+
"normalized": false,
|
| 32 |
+
"rstrip": false,
|
| 33 |
+
"single_word": false,
|
| 34 |
+
"special": true
|
| 35 |
+
},
|
| 36 |
+
"4": {
|
| 37 |
+
"content": "<reponame>",
|
| 38 |
+
"lstrip": false,
|
| 39 |
+
"normalized": false,
|
| 40 |
+
"rstrip": false,
|
| 41 |
+
"single_word": false,
|
| 42 |
+
"special": true
|
| 43 |
+
},
|
| 44 |
+
"5": {
|
| 45 |
+
"content": "<file_sep>",
|
| 46 |
+
"lstrip": false,
|
| 47 |
+
"normalized": false,
|
| 48 |
+
"rstrip": false,
|
| 49 |
+
"single_word": false,
|
| 50 |
+
"special": true
|
| 51 |
+
},
|
| 52 |
+
"6": {
|
| 53 |
+
"content": "<filename>",
|
| 54 |
+
"lstrip": false,
|
| 55 |
+
"normalized": false,
|
| 56 |
+
"rstrip": false,
|
| 57 |
+
"single_word": false,
|
| 58 |
+
"special": true
|
| 59 |
+
},
|
| 60 |
+
"7": {
|
| 61 |
+
"content": "<gh_stars>",
|
| 62 |
+
"lstrip": false,
|
| 63 |
+
"normalized": false,
|
| 64 |
+
"rstrip": false,
|
| 65 |
+
"single_word": false,
|
| 66 |
+
"special": true
|
| 67 |
+
},
|
| 68 |
+
"8": {
|
| 69 |
+
"content": "<issue_start>",
|
| 70 |
+
"lstrip": false,
|
| 71 |
+
"normalized": false,
|
| 72 |
+
"rstrip": false,
|
| 73 |
+
"single_word": false,
|
| 74 |
+
"special": true
|
| 75 |
+
},
|
| 76 |
+
"9": {
|
| 77 |
+
"content": "<issue_comment>",
|
| 78 |
+
"lstrip": false,
|
| 79 |
+
"normalized": false,
|
| 80 |
+
"rstrip": false,
|
| 81 |
+
"single_word": false,
|
| 82 |
+
"special": true
|
| 83 |
+
},
|
| 84 |
+
"10": {
|
| 85 |
+
"content": "<issue_closed>",
|
| 86 |
+
"lstrip": false,
|
| 87 |
+
"normalized": false,
|
| 88 |
+
"rstrip": false,
|
| 89 |
+
"single_word": false,
|
| 90 |
+
"special": true
|
| 91 |
+
},
|
| 92 |
+
"11": {
|
| 93 |
+
"content": "<jupyter_start>",
|
| 94 |
+
"lstrip": false,
|
| 95 |
+
"normalized": false,
|
| 96 |
+
"rstrip": false,
|
| 97 |
+
"single_word": false,
|
| 98 |
+
"special": true
|
| 99 |
+
},
|
| 100 |
+
"12": {
|
| 101 |
+
"content": "<jupyter_text>",
|
| 102 |
+
"lstrip": false,
|
| 103 |
+
"normalized": false,
|
| 104 |
+
"rstrip": false,
|
| 105 |
+
"single_word": false,
|
| 106 |
+
"special": true
|
| 107 |
+
},
|
| 108 |
+
"13": {
|
| 109 |
+
"content": "<jupyter_code>",
|
| 110 |
+
"lstrip": false,
|
| 111 |
+
"normalized": false,
|
| 112 |
+
"rstrip": false,
|
| 113 |
+
"single_word": false,
|
| 114 |
+
"special": true
|
| 115 |
+
},
|
| 116 |
+
"14": {
|
| 117 |
+
"content": "<jupyter_output>",
|
| 118 |
+
"lstrip": false,
|
| 119 |
+
"normalized": false,
|
| 120 |
+
"rstrip": false,
|
| 121 |
+
"single_word": false,
|
| 122 |
+
"special": true
|
| 123 |
+
},
|
| 124 |
+
"15": {
|
| 125 |
+
"content": "<jupyter_script>",
|
| 126 |
+
"lstrip": false,
|
| 127 |
+
"normalized": false,
|
| 128 |
+
"rstrip": false,
|
| 129 |
+
"single_word": false,
|
| 130 |
+
"special": true
|
| 131 |
+
},
|
| 132 |
+
"16": {
|
| 133 |
+
"content": "<empty_output>",
|
| 134 |
+
"lstrip": false,
|
| 135 |
+
"normalized": false,
|
| 136 |
+
"rstrip": false,
|
| 137 |
+
"single_word": false,
|
| 138 |
+
"special": true
|
| 139 |
+
}
|
| 140 |
+
},
|
| 141 |
+
"additional_special_tokens": [
|
| 142 |
+
"<|im_start|>",
|
| 143 |
+
"<|im_end|>"
|
| 144 |
+
],
|
| 145 |
+
"bos_token": "<|im_start|>",
|
| 146 |
+
"chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful AI assistant named SmolLM, trained by Hugging Face<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
| 147 |
+
"clean_up_tokenization_spaces": false,
|
| 148 |
+
"eos_token": "<|im_end|>",
|
| 149 |
+
"extra_special_tokens": {},
|
| 150 |
+
"model_max_length": 8192,
|
| 151 |
+
"pad_token": "<|im_end|>",
|
| 152 |
+
"tokenizer_class": "GPT2Tokenizer",
|
| 153 |
+
"unk_token": "<|endoftext|>",
|
| 154 |
+
"vocab_size": 49152
|
| 155 |
+
}
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/trainer_state.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:08cafd23a9b55fc14f8bfb6a5d68a2a14e18205ef52010be996b64b43b792a47
|
| 3 |
+
size 6008
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/vocab.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/config.json
ADDED
|
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "HuggingFaceTB/SmolLM2-135M-Instruct",
|
| 3 |
+
"action_dim": 2,
|
| 4 |
+
"architectures": [
|
| 5 |
+
"LowdimLlamaForCausalLM"
|
| 6 |
+
],
|
| 7 |
+
"attention_bias": false,
|
| 8 |
+
"attention_dropout": 0.0,
|
| 9 |
+
"bos_token_id": 1,
|
| 10 |
+
"eos_token_id": 2,
|
| 11 |
+
"head_dim": 64,
|
| 12 |
+
"hidden_act": "silu",
|
| 13 |
+
"hidden_size": 576,
|
| 14 |
+
"initializer_range": 0.041666666666666664,
|
| 15 |
+
"intermediate_size": 1536,
|
| 16 |
+
"is_llama_config": true,
|
| 17 |
+
"max_position_embeddings": 8192,
|
| 18 |
+
"mlp_bias": false,
|
| 19 |
+
"model_type": "llama_lowdim",
|
| 20 |
+
"num_attention_heads": 9,
|
| 21 |
+
"num_hidden_layers": 30,
|
| 22 |
+
"num_key_value_heads": 3,
|
| 23 |
+
"obs_dim": 6,
|
| 24 |
+
"pad_token_id": 2,
|
| 25 |
+
"pretraining_tp": 1,
|
| 26 |
+
"rms_norm_eps": 1e-05,
|
| 27 |
+
"rope_interleaved": false,
|
| 28 |
+
"rope_scaling": null,
|
| 29 |
+
"rope_theta": 100000,
|
| 30 |
+
"tie_word_embeddings": true,
|
| 31 |
+
"torch_dtype": "float32",
|
| 32 |
+
"transformers.js_config": {
|
| 33 |
+
"kv_cache_dtype": {
|
| 34 |
+
"fp16": "float16",
|
| 35 |
+
"q4f16": "float16"
|
| 36 |
+
}
|
| 37 |
+
},
|
| 38 |
+
"transformers_version": "4.47.1",
|
| 39 |
+
"use_cache": false,
|
| 40 |
+
"use_joint_mlp_projector": true,
|
| 41 |
+
"vocab_size": 49152
|
| 42 |
+
}
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/generation_config.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"bos_token_id": 1,
|
| 4 |
+
"eos_token_id": 2,
|
| 5 |
+
"pad_token_id": 2,
|
| 6 |
+
"transformers_version": "4.47.1"
|
| 7 |
+
}
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/merges.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/mlp_projector.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4201fb78b8ffcf4d86b37e993658b6b102575e304b5f5464d2fa4b546a20cc81
|
| 3 |
+
size 1352320
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f1e02fdbaeb462e87b69d73abe017d04e05cf93541a3a2d2a716a8f2db7348e1
|
| 3 |
+
size 539441032
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:28f61ded6932f48f586026daba18558bfb1647c8908b8dd7c48987cffc9715f0
|
| 3 |
+
size 1078989882
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:995de6b51117ac658c94aad6b8dea17ddc4d4364b7ecf191600e4948263ea5fa
|
| 3 |
+
size 14244
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dc1290dc5e563ed697fa6534690e877c47e4111b0afadd3cf47f2f7158f0e0f7
|
| 3 |
+
size 1064
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/special_tokens_map.json
ADDED
|
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"additional_special_tokens": [
|
| 3 |
+
"<|im_start|>",
|
| 4 |
+
"<|im_end|>"
|
| 5 |
+
],
|
| 6 |
+
"bos_token": {
|
| 7 |
+
"content": "<|im_start|>",
|
| 8 |
+
"lstrip": false,
|
| 9 |
+
"normalized": false,
|
| 10 |
+
"rstrip": false,
|
| 11 |
+
"single_word": false
|
| 12 |
+
},
|
| 13 |
+
"eos_token": {
|
| 14 |
+
"content": "<|im_end|>",
|
| 15 |
+
"lstrip": false,
|
| 16 |
+
"normalized": false,
|
| 17 |
+
"rstrip": false,
|
| 18 |
+
"single_word": false
|
| 19 |
+
},
|
| 20 |
+
"pad_token": {
|
| 21 |
+
"content": "<|im_end|>",
|
| 22 |
+
"lstrip": false,
|
| 23 |
+
"normalized": false,
|
| 24 |
+
"rstrip": false,
|
| 25 |
+
"single_word": false
|
| 26 |
+
},
|
| 27 |
+
"unk_token": {
|
| 28 |
+
"content": "<|endoftext|>",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false
|
| 33 |
+
}
|
| 34 |
+
}
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/tokenizer_config.json
ADDED
|
@@ -0,0 +1,155 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_prefix_space": false,
|
| 3 |
+
"added_tokens_decoder": {
|
| 4 |
+
"0": {
|
| 5 |
+
"content": "<|endoftext|>",
|
| 6 |
+
"lstrip": false,
|
| 7 |
+
"normalized": false,
|
| 8 |
+
"rstrip": false,
|
| 9 |
+
"single_word": false,
|
| 10 |
+
"special": true
|
| 11 |
+
},
|
| 12 |
+
"1": {
|
| 13 |
+
"content": "<|im_start|>",
|
| 14 |
+
"lstrip": false,
|
| 15 |
+
"normalized": false,
|
| 16 |
+
"rstrip": false,
|
| 17 |
+
"single_word": false,
|
| 18 |
+
"special": true
|
| 19 |
+
},
|
| 20 |
+
"2": {
|
| 21 |
+
"content": "<|im_end|>",
|
| 22 |
+
"lstrip": false,
|
| 23 |
+
"normalized": false,
|
| 24 |
+
"rstrip": false,
|
| 25 |
+
"single_word": false,
|
| 26 |
+
"special": true
|
| 27 |
+
},
|
| 28 |
+
"3": {
|
| 29 |
+
"content": "<repo_name>",
|
| 30 |
+
"lstrip": false,
|
| 31 |
+
"normalized": false,
|
| 32 |
+
"rstrip": false,
|
| 33 |
+
"single_word": false,
|
| 34 |
+
"special": true
|
| 35 |
+
},
|
| 36 |
+
"4": {
|
| 37 |
+
"content": "<reponame>",
|
| 38 |
+
"lstrip": false,
|
| 39 |
+
"normalized": false,
|
| 40 |
+
"rstrip": false,
|
| 41 |
+
"single_word": false,
|
| 42 |
+
"special": true
|
| 43 |
+
},
|
| 44 |
+
"5": {
|
| 45 |
+
"content": "<file_sep>",
|
| 46 |
+
"lstrip": false,
|
| 47 |
+
"normalized": false,
|
| 48 |
+
"rstrip": false,
|
| 49 |
+
"single_word": false,
|
| 50 |
+
"special": true
|
| 51 |
+
},
|
| 52 |
+
"6": {
|
| 53 |
+
"content": "<filename>",
|
| 54 |
+
"lstrip": false,
|
| 55 |
+
"normalized": false,
|
| 56 |
+
"rstrip": false,
|
| 57 |
+
"single_word": false,
|
| 58 |
+
"special": true
|
| 59 |
+
},
|
| 60 |
+
"7": {
|
| 61 |
+
"content": "<gh_stars>",
|
| 62 |
+
"lstrip": false,
|
| 63 |
+
"normalized": false,
|
| 64 |
+
"rstrip": false,
|
| 65 |
+
"single_word": false,
|
| 66 |
+
"special": true
|
| 67 |
+
},
|
| 68 |
+
"8": {
|
| 69 |
+
"content": "<issue_start>",
|
| 70 |
+
"lstrip": false,
|
| 71 |
+
"normalized": false,
|
| 72 |
+
"rstrip": false,
|
| 73 |
+
"single_word": false,
|
| 74 |
+
"special": true
|
| 75 |
+
},
|
| 76 |
+
"9": {
|
| 77 |
+
"content": "<issue_comment>",
|
| 78 |
+
"lstrip": false,
|
| 79 |
+
"normalized": false,
|
| 80 |
+
"rstrip": false,
|
| 81 |
+
"single_word": false,
|
| 82 |
+
"special": true
|
| 83 |
+
},
|
| 84 |
+
"10": {
|
| 85 |
+
"content": "<issue_closed>",
|
| 86 |
+
"lstrip": false,
|
| 87 |
+
"normalized": false,
|
| 88 |
+
"rstrip": false,
|
| 89 |
+
"single_word": false,
|
| 90 |
+
"special": true
|
| 91 |
+
},
|
| 92 |
+
"11": {
|
| 93 |
+
"content": "<jupyter_start>",
|
| 94 |
+
"lstrip": false,
|
| 95 |
+
"normalized": false,
|
| 96 |
+
"rstrip": false,
|
| 97 |
+
"single_word": false,
|
| 98 |
+
"special": true
|
| 99 |
+
},
|
| 100 |
+
"12": {
|
| 101 |
+
"content": "<jupyter_text>",
|
| 102 |
+
"lstrip": false,
|
| 103 |
+
"normalized": false,
|
| 104 |
+
"rstrip": false,
|
| 105 |
+
"single_word": false,
|
| 106 |
+
"special": true
|
| 107 |
+
},
|
| 108 |
+
"13": {
|
| 109 |
+
"content": "<jupyter_code>",
|
| 110 |
+
"lstrip": false,
|
| 111 |
+
"normalized": false,
|
| 112 |
+
"rstrip": false,
|
| 113 |
+
"single_word": false,
|
| 114 |
+
"special": true
|
| 115 |
+
},
|
| 116 |
+
"14": {
|
| 117 |
+
"content": "<jupyter_output>",
|
| 118 |
+
"lstrip": false,
|
| 119 |
+
"normalized": false,
|
| 120 |
+
"rstrip": false,
|
| 121 |
+
"single_word": false,
|
| 122 |
+
"special": true
|
| 123 |
+
},
|
| 124 |
+
"15": {
|
| 125 |
+
"content": "<jupyter_script>",
|
| 126 |
+
"lstrip": false,
|
| 127 |
+
"normalized": false,
|
| 128 |
+
"rstrip": false,
|
| 129 |
+
"single_word": false,
|
| 130 |
+
"special": true
|
| 131 |
+
},
|
| 132 |
+
"16": {
|
| 133 |
+
"content": "<empty_output>",
|
| 134 |
+
"lstrip": false,
|
| 135 |
+
"normalized": false,
|
| 136 |
+
"rstrip": false,
|
| 137 |
+
"single_word": false,
|
| 138 |
+
"special": true
|
| 139 |
+
}
|
| 140 |
+
},
|
| 141 |
+
"additional_special_tokens": [
|
| 142 |
+
"<|im_start|>",
|
| 143 |
+
"<|im_end|>"
|
| 144 |
+
],
|
| 145 |
+
"bos_token": "<|im_start|>",
|
| 146 |
+
"chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful AI assistant named SmolLM, trained by Hugging Face<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
| 147 |
+
"clean_up_tokenization_spaces": false,
|
| 148 |
+
"eos_token": "<|im_end|>",
|
| 149 |
+
"extra_special_tokens": {},
|
| 150 |
+
"model_max_length": 8192,
|
| 151 |
+
"pad_token": "<|im_end|>",
|
| 152 |
+
"tokenizer_class": "GPT2Tokenizer",
|
| 153 |
+
"unk_token": "<|endoftext|>",
|
| 154 |
+
"vocab_size": 49152
|
| 155 |
+
}
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/trainer_state.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:08cafd23a9b55fc14f8bfb6a5d68a2a14e18205ef52010be996b64b43b792a47
|
| 3 |
+
size 6008
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/vocab.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/config.json
ADDED
|
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "HuggingFaceTB/SmolLM2-135M-Instruct",
|
| 3 |
+
"action_dim": 2,
|
| 4 |
+
"architectures": [
|
| 5 |
+
"LowdimLlamaForCausalLM"
|
| 6 |
+
],
|
| 7 |
+
"attention_bias": false,
|
| 8 |
+
"attention_dropout": 0.0,
|
| 9 |
+
"bos_token_id": 1,
|
| 10 |
+
"eos_token_id": 2,
|
| 11 |
+
"head_dim": 64,
|
| 12 |
+
"hidden_act": "silu",
|
| 13 |
+
"hidden_size": 576,
|
| 14 |
+
"initializer_range": 0.041666666666666664,
|
| 15 |
+
"intermediate_size": 1536,
|
| 16 |
+
"is_llama_config": true,
|
| 17 |
+
"max_position_embeddings": 8192,
|
| 18 |
+
"mlp_bias": false,
|
| 19 |
+
"model_type": "llama_lowdim",
|
| 20 |
+
"num_attention_heads": 9,
|
| 21 |
+
"num_hidden_layers": 30,
|
| 22 |
+
"num_key_value_heads": 3,
|
| 23 |
+
"obs_dim": 6,
|
| 24 |
+
"pad_token_id": 2,
|
| 25 |
+
"pretraining_tp": 1,
|
| 26 |
+
"rms_norm_eps": 1e-05,
|
| 27 |
+
"rope_interleaved": false,
|
| 28 |
+
"rope_scaling": null,
|
| 29 |
+
"rope_theta": 100000,
|
| 30 |
+
"tie_word_embeddings": true,
|
| 31 |
+
"torch_dtype": "float32",
|
| 32 |
+
"transformers.js_config": {
|
| 33 |
+
"kv_cache_dtype": {
|
| 34 |
+
"fp16": "float16",
|
| 35 |
+
"q4f16": "float16"
|
| 36 |
+
}
|
| 37 |
+
},
|
| 38 |
+
"transformers_version": "4.47.1",
|
| 39 |
+
"use_cache": false,
|
| 40 |
+
"use_joint_mlp_projector": true,
|
| 41 |
+
"vocab_size": 49152
|
| 42 |
+
}
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/generation_config.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"bos_token_id": 1,
|
| 4 |
+
"eos_token_id": 2,
|
| 5 |
+
"pad_token_id": 2,
|
| 6 |
+
"transformers_version": "4.47.1"
|
| 7 |
+
}
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/merges.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/mlp_projector.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3962ffb194af4d89ac2b4c3569ff58ccb19f1a4a072a6be1082890ad676290d3
|
| 3 |
+
size 1352320
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:13e01b18a04c049e6fd594bbebfe935dc3db9c3ec70c2e99a455c42091b12a95
|
| 3 |
+
size 539441032
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e6456a31875e250c04cc56d391f44f1480a7b1b5683bfd337bc54343908b4559
|
| 3 |
+
size 1078989882
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d006bb71aa92a2f143fe04a3f51f7ea6a043b3d21d0b9682128e80dc617a4c30
|
| 3 |
+
size 14244
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:942282e81212375ad3d3c856a4475ba0961f329fdc35b59c567cb5ab36c8b7f9
|
| 3 |
+
size 1064
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/special_tokens_map.json
ADDED
|
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"additional_special_tokens": [
|
| 3 |
+
"<|im_start|>",
|
| 4 |
+
"<|im_end|>"
|
| 5 |
+
],
|
| 6 |
+
"bos_token": {
|
| 7 |
+
"content": "<|im_start|>",
|
| 8 |
+
"lstrip": false,
|
| 9 |
+
"normalized": false,
|
| 10 |
+
"rstrip": false,
|
| 11 |
+
"single_word": false
|
| 12 |
+
},
|
| 13 |
+
"eos_token": {
|
| 14 |
+
"content": "<|im_end|>",
|
| 15 |
+
"lstrip": false,
|
| 16 |
+
"normalized": false,
|
| 17 |
+
"rstrip": false,
|
| 18 |
+
"single_word": false
|
| 19 |
+
},
|
| 20 |
+
"pad_token": {
|
| 21 |
+
"content": "<|im_end|>",
|
| 22 |
+
"lstrip": false,
|
| 23 |
+
"normalized": false,
|
| 24 |
+
"rstrip": false,
|
| 25 |
+
"single_word": false
|
| 26 |
+
},
|
| 27 |
+
"unk_token": {
|
| 28 |
+
"content": "<|endoftext|>",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false
|
| 33 |
+
}
|
| 34 |
+
}
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/tokenizer_config.json
ADDED
|
@@ -0,0 +1,155 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_prefix_space": false,
|
| 3 |
+
"added_tokens_decoder": {
|
| 4 |
+
"0": {
|
| 5 |
+
"content": "<|endoftext|>",
|
| 6 |
+
"lstrip": false,
|
| 7 |
+
"normalized": false,
|
| 8 |
+
"rstrip": false,
|
| 9 |
+
"single_word": false,
|
| 10 |
+
"special": true
|
| 11 |
+
},
|
| 12 |
+
"1": {
|
| 13 |
+
"content": "<|im_start|>",
|
| 14 |
+
"lstrip": false,
|
| 15 |
+
"normalized": false,
|
| 16 |
+
"rstrip": false,
|
| 17 |
+
"single_word": false,
|
| 18 |
+
"special": true
|
| 19 |
+
},
|
| 20 |
+
"2": {
|
| 21 |
+
"content": "<|im_end|>",
|
| 22 |
+
"lstrip": false,
|
| 23 |
+
"normalized": false,
|
| 24 |
+
"rstrip": false,
|
| 25 |
+
"single_word": false,
|
| 26 |
+
"special": true
|
| 27 |
+
},
|
| 28 |
+
"3": {
|
| 29 |
+
"content": "<repo_name>",
|
| 30 |
+
"lstrip": false,
|
| 31 |
+
"normalized": false,
|
| 32 |
+
"rstrip": false,
|
| 33 |
+
"single_word": false,
|
| 34 |
+
"special": true
|
| 35 |
+
},
|
| 36 |
+
"4": {
|
| 37 |
+
"content": "<reponame>",
|
| 38 |
+
"lstrip": false,
|
| 39 |
+
"normalized": false,
|
| 40 |
+
"rstrip": false,
|
| 41 |
+
"single_word": false,
|
| 42 |
+
"special": true
|
| 43 |
+
},
|
| 44 |
+
"5": {
|
| 45 |
+
"content": "<file_sep>",
|
| 46 |
+
"lstrip": false,
|
| 47 |
+
"normalized": false,
|
| 48 |
+
"rstrip": false,
|
| 49 |
+
"single_word": false,
|
| 50 |
+
"special": true
|
| 51 |
+
},
|
| 52 |
+
"6": {
|
| 53 |
+
"content": "<filename>",
|
| 54 |
+
"lstrip": false,
|
| 55 |
+
"normalized": false,
|
| 56 |
+
"rstrip": false,
|
| 57 |
+
"single_word": false,
|
| 58 |
+
"special": true
|
| 59 |
+
},
|
| 60 |
+
"7": {
|
| 61 |
+
"content": "<gh_stars>",
|
| 62 |
+
"lstrip": false,
|
| 63 |
+
"normalized": false,
|
| 64 |
+
"rstrip": false,
|
| 65 |
+
"single_word": false,
|
| 66 |
+
"special": true
|
| 67 |
+
},
|
| 68 |
+
"8": {
|
| 69 |
+
"content": "<issue_start>",
|
| 70 |
+
"lstrip": false,
|
| 71 |
+
"normalized": false,
|
| 72 |
+
"rstrip": false,
|
| 73 |
+
"single_word": false,
|
| 74 |
+
"special": true
|
| 75 |
+
},
|
| 76 |
+
"9": {
|
| 77 |
+
"content": "<issue_comment>",
|
| 78 |
+
"lstrip": false,
|
| 79 |
+
"normalized": false,
|
| 80 |
+
"rstrip": false,
|
| 81 |
+
"single_word": false,
|
| 82 |
+
"special": true
|
| 83 |
+
},
|
| 84 |
+
"10": {
|
| 85 |
+
"content": "<issue_closed>",
|
| 86 |
+
"lstrip": false,
|
| 87 |
+
"normalized": false,
|
| 88 |
+
"rstrip": false,
|
| 89 |
+
"single_word": false,
|
| 90 |
+
"special": true
|
| 91 |
+
},
|
| 92 |
+
"11": {
|
| 93 |
+
"content": "<jupyter_start>",
|
| 94 |
+
"lstrip": false,
|
| 95 |
+
"normalized": false,
|
| 96 |
+
"rstrip": false,
|
| 97 |
+
"single_word": false,
|
| 98 |
+
"special": true
|
| 99 |
+
},
|
| 100 |
+
"12": {
|
| 101 |
+
"content": "<jupyter_text>",
|
| 102 |
+
"lstrip": false,
|
| 103 |
+
"normalized": false,
|
| 104 |
+
"rstrip": false,
|
| 105 |
+
"single_word": false,
|
| 106 |
+
"special": true
|
| 107 |
+
},
|
| 108 |
+
"13": {
|
| 109 |
+
"content": "<jupyter_code>",
|
| 110 |
+
"lstrip": false,
|
| 111 |
+
"normalized": false,
|
| 112 |
+
"rstrip": false,
|
| 113 |
+
"single_word": false,
|
| 114 |
+
"special": true
|
| 115 |
+
},
|
| 116 |
+
"14": {
|
| 117 |
+
"content": "<jupyter_output>",
|
| 118 |
+
"lstrip": false,
|
| 119 |
+
"normalized": false,
|
| 120 |
+
"rstrip": false,
|
| 121 |
+
"single_word": false,
|
| 122 |
+
"special": true
|
| 123 |
+
},
|
| 124 |
+
"15": {
|
| 125 |
+
"content": "<jupyter_script>",
|
| 126 |
+
"lstrip": false,
|
| 127 |
+
"normalized": false,
|
| 128 |
+
"rstrip": false,
|
| 129 |
+
"single_word": false,
|
| 130 |
+
"special": true
|
| 131 |
+
},
|
| 132 |
+
"16": {
|
| 133 |
+
"content": "<empty_output>",
|
| 134 |
+
"lstrip": false,
|
| 135 |
+
"normalized": false,
|
| 136 |
+
"rstrip": false,
|
| 137 |
+
"single_word": false,
|
| 138 |
+
"special": true
|
| 139 |
+
}
|
| 140 |
+
},
|
| 141 |
+
"additional_special_tokens": [
|
| 142 |
+
"<|im_start|>",
|
| 143 |
+
"<|im_end|>"
|
| 144 |
+
],
|
| 145 |
+
"bos_token": "<|im_start|>",
|
| 146 |
+
"chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful AI assistant named SmolLM, trained by Hugging Face<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
| 147 |
+
"clean_up_tokenization_spaces": false,
|
| 148 |
+
"eos_token": "<|im_end|>",
|
| 149 |
+
"extra_special_tokens": {},
|
| 150 |
+
"model_max_length": 8192,
|
| 151 |
+
"pad_token": "<|im_end|>",
|
| 152 |
+
"tokenizer_class": "GPT2Tokenizer",
|
| 153 |
+
"unk_token": "<|endoftext|>",
|
| 154 |
+
"vocab_size": 49152
|
| 155 |
+
}
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/trainer_state.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:08cafd23a9b55fc14f8bfb6a5d68a2a14e18205ef52010be996b64b43b792a47
|
| 3 |
+
size 6008
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/vocab.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/config.json
ADDED
|
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "HuggingFaceTB/SmolLM2-135M-Instruct",
|
| 3 |
+
"action_dim": 2,
|
| 4 |
+
"architectures": [
|
| 5 |
+
"LowdimLlamaForCausalLM"
|
| 6 |
+
],
|
| 7 |
+
"attention_bias": false,
|
| 8 |
+
"attention_dropout": 0.0,
|
| 9 |
+
"bos_token_id": 1,
|
| 10 |
+
"eos_token_id": 2,
|
| 11 |
+
"head_dim": 64,
|
| 12 |
+
"hidden_act": "silu",
|
| 13 |
+
"hidden_size": 576,
|
| 14 |
+
"initializer_range": 0.041666666666666664,
|
| 15 |
+
"intermediate_size": 1536,
|
| 16 |
+
"is_llama_config": true,
|
| 17 |
+
"max_position_embeddings": 8192,
|
| 18 |
+
"mlp_bias": false,
|
| 19 |
+
"model_type": "llama_lowdim",
|
| 20 |
+
"num_attention_heads": 9,
|
| 21 |
+
"num_hidden_layers": 30,
|
| 22 |
+
"num_key_value_heads": 3,
|
| 23 |
+
"obs_dim": 6,
|
| 24 |
+
"pad_token_id": 2,
|
| 25 |
+
"pretraining_tp": 1,
|
| 26 |
+
"rms_norm_eps": 1e-05,
|
| 27 |
+
"rope_interleaved": false,
|
| 28 |
+
"rope_scaling": null,
|
| 29 |
+
"rope_theta": 100000,
|
| 30 |
+
"tie_word_embeddings": true,
|
| 31 |
+
"torch_dtype": "float32",
|
| 32 |
+
"transformers.js_config": {
|
| 33 |
+
"kv_cache_dtype": {
|
| 34 |
+
"fp16": "float16",
|
| 35 |
+
"q4f16": "float16"
|
| 36 |
+
}
|
| 37 |
+
},
|
| 38 |
+
"transformers_version": "4.47.1",
|
| 39 |
+
"use_cache": true,
|
| 40 |
+
"use_joint_mlp_projector": true,
|
| 41 |
+
"vocab_size": 49152
|
| 42 |
+
}
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/generation_config.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"bos_token_id": 1,
|
| 4 |
+
"eos_token_id": 2,
|
| 5 |
+
"pad_token_id": 2,
|
| 6 |
+
"transformers_version": "4.47.1"
|
| 7 |
+
}
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f1e02fdbaeb462e87b69d73abe017d04e05cf93541a3a2d2a716a8f2db7348e1
|
| 3 |
+
size 539441032
|
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/normalizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7863dd8b1cec54ea35c3840df161a34ee700911175e25f412f15c526ba1f4db9
|
| 3 |
+
size 940
|