CHYang25 commited on
Commit
b951e9f
·
verified ·
1 Parent(s): cb83726

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/.hydra/config.yaml +115 -0
  3. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/.hydra/hydra.yaml +156 -0
  4. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/.hydra/overrides.yaml +1 -0
  5. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/config.json +42 -0
  6. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/generation_config.json +7 -0
  7. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/merges.txt +0 -0
  8. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/mlp_projector.bin +3 -0
  9. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/model.safetensors +3 -0
  10. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/optimizer.pt +3 -0
  11. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/rng_state.pth +3 -0
  12. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/scheduler.pt +3 -0
  13. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/special_tokens_map.json +34 -0
  14. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/tokenizer.json +0 -0
  15. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/tokenizer_config.json +155 -0
  16. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/trainer_state.json +0 -0
  17. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/training_args.bin +3 -0
  18. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/vocab.json +0 -0
  19. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/config.json +42 -0
  20. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/generation_config.json +7 -0
  21. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/merges.txt +0 -0
  22. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/mlp_projector.bin +3 -0
  23. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/model.safetensors +3 -0
  24. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/optimizer.pt +3 -0
  25. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/rng_state.pth +3 -0
  26. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/scheduler.pt +3 -0
  27. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/special_tokens_map.json +34 -0
  28. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/tokenizer.json +0 -0
  29. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/tokenizer_config.json +155 -0
  30. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/trainer_state.json +0 -0
  31. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/training_args.bin +3 -0
  32. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/vocab.json +0 -0
  33. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/config.json +42 -0
  34. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/generation_config.json +7 -0
  35. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/merges.txt +0 -0
  36. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/mlp_projector.bin +3 -0
  37. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/model.safetensors +3 -0
  38. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/optimizer.pt +3 -0
  39. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/rng_state.pth +3 -0
  40. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/scheduler.pt +3 -0
  41. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/special_tokens_map.json +34 -0
  42. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/tokenizer.json +0 -0
  43. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/tokenizer_config.json +155 -0
  44. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/trainer_state.json +0 -0
  45. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/training_args.bin +3 -0
  46. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/vocab.json +0 -0
  47. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/config.json +42 -0
  48. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/generation_config.json +7 -0
  49. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/model.safetensors +3 -0
  50. 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/normalizer.pt +3 -0
.gitattributes CHANGED
@@ -39,3 +39,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
39
  2025.10.01/22.05.45_train_llm_lowdim_PegInsertionSide-v1/wandb/run-20251001_220546-8mqefeab/run-8mqefeab.wandb filter=lfs diff=lfs merge=lfs -text
40
  2025.10.01/22.07.23_train_llm_lowdim_blockpush_lowdim_seed/wandb/run-20251001_220724-uktwy92f/run-uktwy92f.wandb filter=lfs diff=lfs merge=lfs -text
41
  2025.10.03/09.42.26_train_llm_lowdim_push-back-v2/wandb/run-20251003_094227-po9rvkh2/run-po9rvkh2.wandb filter=lfs diff=lfs merge=lfs -text
 
 
39
  2025.10.01/22.05.45_train_llm_lowdim_PegInsertionSide-v1/wandb/run-20251001_220546-8mqefeab/run-8mqefeab.wandb filter=lfs diff=lfs merge=lfs -text
40
  2025.10.01/22.07.23_train_llm_lowdim_blockpush_lowdim_seed/wandb/run-20251001_220724-uktwy92f/run-uktwy92f.wandb filter=lfs diff=lfs merge=lfs -text
41
  2025.10.03/09.42.26_train_llm_lowdim_push-back-v2/wandb/run-20251003_094227-po9rvkh2/run-po9rvkh2.wandb filter=lfs diff=lfs merge=lfs -text
42
+ 2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/wandb/run-20251016_115323-y9v0s119/run-y9v0s119.wandb filter=lfs diff=lfs merge=lfs -text
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/.hydra/config.yaml ADDED
@@ -0,0 +1,115 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: train_llm_lowdim
2
+ _target_: llmbc.workspace.train_llm_workspace.TrainLLMWorkspace
3
+ obs_dim: ${task.obs_dim}
4
+ action_dim: ${task.action_dim}
5
+ horizon: 1
6
+ n_obs_steps: 1
7
+ n_action_steps: 1
8
+ task_name: ${task.name}
9
+ exp_name: train llm
10
+ model_name: ${llm.name}
11
+ use_quantization: ${llm.use_quantization}
12
+ lora_config: ${llm.lora_config}
13
+ dataset:
14
+ test_data_ratio: 0.01
15
+ debug: false
16
+ training:
17
+ seed: 42
18
+ per_device_train_batch_size: 128
19
+ per_device_eval_batch_size: 128
20
+ gradient_accumulation_steps: 1
21
+ optim: paged_adamw_32bit
22
+ num_train_epochs: 10
23
+ eval_strategy: steps
24
+ logging_steps: 1
25
+ warmup_steps: 10
26
+ logging_strategy: steps
27
+ learning_rate: 1.0e-05
28
+ fp16: false
29
+ bf16: true
30
+ tf32: true
31
+ group_by_length: true
32
+ report_to: wandb
33
+ save_steps: 5000
34
+ eval_steps: 10
35
+ use_joint_mlp_projector: ${llm.use_joint_mlp_projector}
36
+ joint_obs_action_mlp_lr: 1.0e-05
37
+ trainer:
38
+ obs_dim: ${obs_dim}
39
+ action_dim: ${action_dim}
40
+ use_joint_mlp_projector: ${llm.use_joint_mlp_projector}
41
+ max_seq_length: ${llm.max_length}
42
+ dataset_text_field: text
43
+ packing: false
44
+ logging:
45
+ project: llm_module_finetuning
46
+ resume: true
47
+ mode: online
48
+ name: ${now:%Y.%m.%d-%H.%M.%S}_${name}_${task_name}
49
+ tags:
50
+ - ${name}
51
+ - ${task_name}
52
+ - ${exp_name}
53
+ id: null
54
+ group: null
55
+ multi_run:
56
+ run_dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
57
+ wandb_name_base: ${now:%Y.%m.%d-%H.%M.%S}_${name}_${task_name}
58
+ task:
59
+ name: maze2d-medium-v0
60
+ obs_dim: 6
61
+ action_dim: 2
62
+ env_runner:
63
+ _target_: llmbc.env_runner.pointmaze_lowdim_runner.PointmazeLowdimRunner
64
+ env_name: llf-pointmaze-maze2d-medium-v0
65
+ n_train: 10
66
+ n_test: 50
67
+ n_envs: 10
68
+ max_steps: 250
69
+ n_obs_steps: ${n_obs_steps}
70
+ n_action_steps: ${n_action_steps}
71
+ instruction_type: b
72
+ feedback_type:
73
+ - hp
74
+ - hn
75
+ - fp
76
+ visual: false
77
+ discount: 0.99
78
+ dataset:
79
+ _target_: llmbc.dataset.pointmaze_lowdim_dataset.PointmazeLowdimDataset
80
+ data_path: datasets/maze2d-medium-v0-general.pt
81
+ data_path2: datasets/maze2d-medium-v0-general.pt
82
+ horizon: ${horizon}
83
+ pad_before: ${eval:'${n_obs_steps}-1'}
84
+ pad_after: ${eval:'${n_action_steps}-1'}
85
+ obs_eef_target: true
86
+ use_manual_normalizer: false
87
+ val_ratio: 0.02
88
+ dummy_normalizer: true
89
+ instructor:
90
+ _target_: llmbc.translator.instructor.pointmaze_instructor.maze2d_medium_v0_instructor.Maze2dMediumV0Instructor
91
+ llm:
92
+ name: HuggingFaceTB/SmolLM2-135M-Instruct
93
+ model_name: SmolLM2-135M-Instruct
94
+ config_target: llmbc.model.llm.llama_lowdim_model.LowdimLlamaConfig
95
+ causal_lm_target: llmbc.model.llm.llama_lowdim_model.LowdimLlamaForCausalLM
96
+ use_quantization: false
97
+ use_joint_mlp_projector: true
98
+ llm_mode: mlp-finetuned
99
+ finetune_mode: orig
100
+ checkpoint: data/outputs/2025.10.14/13.17.23_train_mlp_projector_maze2d-medium-v0/checkpoints/latest.ckpt
101
+ max_length: 100
102
+ lora_config:
103
+ r: 32
104
+ lora_alpha: 64
105
+ lora_dropout: 0.05
106
+ bias: none
107
+ task_type: CAUSAL_LM
108
+ prompter:
109
+ _target_: llmbc.translator.prompter.smollm2_prompter.SmolLM2Prompter
110
+ use_joint_mlp_projector: true
111
+ hydra:
112
+ job:
113
+ override_dirname: ${model_name}
114
+ run:
115
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${model_name}
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/.hydra/hydra.yaml ADDED
@@ -0,0 +1,156 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
4
+ sweep:
5
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task: []
115
+ job:
116
+ name: train_llm_workspace
117
+ chdir: null
118
+ override_dirname: ''
119
+ id: ???
120
+ num: ???
121
+ config_name: train_llm_workspace
122
+ env_set: {}
123
+ env_copy: []
124
+ config:
125
+ override_dirname:
126
+ kv_sep: '='
127
+ item_sep: ','
128
+ exclude_keys: []
129
+ runtime:
130
+ version: 1.2.0
131
+ version_base: '1.2'
132
+ cwd: /home/chyang/workspace/LLM-BC
133
+ config_sources:
134
+ - path: hydra.conf
135
+ schema: pkg
136
+ provider: hydra
137
+ - path: /home/chyang/workspace/LLM-BC/llmbc/config
138
+ schema: file
139
+ provider: main
140
+ - path: ''
141
+ schema: structured
142
+ provider: schema
143
+ output_dir: /home/chyang/workspace/LLM-BC/data/outputs/2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0
144
+ choices:
145
+ llm: smollm2-135m-instruct
146
+ task: maze2d-medium-v0
147
+ hydra/env: default
148
+ hydra/callbacks: null
149
+ hydra/job_logging: default
150
+ hydra/hydra_logging: default
151
+ hydra/hydra_help: default
152
+ hydra/help: default
153
+ hydra/sweeper: basic
154
+ hydra/launcher: basic
155
+ hydra/output: default
156
+ verbose: false
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/.hydra/overrides.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ []
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "HuggingFaceTB/SmolLM2-135M-Instruct",
3
+ "action_dim": 2,
4
+ "architectures": [
5
+ "LowdimLlamaForCausalLM"
6
+ ],
7
+ "attention_bias": false,
8
+ "attention_dropout": 0.0,
9
+ "bos_token_id": 1,
10
+ "eos_token_id": 2,
11
+ "head_dim": 64,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 576,
14
+ "initializer_range": 0.041666666666666664,
15
+ "intermediate_size": 1536,
16
+ "is_llama_config": true,
17
+ "max_position_embeddings": 8192,
18
+ "mlp_bias": false,
19
+ "model_type": "llama_lowdim",
20
+ "num_attention_heads": 9,
21
+ "num_hidden_layers": 30,
22
+ "num_key_value_heads": 3,
23
+ "obs_dim": 6,
24
+ "pad_token_id": 2,
25
+ "pretraining_tp": 1,
26
+ "rms_norm_eps": 1e-05,
27
+ "rope_interleaved": false,
28
+ "rope_scaling": null,
29
+ "rope_theta": 100000,
30
+ "tie_word_embeddings": true,
31
+ "torch_dtype": "float32",
32
+ "transformers.js_config": {
33
+ "kv_cache_dtype": {
34
+ "fp16": "float16",
35
+ "q4f16": "float16"
36
+ }
37
+ },
38
+ "transformers_version": "4.47.1",
39
+ "use_cache": false,
40
+ "use_joint_mlp_projector": true,
41
+ "vocab_size": 49152
42
+ }
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 2,
6
+ "transformers_version": "4.47.1"
7
+ }
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/mlp_projector.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5772a53afc6bd0228788f1070597710ed5c2fd894fdb8e0c60504896bcfb4590
3
+ size 1352320
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f631a5ae310a208d9c7bf046cff8cc7f455840e8aca03fccee52a8470109d8ec
3
+ size 539441032
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:260bf78e9b154b4467f9291eb5de3d59a41dee14778b8f2287f2bff464c9a197
3
+ size 1078989882
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23f694cc049a249c7913981f2ac716d798be392b5b087b08cf3ecdcb75397653
3
+ size 14244
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8c9ac5a49fa3e3d4135d2414df5cdedecb4ad2b52224edc0a38352c38e84eb7
3
+ size 1064
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/special_tokens_map.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "bos_token": {
7
+ "content": "<|im_start|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "eos_token": {
14
+ "content": "<|im_end|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ },
20
+ "pad_token": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false
26
+ },
27
+ "unk_token": {
28
+ "content": "<|endoftext|>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false
33
+ }
34
+ }
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/tokenizer_config.json ADDED
@@ -0,0 +1,155 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<repo_name>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "4": {
37
+ "content": "<reponame>",
38
+ "lstrip": false,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ },
44
+ "5": {
45
+ "content": "<file_sep>",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false,
50
+ "special": true
51
+ },
52
+ "6": {
53
+ "content": "<filename>",
54
+ "lstrip": false,
55
+ "normalized": false,
56
+ "rstrip": false,
57
+ "single_word": false,
58
+ "special": true
59
+ },
60
+ "7": {
61
+ "content": "<gh_stars>",
62
+ "lstrip": false,
63
+ "normalized": false,
64
+ "rstrip": false,
65
+ "single_word": false,
66
+ "special": true
67
+ },
68
+ "8": {
69
+ "content": "<issue_start>",
70
+ "lstrip": false,
71
+ "normalized": false,
72
+ "rstrip": false,
73
+ "single_word": false,
74
+ "special": true
75
+ },
76
+ "9": {
77
+ "content": "<issue_comment>",
78
+ "lstrip": false,
79
+ "normalized": false,
80
+ "rstrip": false,
81
+ "single_word": false,
82
+ "special": true
83
+ },
84
+ "10": {
85
+ "content": "<issue_closed>",
86
+ "lstrip": false,
87
+ "normalized": false,
88
+ "rstrip": false,
89
+ "single_word": false,
90
+ "special": true
91
+ },
92
+ "11": {
93
+ "content": "<jupyter_start>",
94
+ "lstrip": false,
95
+ "normalized": false,
96
+ "rstrip": false,
97
+ "single_word": false,
98
+ "special": true
99
+ },
100
+ "12": {
101
+ "content": "<jupyter_text>",
102
+ "lstrip": false,
103
+ "normalized": false,
104
+ "rstrip": false,
105
+ "single_word": false,
106
+ "special": true
107
+ },
108
+ "13": {
109
+ "content": "<jupyter_code>",
110
+ "lstrip": false,
111
+ "normalized": false,
112
+ "rstrip": false,
113
+ "single_word": false,
114
+ "special": true
115
+ },
116
+ "14": {
117
+ "content": "<jupyter_output>",
118
+ "lstrip": false,
119
+ "normalized": false,
120
+ "rstrip": false,
121
+ "single_word": false,
122
+ "special": true
123
+ },
124
+ "15": {
125
+ "content": "<jupyter_script>",
126
+ "lstrip": false,
127
+ "normalized": false,
128
+ "rstrip": false,
129
+ "single_word": false,
130
+ "special": true
131
+ },
132
+ "16": {
133
+ "content": "<empty_output>",
134
+ "lstrip": false,
135
+ "normalized": false,
136
+ "rstrip": false,
137
+ "single_word": false,
138
+ "special": true
139
+ }
140
+ },
141
+ "additional_special_tokens": [
142
+ "<|im_start|>",
143
+ "<|im_end|>"
144
+ ],
145
+ "bos_token": "<|im_start|>",
146
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful AI assistant named SmolLM, trained by Hugging Face<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
147
+ "clean_up_tokenization_spaces": false,
148
+ "eos_token": "<|im_end|>",
149
+ "extra_special_tokens": {},
150
+ "model_max_length": 8192,
151
+ "pad_token": "<|im_end|>",
152
+ "tokenizer_class": "GPT2Tokenizer",
153
+ "unk_token": "<|endoftext|>",
154
+ "vocab_size": 49152
155
+ }
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08cafd23a9b55fc14f8bfb6a5d68a2a14e18205ef52010be996b64b43b792a47
3
+ size 6008
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-10000/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "HuggingFaceTB/SmolLM2-135M-Instruct",
3
+ "action_dim": 2,
4
+ "architectures": [
5
+ "LowdimLlamaForCausalLM"
6
+ ],
7
+ "attention_bias": false,
8
+ "attention_dropout": 0.0,
9
+ "bos_token_id": 1,
10
+ "eos_token_id": 2,
11
+ "head_dim": 64,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 576,
14
+ "initializer_range": 0.041666666666666664,
15
+ "intermediate_size": 1536,
16
+ "is_llama_config": true,
17
+ "max_position_embeddings": 8192,
18
+ "mlp_bias": false,
19
+ "model_type": "llama_lowdim",
20
+ "num_attention_heads": 9,
21
+ "num_hidden_layers": 30,
22
+ "num_key_value_heads": 3,
23
+ "obs_dim": 6,
24
+ "pad_token_id": 2,
25
+ "pretraining_tp": 1,
26
+ "rms_norm_eps": 1e-05,
27
+ "rope_interleaved": false,
28
+ "rope_scaling": null,
29
+ "rope_theta": 100000,
30
+ "tie_word_embeddings": true,
31
+ "torch_dtype": "float32",
32
+ "transformers.js_config": {
33
+ "kv_cache_dtype": {
34
+ "fp16": "float16",
35
+ "q4f16": "float16"
36
+ }
37
+ },
38
+ "transformers_version": "4.47.1",
39
+ "use_cache": false,
40
+ "use_joint_mlp_projector": true,
41
+ "vocab_size": 49152
42
+ }
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 2,
6
+ "transformers_version": "4.47.1"
7
+ }
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/mlp_projector.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4201fb78b8ffcf4d86b37e993658b6b102575e304b5f5464d2fa4b546a20cc81
3
+ size 1352320
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1e02fdbaeb462e87b69d73abe017d04e05cf93541a3a2d2a716a8f2db7348e1
3
+ size 539441032
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28f61ded6932f48f586026daba18558bfb1647c8908b8dd7c48987cffc9715f0
3
+ size 1078989882
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:995de6b51117ac658c94aad6b8dea17ddc4d4364b7ecf191600e4948263ea5fa
3
+ size 14244
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc1290dc5e563ed697fa6534690e877c47e4111b0afadd3cf47f2f7158f0e0f7
3
+ size 1064
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/special_tokens_map.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "bos_token": {
7
+ "content": "<|im_start|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "eos_token": {
14
+ "content": "<|im_end|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ },
20
+ "pad_token": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false
26
+ },
27
+ "unk_token": {
28
+ "content": "<|endoftext|>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false
33
+ }
34
+ }
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/tokenizer_config.json ADDED
@@ -0,0 +1,155 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<repo_name>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "4": {
37
+ "content": "<reponame>",
38
+ "lstrip": false,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ },
44
+ "5": {
45
+ "content": "<file_sep>",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false,
50
+ "special": true
51
+ },
52
+ "6": {
53
+ "content": "<filename>",
54
+ "lstrip": false,
55
+ "normalized": false,
56
+ "rstrip": false,
57
+ "single_word": false,
58
+ "special": true
59
+ },
60
+ "7": {
61
+ "content": "<gh_stars>",
62
+ "lstrip": false,
63
+ "normalized": false,
64
+ "rstrip": false,
65
+ "single_word": false,
66
+ "special": true
67
+ },
68
+ "8": {
69
+ "content": "<issue_start>",
70
+ "lstrip": false,
71
+ "normalized": false,
72
+ "rstrip": false,
73
+ "single_word": false,
74
+ "special": true
75
+ },
76
+ "9": {
77
+ "content": "<issue_comment>",
78
+ "lstrip": false,
79
+ "normalized": false,
80
+ "rstrip": false,
81
+ "single_word": false,
82
+ "special": true
83
+ },
84
+ "10": {
85
+ "content": "<issue_closed>",
86
+ "lstrip": false,
87
+ "normalized": false,
88
+ "rstrip": false,
89
+ "single_word": false,
90
+ "special": true
91
+ },
92
+ "11": {
93
+ "content": "<jupyter_start>",
94
+ "lstrip": false,
95
+ "normalized": false,
96
+ "rstrip": false,
97
+ "single_word": false,
98
+ "special": true
99
+ },
100
+ "12": {
101
+ "content": "<jupyter_text>",
102
+ "lstrip": false,
103
+ "normalized": false,
104
+ "rstrip": false,
105
+ "single_word": false,
106
+ "special": true
107
+ },
108
+ "13": {
109
+ "content": "<jupyter_code>",
110
+ "lstrip": false,
111
+ "normalized": false,
112
+ "rstrip": false,
113
+ "single_word": false,
114
+ "special": true
115
+ },
116
+ "14": {
117
+ "content": "<jupyter_output>",
118
+ "lstrip": false,
119
+ "normalized": false,
120
+ "rstrip": false,
121
+ "single_word": false,
122
+ "special": true
123
+ },
124
+ "15": {
125
+ "content": "<jupyter_script>",
126
+ "lstrip": false,
127
+ "normalized": false,
128
+ "rstrip": false,
129
+ "single_word": false,
130
+ "special": true
131
+ },
132
+ "16": {
133
+ "content": "<empty_output>",
134
+ "lstrip": false,
135
+ "normalized": false,
136
+ "rstrip": false,
137
+ "single_word": false,
138
+ "special": true
139
+ }
140
+ },
141
+ "additional_special_tokens": [
142
+ "<|im_start|>",
143
+ "<|im_end|>"
144
+ ],
145
+ "bos_token": "<|im_start|>",
146
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful AI assistant named SmolLM, trained by Hugging Face<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
147
+ "clean_up_tokenization_spaces": false,
148
+ "eos_token": "<|im_end|>",
149
+ "extra_special_tokens": {},
150
+ "model_max_length": 8192,
151
+ "pad_token": "<|im_end|>",
152
+ "tokenizer_class": "GPT2Tokenizer",
153
+ "unk_token": "<|endoftext|>",
154
+ "vocab_size": 49152
155
+ }
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08cafd23a9b55fc14f8bfb6a5d68a2a14e18205ef52010be996b64b43b792a47
3
+ size 6008
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-11850/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "HuggingFaceTB/SmolLM2-135M-Instruct",
3
+ "action_dim": 2,
4
+ "architectures": [
5
+ "LowdimLlamaForCausalLM"
6
+ ],
7
+ "attention_bias": false,
8
+ "attention_dropout": 0.0,
9
+ "bos_token_id": 1,
10
+ "eos_token_id": 2,
11
+ "head_dim": 64,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 576,
14
+ "initializer_range": 0.041666666666666664,
15
+ "intermediate_size": 1536,
16
+ "is_llama_config": true,
17
+ "max_position_embeddings": 8192,
18
+ "mlp_bias": false,
19
+ "model_type": "llama_lowdim",
20
+ "num_attention_heads": 9,
21
+ "num_hidden_layers": 30,
22
+ "num_key_value_heads": 3,
23
+ "obs_dim": 6,
24
+ "pad_token_id": 2,
25
+ "pretraining_tp": 1,
26
+ "rms_norm_eps": 1e-05,
27
+ "rope_interleaved": false,
28
+ "rope_scaling": null,
29
+ "rope_theta": 100000,
30
+ "tie_word_embeddings": true,
31
+ "torch_dtype": "float32",
32
+ "transformers.js_config": {
33
+ "kv_cache_dtype": {
34
+ "fp16": "float16",
35
+ "q4f16": "float16"
36
+ }
37
+ },
38
+ "transformers_version": "4.47.1",
39
+ "use_cache": false,
40
+ "use_joint_mlp_projector": true,
41
+ "vocab_size": 49152
42
+ }
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 2,
6
+ "transformers_version": "4.47.1"
7
+ }
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/mlp_projector.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3962ffb194af4d89ac2b4c3569ff58ccb19f1a4a072a6be1082890ad676290d3
3
+ size 1352320
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13e01b18a04c049e6fd594bbebfe935dc3db9c3ec70c2e99a455c42091b12a95
3
+ size 539441032
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6456a31875e250c04cc56d391f44f1480a7b1b5683bfd337bc54343908b4559
3
+ size 1078989882
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d006bb71aa92a2f143fe04a3f51f7ea6a043b3d21d0b9682128e80dc617a4c30
3
+ size 14244
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:942282e81212375ad3d3c856a4475ba0961f329fdc35b59c567cb5ab36c8b7f9
3
+ size 1064
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/special_tokens_map.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "bos_token": {
7
+ "content": "<|im_start|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "eos_token": {
14
+ "content": "<|im_end|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ },
20
+ "pad_token": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false
26
+ },
27
+ "unk_token": {
28
+ "content": "<|endoftext|>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false
33
+ }
34
+ }
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/tokenizer_config.json ADDED
@@ -0,0 +1,155 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<repo_name>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "4": {
37
+ "content": "<reponame>",
38
+ "lstrip": false,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ },
44
+ "5": {
45
+ "content": "<file_sep>",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false,
50
+ "special": true
51
+ },
52
+ "6": {
53
+ "content": "<filename>",
54
+ "lstrip": false,
55
+ "normalized": false,
56
+ "rstrip": false,
57
+ "single_word": false,
58
+ "special": true
59
+ },
60
+ "7": {
61
+ "content": "<gh_stars>",
62
+ "lstrip": false,
63
+ "normalized": false,
64
+ "rstrip": false,
65
+ "single_word": false,
66
+ "special": true
67
+ },
68
+ "8": {
69
+ "content": "<issue_start>",
70
+ "lstrip": false,
71
+ "normalized": false,
72
+ "rstrip": false,
73
+ "single_word": false,
74
+ "special": true
75
+ },
76
+ "9": {
77
+ "content": "<issue_comment>",
78
+ "lstrip": false,
79
+ "normalized": false,
80
+ "rstrip": false,
81
+ "single_word": false,
82
+ "special": true
83
+ },
84
+ "10": {
85
+ "content": "<issue_closed>",
86
+ "lstrip": false,
87
+ "normalized": false,
88
+ "rstrip": false,
89
+ "single_word": false,
90
+ "special": true
91
+ },
92
+ "11": {
93
+ "content": "<jupyter_start>",
94
+ "lstrip": false,
95
+ "normalized": false,
96
+ "rstrip": false,
97
+ "single_word": false,
98
+ "special": true
99
+ },
100
+ "12": {
101
+ "content": "<jupyter_text>",
102
+ "lstrip": false,
103
+ "normalized": false,
104
+ "rstrip": false,
105
+ "single_word": false,
106
+ "special": true
107
+ },
108
+ "13": {
109
+ "content": "<jupyter_code>",
110
+ "lstrip": false,
111
+ "normalized": false,
112
+ "rstrip": false,
113
+ "single_word": false,
114
+ "special": true
115
+ },
116
+ "14": {
117
+ "content": "<jupyter_output>",
118
+ "lstrip": false,
119
+ "normalized": false,
120
+ "rstrip": false,
121
+ "single_word": false,
122
+ "special": true
123
+ },
124
+ "15": {
125
+ "content": "<jupyter_script>",
126
+ "lstrip": false,
127
+ "normalized": false,
128
+ "rstrip": false,
129
+ "single_word": false,
130
+ "special": true
131
+ },
132
+ "16": {
133
+ "content": "<empty_output>",
134
+ "lstrip": false,
135
+ "normalized": false,
136
+ "rstrip": false,
137
+ "single_word": false,
138
+ "special": true
139
+ }
140
+ },
141
+ "additional_special_tokens": [
142
+ "<|im_start|>",
143
+ "<|im_end|>"
144
+ ],
145
+ "bos_token": "<|im_start|>",
146
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful AI assistant named SmolLM, trained by Hugging Face<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
147
+ "clean_up_tokenization_spaces": false,
148
+ "eos_token": "<|im_end|>",
149
+ "extra_special_tokens": {},
150
+ "model_max_length": 8192,
151
+ "pad_token": "<|im_end|>",
152
+ "tokenizer_class": "GPT2Tokenizer",
153
+ "unk_token": "<|endoftext|>",
154
+ "vocab_size": 49152
155
+ }
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08cafd23a9b55fc14f8bfb6a5d68a2a14e18205ef52010be996b64b43b792a47
3
+ size 6008
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/checkpoint-5000/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "HuggingFaceTB/SmolLM2-135M-Instruct",
3
+ "action_dim": 2,
4
+ "architectures": [
5
+ "LowdimLlamaForCausalLM"
6
+ ],
7
+ "attention_bias": false,
8
+ "attention_dropout": 0.0,
9
+ "bos_token_id": 1,
10
+ "eos_token_id": 2,
11
+ "head_dim": 64,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 576,
14
+ "initializer_range": 0.041666666666666664,
15
+ "intermediate_size": 1536,
16
+ "is_llama_config": true,
17
+ "max_position_embeddings": 8192,
18
+ "mlp_bias": false,
19
+ "model_type": "llama_lowdim",
20
+ "num_attention_heads": 9,
21
+ "num_hidden_layers": 30,
22
+ "num_key_value_heads": 3,
23
+ "obs_dim": 6,
24
+ "pad_token_id": 2,
25
+ "pretraining_tp": 1,
26
+ "rms_norm_eps": 1e-05,
27
+ "rope_interleaved": false,
28
+ "rope_scaling": null,
29
+ "rope_theta": 100000,
30
+ "tie_word_embeddings": true,
31
+ "torch_dtype": "float32",
32
+ "transformers.js_config": {
33
+ "kv_cache_dtype": {
34
+ "fp16": "float16",
35
+ "q4f16": "float16"
36
+ }
37
+ },
38
+ "transformers_version": "4.47.1",
39
+ "use_cache": true,
40
+ "use_joint_mlp_projector": true,
41
+ "vocab_size": 49152
42
+ }
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 2,
6
+ "transformers_version": "4.47.1"
7
+ }
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1e02fdbaeb462e87b69d73abe017d04e05cf93541a3a2d2a716a8f2db7348e1
3
+ size 539441032
2025.10.16/11.53.22_train_llm_lowdim_maze2d-medium-v0/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-maze2d-medium-v0/normalizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7863dd8b1cec54ea35c3840df161a34ee700911175e25f412f15c526ba1f4db9
3
+ size 940