CHYang25 commited on
Commit
121dd45
·
verified ·
1 Parent(s): 32468ac

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +9 -0
  2. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/.hydra/config.yaml +116 -0
  3. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/.hydra/hydra.yaml +156 -0
  4. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/.hydra/overrides.yaml +1 -0
  5. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/README.md +202 -0
  6. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/adapter_config.json +37 -0
  7. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/adapter_model.safetensors +3 -0
  8. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/README.md +202 -0
  9. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/adapter_config.json +37 -0
  10. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/adapter_model.safetensors +3 -0
  11. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/added_tokens.json +5 -0
  12. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/config.json +33 -0
  13. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/merges.txt +0 -0
  14. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/mlp_projector.bin +3 -0
  15. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/optimizer.pt +3 -0
  16. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/rng_state.pth +3 -0
  17. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/scheduler.pt +3 -0
  18. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/special_tokens_map.json +20 -0
  19. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/tokenizer.json +3 -0
  20. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/tokenizer_config.json +44 -0
  21. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/trainer_state.json +0 -0
  22. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/training_args.bin +3 -0
  23. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/vocab.json +0 -0
  24. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/README.md +202 -0
  25. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/adapter_config.json +37 -0
  26. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/adapter_model.safetensors +3 -0
  27. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/added_tokens.json +5 -0
  28. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/config.json +33 -0
  29. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/merges.txt +0 -0
  30. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/mlp_projector.bin +3 -0
  31. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/optimizer.pt +3 -0
  32. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/rng_state.pth +3 -0
  33. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/scheduler.pt +3 -0
  34. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/special_tokens_map.json +20 -0
  35. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/tokenizer.json +3 -0
  36. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/tokenizer_config.json +44 -0
  37. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/trainer_state.json +0 -0
  38. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/training_args.bin +3 -0
  39. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/vocab.json +0 -0
  40. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/README.md +202 -0
  41. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/adapter_config.json +37 -0
  42. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/adapter_model.safetensors +3 -0
  43. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/added_tokens.json +5 -0
  44. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/config.json +33 -0
  45. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/merges.txt +0 -0
  46. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/mlp_projector.bin +3 -0
  47. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/optimizer.pt +3 -0
  48. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/rng_state.pth +3 -0
  49. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/scheduler.pt +3 -0
  50. 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/special_tokens_map.json +20 -0
.gitattributes CHANGED
@@ -229,3 +229,12 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
229
  2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-14000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
230
  2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-14562/tokenizer.json filter=lfs diff=lfs merge=lfs -text
231
  2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/wandb/run-20250409_155335-vtnkssol/run-vtnkssol.wandb filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
229
  2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-14000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
230
  2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-14562/tokenizer.json filter=lfs diff=lfs merge=lfs -text
231
  2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/wandb/run-20250409_155335-vtnkssol/run-vtnkssol.wandb filter=lfs diff=lfs merge=lfs -text
232
+ 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
233
+ 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
234
+ 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
235
+ 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-14658/tokenizer.json filter=lfs diff=lfs merge=lfs -text
236
+ 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-3000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
237
+ 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-5000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
238
+ 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-7000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
239
+ 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-9000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
240
+ 2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/wandb/run-20250421_093901-kdsfeucx/run-kdsfeucx.wandb filter=lfs diff=lfs merge=lfs -text
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/.hydra/config.yaml ADDED
@@ -0,0 +1,116 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: train_llm_lowdim
2
+ _target_: llmbc.workspace.train_llm_workspace.TrainLLMWorkspace
3
+ obs_dim: ${task.obs_dim}
4
+ action_dim: ${task.action_dim}
5
+ horizon: 1
6
+ n_obs_steps: 1
7
+ n_action_steps: 1
8
+ task_name: ${task.name}
9
+ exp_name: train llm
10
+ model_name: ${llm.name}
11
+ use_quantization: ${llm.use_quantization}
12
+ lora_config: ${llm.lora_config}
13
+ dataset:
14
+ test_data_ratio: 0.01
15
+ debug: false
16
+ training:
17
+ seed: 42
18
+ per_device_train_batch_size: 1
19
+ per_device_eval_batch_size: 1
20
+ gradient_accumulation_steps: 32
21
+ optim: paged_adamw_32bit
22
+ num_train_epochs: 3
23
+ eval_strategy: steps
24
+ logging_steps: 1
25
+ warmup_steps: 10
26
+ logging_strategy: steps
27
+ learning_rate: 5.0e-05
28
+ fp16: false
29
+ bf16: true
30
+ tf32: true
31
+ group_by_length: true
32
+ report_to: wandb
33
+ save_steps: 1000
34
+ eval_steps: 10
35
+ use_joint_mlp_projector: ${llm.use_joint_mlp_projector}
36
+ joint_obs_action_mlp_lr: 1.0e-06
37
+ trainer:
38
+ obs_dim: ${obs_dim}
39
+ action_dim: ${action_dim}
40
+ use_joint_mlp_projector: ${llm.use_joint_mlp_projector}
41
+ max_seq_length: ${llm.max_length}
42
+ dataset_text_field: text
43
+ packing: false
44
+ logging:
45
+ project: llm_module_finetuning
46
+ resume: true
47
+ mode: online
48
+ name: ${now:%Y.%m.%d-%H.%M.%S}_${name}_${task_name}
49
+ tags:
50
+ - ${name}
51
+ - ${task_name}
52
+ - ${exp_name}
53
+ id: null
54
+ group: null
55
+ multi_run:
56
+ run_dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
57
+ wandb_name_base: ${now:%Y.%m.%d-%H.%M.%S}_${name}_${task_name}
58
+ task:
59
+ name: blockpush_lowdim_seed
60
+ obs_dim: 16
61
+ action_dim: 2
62
+ keypoint_dim: 2
63
+ obs_eef_target: true
64
+ env_runner:
65
+ _target_: llmbc.env_runner.blockpushing_lowdim_runner.BlockPushingLowdimRunner
66
+ env_name: llf-blockpushing-BlockPushMultimodal-v0
67
+ instruction_type: b
68
+ feedback_type:
69
+ - hp
70
+ - hn
71
+ - fp
72
+ visual: false
73
+ max_steps: 350
74
+ n_obs_steps: ${n_obs_steps}
75
+ n_action_steps: ${n_action_steps}
76
+ dataset:
77
+ _target_: llmbc.dataset.blockpush_lowdim_dataset.BlockPushLowdimDataset
78
+ data_path: datasets/BlockPushMultimodal-v0-general.pt
79
+ data_path2: datasets/BlockPushMultimodal-v0.pt
80
+ horizon: ${horizon}
81
+ pad_before: ${eval:'${n_obs_steps}-1'}
82
+ pad_after: ${eval:'${n_action_steps}-1'}
83
+ obs_eef_target: ${task.obs_eef_target}
84
+ use_manual_normalizer: false
85
+ seed: 42
86
+ val_ratio: 0.02
87
+ dummy_normalizer: false
88
+ instructor:
89
+ _target_: llmbc.translator.instructor.blockpush_instructor.BlockPushMultimodal_v0_instructor.BlockPushMultimodalV0Instructor
90
+ llm:
91
+ name: Qwen/Qwen2-0.5B-Instruct
92
+ model_name: Qwen2-0.5B-Instruct
93
+ use_quantization: false
94
+ load_from_checkpoint: false
95
+ adaptor_path: /home/chyang/workspace/LLM-BC/data/outputs/2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-14562
96
+ use_orig_model: false
97
+ use_joint_mlp_projector: true
98
+ load_from_mlp_projector_checkpoint: true
99
+ mlp_projector_checkpoint_path: /home/chyang/workspace/LLM-BC/data/outputs/2025.04.20/16.50.02_train_mlp_projector_blockpush_lowdim_seed/checkpoints/latest.ckpt
100
+ max_length: 100
101
+ config_target: llmbc.model.llm.qwen2_lowdim_model.LowdimQwen2Config
102
+ causal_lm_target: llmbc.model.llm.qwen2_lowdim_model.LowdimQwen2ForCausalLM
103
+ lora_config:
104
+ r: 16
105
+ lora_alpha: 32
106
+ lora_dropout: 0.05
107
+ bias: none
108
+ task_type: CAUSAL_LM
109
+ prompter:
110
+ _target_: llmbc.translator.prompter.qwen2_prompter.Qwen2Prompter
111
+ use_joint_mlp_projector: true
112
+ hydra:
113
+ job:
114
+ override_dirname: ${model_name}
115
+ run:
116
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${model_name}
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/.hydra/hydra.yaml ADDED
@@ -0,0 +1,156 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
4
+ sweep:
5
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task: []
115
+ job:
116
+ name: train_llm_workspace
117
+ chdir: null
118
+ override_dirname: ''
119
+ id: ???
120
+ num: ???
121
+ config_name: train_llm_workspace
122
+ env_set: {}
123
+ env_copy: []
124
+ config:
125
+ override_dirname:
126
+ kv_sep: '='
127
+ item_sep: ','
128
+ exclude_keys: []
129
+ runtime:
130
+ version: 1.2.0
131
+ version_base: '1.2'
132
+ cwd: /home/chyang/workspace/LLM-BC
133
+ config_sources:
134
+ - path: hydra.conf
135
+ schema: pkg
136
+ provider: hydra
137
+ - path: /home/chyang/workspace/LLM-BC/llmbc/config
138
+ schema: file
139
+ provider: main
140
+ - path: ''
141
+ schema: structured
142
+ provider: schema
143
+ output_dir: /home/chyang/workspace/LLM-BC/data/outputs/2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed
144
+ choices:
145
+ llm: qwen-2-0.5b-instruct
146
+ task: BlockPushMultimodal-v0
147
+ hydra/env: default
148
+ hydra/callbacks: null
149
+ hydra/job_logging: default
150
+ hydra/hydra_logging: default
151
+ hydra/hydra_help: default
152
+ hydra/help: default
153
+ hydra/sweeper: basic
154
+ hydra/launcher: basic
155
+ hydra/output: default
156
+ verbose: false
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/.hydra/overrides.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ []
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen2-0.5B-Instruct
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.14.0
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/adapter_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
5
+ "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layer_replication": null,
12
+ "layers_pattern": null,
13
+ "layers_to_transform": null,
14
+ "loftq_config": {},
15
+ "lora_alpha": 32,
16
+ "lora_bias": false,
17
+ "lora_dropout": 0.05,
18
+ "megatron_config": null,
19
+ "megatron_core": "megatron.core",
20
+ "modules_to_save": null,
21
+ "peft_type": "LORA",
22
+ "r": 16,
23
+ "rank_pattern": {},
24
+ "revision": null,
25
+ "target_modules": [
26
+ "gate_proj",
27
+ "k_proj",
28
+ "v_proj",
29
+ "q_proj",
30
+ "down_proj",
31
+ "up_proj",
32
+ "o_proj"
33
+ ],
34
+ "task_type": "CAUSAL_LM",
35
+ "use_dora": false,
36
+ "use_rslora": false
37
+ }
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:254048edc2a383e2f181b205ae0523d9aff1f285c616c36b1746f5e172d2ec6d
3
+ size 35237104
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen2-0.5B-Instruct
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.14.0
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/adapter_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
5
+ "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layer_replication": null,
12
+ "layers_pattern": null,
13
+ "layers_to_transform": null,
14
+ "loftq_config": {},
15
+ "lora_alpha": 32,
16
+ "lora_bias": false,
17
+ "lora_dropout": 0.05,
18
+ "megatron_config": null,
19
+ "megatron_core": "megatron.core",
20
+ "modules_to_save": null,
21
+ "peft_type": "LORA",
22
+ "r": 16,
23
+ "rank_pattern": {},
24
+ "revision": null,
25
+ "target_modules": [
26
+ "gate_proj",
27
+ "k_proj",
28
+ "v_proj",
29
+ "q_proj",
30
+ "down_proj",
31
+ "up_proj",
32
+ "o_proj"
33
+ ],
34
+ "task_type": "CAUSAL_LM",
35
+ "use_dora": false,
36
+ "use_rslora": false
37
+ }
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6f7a109d006108deeab42c34f30ff2ecd0a893d9249a0cf2215a4685502fb78
3
+ size 35237104
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
4
+ "action_dim": 2,
5
+ "architectures": [
6
+ "Qwen2ForCausalLM"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "bos_token_id": 151643,
10
+ "eos_token_id": 151645,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 896,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 4864,
15
+ "max_position_embeddings": 32768,
16
+ "max_window_layers": 24,
17
+ "model_type": "qwen2_lowdim",
18
+ "num_attention_heads": 14,
19
+ "num_hidden_layers": 24,
20
+ "num_key_value_heads": 2,
21
+ "obs_dim": 16,
22
+ "rms_norm_eps": 1e-06,
23
+ "rope_scaling": null,
24
+ "rope_theta": 1000000.0,
25
+ "sliding_window": null,
26
+ "tie_word_embeddings": true,
27
+ "torch_dtype": "bfloat16",
28
+ "transformers_version": "4.47.1",
29
+ "use_cache": false,
30
+ "use_joint_mlp_projector": true,
31
+ "use_sliding_window": false,
32
+ "vocab_size": 151936
33
+ }
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/mlp_projector.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ccd97cd9fa829329f834ffb83a6cbfe6dc49c1826315812d2a7625f0984b514
3
+ size 3285184
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c02afd49955364ea22b159a382df29d2e94aa30729df96a9d8bf56d4d014772
3
+ size 77153338
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:301ce654de00e3d4781e43abb88fc593cf04d760d0f11fc436a64c6513dbcdc3
3
+ size 14244
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e2d60bb72b0c0a1afca673ac1f16a5fd40f24dc867a5696aeda8c45b1ced56f
3
+ size 1064
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|im_end|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "pad_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ }
20
+ }
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcfe42da0a4497e8b2b172c1f9f4ec423a46dc12907f4349c55025f670422ba9
3
+ size 11418266
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/tokenizer_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
+ "clean_up_tokenization_spaces": false,
36
+ "eos_token": "<|im_end|>",
37
+ "errors": "replace",
38
+ "extra_special_tokens": {},
39
+ "model_max_length": 32768,
40
+ "pad_token": "<|endoftext|>",
41
+ "split_special_tokens": false,
42
+ "tokenizer_class": "Qwen2Tokenizer",
43
+ "unk_token": null
44
+ }
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ed0c811af16994cd40dd5048018038dea5d339fa52de63dfe4a4e30f62734df
3
+ size 6008
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-1000/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen2-0.5B-Instruct
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.14.0
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/adapter_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
5
+ "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layer_replication": null,
12
+ "layers_pattern": null,
13
+ "layers_to_transform": null,
14
+ "loftq_config": {},
15
+ "lora_alpha": 32,
16
+ "lora_bias": false,
17
+ "lora_dropout": 0.05,
18
+ "megatron_config": null,
19
+ "megatron_core": "megatron.core",
20
+ "modules_to_save": null,
21
+ "peft_type": "LORA",
22
+ "r": 16,
23
+ "rank_pattern": {},
24
+ "revision": null,
25
+ "target_modules": [
26
+ "gate_proj",
27
+ "k_proj",
28
+ "v_proj",
29
+ "q_proj",
30
+ "down_proj",
31
+ "up_proj",
32
+ "o_proj"
33
+ ],
34
+ "task_type": "CAUSAL_LM",
35
+ "use_dora": false,
36
+ "use_rslora": false
37
+ }
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd1f9184e723ff57c1a701218e755f47e2c82cac78c9a89ac7d89bb4c9576629
3
+ size 35237104
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
4
+ "action_dim": 2,
5
+ "architectures": [
6
+ "Qwen2ForCausalLM"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "bos_token_id": 151643,
10
+ "eos_token_id": 151645,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 896,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 4864,
15
+ "max_position_embeddings": 32768,
16
+ "max_window_layers": 24,
17
+ "model_type": "qwen2_lowdim",
18
+ "num_attention_heads": 14,
19
+ "num_hidden_layers": 24,
20
+ "num_key_value_heads": 2,
21
+ "obs_dim": 16,
22
+ "rms_norm_eps": 1e-06,
23
+ "rope_scaling": null,
24
+ "rope_theta": 1000000.0,
25
+ "sliding_window": null,
26
+ "tie_word_embeddings": true,
27
+ "torch_dtype": "bfloat16",
28
+ "transformers_version": "4.47.1",
29
+ "use_cache": false,
30
+ "use_joint_mlp_projector": true,
31
+ "use_sliding_window": false,
32
+ "vocab_size": 151936
33
+ }
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/mlp_projector.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bed0b4f82594db4bfd2c5a79a91194beaf17176b1f7f5b9e3279fd64eb5ee9ba
3
+ size 3285184
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4155fea7637ad668e320e4288e4e25d4d88e04b85849ca94224328df10e4b8c8
3
+ size 77153338
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0967e7342b73cc8d48d6c7fd516c0e8f3f6c9d016138bb7ba443fa4ebc7f3593
3
+ size 14244
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f462e2a7559cc4be212e52a0c0a754b462be6a5610c7e342cbac42803c3b8300
3
+ size 1064
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|im_end|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "pad_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ }
20
+ }
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcfe42da0a4497e8b2b172c1f9f4ec423a46dc12907f4349c55025f670422ba9
3
+ size 11418266
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/tokenizer_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
+ "clean_up_tokenization_spaces": false,
36
+ "eos_token": "<|im_end|>",
37
+ "errors": "replace",
38
+ "extra_special_tokens": {},
39
+ "model_max_length": 32768,
40
+ "pad_token": "<|endoftext|>",
41
+ "split_special_tokens": false,
42
+ "tokenizer_class": "Qwen2Tokenizer",
43
+ "unk_token": null
44
+ }
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ed0c811af16994cd40dd5048018038dea5d339fa52de63dfe4a4e30f62734df
3
+ size 6008
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen2-0.5B-Instruct
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.14.0
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/adapter_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
5
+ "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layer_replication": null,
12
+ "layers_pattern": null,
13
+ "layers_to_transform": null,
14
+ "loftq_config": {},
15
+ "lora_alpha": 32,
16
+ "lora_bias": false,
17
+ "lora_dropout": 0.05,
18
+ "megatron_config": null,
19
+ "megatron_core": "megatron.core",
20
+ "modules_to_save": null,
21
+ "peft_type": "LORA",
22
+ "r": 16,
23
+ "rank_pattern": {},
24
+ "revision": null,
25
+ "target_modules": [
26
+ "gate_proj",
27
+ "k_proj",
28
+ "v_proj",
29
+ "q_proj",
30
+ "down_proj",
31
+ "up_proj",
32
+ "o_proj"
33
+ ],
34
+ "task_type": "CAUSAL_LM",
35
+ "use_dora": false,
36
+ "use_rslora": false
37
+ }
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5454c5eb42acfa48b0c8836c010c6ec73e45fc784b116dcac11e7a20fb04419e
3
+ size 35237104
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
4
+ "action_dim": 2,
5
+ "architectures": [
6
+ "Qwen2ForCausalLM"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "bos_token_id": 151643,
10
+ "eos_token_id": 151645,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 896,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 4864,
15
+ "max_position_embeddings": 32768,
16
+ "max_window_layers": 24,
17
+ "model_type": "qwen2_lowdim",
18
+ "num_attention_heads": 14,
19
+ "num_hidden_layers": 24,
20
+ "num_key_value_heads": 2,
21
+ "obs_dim": 16,
22
+ "rms_norm_eps": 1e-06,
23
+ "rope_scaling": null,
24
+ "rope_theta": 1000000.0,
25
+ "sliding_window": null,
26
+ "tie_word_embeddings": true,
27
+ "torch_dtype": "bfloat16",
28
+ "transformers_version": "4.47.1",
29
+ "use_cache": false,
30
+ "use_joint_mlp_projector": true,
31
+ "use_sliding_window": false,
32
+ "vocab_size": 151936
33
+ }
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/mlp_projector.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13c5154b75cb759ae0a844dc1342e0b55d3f5305efd03acc0ae79f055c406799
3
+ size 3285184
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c13bf6abdfd0f103b85908e8c78c60188d882b696495c52f0fc970b6fcb65bc3
3
+ size 77153338
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e35de1aa4a7a0c70cf7926eb01cbfb1609e6a8b5d8df841aa1ac21e0e8be7a9
3
+ size 14244
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b73c5a16b59ac31281693ac5c6e06d8b5a19d91245e20285c7f08976f93c578d
3
+ size 1064
2025.04.21/09.39.00_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|im_end|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "pad_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ }
20
+ }