CHYang25 commited on
Commit
32468ac
·
verified ·
1 Parent(s): ea9a3be

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +6 -0
  2. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/.hydra/config.yaml +116 -0
  3. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/.hydra/hydra.yaml +156 -0
  4. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/.hydra/overrides.yaml +1 -0
  5. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/README.md +202 -0
  6. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/adapter_config.json +37 -0
  7. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/adapter_model.safetensors +3 -0
  8. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/README.md +202 -0
  9. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/adapter_config.json +37 -0
  10. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/adapter_model.safetensors +3 -0
  11. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/added_tokens.json +5 -0
  12. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/config.json +33 -0
  13. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/merges.txt +0 -0
  14. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/mlp_projector.bin +3 -0
  15. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/optimizer.pt +3 -0
  16. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/rng_state.pth +3 -0
  17. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/scheduler.pt +3 -0
  18. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/special_tokens_map.json +20 -0
  19. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/tokenizer.json +3 -0
  20. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/tokenizer_config.json +44 -0
  21. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/trainer_state.json +0 -0
  22. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/training_args.bin +3 -0
  23. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/vocab.json +0 -0
  24. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/README.md +202 -0
  25. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/adapter_config.json +37 -0
  26. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/adapter_model.safetensors +3 -0
  27. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/added_tokens.json +5 -0
  28. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/config.json +33 -0
  29. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/merges.txt +0 -0
  30. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/mlp_projector.bin +3 -0
  31. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/optimizer.pt +3 -0
  32. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/rng_state.pth +3 -0
  33. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/scheduler.pt +3 -0
  34. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/special_tokens_map.json +20 -0
  35. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/tokenizer.json +3 -0
  36. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/tokenizer_config.json +44 -0
  37. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/trainer_state.json +0 -0
  38. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/training_args.bin +3 -0
  39. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/vocab.json +0 -0
  40. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/README.md +202 -0
  41. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/adapter_config.json +37 -0
  42. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/adapter_model.safetensors +3 -0
  43. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/added_tokens.json +5 -0
  44. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/config.json +33 -0
  45. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/merges.txt +0 -0
  46. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/mlp_projector.bin +3 -0
  47. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/optimizer.pt +3 -0
  48. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/rng_state.pth +3 -0
  49. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/scheduler.pt +3 -0
  50. 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/special_tokens_map.json +20 -0
.gitattributes CHANGED
@@ -223,3 +223,9 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
223
  2025.03.28/22.12.17_train_llm_lowdim_blockpush_lowdim_seed/wandb/run-20250328_221218-xr6q3ag2/run-xr6q3ag2.wandb filter=lfs diff=lfs merge=lfs -text
224
  2025.04.07/23.19.20_train_mlp_projector_blockpush_lowdim_seed/logs.json.txt filter=lfs diff=lfs merge=lfs -text
225
  2025.04.07/23.19.20_train_mlp_projector_blockpush_lowdim_seed/wandb/run-20250407_231925-7vg30p6j/run-7vg30p6j.wandb filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
223
  2025.03.28/22.12.17_train_llm_lowdim_blockpush_lowdim_seed/wandb/run-20250328_221218-xr6q3ag2/run-xr6q3ag2.wandb filter=lfs diff=lfs merge=lfs -text
224
  2025.04.07/23.19.20_train_mlp_projector_blockpush_lowdim_seed/logs.json.txt filter=lfs diff=lfs merge=lfs -text
225
  2025.04.07/23.19.20_train_mlp_projector_blockpush_lowdim_seed/wandb/run-20250407_231925-7vg30p6j/run-7vg30p6j.wandb filter=lfs diff=lfs merge=lfs -text
226
+ 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
227
+ 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
228
+ 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
229
+ 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-14000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
230
+ 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-14562/tokenizer.json filter=lfs diff=lfs merge=lfs -text
231
+ 2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/wandb/run-20250409_155335-vtnkssol/run-vtnkssol.wandb filter=lfs diff=lfs merge=lfs -text
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/.hydra/config.yaml ADDED
@@ -0,0 +1,116 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: train_llm_lowdim
2
+ _target_: llmbc.workspace.train_llm_workspace.TrainLLMWorkspace
3
+ obs_dim: ${task.obs_dim}
4
+ action_dim: ${task.action_dim}
5
+ horizon: 1
6
+ n_obs_steps: 1
7
+ n_action_steps: 1
8
+ task_name: ${task.name}
9
+ exp_name: train llm
10
+ model_name: ${llm.name}
11
+ use_quantization: ${llm.use_quantization}
12
+ lora_config: ${llm.lora_config}
13
+ dataset:
14
+ test_data_ratio: 0.01
15
+ debug: false
16
+ training:
17
+ seed: 42
18
+ per_device_train_batch_size: 1
19
+ per_device_eval_batch_size: 1
20
+ gradient_accumulation_steps: 32
21
+ optim: paged_adamw_32bit
22
+ num_train_epochs: 3
23
+ eval_strategy: steps
24
+ logging_steps: 1
25
+ warmup_steps: 10
26
+ logging_strategy: steps
27
+ learning_rate: 5.0e-05
28
+ fp16: false
29
+ bf16: true
30
+ tf32: true
31
+ group_by_length: true
32
+ report_to: wandb
33
+ save_steps: 1000
34
+ eval_steps: 10
35
+ use_joint_mlp_projector: ${llm.use_joint_mlp_projector}
36
+ joint_obs_action_mlp_lr: 1.0e-06
37
+ trainer:
38
+ obs_dim: ${obs_dim}
39
+ action_dim: ${action_dim}
40
+ use_joint_mlp_projector: ${llm.use_joint_mlp_projector}
41
+ max_seq_length: ${llm.max_length}
42
+ dataset_text_field: text
43
+ packing: false
44
+ logging:
45
+ project: llm_module_finetuning
46
+ resume: true
47
+ mode: online
48
+ name: ${now:%Y.%m.%d-%H.%M.%S}_${name}_${task_name}
49
+ tags:
50
+ - ${name}
51
+ - ${task_name}
52
+ - ${exp_name}
53
+ id: null
54
+ group: null
55
+ multi_run:
56
+ run_dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
57
+ wandb_name_base: ${now:%Y.%m.%d-%H.%M.%S}_${name}_${task_name}
58
+ task:
59
+ name: blockpush_lowdim_seed
60
+ obs_dim: 16
61
+ action_dim: 2
62
+ keypoint_dim: 2
63
+ obs_eef_target: true
64
+ env_runner:
65
+ _target_: llmbc.env_runner.blockpushing_lowdim_runner.BlockPushingLowdimRunner
66
+ env_name: llf-blockpushing-BlockPushMultimodal-v0
67
+ instruction_type: b
68
+ feedback_type:
69
+ - hp
70
+ - hn
71
+ - fp
72
+ visual: false
73
+ max_steps: 350
74
+ n_obs_steps: ${n_obs_steps}
75
+ n_action_steps: ${n_action_steps}
76
+ dataset:
77
+ _target_: llmbc.dataset.blockpush_lowdim_dataset.BlockPushLowdimDataset
78
+ data_path: datasets/BlockPushMultimodal-v0-general.pt
79
+ data_path2: datasets/BlockPushMultimodal-v0.pt
80
+ horizon: ${horizon}
81
+ pad_before: ${eval:'${n_obs_steps}-1'}
82
+ pad_after: ${eval:'${n_action_steps}-1'}
83
+ obs_eef_target: ${task.obs_eef_target}
84
+ use_manual_normalizer: false
85
+ seed: 42
86
+ val_ratio: 0.02
87
+ dummy_normalizer: false
88
+ instructor:
89
+ _target_: llmbc.translator.instructor.blockpush_instructor.BlockPushMultimodal_v0_instructor.BlockPushMultimodalV0Instructor
90
+ llm:
91
+ name: Qwen/Qwen2-0.5B-Instruct
92
+ model_name: Qwen2-0.5B-Instruct
93
+ use_quantization: false
94
+ load_from_checkpoint: false
95
+ adaptor_path: /home/chyang/workspace/LLM-BC/data/outputs/2025.04.07/12.55.38_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-4000
96
+ use_orig_model: false
97
+ use_joint_mlp_projector: true
98
+ load_from_mlp_projector_checkpoint: true
99
+ mlp_projector_checkpoint_path: /home/chyang/workspace/LLM-BC/data/outputs/2025.04.07/23.19.20_train_mlp_projector_blockpush_lowdim_seed/checkpoints/latest.ckpt
100
+ max_length: 100
101
+ config_target: llmbc.model.llm.qwen2_lowdim_model.LowdimQwen2Config
102
+ causal_lm_target: llmbc.model.llm.qwen2_lowdim_model.LowdimQwen2ForCausalLM
103
+ lora_config:
104
+ r: 64
105
+ lora_alpha: 128
106
+ lora_dropout: 0.05
107
+ bias: none
108
+ task_type: CAUSAL_LM
109
+ prompter:
110
+ _target_: llmbc.translator.prompter.qwen2_prompter.Qwen2Prompter
111
+ use_joint_mlp_projector: true
112
+ hydra:
113
+ job:
114
+ override_dirname: ${model_name}
115
+ run:
116
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${model_name}
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/.hydra/hydra.yaml ADDED
@@ -0,0 +1,156 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
4
+ sweep:
5
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task: []
115
+ job:
116
+ name: train_llm_workspace
117
+ chdir: null
118
+ override_dirname: ''
119
+ id: ???
120
+ num: ???
121
+ config_name: train_llm_workspace
122
+ env_set: {}
123
+ env_copy: []
124
+ config:
125
+ override_dirname:
126
+ kv_sep: '='
127
+ item_sep: ','
128
+ exclude_keys: []
129
+ runtime:
130
+ version: 1.2.0
131
+ version_base: '1.2'
132
+ cwd: /home/chyang/workspace/LLM-BC
133
+ config_sources:
134
+ - path: hydra.conf
135
+ schema: pkg
136
+ provider: hydra
137
+ - path: /home/chyang/workspace/LLM-BC/llmbc/config
138
+ schema: file
139
+ provider: main
140
+ - path: ''
141
+ schema: structured
142
+ provider: schema
143
+ output_dir: /home/chyang/workspace/LLM-BC/data/outputs/2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed
144
+ choices:
145
+ llm: qwen-2-0.5b-instruct
146
+ task: BlockPushMultimodal-v0
147
+ hydra/env: default
148
+ hydra/callbacks: null
149
+ hydra/job_logging: default
150
+ hydra/hydra_logging: default
151
+ hydra/hydra_help: default
152
+ hydra/help: default
153
+ hydra/sweeper: basic
154
+ hydra/launcher: basic
155
+ hydra/output: default
156
+ verbose: false
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/.hydra/overrides.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ []
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen2-0.5B-Instruct
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.14.0
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/adapter_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
5
+ "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layer_replication": null,
12
+ "layers_pattern": null,
13
+ "layers_to_transform": null,
14
+ "loftq_config": {},
15
+ "lora_alpha": 128,
16
+ "lora_bias": false,
17
+ "lora_dropout": 0.05,
18
+ "megatron_config": null,
19
+ "megatron_core": "megatron.core",
20
+ "modules_to_save": null,
21
+ "peft_type": "LORA",
22
+ "r": 64,
23
+ "rank_pattern": {},
24
+ "revision": null,
25
+ "target_modules": [
26
+ "down_proj",
27
+ "k_proj",
28
+ "q_proj",
29
+ "up_proj",
30
+ "o_proj",
31
+ "v_proj",
32
+ "gate_proj"
33
+ ],
34
+ "task_type": "CAUSAL_LM",
35
+ "use_dora": false,
36
+ "use_rslora": false
37
+ }
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e25c85589660804a03363df61ee0de2eae38fedef42dbc80c230e627b2b2c04
3
+ size 140815952
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen2-0.5B-Instruct
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.14.0
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/adapter_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
5
+ "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layer_replication": null,
12
+ "layers_pattern": null,
13
+ "layers_to_transform": null,
14
+ "loftq_config": {},
15
+ "lora_alpha": 128,
16
+ "lora_bias": false,
17
+ "lora_dropout": 0.05,
18
+ "megatron_config": null,
19
+ "megatron_core": "megatron.core",
20
+ "modules_to_save": null,
21
+ "peft_type": "LORA",
22
+ "r": 64,
23
+ "rank_pattern": {},
24
+ "revision": null,
25
+ "target_modules": [
26
+ "down_proj",
27
+ "k_proj",
28
+ "q_proj",
29
+ "up_proj",
30
+ "o_proj",
31
+ "v_proj",
32
+ "gate_proj"
33
+ ],
34
+ "task_type": "CAUSAL_LM",
35
+ "use_dora": false,
36
+ "use_rslora": false
37
+ }
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:381770d496029197905cd45369aadaa45f5934880c2d1acd44f1fa73ec76cfd3
3
+ size 140815952
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
4
+ "action_dim": 2,
5
+ "architectures": [
6
+ "Qwen2ForCausalLM"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "bos_token_id": 151643,
10
+ "eos_token_id": 151645,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 896,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 4864,
15
+ "max_position_embeddings": 32768,
16
+ "max_window_layers": 24,
17
+ "model_type": "qwen2_lowdim",
18
+ "num_attention_heads": 14,
19
+ "num_hidden_layers": 24,
20
+ "num_key_value_heads": 2,
21
+ "obs_dim": 16,
22
+ "rms_norm_eps": 1e-06,
23
+ "rope_scaling": null,
24
+ "rope_theta": 1000000.0,
25
+ "sliding_window": null,
26
+ "tie_word_embeddings": true,
27
+ "torch_dtype": "bfloat16",
28
+ "transformers_version": "4.47.1",
29
+ "use_cache": false,
30
+ "use_joint_mlp_projector": true,
31
+ "use_sliding_window": false,
32
+ "vocab_size": 151936
33
+ }
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/mlp_projector.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc53d1ee503a33da68296175178d08b9cb732f4060a9f6f3bff753f8c672f010
3
+ size 3285184
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ecca18083f3eb9c65ebf61efac4f19b7e5f3f7b6a60eaee6c775e87e643235b
3
+ size 288312058
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1710bacb8bab1f9789a3d97616d96b84112ef927751356e0c10d06430efd370f
3
+ size 14244
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1b18dab89614bee93a9f4e2e5aa2557c4e025d89cf18feeb2f38436d86185cb
3
+ size 1064
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|im_end|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "pad_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ }
20
+ }
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcfe42da0a4497e8b2b172c1f9f4ec423a46dc12907f4349c55025f670422ba9
3
+ size 11418266
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/tokenizer_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
+ "clean_up_tokenization_spaces": false,
36
+ "eos_token": "<|im_end|>",
37
+ "errors": "replace",
38
+ "extra_special_tokens": {},
39
+ "model_max_length": 32768,
40
+ "pad_token": "<|endoftext|>",
41
+ "split_special_tokens": false,
42
+ "tokenizer_class": "Qwen2Tokenizer",
43
+ "unk_token": null
44
+ }
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:119779b0b00171e3898c42ab56716b6de6a610665be886335f58be331fbef038
3
+ size 6008
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-11000/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen2-0.5B-Instruct
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.14.0
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/adapter_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
5
+ "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layer_replication": null,
12
+ "layers_pattern": null,
13
+ "layers_to_transform": null,
14
+ "loftq_config": {},
15
+ "lora_alpha": 128,
16
+ "lora_bias": false,
17
+ "lora_dropout": 0.05,
18
+ "megatron_config": null,
19
+ "megatron_core": "megatron.core",
20
+ "modules_to_save": null,
21
+ "peft_type": "LORA",
22
+ "r": 64,
23
+ "rank_pattern": {},
24
+ "revision": null,
25
+ "target_modules": [
26
+ "down_proj",
27
+ "k_proj",
28
+ "q_proj",
29
+ "up_proj",
30
+ "o_proj",
31
+ "v_proj",
32
+ "gate_proj"
33
+ ],
34
+ "task_type": "CAUSAL_LM",
35
+ "use_dora": false,
36
+ "use_rslora": false
37
+ }
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:726815b2b42e2c2ec14aec643abd1920fc8eabf554798a879b28a58cb5c58cde
3
+ size 140815952
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
4
+ "action_dim": 2,
5
+ "architectures": [
6
+ "Qwen2ForCausalLM"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "bos_token_id": 151643,
10
+ "eos_token_id": 151645,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 896,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 4864,
15
+ "max_position_embeddings": 32768,
16
+ "max_window_layers": 24,
17
+ "model_type": "qwen2_lowdim",
18
+ "num_attention_heads": 14,
19
+ "num_hidden_layers": 24,
20
+ "num_key_value_heads": 2,
21
+ "obs_dim": 16,
22
+ "rms_norm_eps": 1e-06,
23
+ "rope_scaling": null,
24
+ "rope_theta": 1000000.0,
25
+ "sliding_window": null,
26
+ "tie_word_embeddings": true,
27
+ "torch_dtype": "bfloat16",
28
+ "transformers_version": "4.47.1",
29
+ "use_cache": false,
30
+ "use_joint_mlp_projector": true,
31
+ "use_sliding_window": false,
32
+ "vocab_size": 151936
33
+ }
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/mlp_projector.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3cd90ce8f97ac9a2da1cfd879557ba62e9848f9ce64bf335a977e4c5d9271dc
3
+ size 3285184
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66dcc1514d426ee59f903a3bbb14c22ef6b8162526c39c360ef331452c90a5e5
3
+ size 288312058
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e9834a0327328cbe4998064de1c68968d9efa1100c7d78ef4b27f54c02f2917
3
+ size 14244
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a78b8f1b0140c3bd431e8bb086bde8c6605dded4d914e74610bf50022557e05c
3
+ size 1064
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|im_end|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "pad_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ }
20
+ }
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcfe42da0a4497e8b2b172c1f9f4ec423a46dc12907f4349c55025f670422ba9
3
+ size 11418266
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/tokenizer_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
+ "clean_up_tokenization_spaces": false,
36
+ "eos_token": "<|im_end|>",
37
+ "errors": "replace",
38
+ "extra_special_tokens": {},
39
+ "model_max_length": 32768,
40
+ "pad_token": "<|endoftext|>",
41
+ "split_special_tokens": false,
42
+ "tokenizer_class": "Qwen2Tokenizer",
43
+ "unk_token": null
44
+ }
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:119779b0b00171e3898c42ab56716b6de6a610665be886335f58be331fbef038
3
+ size 6008
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-12000/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen2-0.5B-Instruct
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.14.0
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/adapter_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
5
+ "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layer_replication": null,
12
+ "layers_pattern": null,
13
+ "layers_to_transform": null,
14
+ "loftq_config": {},
15
+ "lora_alpha": 128,
16
+ "lora_bias": false,
17
+ "lora_dropout": 0.05,
18
+ "megatron_config": null,
19
+ "megatron_core": "megatron.core",
20
+ "modules_to_save": null,
21
+ "peft_type": "LORA",
22
+ "r": 64,
23
+ "rank_pattern": {},
24
+ "revision": null,
25
+ "target_modules": [
26
+ "down_proj",
27
+ "k_proj",
28
+ "q_proj",
29
+ "up_proj",
30
+ "o_proj",
31
+ "v_proj",
32
+ "gate_proj"
33
+ ],
34
+ "task_type": "CAUSAL_LM",
35
+ "use_dora": false,
36
+ "use_rslora": false
37
+ }
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:720170f61aa1d58c68511513d7c6f19d32ba6b5385cbc37bd528a9c147983913
3
+ size 140815952
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
4
+ "action_dim": 2,
5
+ "architectures": [
6
+ "Qwen2ForCausalLM"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "bos_token_id": 151643,
10
+ "eos_token_id": 151645,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 896,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 4864,
15
+ "max_position_embeddings": 32768,
16
+ "max_window_layers": 24,
17
+ "model_type": "qwen2_lowdim",
18
+ "num_attention_heads": 14,
19
+ "num_hidden_layers": 24,
20
+ "num_key_value_heads": 2,
21
+ "obs_dim": 16,
22
+ "rms_norm_eps": 1e-06,
23
+ "rope_scaling": null,
24
+ "rope_theta": 1000000.0,
25
+ "sliding_window": null,
26
+ "tie_word_embeddings": true,
27
+ "torch_dtype": "bfloat16",
28
+ "transformers_version": "4.47.1",
29
+ "use_cache": false,
30
+ "use_joint_mlp_projector": true,
31
+ "use_sliding_window": false,
32
+ "vocab_size": 151936
33
+ }
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/mlp_projector.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75a47e450da8631a60cfa5ef7c0c1f869949147684fb6618440a765b6a84bc65
3
+ size 3285184
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:625001e9cc539b7266a1e72c47c68b5305327fc4e92d88a66fbedb77ae2125b8
3
+ size 288312058
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6fc8b5b9c429d2bd7d427e7b3088118380a396fb214ce44f8a23859b42678f3
3
+ size 14244
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d8e818bc34b8785efc085a667b85940085581e7d52bbc2f57a06af0cb7baec3
3
+ size 1064
2025.04.09/15.53.34_train_llm_lowdim_blockpush_lowdim_seed/Qwen/Qwen2-0.5B-Instruct-finetuned-blockpush_lowdim_seed/checkpoint-13000/special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|im_end|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "pad_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ }
20
+ }