CHYang25 commited on
Commit
8429b42
·
verified ·
1 Parent(s): 4423dd5

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +7 -0
  2. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/.hydra/config.yaml +115 -0
  3. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/.hydra/hydra.yaml +154 -0
  4. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/.hydra/overrides.yaml +1 -0
  5. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/added_tokens.json +5 -0
  6. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/config.json +32 -0
  7. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/generation_config.json +14 -0
  8. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/merges.txt +0 -0
  9. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/mlp_projector.bin +3 -0
  10. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/model.safetensors +3 -0
  11. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/optimizer.pt +3 -0
  12. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/rng_state.pth +3 -0
  13. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/scheduler.pt +3 -0
  14. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/special_tokens_map.json +20 -0
  15. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/tokenizer.json +3 -0
  16. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/tokenizer_config.json +44 -0
  17. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/trainer_state.json +0 -0
  18. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/training_args.bin +3 -0
  19. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/vocab.json +0 -0
  20. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/added_tokens.json +5 -0
  21. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/config.json +32 -0
  22. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/generation_config.json +14 -0
  23. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/merges.txt +0 -0
  24. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/mlp_projector.bin +3 -0
  25. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/model.safetensors +3 -0
  26. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/optimizer.pt +3 -0
  27. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/rng_state.pth +3 -0
  28. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/scheduler.pt +3 -0
  29. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/special_tokens_map.json +20 -0
  30. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/tokenizer.json +3 -0
  31. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/tokenizer_config.json +44 -0
  32. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/trainer_state.json +0 -0
  33. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/training_args.bin +3 -0
  34. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/vocab.json +0 -0
  35. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/added_tokens.json +5 -0
  36. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/config.json +32 -0
  37. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/generation_config.json +14 -0
  38. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/merges.txt +0 -0
  39. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/mlp_projector.bin +3 -0
  40. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/model.safetensors +3 -0
  41. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/optimizer.pt +3 -0
  42. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/rng_state.pth +3 -0
  43. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/scheduler.pt +3 -0
  44. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/special_tokens_map.json +20 -0
  45. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/tokenizer.json +3 -0
  46. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/tokenizer_config.json +44 -0
  47. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/trainer_state.json +0 -0
  48. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/training_args.bin +3 -0
  49. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/vocab.json +0 -0
  50. 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-5000/added_tokens.json +5 -0
.gitattributes CHANGED
@@ -107,3 +107,10 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
107
  2026.02.15/12.44.58_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-5000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
108
  2026.02.15/12.44.58_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-5890/tokenizer.json filter=lfs diff=lfs merge=lfs -text
109
  2026.02.15/12.44.58_train_llm_lowdim_box-close-v2/wandb/run-20260215_124507-yy4t6x0o/run-yy4t6x0o.wandb filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
107
  2026.02.15/12.44.58_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-5000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
108
  2026.02.15/12.44.58_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-5890/tokenizer.json filter=lfs diff=lfs merge=lfs -text
109
  2026.02.15/12.44.58_train_llm_lowdim_box-close-v2/wandb/run-20260215_124507-yy4t6x0o/run-yy4t6x0o.wandb filter=lfs diff=lfs merge=lfs -text
110
+ 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
111
+ 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
112
+ 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/tokenizer.json filter=lfs diff=lfs merge=lfs -text
113
+ 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-5000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
114
+ 2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/wandb/run-20260224_231158-zzc3dyq3/run-zzc3dyq3.wandb filter=lfs diff=lfs merge=lfs -text
115
+ 2026.02.24/23.25.46_train_llmbc_lowdim_box-close-v2/logs.json.txt filter=lfs diff=lfs merge=lfs -text
116
+ 2026.02.24/23.25.46_train_llmbc_lowdim_box-close-v2/wandb/run-20260224_232605-x2fbtqtp/run-x2fbtqtp.wandb filter=lfs diff=lfs merge=lfs -text
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/.hydra/config.yaml ADDED
@@ -0,0 +1,115 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: train_llm_lowdim
2
+ _target_: llmbc.workspace.train_llm_workspace.TrainLLMWorkspace
3
+ obs_dim: ${task.obs_dim}
4
+ action_dim: ${task.action_dim}
5
+ horizon: 1
6
+ n_obs_steps: 1
7
+ n_action_steps: 1
8
+ task_name: ${task.name}
9
+ exp_name: train llm
10
+ model_name: ${llm.name}
11
+ use_quantization: ${llm.use_quantization}
12
+ lora_config: ${llm.lora_config}
13
+ dataset:
14
+ test_data_ratio: 0.01
15
+ debug: false
16
+ training:
17
+ seed: 42
18
+ per_device_train_batch_size: 64
19
+ per_device_eval_batch_size: 64
20
+ gradient_accumulation_steps: 2
21
+ optim: paged_adamw_32bit
22
+ num_train_epochs: 30
23
+ eval_strategy: steps
24
+ logging_steps: 1
25
+ warmup_steps: 500
26
+ logging_strategy: steps
27
+ learning_rate: 0.0001
28
+ fp16: false
29
+ bf16: true
30
+ tf32: true
31
+ group_by_length: true
32
+ report_to: wandb
33
+ save_steps: 5000
34
+ eval_steps: 10
35
+ use_joint_mlp_projector: ${llm.use_joint_mlp_projector}
36
+ joint_obs_action_mlp_lr: 0.0001
37
+ trainer:
38
+ obs_dim: ${obs_dim}
39
+ action_dim: ${action_dim}
40
+ use_joint_mlp_projector: ${llm.use_joint_mlp_projector}
41
+ max_seq_length: ${llm.max_length}
42
+ dataset_text_field: text
43
+ packing: false
44
+ logging:
45
+ project: llm_module_finetuning
46
+ resume: true
47
+ mode: online
48
+ name: ${now:%Y.%m.%d-%H.%M.%S}_${name}_${task_name}
49
+ tags:
50
+ - ${name}
51
+ - ${task_name}
52
+ - ${exp_name}
53
+ id: null
54
+ group: null
55
+ multi_run:
56
+ run_dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
57
+ wandb_name_base: ${now:%Y.%m.%d-%H.%M.%S}_${name}_${task_name}
58
+ task:
59
+ name: box-close-v2
60
+ obs_dim: 9
61
+ action_dim: 4
62
+ env_runner:
63
+ _target_: llmbc.env_runner.metaworld_lowdim_runner.MetaworldLowdimRunner
64
+ env_name: llf-metaworld-box-close-v2
65
+ n_train: 10
66
+ n_test: 50
67
+ n_envs: 10
68
+ max_steps: 30
69
+ n_obs_steps: ${n_obs_steps}
70
+ n_action_steps: ${n_action_steps}
71
+ instruction_type: b
72
+ feedback_type:
73
+ - hp
74
+ - hn
75
+ - fp
76
+ visual: false
77
+ discount: 0.9
78
+ dataset:
79
+ _target_: llmbc.dataset.metaworld_lowdim_dataset.MetaworldLowdimDataset
80
+ data_path: datasets/box-close-v2-general-o4-mini.pt
81
+ data_path2: datasets/box-close-v2-o4-mini.pt
82
+ horizon: ${horizon}
83
+ pad_before: ${eval:'${n_obs_steps}-1'}
84
+ pad_after: ${eval:'${n_action_steps}-1'}
85
+ obs_eef_target: true
86
+ use_manual_normalizer: false
87
+ val_ratio: 0.05
88
+ dummy_normalizer: false
89
+ instructor:
90
+ _target_: llmbc.translator.instructor.metaworld_instructor.box_close_v2_instructor.BoxCloseV2Instructor
91
+ llm:
92
+ name: Qwen/Qwen2-0.5B-Instruct
93
+ model_name: Qwen2-0.5B-Instruct
94
+ config_target: llmbc.model.llm.qwen2_lowdim_model.LowdimQwen2Config
95
+ causal_lm_target: llmbc.model.llm.qwen2_lowdim_model.LowdimQwen2ForCausalLM
96
+ use_quantization: false
97
+ use_joint_mlp_projector: true
98
+ llm_mode: mlp-finetuned
99
+ finetune_mode: orig
100
+ checkpoint: data/outputs/2026.02.11/22.57.59_train_mlp_projector_box-close-v2/checkpoints/latest.ckpt
101
+ max_length: 100
102
+ lora_config:
103
+ r: 32
104
+ lora_alpha: 64
105
+ lora_dropout: 0.05
106
+ bias: none
107
+ task_type: CAUSAL_LM
108
+ prompter:
109
+ _target_: llmbc.translator.prompter.qwen2_prompter.Qwen2Prompter
110
+ use_joint_mlp_projector: true
111
+ hydra:
112
+ job:
113
+ override_dirname: ${model_name}
114
+ run:
115
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${model_name}
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/.hydra/hydra.yaml ADDED
@@ -0,0 +1,154 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
4
+ sweep:
5
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task: []
115
+ job:
116
+ name: train
117
+ chdir: null
118
+ override_dirname: ''
119
+ id: ???
120
+ num: ???
121
+ config_name: llmdp_llm_box-close-v2_o4-mini_qwen2-0.5b-instruct.yaml
122
+ env_set: {}
123
+ env_copy: []
124
+ config:
125
+ override_dirname:
126
+ kv_sep: '='
127
+ item_sep: ','
128
+ exclude_keys: []
129
+ runtime:
130
+ version: 1.2.0
131
+ version_base: '1.2'
132
+ cwd: /work/u1131674/LLM-BC
133
+ config_sources:
134
+ - path: hydra.conf
135
+ schema: pkg
136
+ provider: hydra
137
+ - path: /work/u1131674/LLM-BC/config/llm_backbone
138
+ schema: file
139
+ provider: main
140
+ - path: ''
141
+ schema: structured
142
+ provider: schema
143
+ output_dir: /work/u1131674/LLM-BC/data/outputs/2026.02.24/23.11.45_train_llm_lowdim_box-close-v2
144
+ choices:
145
+ hydra/env: default
146
+ hydra/callbacks: null
147
+ hydra/job_logging: default
148
+ hydra/hydra_logging: default
149
+ hydra/hydra_help: default
150
+ hydra/help: default
151
+ hydra/sweeper: basic
152
+ hydra/launcher: basic
153
+ hydra/output: default
154
+ verbose: false
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/.hydra/overrides.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ []
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
3
+ "action_dim": 4,
4
+ "architectures": [
5
+ "LowdimQwen2ForCausalLM"
6
+ ],
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 151643,
9
+ "eos_token_id": 151645,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 896,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 4864,
14
+ "max_position_embeddings": 32768,
15
+ "max_window_layers": 24,
16
+ "model_type": "qwen2_lowdim",
17
+ "num_attention_heads": 14,
18
+ "num_hidden_layers": 24,
19
+ "num_key_value_heads": 2,
20
+ "obs_dim": 9,
21
+ "rms_norm_eps": 1e-06,
22
+ "rope_scaling": null,
23
+ "rope_theta": 1000000.0,
24
+ "sliding_window": null,
25
+ "tie_word_embeddings": true,
26
+ "torch_dtype": "float32",
27
+ "transformers_version": "4.47.1",
28
+ "use_cache": false,
29
+ "use_joint_mlp_projector": true,
30
+ "use_sliding_window": false,
31
+ "vocab_size": 151936
32
+ }
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/generation_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.1,
10
+ "temperature": 0.7,
11
+ "top_k": 20,
12
+ "top_p": 0.8,
13
+ "transformers_version": "4.47.1"
14
+ }
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/mlp_projector.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d017c8218ec0a7d2387adad36baf50d313656ba4bd26bfe6047bbe44d9b89caa
3
+ size 3267200
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c19e36c2c3dc187a3ead6069a84b69bc4811a169605d146a1a042edc5e9bc6aa
3
+ size 1979428992
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23778b1f1b3c3251ee7e2314e47a786cca68cef9921e6711d0b9c6a196994780
3
+ size 3958970426
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea7a6a9f9b1bf8fef6e006b9abdee0ea6dcd7eed8a3a81a3822b1f2b7bd1d0ec
3
+ size 14244
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d8b080bf4d9132bd5dc533200edda4097ce7f9c8657c1eeb94c749f81f025ca
3
+ size 1064
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|im_end|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "pad_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ }
20
+ }
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcfe42da0a4497e8b2b172c1f9f4ec423a46dc12907f4349c55025f670422ba9
3
+ size 11418266
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/tokenizer_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
+ "clean_up_tokenization_spaces": false,
36
+ "eos_token": "<|im_end|>",
37
+ "errors": "replace",
38
+ "extra_special_tokens": {},
39
+ "model_max_length": 32768,
40
+ "pad_token": "<|endoftext|>",
41
+ "split_special_tokens": false,
42
+ "tokenizer_class": "Qwen2Tokenizer",
43
+ "unk_token": null
44
+ }
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e352bbf79fe71f7b32e56f82547a77a248dd86186015e71486b983673d6022f3
3
+ size 5944
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-10000/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
3
+ "action_dim": 4,
4
+ "architectures": [
5
+ "LowdimQwen2ForCausalLM"
6
+ ],
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 151643,
9
+ "eos_token_id": 151645,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 896,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 4864,
14
+ "max_position_embeddings": 32768,
15
+ "max_window_layers": 24,
16
+ "model_type": "qwen2_lowdim",
17
+ "num_attention_heads": 14,
18
+ "num_hidden_layers": 24,
19
+ "num_key_value_heads": 2,
20
+ "obs_dim": 9,
21
+ "rms_norm_eps": 1e-06,
22
+ "rope_scaling": null,
23
+ "rope_theta": 1000000.0,
24
+ "sliding_window": null,
25
+ "tie_word_embeddings": true,
26
+ "torch_dtype": "float32",
27
+ "transformers_version": "4.47.1",
28
+ "use_cache": false,
29
+ "use_joint_mlp_projector": true,
30
+ "use_sliding_window": false,
31
+ "vocab_size": 151936
32
+ }
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/generation_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.1,
10
+ "temperature": 0.7,
11
+ "top_k": 20,
12
+ "top_p": 0.8,
13
+ "transformers_version": "4.47.1"
14
+ }
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/mlp_projector.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1949e3f69f34544ade64904e380ebb265a3ccbfd4e57c32c86cb0f0a911aa28e
3
+ size 3267200
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2d6a1d4a74b31ee77e75ee23cddc7a44995fe4166910e5109621d01c7534e24
3
+ size 1979428992
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83351d7f1eb26474d8d776225d64b29d1f22eb626172aea836baa21022ec9b9a
3
+ size 3958970426
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c0449e6e0cd007dc6ea5d34ae43ef270855a286ad44e4514c4010c473eee2b6
3
+ size 14244
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72faacca39addc62cca7c330084846f9763fbda351a9eef5b29cb4490c2c9143
3
+ size 1064
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|im_end|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "pad_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ }
20
+ }
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcfe42da0a4497e8b2b172c1f9f4ec423a46dc12907f4349c55025f670422ba9
3
+ size 11418266
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/tokenizer_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
+ "clean_up_tokenization_spaces": false,
36
+ "eos_token": "<|im_end|>",
37
+ "errors": "replace",
38
+ "extra_special_tokens": {},
39
+ "model_max_length": 32768,
40
+ "pad_token": "<|endoftext|>",
41
+ "split_special_tokens": false,
42
+ "tokenizer_class": "Qwen2Tokenizer",
43
+ "unk_token": null
44
+ }
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e352bbf79fe71f7b32e56f82547a77a248dd86186015e71486b983673d6022f3
3
+ size 5944
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-15000/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
3
+ "action_dim": 4,
4
+ "architectures": [
5
+ "LowdimQwen2ForCausalLM"
6
+ ],
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 151643,
9
+ "eos_token_id": 151645,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 896,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 4864,
14
+ "max_position_embeddings": 32768,
15
+ "max_window_layers": 24,
16
+ "model_type": "qwen2_lowdim",
17
+ "num_attention_heads": 14,
18
+ "num_hidden_layers": 24,
19
+ "num_key_value_heads": 2,
20
+ "obs_dim": 9,
21
+ "rms_norm_eps": 1e-06,
22
+ "rope_scaling": null,
23
+ "rope_theta": 1000000.0,
24
+ "sliding_window": null,
25
+ "tie_word_embeddings": true,
26
+ "torch_dtype": "float32",
27
+ "transformers_version": "4.47.1",
28
+ "use_cache": false,
29
+ "use_joint_mlp_projector": true,
30
+ "use_sliding_window": false,
31
+ "vocab_size": 151936
32
+ }
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/generation_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.1,
10
+ "temperature": 0.7,
11
+ "top_k": 20,
12
+ "top_p": 0.8,
13
+ "transformers_version": "4.47.1"
14
+ }
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/mlp_projector.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:254338976389ad6401d2b1be9eac419b3f1418e1aafdb9b4adb18707d9baaed4
3
+ size 3267200
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:823d212d2b3cd2a22e533675fb1ecdfd36dff77973cf704172081f4ac19ce296
3
+ size 1979428992
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ae351c8f2a7ff438249f68d2c04e09a9889def6eb38897c96f5323d22df0ca9
3
+ size 3958970426
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a35eab652c443dd1b6a68e6e80334a969846b068e7ba84d71b7415fc514dd748
3
+ size 14244
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:798050277260c8ea3fe4950e8dd7d05d50dbcb666d3589507b86e208056b7f4e
3
+ size 1064
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|im_end|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "pad_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ }
20
+ }
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcfe42da0a4497e8b2b172c1f9f4ec423a46dc12907f4349c55025f670422ba9
3
+ size 11418266
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/tokenizer_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
+ "clean_up_tokenization_spaces": false,
36
+ "eos_token": "<|im_end|>",
37
+ "errors": "replace",
38
+ "extra_special_tokens": {},
39
+ "model_max_length": 32768,
40
+ "pad_token": "<|endoftext|>",
41
+ "split_special_tokens": false,
42
+ "tokenizer_class": "Qwen2Tokenizer",
43
+ "unk_token": null
44
+ }
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e352bbf79fe71f7b32e56f82547a77a248dd86186015e71486b983673d6022f3
3
+ size 5944
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-17670/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
2026.02.24/23.11.45_train_llm_lowdim_box-close-v2/Qwen/Qwen2-0.5B-Instruct-finetuned-box-close-v2/checkpoint-5000/added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }