Jacklu0831 commited on
Commit
e39ff67
·
verified ·
1 Parent(s): 6256253

Append 1 new entries to hyperparameters__rl (total: 1)

Browse files
README.md CHANGED
@@ -1,5 +1,149 @@
1
  ---
2
  dataset_info:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  - config_name: logs__verl_rl
4
  features:
5
  - name: timestamp
@@ -79,6 +223,10 @@ dataset_info:
79
  download_size: 4991
80
  dataset_size: 202
81
  configs:
 
 
 
 
82
  - config_name: logs__verl_rl
83
  data_files:
84
  - split: train
 
1
  ---
2
  dataset_info:
3
+ - config_name: hyperparameters__rl
4
+ features:
5
+ - name: stage_name
6
+ dtype: string
7
+ - name: stage_number
8
+ dtype: int64
9
+ - name: stage_type
10
+ dtype: string
11
+ - name: model_repo_id
12
+ dtype: string
13
+ - name: base_model
14
+ dtype: string
15
+ - name: timestamp
16
+ dtype: string
17
+ - name: verl_parameter_config
18
+ struct:
19
+ - name: actor_rollout_ref.actor.fsdp_config.forward_prefetch
20
+ dtype: bool
21
+ - name: actor_rollout_ref.actor.optim.lr
22
+ dtype: float64
23
+ - name: actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu
24
+ dtype: int64
25
+ - name: actor_rollout_ref.actor.ppo_mini_batch_size
26
+ dtype: int64
27
+ - name: actor_rollout_ref.actor.strategy
28
+ dtype: string
29
+ - name: actor_rollout_ref.model.enable_activation_offload
30
+ dtype: bool
31
+ - name: actor_rollout_ref.model.enable_gradient_checkpointing
32
+ dtype: bool
33
+ - name: actor_rollout_ref.model.path
34
+ dtype: string
35
+ - name: actor_rollout_ref.model.trust_remote_code
36
+ dtype: bool
37
+ - name: actor_rollout_ref.model.use_remove_padding
38
+ dtype: bool
39
+ - name: actor_rollout_ref.ref.fsdp_config.forward_prefetch
40
+ dtype: bool
41
+ - name: actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu
42
+ dtype: int64
43
+ - name: actor_rollout_ref.rollout.dtype
44
+ dtype: string
45
+ - name: actor_rollout_ref.rollout.gpu_memory_utilization
46
+ dtype: float64
47
+ - name: actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu
48
+ dtype: int64
49
+ - name: actor_rollout_ref.rollout.max_num_batched_tokens
50
+ dtype: int64
51
+ - name: actor_rollout_ref.rollout.max_num_seqs
52
+ dtype: int64
53
+ - name: actor_rollout_ref.rollout.n
54
+ dtype: int64
55
+ - name: actor_rollout_ref.rollout.tensor_model_parallel_size
56
+ dtype: int64
57
+ - name: algorithm.adv_estimator
58
+ dtype: string
59
+ - name: algorithm.kl_ctrl.kl_coef
60
+ dtype: float64
61
+ - name: critic.model.path
62
+ dtype: string
63
+ - name: critic.model.trust_remote_code
64
+ dtype: bool
65
+ - name: critic.optim.lr
66
+ dtype: float64
67
+ - name: critic.ppo_micro_batch_size_per_gpu
68
+ dtype: int64
69
+ - name: custom_reward_function.name
70
+ dtype: string
71
+ - name: custom_reward_function.path
72
+ dtype: string
73
+ - name: custom_reward_function.reward_kwargs.complex_format_reward_weight
74
+ dtype: float64
75
+ - name: custom_reward_function.reward_kwargs.final_answer_in_samples_reward_weight
76
+ dtype: float64
77
+ - name: custom_reward_function.reward_kwargs.reflection_correctness_reward_weight
78
+ dtype: float64
79
+ - name: custom_reward_function.reward_kwargs.response_or_sample
80
+ dtype: string
81
+ - name: custom_reward_function.reward_kwargs.reward_max
82
+ dtype: float64
83
+ - name: custom_reward_function.reward_kwargs.reward_min
84
+ dtype: float64
85
+ - name: custom_reward_function.reward_kwargs.sample_correctness_reward_weight
86
+ dtype: float64
87
+ - name: custom_reward_function.reward_kwargs.sample_count_penalty_weight
88
+ dtype: float64
89
+ - name: custom_reward_function.reward_kwargs.similarity_penalty_weight
90
+ dtype: float64
91
+ - name: custom_reward_function.reward_kwargs.simple_format_reward_weight
92
+ dtype: float64
93
+ - name: custom_reward_function.reward_kwargs.transition_penalty_weight
94
+ dtype: float64
95
+ - name: custom_reward_function.reward_kwargs.verdict_correctness_reward_weight
96
+ dtype: float64
97
+ - name: data.max_prompt_length
98
+ dtype: int64
99
+ - name: data.max_response_length
100
+ dtype: int64
101
+ - name: data.train_batch_size
102
+ dtype: int64
103
+ - name: data.train_files
104
+ dtype: string
105
+ - name: data.val_files
106
+ dtype: string
107
+ - name: hydra.job.chdir
108
+ dtype: bool
109
+ - name: hydra.output_subdir
110
+ dtype: string
111
+ - name: hydra.run.dir
112
+ dtype: string
113
+ - name: reward_model.launch_reward_fn_async
114
+ dtype: bool
115
+ - name: reward_model.model.fsdp_config.forward_prefetch
116
+ dtype: bool
117
+ - name: reward_model.reward_manager
118
+ dtype: string
119
+ - name: trainer.default_local_dir
120
+ dtype: string
121
+ - name: trainer.experiment_name
122
+ dtype: string
123
+ - name: trainer.logger
124
+ dtype: string
125
+ - name: trainer.n_gpus_per_node
126
+ dtype: int64
127
+ - name: trainer.nnodes
128
+ dtype: int64
129
+ - name: trainer.project_name
130
+ dtype: string
131
+ - name: trainer.resume_mode
132
+ dtype: string
133
+ - name: trainer.save_freq
134
+ dtype: int64
135
+ - name: trainer.test_freq
136
+ dtype: int64
137
+ - name: trainer.total_epochs
138
+ dtype: int64
139
+ - name: trainer.val_before_train
140
+ dtype: bool
141
+ splits:
142
+ - name: train
143
+ num_bytes: 1382
144
+ num_examples: 1
145
+ download_size: 41655
146
+ dataset_size: 1382
147
  - config_name: logs__verl_rl
148
  features:
149
  - name: timestamp
 
223
  download_size: 4991
224
  dataset_size: 202
225
  configs:
226
+ - config_name: hyperparameters__rl
227
+ data_files:
228
+ - split: train
229
+ path: hyperparameters__rl/train-*
230
  - config_name: logs__verl_rl
231
  data_files:
232
  - split: train
hyperparameters__rl/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fcbffec8f6b73c547483d56a8df28abba41885984f5e779c2f14dad7d1e5cdd4
3
+ size 41655