aliangdw commited on
Commit
474a8d9
·
verified ·
1 Parent(s): 8addb9c

Upload config.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. config.yaml +212 -0
config.yaml ADDED
@@ -0,0 +1,212 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ custom_eval:
2
+ confusion_matrix:
3
+ - aliangdw_metaworld_metaworld_eval
4
+ eval_types:
5
+ - policy_ranking
6
+ - confusion_matrix
7
+ - reward_alignment
8
+ policy_ranking:
9
+ - aliangdw_metaworld_metaworld_eval
10
+ reward_alignment:
11
+ - aliangdw_metaworld_metaworld_eval
12
+ data:
13
+ data_source_weights:
14
+ libero256_90: 1.0
15
+ libero_90_failure: 1.0
16
+ metaworld_train: 5.0
17
+ dataloader_num_workers: 0
18
+ dataloader_pin_memory: false
19
+ dataset_preference_ratio: 0.7
20
+ dataset_success_cutoff_file: rfm/data/dataset_success_cutoff.txt
21
+ dataset_type: balanced_mixed
22
+ eval_datasets:
23
+ - abraranwar_libero_rfm_libero256_10
24
+ eval_subset_size: 500
25
+ fps: 10
26
+ load_embeddings: true
27
+ max_frames: 16
28
+ max_frames_after_preprocessing: 64
29
+ max_success: 0.95
30
+ max_trajectories: -1
31
+ min_success: 0.7
32
+ n_wrong_tasks: 5
33
+ num_bins: 10
34
+ pairwise_progress: false
35
+ preference_strategy_ratio:
36
+ - 6.0
37
+ - 1.0
38
+ - 1.0
39
+ - 0.0
40
+ progress_pred_type: absolute
41
+ progress_strategy_ratio:
42
+ - 1.0
43
+ - 6.0
44
+ - 1.0
45
+ resized_height: 128
46
+ resized_width: 128
47
+ rewind_lengths: null
48
+ sample_type_ratio:
49
+ - 0.0
50
+ - 1.0
51
+ - 0.0
52
+ samples_per_trajectory: 1
53
+ seed: 42
54
+ shuffle: true
55
+ similarity_strategy_ratio:
56
+ - 1.0
57
+ - 1.0
58
+ train_datasets:
59
+ - aliangdw_metaworld_metaworld_train
60
+ debug: false
61
+ logging:
62
+ save_best:
63
+ greater_is_better:
64
+ - true
65
+ - true
66
+ hub_model_id: null
67
+ hub_private: false
68
+ hub_token: null
69
+ keep_top_k: 5
70
+ metric_names:
71
+ - custom_eval/p_rank_spearman_mw_eval
72
+ - custom_eval/rew_align_spearman_mw_eval
73
+ upload_to_hub: true
74
+ save_model: true
75
+ save_processor: true
76
+ use_wandb: true
77
+ wandb_entity: clvr
78
+ wandb_project: rfm
79
+ wandb_run_name: rfm
80
+ mode: train
81
+ model:
82
+ base_model_id: rewind_transformer
83
+ model_type: default
84
+ peft_vision_encoder: false
85
+ quantization: false
86
+ rewind: !!python/object:rfm.models.rewind_transformer.ReWINDTransformerConfig
87
+ _attn_implementation_autoset: true
88
+ _attn_implementation_internal: null
89
+ _commit_hash: null
90
+ _name_or_path: ''
91
+ _output_attentions: false
92
+ add_cross_attention: false
93
+ architectures: null
94
+ bad_words_ids: null
95
+ begin_suppress_tokens: null
96
+ bos_token_id: null
97
+ chunk_size_feed_forward: 0
98
+ cross_attention_hidden_size: null
99
+ decoder_start_token_id: null
100
+ diversity_penalty: 0.0
101
+ do_sample: false
102
+ dropout: 0.1
103
+ early_stopping: false
104
+ encoder_no_repeat_ngram_size: 0
105
+ eos_token_id: null
106
+ exponential_decay_length_penalty: null
107
+ finetuning_task: null
108
+ forced_bos_token_id: null
109
+ forced_eos_token_id: null
110
+ hidden_dim: 512
111
+ id2label:
112
+ 0: LABEL_0
113
+ 1: LABEL_1
114
+ is_decoder: false
115
+ is_encoder_decoder: false
116
+ label2id:
117
+ LABEL_0: 0
118
+ LABEL_1: 1
119
+ length_penalty: 1.0
120
+ max_len: 16
121
+ max_length: 20
122
+ min_length: 0
123
+ no_repeat_ngram_size: 0
124
+ num_attention_heads: 8
125
+ num_beam_groups: 1
126
+ num_beams: 1
127
+ num_layers: 4
128
+ num_return_sequences: 1
129
+ output_hidden_states: false
130
+ output_scores: false
131
+ pad_token_id: null
132
+ prefix: null
133
+ problem_type: null
134
+ pruned_heads: {}
135
+ remove_invalid_values: false
136
+ repetition_penalty: 1.0
137
+ return_dict: true
138
+ return_dict_in_generate: false
139
+ sep_token_id: null
140
+ suppress_tokens: null
141
+ task_specific_params: null
142
+ temperature: 1.0
143
+ text_feature_dim: 384
144
+ tf_legacy_loss: false
145
+ tie_encoder_decoder: false
146
+ tie_word_embeddings: true
147
+ tokenizer_class: null
148
+ top_k: 50
149
+ top_p: 1.0
150
+ torch_dtype: null
151
+ torchscript: false
152
+ transformers_version: null
153
+ typical_p: 1.0
154
+ use_bfloat16: false
155
+ video_feature_dim: 768
156
+ torch_dtype: bfloat16
157
+ train_language_model: false
158
+ train_preference_head: false
159
+ train_progress_head: true
160
+ train_similarity_head: false
161
+ train_success_head: false
162
+ train_vision_encoder: false
163
+ trust_remote_code: true
164
+ use_peft: false
165
+ peft:
166
+ bias: none
167
+ lora_alpha: 64
168
+ lora_dropout: 0.05
169
+ r: 32
170
+ target_modules:
171
+ - q_proj
172
+ - k_proj
173
+ - v_proj
174
+ - o_proj
175
+ - gate_proj
176
+ - up_proj
177
+ - down_proj
178
+ trainer_cls: rewind_transformer
179
+ training:
180
+ beta: 0.1
181
+ bf16: false
182
+ custom_eval_steps: 100
183
+ ddp_bucket_cap_mb: 25
184
+ ddp_find_unused_parameters: true
185
+ do_eval: true
186
+ eval_steps: 100
187
+ evaluation_strategy: steps
188
+ exp_name: rewind_base_mw
189
+ fp16: true
190
+ gradient_accumulation_steps: 1
191
+ gradient_checkpointing: false
192
+ learning_rate: 0.0001
193
+ logging_steps: 1
194
+ lr_scheduler_type: cosine
195
+ max_grad_norm: 10.0
196
+ max_seq_length: 1024
197
+ max_steps: 10000
198
+ num_gpus: 2
199
+ num_train_epochs: -1
200
+ output_dir: ./logs
201
+ per_device_eval_batch_size: 512
202
+ per_device_train_batch_size: 1024
203
+ predict_pref_progress: false
204
+ predict_sim_progress: false
205
+ prediction_loss_only: true
206
+ remove_unused_columns: false
207
+ resume_from_checkpoint: null
208
+ run_default_eval: false
209
+ save_steps: 200
210
+ save_strategy: steps
211
+ warmup_ratio: 0.01
212
+ warmup_steps: 0