artpli commited on
Commit
c3c1622
·
verified ·
1 Parent(s): 5189bc6

Upload 3 files

Browse files
hgpt/motionx/VQVAE_MotionX_2Kx1K_2024-12-20-16-56-10/checkpoints/last.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c705e6cb86806658c20fefdd4630c4c8b47dba9ec062ef186b8aa5b9cda39fa
3
+ size 964443382
hgpt/motionx/VQVAE_MotionX_2Kx1K_2024-12-20-16-56-10/config_2024-12-19-22-23-43_train.yaml ADDED
@@ -0,0 +1,225 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ SEED_VALUE: 1234
2
+ DEBUG: false
3
+ NAME: VQVAE_MotionX_2Kx1K
4
+ ACCELERATOR: gpu
5
+ NUM_NODES: 1
6
+ DEVICE:
7
+ - 0
8
+ - 1
9
+ - 2
10
+ - 3
11
+ - 4
12
+ - 5
13
+ - 6
14
+ - 7
15
+ DATASET:
16
+ target: hGPT.data.MotionX.MotionXDataModule
17
+ NFEATS: 623
18
+ JOINT_TYPE: motionx
19
+ MOTIONX:
20
+ FRAME_RATE: 30.0
21
+ MAX_MOTION_LEN: 400
22
+ MIN_MOTION_LEN: 40
23
+ MAX_TEXT_LEN: 20
24
+ PICK_ONE_TEXT: true
25
+ UNIT_LEN: 4
26
+ STD_TEXT: false
27
+ ROOT: datasets/motionx/data
28
+ SPLIT_ROOT: datasets/motionx/data/split
29
+ SEMANTIC_TEXT_ROOT: datasets/motionx/data/texts/semantic_texts
30
+ FACE_TEXT_ROOT: datasets/motionx/data/texts/face_texts
31
+ WORD_VERTILIZER_PATH: deps/glove/
32
+ HUMANML3D:
33
+ ROOT: datasets/humanml3d
34
+ SPLIT_ROOT: datasets/humanml3d
35
+ model:
36
+ target: hGPT.models.hgpt.HumanoidGPT
37
+ params:
38
+ lm: ''
39
+ motion_vae: ${vq.vqvae_2k_1k}
40
+ task: vae
41
+ condition: ''
42
+ stage: ${TRAIN.STAGE}
43
+ codebook_size: ${model.params.motion_vae.params.code_num}
44
+ metrics_dict: ${METRIC.TYPE}
45
+ debug: ${DEBUG}
46
+ TRAIN:
47
+ STAGE: vae
48
+ SPLIT: train
49
+ PRECISION: '32'
50
+ NUM_JOINTS: 52
51
+ NUM_WORKERS: 16
52
+ BATCH_SIZE: 2048
53
+ END_EPOCH: 3000
54
+ RESUME: ''
55
+ PRETRAINED: ''
56
+ PRETRAINED_VAE: ''
57
+ OPTIM:
58
+ target: AdamW
59
+ params:
60
+ lr: 0.0002
61
+ betas:
62
+ - 0.9
63
+ - 0.99
64
+ weight_decay: 0.0
65
+ LR_SCHEDULER:
66
+ target: CosineAnnealingLR
67
+ params:
68
+ T_max: ${eval:${LOGGER.VAL_EVERY_STEPS} * 100}
69
+ eta_min: 1.0e-06
70
+ EVAL:
71
+ SPLIT: val
72
+ BATCH_SIZE: 32
73
+ NUM_WORKERS: 8
74
+ TEST:
75
+ DATASETS:
76
+ - MotionX
77
+ CHECKPOINTS: experiments/hgpt/VQVAE_MotionX_Debug/checkpoints/last.ckpt
78
+ SPLIT: test
79
+ BATCH_SIZE: 512
80
+ NUM_WORKERS: 8
81
+ SAVE_PREDICTIONS: true
82
+ COUNT_TIME: true
83
+ REPLICATION_TIMES: 1
84
+ REP_I: 0
85
+ FOLDER: results
86
+ PREDICT:
87
+ SPLIT: test
88
+ CHECKPOINTS: experiments/hgpt/VQVAE_MotionX_Debug/checkpoints/last.ckpt
89
+ BATCH_SIZE: 512
90
+ LOSS:
91
+ LAMBDA_FEATURE: 1.0
92
+ LAMBDA_VELOCITY: 0.5
93
+ LAMBDA_COMMIT: 0.02
94
+ LAMBDA_CLS: 1.0
95
+ ABLATION:
96
+ RECONS_LOSS: l1_smooth
97
+ METRIC:
98
+ TYPE:
99
+ - MRMetrics
100
+ FORCE_IN_METER: true
101
+ DIST_SYNC_ON_STEP: true
102
+ MM_NUM_SAMPLES: 100
103
+ MM_NUM_REPEATS: 30
104
+ MM_NUM_TIMES: 10
105
+ LOGGER:
106
+ VAL_EVERY_STEPS: 10
107
+ TYPE:
108
+ - tensorboard
109
+ - wandb
110
+ TENSORBOARD:
111
+ target: pytorch_lightning.loggers.TensorBoardLogger
112
+ params:
113
+ save_dir: ${FOLDER_EXP}
114
+ name: tensorboard
115
+ version: ''
116
+ WANDB:
117
+ target: pytorch_lightning.loggers.WandbLogger
118
+ params:
119
+ project: humanoidgpt
120
+ offline: true
121
+ id: null
122
+ version: ''
123
+ name: ${NAME}
124
+ save_dir: ${FOLDER_EXP}
125
+ eval:
126
+ tm2t:
127
+ t2m_path: deps/t2m/
128
+ t2m_textencoder:
129
+ target: hGPT.archs.tm2t_evaluator.TextEncoderBiGRUCo
130
+ params:
131
+ word_size: 300
132
+ pos_size: 15
133
+ hidden_size: 512
134
+ output_size: 512
135
+ t2m_moveencoder:
136
+ target: hGPT.archs.tm2t_evaluator.MovementConvEncoder
137
+ params:
138
+ input_size: 259
139
+ hidden_size: 512
140
+ output_size: 512
141
+ t2m_motionencoder:
142
+ target: hGPT.archs.tm2t_evaluator.MotionEncoderBiGRUCo
143
+ params:
144
+ input_size: ${eval.tm2t.t2m_moveencoder.params.output_size}
145
+ hidden_size: 1024
146
+ output_size: 512
147
+ vq:
148
+ default:
149
+ target: hGPT.archs.hgpt_vq.VQVae
150
+ params:
151
+ quantizer: ema_reset
152
+ code_num: 512
153
+ code_dim: 512
154
+ output_emb_width: 512
155
+ down_t: 2
156
+ stride_t: 2
157
+ width: 512
158
+ depth: 3
159
+ dilation_growth_rate: 3
160
+ norm: None
161
+ activation: relu
162
+ nfeats: ${DATASET.NFEATS}
163
+ vqvae_2k_1k:
164
+ target: hGPT.archs.hgpt_vq.VQVae
165
+ params:
166
+ quantizer: ema_reset
167
+ n_quantizer: 1
168
+ code_num: 2048
169
+ code_dim: 1024
170
+ output_emb_width: 1024
171
+ width: 1024
172
+ down_t: 2
173
+ stride_t: 2
174
+ depth: 3
175
+ dilation_growth_rate: 3
176
+ norm: None
177
+ activation: relu
178
+ nfeats: ${DATASET.NFEATS}
179
+ rvqvae_2k_1k_2layer:
180
+ target: hGPT.archs.hgpt_vq.RVQVae
181
+ params:
182
+ quantizer: ema_reset
183
+ n_quantizer: 2
184
+ code_num: 2048
185
+ code_dim: 1024
186
+ output_emb_width: 1024
187
+ width: 1024
188
+ down_t: 2
189
+ stride_t: 2
190
+ depth: 3
191
+ dilation_growth_rate: 3
192
+ norm: None
193
+ activation: relu
194
+ nfeats: ${DATASET.NFEATS}
195
+ lm:
196
+ default:
197
+ target: hGPT.archs.hgpt_lm.MLM
198
+ params:
199
+ model_type: t5
200
+ model_path: ./deps/flan-t5-base
201
+ stage: ${TRAIN.STAGE}
202
+ motion_codebook_size: ${model.params.codebook_size}
203
+ llama:
204
+ target: hGPT.archs.hgpt_lm.MLM
205
+ params:
206
+ model_type: llama
207
+ model_path: ./deps/Meta-Llama-3.1-8B-Instruct
208
+ stage: ${TRAIN.STAGE}
209
+ motion_codebook_size: ${model.params.codebook_size}
210
+ max_length: 1500
211
+ lora: true
212
+ train_strategy: pt
213
+ new_token_type: insert
214
+ framerate: 30.0
215
+ down_t: 4
216
+ predict_ratio: 0.2
217
+ inbetween_ratio: 0.25
218
+ quota_ratio: 0.5
219
+ noise_density: 0.15
220
+ mean_noise_span_length: 3
221
+ CONFIG_FOLDER: configs
222
+ ARCHS_FOLDER: configs/archs
223
+ FOLDER: experiments
224
+ FOLDER_EXP: experiments/hgpt/VQVAE_MotionX_2Kx1K
225
+ TIME: 2024-12-19-22-23-43
hgpt/motionx/VQVAE_MotionX_2Kx1K_2024-12-20-16-56-10/log_2024-12-19-22-23-43_train.log ADDED
The diff for this file is too large to render. See raw diff