halffried commited on
Commit
f346a0f
·
1 Parent(s): ccaa0cb

Upload 12 files

Browse files

Add keypose models (object detector and top-down COCO pose estimator models)

keypose_mmdet_fasterrcnn_r50/README.md ADDED
@@ -0,0 +1 @@
 
 
1
+ Originally from https://github.com/open-mmlab/mmdetection, converted to safetensors, used under Apache-2.0
keypose_mmdet_fasterrcnn_r50/faster_rcnn_r50_fpn_1x_coco_20200130-047c8118.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d5d2dd5c9b96c55ab516dacc2bbe6293ff730e5b7bdcff45bef9130ab97d9ce
3
+ size 167266866
keypose_mmdet_fasterrcnn_r50/faster_rcnn_r50_fpn_coco.yaml ADDED
@@ -0,0 +1,359 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ checkpoint_config:
2
+ interval: 1
3
+ data:
4
+ samples_per_gpu: 2
5
+ test:
6
+ ann_file: data/coco/annotations/instances_val2017.json
7
+ img_prefix: data/coco/val2017/
8
+ pipeline:
9
+ - type: LoadImageFromFile
10
+ - flip: false
11
+ img_scale: !!python/tuple
12
+ - 1333
13
+ - 800
14
+ transforms:
15
+ - keep_ratio: true
16
+ type: Resize
17
+ - type: RandomFlip
18
+ - mean:
19
+ - 123.675
20
+ - 116.28
21
+ - 103.53
22
+ std:
23
+ - 58.395
24
+ - 57.12
25
+ - 57.375
26
+ to_rgb: true
27
+ type: Normalize
28
+ - size_divisor: 32
29
+ type: Pad
30
+ - type: DefaultFormatBundle
31
+ - keys:
32
+ - img
33
+ type: Collect
34
+ type: MultiScaleFlipAug
35
+ type: CocoDataset
36
+ train:
37
+ ann_file: data/coco/annotations/instances_train2017.json
38
+ img_prefix: data/coco/train2017/
39
+ pipeline:
40
+ - type: LoadImageFromFile
41
+ - type: LoadAnnotations
42
+ with_bbox: true
43
+ - img_scale: !!python/tuple
44
+ - 1333
45
+ - 800
46
+ keep_ratio: true
47
+ type: Resize
48
+ - flip_ratio: 0.5
49
+ type: RandomFlip
50
+ - mean:
51
+ - 123.675
52
+ - 116.28
53
+ - 103.53
54
+ std:
55
+ - 58.395
56
+ - 57.12
57
+ - 57.375
58
+ to_rgb: true
59
+ type: Normalize
60
+ - size_divisor: 32
61
+ type: Pad
62
+ - type: DefaultFormatBundle
63
+ - keys:
64
+ - img
65
+ - gt_bboxes
66
+ - gt_labels
67
+ type: Collect
68
+ type: CocoDataset
69
+ val:
70
+ ann_file: data/coco/annotations/instances_val2017.json
71
+ img_prefix: data/coco/val2017/
72
+ pipeline:
73
+ - type: LoadImageFromFile
74
+ - flip: false
75
+ img_scale: !!python/tuple
76
+ - 1333
77
+ - 800
78
+ transforms:
79
+ - keep_ratio: true
80
+ type: Resize
81
+ - type: RandomFlip
82
+ - mean:
83
+ - 123.675
84
+ - 116.28
85
+ - 103.53
86
+ std:
87
+ - 58.395
88
+ - 57.12
89
+ - 57.375
90
+ to_rgb: true
91
+ type: Normalize
92
+ - size_divisor: 32
93
+ type: Pad
94
+ - type: DefaultFormatBundle
95
+ - keys:
96
+ - img
97
+ type: Collect
98
+ type: MultiScaleFlipAug
99
+ type: CocoDataset
100
+ workers_per_gpu: 2
101
+ data_root: data/coco
102
+ dataset_type: CocoDataset
103
+ dist_params:
104
+ backend: nccl
105
+ evaluation:
106
+ interval: 1
107
+ metric: bbox
108
+ img_norm_cfg:
109
+ mean:
110
+ - 123.675
111
+ - 116.28
112
+ - 103.53
113
+ std:
114
+ - 58.395
115
+ - 57.12
116
+ - 57.375
117
+ to_rgb: true
118
+ load_from: null
119
+ log_config:
120
+ hooks:
121
+ - type: TextLoggerHook
122
+ interval: 50
123
+ log_level: INFO
124
+ lr_config:
125
+ policy: step
126
+ step:
127
+ - 8
128
+ - 11
129
+ warmup: linear
130
+ warmup_iters: 500
131
+ warmup_ratio: 0.001
132
+ model:
133
+ backbone:
134
+ depth: 50
135
+ frozen_stages: 1
136
+ norm_cfg:
137
+ requires_grad: true
138
+ type: BN
139
+ norm_eval: true
140
+ num_stages: 4
141
+ out_indices: !!python/tuple
142
+ - 0
143
+ - 1
144
+ - 2
145
+ - 3
146
+ style: pytorch
147
+ type: ResNet
148
+ neck:
149
+ in_channels:
150
+ - 256
151
+ - 512
152
+ - 1024
153
+ - 2048
154
+ num_outs: 5
155
+ out_channels: 256
156
+ type: FPN
157
+ pretrained: torchvision://resnet50
158
+ roi_head:
159
+ bbox_head:
160
+ bbox_coder:
161
+ target_means:
162
+ - 0.0
163
+ - 0.0
164
+ - 0.0
165
+ - 0.0
166
+ target_stds:
167
+ - 0.1
168
+ - 0.1
169
+ - 0.2
170
+ - 0.2
171
+ type: DeltaXYWHBBoxCoder
172
+ fc_out_channels: 1024
173
+ in_channels: 256
174
+ loss_bbox:
175
+ loss_weight: 1.0
176
+ type: L1Loss
177
+ loss_cls:
178
+ loss_weight: 1.0
179
+ type: CrossEntropyLoss
180
+ use_sigmoid: false
181
+ num_classes: 80
182
+ reg_class_agnostic: false
183
+ roi_feat_size: 7
184
+ type: Shared2FCBBoxHead
185
+ bbox_roi_extractor:
186
+ featmap_strides:
187
+ - 4
188
+ - 8
189
+ - 16
190
+ - 32
191
+ out_channels: 256
192
+ roi_layer:
193
+ output_size: 7
194
+ sampling_ratio: 0
195
+ type: RoIAlign
196
+ type: SingleRoIExtractor
197
+ type: StandardRoIHead
198
+ rpn_head:
199
+ anchor_generator:
200
+ ratios:
201
+ - 0.5
202
+ - 1.0
203
+ - 2.0
204
+ scales:
205
+ - 8
206
+ strides:
207
+ - 4
208
+ - 8
209
+ - 16
210
+ - 32
211
+ - 64
212
+ type: AnchorGenerator
213
+ bbox_coder:
214
+ target_means:
215
+ - 0.0
216
+ - 0.0
217
+ - 0.0
218
+ - 0.0
219
+ target_stds:
220
+ - 1.0
221
+ - 1.0
222
+ - 1.0
223
+ - 1.0
224
+ type: DeltaXYWHBBoxCoder
225
+ feat_channels: 256
226
+ in_channels: 256
227
+ loss_bbox:
228
+ loss_weight: 1.0
229
+ type: L1Loss
230
+ loss_cls:
231
+ loss_weight: 1.0
232
+ type: CrossEntropyLoss
233
+ use_sigmoid: true
234
+ type: RPNHead
235
+ test_cfg:
236
+ rcnn:
237
+ max_per_img: 100
238
+ nms:
239
+ iou_threshold: 0.5
240
+ type: nms
241
+ score_thr: 0.05
242
+ rpn:
243
+ max_per_img: 1000
244
+ min_bbox_size: 0
245
+ nms:
246
+ iou_threshold: 0.7
247
+ type: nms
248
+ nms_pre: 1000
249
+ train_cfg:
250
+ rcnn:
251
+ assigner:
252
+ ignore_iof_thr: -1
253
+ match_low_quality: false
254
+ min_pos_iou: 0.5
255
+ neg_iou_thr: 0.5
256
+ pos_iou_thr: 0.5
257
+ type: MaxIoUAssigner
258
+ debug: false
259
+ pos_weight: -1
260
+ sampler:
261
+ add_gt_as_proposals: true
262
+ neg_pos_ub: -1
263
+ num: 512
264
+ pos_fraction: 0.25
265
+ type: RandomSampler
266
+ rpn:
267
+ allowed_border: -1
268
+ assigner:
269
+ ignore_iof_thr: -1
270
+ match_low_quality: true
271
+ min_pos_iou: 0.3
272
+ neg_iou_thr: 0.3
273
+ pos_iou_thr: 0.7
274
+ type: MaxIoUAssigner
275
+ debug: false
276
+ pos_weight: -1
277
+ sampler:
278
+ add_gt_as_proposals: false
279
+ neg_pos_ub: -1
280
+ num: 256
281
+ pos_fraction: 0.5
282
+ type: RandomSampler
283
+ rpn_proposal:
284
+ max_per_img: 1000
285
+ min_bbox_size: 0
286
+ nms:
287
+ iou_threshold: 0.7
288
+ type: nms
289
+ nms_pre: 2000
290
+ type: FasterRCNN
291
+ optimizer:
292
+ lr: 0.02
293
+ momentum: 0.9
294
+ type: SGD
295
+ weight_decay: 0.0001
296
+ optimizer_config:
297
+ grad_clip: null
298
+ resume_from: null
299
+ test_pipeline:
300
+ - type: LoadImageFromFile
301
+ - flip: false
302
+ img_scale: !!python/tuple
303
+ - 1333
304
+ - 800
305
+ transforms:
306
+ - keep_ratio: true
307
+ type: Resize
308
+ - type: RandomFlip
309
+ - mean:
310
+ - 123.675
311
+ - 116.28
312
+ - 103.53
313
+ std:
314
+ - 58.395
315
+ - 57.12
316
+ - 57.375
317
+ to_rgb: true
318
+ type: Normalize
319
+ - size_divisor: 32
320
+ type: Pad
321
+ - type: DefaultFormatBundle
322
+ - keys:
323
+ - img
324
+ type: Collect
325
+ type: MultiScaleFlipAug
326
+ total_epochs: 12
327
+ train_pipeline:
328
+ - type: LoadImageFromFile
329
+ - type: LoadAnnotations
330
+ with_bbox: true
331
+ - img_scale: !!python/tuple
332
+ - 1333
333
+ - 800
334
+ keep_ratio: true
335
+ type: Resize
336
+ - flip_ratio: 0.5
337
+ type: RandomFlip
338
+ - mean:
339
+ - 123.675
340
+ - 116.28
341
+ - 103.53
342
+ std:
343
+ - 58.395
344
+ - 57.12
345
+ - 57.375
346
+ to_rgb: true
347
+ type: Normalize
348
+ - size_divisor: 32
349
+ type: Pad
350
+ - type: DefaultFormatBundle
351
+ - keys:
352
+ - img
353
+ - gt_bboxes
354
+ - gt_labels
355
+ type: Collect
356
+ workflow:
357
+ - !!python/tuple
358
+ - train
359
+ - 1
keypose_mmdet_fasterrcnn_x101/README.md ADDED
@@ -0,0 +1 @@
 
 
1
+ Originally from https://github.com/open-mmlab/mmdetection, converted to safetensors, used under Apache-2.0
keypose_mmdet_fasterrcnn_x101/faster_rcnn_x101_64x4d_fpn_mstrain_3x_coco.yaml ADDED
@@ -0,0 +1,391 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ auto_scale_lr:
2
+ base_batch_size: 16
3
+ enable: false
4
+ checkpoint_config:
5
+ interval: 1
6
+ custom_hooks:
7
+ - type: NumClassCheckHook
8
+ data:
9
+ samples_per_gpu: 2
10
+ test:
11
+ ann_file: data/coco/annotations/instances_val2017.json
12
+ img_prefix: data/coco/val2017/
13
+ pipeline:
14
+ - type: LoadImageFromFile
15
+ - flip: false
16
+ img_scale: !!python/tuple
17
+ - 1333
18
+ - 800
19
+ transforms:
20
+ - keep_ratio: true
21
+ type: Resize
22
+ - type: RandomFlip
23
+ - mean:
24
+ - 123.675
25
+ - 116.28
26
+ - 103.53
27
+ std:
28
+ - 58.395
29
+ - 57.12
30
+ - 57.375
31
+ to_rgb: true
32
+ type: Normalize
33
+ - size_divisor: 32
34
+ type: Pad
35
+ - keys:
36
+ - img
37
+ type: ImageToTensor
38
+ - keys:
39
+ - img
40
+ type: Collect
41
+ type: MultiScaleFlipAug
42
+ type: CocoDataset
43
+ train:
44
+ dataset:
45
+ ann_file: data/coco/annotations/instances_train2017.json
46
+ img_prefix: data/coco/train2017/
47
+ pipeline:
48
+ - type: LoadImageFromFile
49
+ - type: LoadAnnotations
50
+ with_bbox: true
51
+ - img_scale:
52
+ - !!python/tuple
53
+ - 1333
54
+ - 640
55
+ - !!python/tuple
56
+ - 1333
57
+ - 800
58
+ keep_ratio: true
59
+ multiscale_mode: range
60
+ type: Resize
61
+ - flip_ratio: 0.5
62
+ type: RandomFlip
63
+ - mean:
64
+ - 123.675
65
+ - 116.28
66
+ - 103.53
67
+ std:
68
+ - 58.395
69
+ - 57.12
70
+ - 57.375
71
+ to_rgb: true
72
+ type: Normalize
73
+ - size_divisor: 32
74
+ type: Pad
75
+ - type: DefaultFormatBundle
76
+ - keys:
77
+ - img
78
+ - gt_bboxes
79
+ - gt_labels
80
+ type: Collect
81
+ type: CocoDataset
82
+ times: 3
83
+ type: RepeatDataset
84
+ val:
85
+ ann_file: data/coco/annotations/instances_val2017.json
86
+ img_prefix: data/coco/val2017/
87
+ pipeline:
88
+ - type: LoadImageFromFile
89
+ - flip: false
90
+ img_scale: !!python/tuple
91
+ - 1333
92
+ - 800
93
+ transforms:
94
+ - keep_ratio: true
95
+ type: Resize
96
+ - type: RandomFlip
97
+ - mean:
98
+ - 123.675
99
+ - 116.28
100
+ - 103.53
101
+ std:
102
+ - 58.395
103
+ - 57.12
104
+ - 57.375
105
+ to_rgb: true
106
+ type: Normalize
107
+ - size_divisor: 32
108
+ type: Pad
109
+ - keys:
110
+ - img
111
+ type: ImageToTensor
112
+ - keys:
113
+ - img
114
+ type: Collect
115
+ type: MultiScaleFlipAug
116
+ type: CocoDataset
117
+ workers_per_gpu: 2
118
+ data_root: data/coco/
119
+ dataset_type: CocoDataset
120
+ dist_params:
121
+ backend: nccl
122
+ evaluation:
123
+ interval: 1
124
+ metric: bbox
125
+ img_norm_cfg:
126
+ mean:
127
+ - 123.675
128
+ - 116.28
129
+ - 103.53
130
+ std:
131
+ - 58.395
132
+ - 57.12
133
+ - 57.375
134
+ to_rgb: true
135
+ load_from: null
136
+ log_config:
137
+ hooks:
138
+ - type: TextLoggerHook
139
+ interval: 50
140
+ log_level: INFO
141
+ lr_config:
142
+ policy: step
143
+ step:
144
+ - 9
145
+ - 11
146
+ warmup: linear
147
+ warmup_iters: 500
148
+ warmup_ratio: 0.001
149
+ model:
150
+ backbone:
151
+ base_width: 4
152
+ depth: 101
153
+ frozen_stages: 1
154
+ groups: 64
155
+ init_cfg:
156
+ checkpoint: open-mmlab://resnext101_64x4d
157
+ type: Pretrained
158
+ norm_cfg:
159
+ requires_grad: true
160
+ type: BN
161
+ norm_eval: true
162
+ num_stages: 4
163
+ out_indices: !!python/tuple
164
+ - 0
165
+ - 1
166
+ - 2
167
+ - 3
168
+ style: pytorch
169
+ type: ResNeXt
170
+ neck:
171
+ in_channels:
172
+ - 256
173
+ - 512
174
+ - 1024
175
+ - 2048
176
+ num_outs: 5
177
+ out_channels: 256
178
+ type: FPN
179
+ roi_head:
180
+ bbox_head:
181
+ bbox_coder:
182
+ target_means:
183
+ - 0.0
184
+ - 0.0
185
+ - 0.0
186
+ - 0.0
187
+ target_stds:
188
+ - 0.1
189
+ - 0.1
190
+ - 0.2
191
+ - 0.2
192
+ type: DeltaXYWHBBoxCoder
193
+ fc_out_channels: 1024
194
+ in_channels: 256
195
+ loss_bbox:
196
+ loss_weight: 1.0
197
+ type: L1Loss
198
+ loss_cls:
199
+ loss_weight: 1.0
200
+ type: CrossEntropyLoss
201
+ use_sigmoid: false
202
+ num_classes: 80
203
+ reg_class_agnostic: false
204
+ roi_feat_size: 7
205
+ type: Shared2FCBBoxHead
206
+ bbox_roi_extractor:
207
+ featmap_strides:
208
+ - 4
209
+ - 8
210
+ - 16
211
+ - 32
212
+ out_channels: 256
213
+ roi_layer:
214
+ output_size: 7
215
+ sampling_ratio: 0
216
+ type: RoIAlign
217
+ type: SingleRoIExtractor
218
+ type: StandardRoIHead
219
+ rpn_head:
220
+ anchor_generator:
221
+ ratios:
222
+ - 0.5
223
+ - 1.0
224
+ - 2.0
225
+ scales:
226
+ - 8
227
+ strides:
228
+ - 4
229
+ - 8
230
+ - 16
231
+ - 32
232
+ - 64
233
+ type: AnchorGenerator
234
+ bbox_coder:
235
+ target_means:
236
+ - 0.0
237
+ - 0.0
238
+ - 0.0
239
+ - 0.0
240
+ target_stds:
241
+ - 1.0
242
+ - 1.0
243
+ - 1.0
244
+ - 1.0
245
+ type: DeltaXYWHBBoxCoder
246
+ feat_channels: 256
247
+ in_channels: 256
248
+ loss_bbox:
249
+ loss_weight: 1.0
250
+ type: L1Loss
251
+ loss_cls:
252
+ loss_weight: 1.0
253
+ type: CrossEntropyLoss
254
+ use_sigmoid: true
255
+ type: RPNHead
256
+ test_cfg:
257
+ rcnn:
258
+ max_per_img: 100
259
+ nms:
260
+ iou_threshold: 0.5
261
+ type: nms
262
+ score_thr: 0.05
263
+ rpn:
264
+ max_per_img: 1000
265
+ min_bbox_size: 0
266
+ nms:
267
+ iou_threshold: 0.7
268
+ type: nms
269
+ nms_pre: 1000
270
+ train_cfg:
271
+ rcnn:
272
+ assigner:
273
+ ignore_iof_thr: -1
274
+ match_low_quality: false
275
+ min_pos_iou: 0.5
276
+ neg_iou_thr: 0.5
277
+ pos_iou_thr: 0.5
278
+ type: MaxIoUAssigner
279
+ debug: false
280
+ pos_weight: -1
281
+ sampler:
282
+ add_gt_as_proposals: true
283
+ neg_pos_ub: -1
284
+ num: 512
285
+ pos_fraction: 0.25
286
+ type: RandomSampler
287
+ rpn:
288
+ allowed_border: -1
289
+ assigner:
290
+ ignore_iof_thr: -1
291
+ match_low_quality: true
292
+ min_pos_iou: 0.3
293
+ neg_iou_thr: 0.3
294
+ pos_iou_thr: 0.7
295
+ type: MaxIoUAssigner
296
+ debug: false
297
+ pos_weight: -1
298
+ sampler:
299
+ add_gt_as_proposals: false
300
+ neg_pos_ub: -1
301
+ num: 256
302
+ pos_fraction: 0.5
303
+ type: RandomSampler
304
+ rpn_proposal:
305
+ max_per_img: 1000
306
+ min_bbox_size: 0
307
+ nms:
308
+ iou_threshold: 0.7
309
+ type: nms
310
+ nms_pre: 2000
311
+ type: FasterRCNN
312
+ mp_start_method: fork
313
+ opencv_num_threads: 0
314
+ optimizer:
315
+ lr: 0.02
316
+ momentum: 0.9
317
+ type: SGD
318
+ weight_decay: 0.0001
319
+ optimizer_config:
320
+ grad_clip: null
321
+ resume_from: null
322
+ runner:
323
+ max_epochs: 12
324
+ type: EpochBasedRunner
325
+ test_pipeline:
326
+ - type: LoadImageFromFile
327
+ - flip: false
328
+ img_scale: !!python/tuple
329
+ - 1333
330
+ - 800
331
+ transforms:
332
+ - keep_ratio: true
333
+ type: Resize
334
+ - type: RandomFlip
335
+ - mean:
336
+ - 123.675
337
+ - 116.28
338
+ - 103.53
339
+ std:
340
+ - 58.395
341
+ - 57.12
342
+ - 57.375
343
+ to_rgb: true
344
+ type: Normalize
345
+ - size_divisor: 32
346
+ type: Pad
347
+ - keys:
348
+ - img
349
+ type: ImageToTensor
350
+ - keys:
351
+ - img
352
+ type: Collect
353
+ type: MultiScaleFlipAug
354
+ train_pipeline:
355
+ - type: LoadImageFromFile
356
+ - type: LoadAnnotations
357
+ with_bbox: true
358
+ - img_scale:
359
+ - !!python/tuple
360
+ - 1333
361
+ - 640
362
+ - !!python/tuple
363
+ - 1333
364
+ - 800
365
+ keep_ratio: true
366
+ multiscale_mode: range
367
+ type: Resize
368
+ - flip_ratio: 0.5
369
+ type: RandomFlip
370
+ - mean:
371
+ - 123.675
372
+ - 116.28
373
+ - 103.53
374
+ std:
375
+ - 58.395
376
+ - 57.12
377
+ - 57.375
378
+ to_rgb: true
379
+ type: Normalize
380
+ - size_divisor: 32
381
+ type: Pad
382
+ - type: DefaultFormatBundle
383
+ - keys:
384
+ - img
385
+ - gt_bboxes
386
+ - gt_labels
387
+ type: Collect
388
+ workflow:
389
+ - !!python/tuple
390
+ - train
391
+ - 1
keypose_mmdet_fasterrcnn_x101/faster_rcnn_x101_64x4d_fpn_mstrain_3x_coco_20210524_124528-26c63de6.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e21d152cd52ccc48e037e365d808a6f2d032f99dc918afeeacdb62c1a553e462
3
+ size 399496787
keypose_mmpose_hrnet/README.md ADDED
@@ -0,0 +1 @@
 
 
1
+ Originally from https://github.com/open-mmlab/mmpose, converted to safetensors, used under Apache-2.0
keypose_mmpose_hrnet/hrnet_w48_coco_256x192-b9e0b3ab_20200708.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5213d35534129a5201dabb278ae699318f27f84f69b91b12ca9cc7cadc493a3d
3
+ size 254911178
keypose_mmpose_hrnet/hrnet_w48_coco_256x192.yaml ADDED
@@ -0,0 +1,2040 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ channel_cfg:
2
+ dataset_channel:
3
+ - - 0
4
+ - 1
5
+ - 2
6
+ - 3
7
+ - 4
8
+ - 5
9
+ - 6
10
+ - 7
11
+ - 8
12
+ - 9
13
+ - 10
14
+ - 11
15
+ - 12
16
+ - 13
17
+ - 14
18
+ - 15
19
+ - 16
20
+ dataset_joints: 17
21
+ inference_channel:
22
+ - 0
23
+ - 1
24
+ - 2
25
+ - 3
26
+ - 4
27
+ - 5
28
+ - 6
29
+ - 7
30
+ - 8
31
+ - 9
32
+ - 10
33
+ - 11
34
+ - 12
35
+ - 13
36
+ - 14
37
+ - 15
38
+ - 16
39
+ num_output_channels: 17
40
+ checkpoint_config:
41
+ interval: 10
42
+ data:
43
+ samples_per_gpu: 32
44
+ test:
45
+ ann_file: data/coco/annotations/person_keypoints_val2017.json
46
+ data_cfg:
47
+ bbox_file: data/coco/person_detection_results/COCO_val2017_detections_AP_H_56_person.json
48
+ dataset_channel:
49
+ - - 0
50
+ - 1
51
+ - 2
52
+ - 3
53
+ - 4
54
+ - 5
55
+ - 6
56
+ - 7
57
+ - 8
58
+ - 9
59
+ - 10
60
+ - 11
61
+ - 12
62
+ - 13
63
+ - 14
64
+ - 15
65
+ - 16
66
+ det_bbox_thr: 0.0
67
+ heatmap_size:
68
+ - 48
69
+ - 64
70
+ image_size:
71
+ - 192
72
+ - 256
73
+ inference_channel:
74
+ - 0
75
+ - 1
76
+ - 2
77
+ - 3
78
+ - 4
79
+ - 5
80
+ - 6
81
+ - 7
82
+ - 8
83
+ - 9
84
+ - 10
85
+ - 11
86
+ - 12
87
+ - 13
88
+ - 14
89
+ - 15
90
+ - 16
91
+ nms_thr: 1.0
92
+ num_joints: 17
93
+ num_output_channels: 17
94
+ oks_thr: 0.9
95
+ soft_nms: false
96
+ use_gt_bbox: false
97
+ vis_thr: 0.2
98
+ dataset_info:
99
+ dataset_name: coco
100
+ joint_weights:
101
+ - 1.0
102
+ - 1.0
103
+ - 1.0
104
+ - 1.0
105
+ - 1.0
106
+ - 1.0
107
+ - 1.0
108
+ - 1.2
109
+ - 1.2
110
+ - 1.5
111
+ - 1.5
112
+ - 1.0
113
+ - 1.0
114
+ - 1.2
115
+ - 1.2
116
+ - 1.5
117
+ - 1.5
118
+ keypoint_info:
119
+ 0:
120
+ color:
121
+ - 51
122
+ - 153
123
+ - 255
124
+ id: 0
125
+ name: nose
126
+ swap: ''
127
+ type: upper
128
+ 1:
129
+ color:
130
+ - 51
131
+ - 153
132
+ - 255
133
+ id: 1
134
+ name: left_eye
135
+ swap: right_eye
136
+ type: upper
137
+ 2:
138
+ color:
139
+ - 51
140
+ - 153
141
+ - 255
142
+ id: 2
143
+ name: right_eye
144
+ swap: left_eye
145
+ type: upper
146
+ 3:
147
+ color:
148
+ - 51
149
+ - 153
150
+ - 255
151
+ id: 3
152
+ name: left_ear
153
+ swap: right_ear
154
+ type: upper
155
+ 4:
156
+ color:
157
+ - 51
158
+ - 153
159
+ - 255
160
+ id: 4
161
+ name: right_ear
162
+ swap: left_ear
163
+ type: upper
164
+ 5:
165
+ color:
166
+ - 0
167
+ - 255
168
+ - 0
169
+ id: 5
170
+ name: left_shoulder
171
+ swap: right_shoulder
172
+ type: upper
173
+ 6:
174
+ color:
175
+ - 255
176
+ - 128
177
+ - 0
178
+ id: 6
179
+ name: right_shoulder
180
+ swap: left_shoulder
181
+ type: upper
182
+ 7:
183
+ color:
184
+ - 0
185
+ - 255
186
+ - 0
187
+ id: 7
188
+ name: left_elbow
189
+ swap: right_elbow
190
+ type: upper
191
+ 8:
192
+ color:
193
+ - 255
194
+ - 128
195
+ - 0
196
+ id: 8
197
+ name: right_elbow
198
+ swap: left_elbow
199
+ type: upper
200
+ 9:
201
+ color:
202
+ - 0
203
+ - 255
204
+ - 0
205
+ id: 9
206
+ name: left_wrist
207
+ swap: right_wrist
208
+ type: upper
209
+ 10:
210
+ color:
211
+ - 255
212
+ - 128
213
+ - 0
214
+ id: 10
215
+ name: right_wrist
216
+ swap: left_wrist
217
+ type: upper
218
+ 11:
219
+ color:
220
+ - 0
221
+ - 255
222
+ - 0
223
+ id: 11
224
+ name: left_hip
225
+ swap: right_hip
226
+ type: lower
227
+ 12:
228
+ color:
229
+ - 255
230
+ - 128
231
+ - 0
232
+ id: 12
233
+ name: right_hip
234
+ swap: left_hip
235
+ type: lower
236
+ 13:
237
+ color:
238
+ - 0
239
+ - 255
240
+ - 0
241
+ id: 13
242
+ name: left_knee
243
+ swap: right_knee
244
+ type: lower
245
+ 14:
246
+ color:
247
+ - 255
248
+ - 128
249
+ - 0
250
+ id: 14
251
+ name: right_knee
252
+ swap: left_knee
253
+ type: lower
254
+ 15:
255
+ color:
256
+ - 0
257
+ - 255
258
+ - 0
259
+ id: 15
260
+ name: left_ankle
261
+ swap: right_ankle
262
+ type: lower
263
+ 16:
264
+ color:
265
+ - 255
266
+ - 128
267
+ - 0
268
+ id: 16
269
+ name: right_ankle
270
+ swap: left_ankle
271
+ type: lower
272
+ paper_info:
273
+ author: Lin, Tsung-Yi and Maire, Michael and Belongie, Serge and Hays, James
274
+ and Perona, Pietro and Ramanan, Deva and Doll{\'a}r, Piotr and Zitnick,
275
+ C Lawrence
276
+ container: European conference on computer vision
277
+ homepage: http://cocodataset.org/
278
+ title: 'Microsoft coco: Common objects in context'
279
+ year: '2014'
280
+ sigmas:
281
+ - 0.026
282
+ - 0.025
283
+ - 0.025
284
+ - 0.035
285
+ - 0.035
286
+ - 0.079
287
+ - 0.079
288
+ - 0.072
289
+ - 0.072
290
+ - 0.062
291
+ - 0.062
292
+ - 0.107
293
+ - 0.107
294
+ - 0.087
295
+ - 0.087
296
+ - 0.089
297
+ - 0.089
298
+ skeleton_info:
299
+ 0:
300
+ color:
301
+ - 0
302
+ - 255
303
+ - 0
304
+ id: 0
305
+ link: !!python/tuple
306
+ - left_ankle
307
+ - left_knee
308
+ 1:
309
+ color:
310
+ - 0
311
+ - 255
312
+ - 0
313
+ id: 1
314
+ link: !!python/tuple
315
+ - left_knee
316
+ - left_hip
317
+ 2:
318
+ color:
319
+ - 255
320
+ - 128
321
+ - 0
322
+ id: 2
323
+ link: !!python/tuple
324
+ - right_ankle
325
+ - right_knee
326
+ 3:
327
+ color:
328
+ - 255
329
+ - 128
330
+ - 0
331
+ id: 3
332
+ link: !!python/tuple
333
+ - right_knee
334
+ - right_hip
335
+ 4:
336
+ color:
337
+ - 51
338
+ - 153
339
+ - 255
340
+ id: 4
341
+ link: !!python/tuple
342
+ - left_hip
343
+ - right_hip
344
+ 5:
345
+ color:
346
+ - 51
347
+ - 153
348
+ - 255
349
+ id: 5
350
+ link: !!python/tuple
351
+ - left_shoulder
352
+ - left_hip
353
+ 6:
354
+ color:
355
+ - 51
356
+ - 153
357
+ - 255
358
+ id: 6
359
+ link: !!python/tuple
360
+ - right_shoulder
361
+ - right_hip
362
+ 7:
363
+ color:
364
+ - 51
365
+ - 153
366
+ - 255
367
+ id: 7
368
+ link: !!python/tuple
369
+ - left_shoulder
370
+ - right_shoulder
371
+ 8:
372
+ color:
373
+ - 0
374
+ - 255
375
+ - 0
376
+ id: 8
377
+ link: !!python/tuple
378
+ - left_shoulder
379
+ - left_elbow
380
+ 9:
381
+ color:
382
+ - 255
383
+ - 128
384
+ - 0
385
+ id: 9
386
+ link: !!python/tuple
387
+ - right_shoulder
388
+ - right_elbow
389
+ 10:
390
+ color:
391
+ - 0
392
+ - 255
393
+ - 0
394
+ id: 10
395
+ link: !!python/tuple
396
+ - left_elbow
397
+ - left_wrist
398
+ 11:
399
+ color:
400
+ - 255
401
+ - 128
402
+ - 0
403
+ id: 11
404
+ link: !!python/tuple
405
+ - right_elbow
406
+ - right_wrist
407
+ 12:
408
+ color:
409
+ - 51
410
+ - 153
411
+ - 255
412
+ id: 12
413
+ link: !!python/tuple
414
+ - left_eye
415
+ - right_eye
416
+ 13:
417
+ color:
418
+ - 51
419
+ - 153
420
+ - 255
421
+ id: 13
422
+ link: !!python/tuple
423
+ - nose
424
+ - left_eye
425
+ 14:
426
+ color:
427
+ - 51
428
+ - 153
429
+ - 255
430
+ id: 14
431
+ link: !!python/tuple
432
+ - nose
433
+ - right_eye
434
+ 15:
435
+ color:
436
+ - 51
437
+ - 153
438
+ - 255
439
+ id: 15
440
+ link: !!python/tuple
441
+ - left_eye
442
+ - left_ear
443
+ 16:
444
+ color:
445
+ - 51
446
+ - 153
447
+ - 255
448
+ id: 16
449
+ link: !!python/tuple
450
+ - right_eye
451
+ - right_ear
452
+ 17:
453
+ color:
454
+ - 51
455
+ - 153
456
+ - 255
457
+ id: 17
458
+ link: !!python/tuple
459
+ - left_ear
460
+ - left_shoulder
461
+ 18:
462
+ color:
463
+ - 51
464
+ - 153
465
+ - 255
466
+ id: 18
467
+ link: !!python/tuple
468
+ - right_ear
469
+ - right_shoulder
470
+ img_prefix: data/coco/val2017/
471
+ pipeline:
472
+ - type: LoadImageFromFile
473
+ - padding: 1.25
474
+ type: TopDownGetBboxCenterScale
475
+ - type: TopDownAffine
476
+ - type: ToTensor
477
+ - mean:
478
+ - 0.485
479
+ - 0.456
480
+ - 0.406
481
+ std:
482
+ - 0.229
483
+ - 0.224
484
+ - 0.225
485
+ type: NormalizeTensor
486
+ - keys:
487
+ - img
488
+ meta_keys:
489
+ - image_file
490
+ - center
491
+ - scale
492
+ - rotation
493
+ - bbox_score
494
+ - flip_pairs
495
+ type: Collect
496
+ type: TopDownCocoDataset
497
+ test_dataloader:
498
+ samples_per_gpu: 32
499
+ train:
500
+ ann_file: data/coco/annotations/person_keypoints_train2017.json
501
+ data_cfg:
502
+ bbox_file: data/coco/person_detection_results/COCO_val2017_detections_AP_H_56_person.json
503
+ dataset_channel:
504
+ - - 0
505
+ - 1
506
+ - 2
507
+ - 3
508
+ - 4
509
+ - 5
510
+ - 6
511
+ - 7
512
+ - 8
513
+ - 9
514
+ - 10
515
+ - 11
516
+ - 12
517
+ - 13
518
+ - 14
519
+ - 15
520
+ - 16
521
+ det_bbox_thr: 0.0
522
+ heatmap_size:
523
+ - 48
524
+ - 64
525
+ image_size:
526
+ - 192
527
+ - 256
528
+ inference_channel:
529
+ - 0
530
+ - 1
531
+ - 2
532
+ - 3
533
+ - 4
534
+ - 5
535
+ - 6
536
+ - 7
537
+ - 8
538
+ - 9
539
+ - 10
540
+ - 11
541
+ - 12
542
+ - 13
543
+ - 14
544
+ - 15
545
+ - 16
546
+ nms_thr: 1.0
547
+ num_joints: 17
548
+ num_output_channels: 17
549
+ oks_thr: 0.9
550
+ soft_nms: false
551
+ use_gt_bbox: false
552
+ vis_thr: 0.2
553
+ dataset_info:
554
+ dataset_name: coco
555
+ joint_weights:
556
+ - 1.0
557
+ - 1.0
558
+ - 1.0
559
+ - 1.0
560
+ - 1.0
561
+ - 1.0
562
+ - 1.0
563
+ - 1.2
564
+ - 1.2
565
+ - 1.5
566
+ - 1.5
567
+ - 1.0
568
+ - 1.0
569
+ - 1.2
570
+ - 1.2
571
+ - 1.5
572
+ - 1.5
573
+ keypoint_info:
574
+ 0:
575
+ color:
576
+ - 51
577
+ - 153
578
+ - 255
579
+ id: 0
580
+ name: nose
581
+ swap: ''
582
+ type: upper
583
+ 1:
584
+ color:
585
+ - 51
586
+ - 153
587
+ - 255
588
+ id: 1
589
+ name: left_eye
590
+ swap: right_eye
591
+ type: upper
592
+ 2:
593
+ color:
594
+ - 51
595
+ - 153
596
+ - 255
597
+ id: 2
598
+ name: right_eye
599
+ swap: left_eye
600
+ type: upper
601
+ 3:
602
+ color:
603
+ - 51
604
+ - 153
605
+ - 255
606
+ id: 3
607
+ name: left_ear
608
+ swap: right_ear
609
+ type: upper
610
+ 4:
611
+ color:
612
+ - 51
613
+ - 153
614
+ - 255
615
+ id: 4
616
+ name: right_ear
617
+ swap: left_ear
618
+ type: upper
619
+ 5:
620
+ color:
621
+ - 0
622
+ - 255
623
+ - 0
624
+ id: 5
625
+ name: left_shoulder
626
+ swap: right_shoulder
627
+ type: upper
628
+ 6:
629
+ color:
630
+ - 255
631
+ - 128
632
+ - 0
633
+ id: 6
634
+ name: right_shoulder
635
+ swap: left_shoulder
636
+ type: upper
637
+ 7:
638
+ color:
639
+ - 0
640
+ - 255
641
+ - 0
642
+ id: 7
643
+ name: left_elbow
644
+ swap: right_elbow
645
+ type: upper
646
+ 8:
647
+ color:
648
+ - 255
649
+ - 128
650
+ - 0
651
+ id: 8
652
+ name: right_elbow
653
+ swap: left_elbow
654
+ type: upper
655
+ 9:
656
+ color:
657
+ - 0
658
+ - 255
659
+ - 0
660
+ id: 9
661
+ name: left_wrist
662
+ swap: right_wrist
663
+ type: upper
664
+ 10:
665
+ color:
666
+ - 255
667
+ - 128
668
+ - 0
669
+ id: 10
670
+ name: right_wrist
671
+ swap: left_wrist
672
+ type: upper
673
+ 11:
674
+ color:
675
+ - 0
676
+ - 255
677
+ - 0
678
+ id: 11
679
+ name: left_hip
680
+ swap: right_hip
681
+ type: lower
682
+ 12:
683
+ color:
684
+ - 255
685
+ - 128
686
+ - 0
687
+ id: 12
688
+ name: right_hip
689
+ swap: left_hip
690
+ type: lower
691
+ 13:
692
+ color:
693
+ - 0
694
+ - 255
695
+ - 0
696
+ id: 13
697
+ name: left_knee
698
+ swap: right_knee
699
+ type: lower
700
+ 14:
701
+ color:
702
+ - 255
703
+ - 128
704
+ - 0
705
+ id: 14
706
+ name: right_knee
707
+ swap: left_knee
708
+ type: lower
709
+ 15:
710
+ color:
711
+ - 0
712
+ - 255
713
+ - 0
714
+ id: 15
715
+ name: left_ankle
716
+ swap: right_ankle
717
+ type: lower
718
+ 16:
719
+ color:
720
+ - 255
721
+ - 128
722
+ - 0
723
+ id: 16
724
+ name: right_ankle
725
+ swap: left_ankle
726
+ type: lower
727
+ paper_info:
728
+ author: Lin, Tsung-Yi and Maire, Michael and Belongie, Serge and Hays, James
729
+ and Perona, Pietro and Ramanan, Deva and Doll{\'a}r, Piotr and Zitnick,
730
+ C Lawrence
731
+ container: European conference on computer vision
732
+ homepage: http://cocodataset.org/
733
+ title: 'Microsoft coco: Common objects in context'
734
+ year: '2014'
735
+ sigmas:
736
+ - 0.026
737
+ - 0.025
738
+ - 0.025
739
+ - 0.035
740
+ - 0.035
741
+ - 0.079
742
+ - 0.079
743
+ - 0.072
744
+ - 0.072
745
+ - 0.062
746
+ - 0.062
747
+ - 0.107
748
+ - 0.107
749
+ - 0.087
750
+ - 0.087
751
+ - 0.089
752
+ - 0.089
753
+ skeleton_info:
754
+ 0:
755
+ color:
756
+ - 0
757
+ - 255
758
+ - 0
759
+ id: 0
760
+ link: !!python/tuple
761
+ - left_ankle
762
+ - left_knee
763
+ 1:
764
+ color:
765
+ - 0
766
+ - 255
767
+ - 0
768
+ id: 1
769
+ link: !!python/tuple
770
+ - left_knee
771
+ - left_hip
772
+ 2:
773
+ color:
774
+ - 255
775
+ - 128
776
+ - 0
777
+ id: 2
778
+ link: !!python/tuple
779
+ - right_ankle
780
+ - right_knee
781
+ 3:
782
+ color:
783
+ - 255
784
+ - 128
785
+ - 0
786
+ id: 3
787
+ link: !!python/tuple
788
+ - right_knee
789
+ - right_hip
790
+ 4:
791
+ color:
792
+ - 51
793
+ - 153
794
+ - 255
795
+ id: 4
796
+ link: !!python/tuple
797
+ - left_hip
798
+ - right_hip
799
+ 5:
800
+ color:
801
+ - 51
802
+ - 153
803
+ - 255
804
+ id: 5
805
+ link: !!python/tuple
806
+ - left_shoulder
807
+ - left_hip
808
+ 6:
809
+ color:
810
+ - 51
811
+ - 153
812
+ - 255
813
+ id: 6
814
+ link: !!python/tuple
815
+ - right_shoulder
816
+ - right_hip
817
+ 7:
818
+ color:
819
+ - 51
820
+ - 153
821
+ - 255
822
+ id: 7
823
+ link: !!python/tuple
824
+ - left_shoulder
825
+ - right_shoulder
826
+ 8:
827
+ color:
828
+ - 0
829
+ - 255
830
+ - 0
831
+ id: 8
832
+ link: !!python/tuple
833
+ - left_shoulder
834
+ - left_elbow
835
+ 9:
836
+ color:
837
+ - 255
838
+ - 128
839
+ - 0
840
+ id: 9
841
+ link: !!python/tuple
842
+ - right_shoulder
843
+ - right_elbow
844
+ 10:
845
+ color:
846
+ - 0
847
+ - 255
848
+ - 0
849
+ id: 10
850
+ link: !!python/tuple
851
+ - left_elbow
852
+ - left_wrist
853
+ 11:
854
+ color:
855
+ - 255
856
+ - 128
857
+ - 0
858
+ id: 11
859
+ link: !!python/tuple
860
+ - right_elbow
861
+ - right_wrist
862
+ 12:
863
+ color:
864
+ - 51
865
+ - 153
866
+ - 255
867
+ id: 12
868
+ link: !!python/tuple
869
+ - left_eye
870
+ - right_eye
871
+ 13:
872
+ color:
873
+ - 51
874
+ - 153
875
+ - 255
876
+ id: 13
877
+ link: !!python/tuple
878
+ - nose
879
+ - left_eye
880
+ 14:
881
+ color:
882
+ - 51
883
+ - 153
884
+ - 255
885
+ id: 14
886
+ link: !!python/tuple
887
+ - nose
888
+ - right_eye
889
+ 15:
890
+ color:
891
+ - 51
892
+ - 153
893
+ - 255
894
+ id: 15
895
+ link: !!python/tuple
896
+ - left_eye
897
+ - left_ear
898
+ 16:
899
+ color:
900
+ - 51
901
+ - 153
902
+ - 255
903
+ id: 16
904
+ link: !!python/tuple
905
+ - right_eye
906
+ - right_ear
907
+ 17:
908
+ color:
909
+ - 51
910
+ - 153
911
+ - 255
912
+ id: 17
913
+ link: !!python/tuple
914
+ - left_ear
915
+ - left_shoulder
916
+ 18:
917
+ color:
918
+ - 51
919
+ - 153
920
+ - 255
921
+ id: 18
922
+ link: !!python/tuple
923
+ - right_ear
924
+ - right_shoulder
925
+ img_prefix: data/coco/train2017/
926
+ pipeline:
927
+ - type: LoadImageFromFile
928
+ - padding: 1.25
929
+ type: TopDownGetBboxCenterScale
930
+ - prob: 0.3
931
+ shift_factor: 0.16
932
+ type: TopDownRandomShiftBboxCenter
933
+ - flip_prob: 0.5
934
+ type: TopDownRandomFlip
935
+ - num_joints_half_body: 8
936
+ prob_half_body: 0.3
937
+ type: TopDownHalfBodyTransform
938
+ - rot_factor: 40
939
+ scale_factor: 0.5
940
+ type: TopDownGetRandomScaleRotation
941
+ - type: TopDownAffine
942
+ - type: ToTensor
943
+ - mean:
944
+ - 0.485
945
+ - 0.456
946
+ - 0.406
947
+ std:
948
+ - 0.229
949
+ - 0.224
950
+ - 0.225
951
+ type: NormalizeTensor
952
+ - sigma: 2
953
+ type: TopDownGenerateTarget
954
+ - keys:
955
+ - img
956
+ - target
957
+ - target_weight
958
+ meta_keys:
959
+ - image_file
960
+ - joints_3d
961
+ - joints_3d_visible
962
+ - center
963
+ - scale
964
+ - rotation
965
+ - bbox_score
966
+ - flip_pairs
967
+ type: Collect
968
+ type: TopDownCocoDataset
969
+ val:
970
+ ann_file: data/coco/annotations/person_keypoints_val2017.json
971
+ data_cfg:
972
+ bbox_file: data/coco/person_detection_results/COCO_val2017_detections_AP_H_56_person.json
973
+ dataset_channel:
974
+ - - 0
975
+ - 1
976
+ - 2
977
+ - 3
978
+ - 4
979
+ - 5
980
+ - 6
981
+ - 7
982
+ - 8
983
+ - 9
984
+ - 10
985
+ - 11
986
+ - 12
987
+ - 13
988
+ - 14
989
+ - 15
990
+ - 16
991
+ det_bbox_thr: 0.0
992
+ heatmap_size:
993
+ - 48
994
+ - 64
995
+ image_size:
996
+ - 192
997
+ - 256
998
+ inference_channel:
999
+ - 0
1000
+ - 1
1001
+ - 2
1002
+ - 3
1003
+ - 4
1004
+ - 5
1005
+ - 6
1006
+ - 7
1007
+ - 8
1008
+ - 9
1009
+ - 10
1010
+ - 11
1011
+ - 12
1012
+ - 13
1013
+ - 14
1014
+ - 15
1015
+ - 16
1016
+ nms_thr: 1.0
1017
+ num_joints: 17
1018
+ num_output_channels: 17
1019
+ oks_thr: 0.9
1020
+ soft_nms: false
1021
+ use_gt_bbox: false
1022
+ vis_thr: 0.2
1023
+ dataset_info:
1024
+ dataset_name: coco
1025
+ joint_weights:
1026
+ - 1.0
1027
+ - 1.0
1028
+ - 1.0
1029
+ - 1.0
1030
+ - 1.0
1031
+ - 1.0
1032
+ - 1.0
1033
+ - 1.2
1034
+ - 1.2
1035
+ - 1.5
1036
+ - 1.5
1037
+ - 1.0
1038
+ - 1.0
1039
+ - 1.2
1040
+ - 1.2
1041
+ - 1.5
1042
+ - 1.5
1043
+ keypoint_info:
1044
+ 0:
1045
+ color:
1046
+ - 51
1047
+ - 153
1048
+ - 255
1049
+ id: 0
1050
+ name: nose
1051
+ swap: ''
1052
+ type: upper
1053
+ 1:
1054
+ color:
1055
+ - 51
1056
+ - 153
1057
+ - 255
1058
+ id: 1
1059
+ name: left_eye
1060
+ swap: right_eye
1061
+ type: upper
1062
+ 2:
1063
+ color:
1064
+ - 51
1065
+ - 153
1066
+ - 255
1067
+ id: 2
1068
+ name: right_eye
1069
+ swap: left_eye
1070
+ type: upper
1071
+ 3:
1072
+ color:
1073
+ - 51
1074
+ - 153
1075
+ - 255
1076
+ id: 3
1077
+ name: left_ear
1078
+ swap: right_ear
1079
+ type: upper
1080
+ 4:
1081
+ color:
1082
+ - 51
1083
+ - 153
1084
+ - 255
1085
+ id: 4
1086
+ name: right_ear
1087
+ swap: left_ear
1088
+ type: upper
1089
+ 5:
1090
+ color:
1091
+ - 0
1092
+ - 255
1093
+ - 0
1094
+ id: 5
1095
+ name: left_shoulder
1096
+ swap: right_shoulder
1097
+ type: upper
1098
+ 6:
1099
+ color:
1100
+ - 255
1101
+ - 128
1102
+ - 0
1103
+ id: 6
1104
+ name: right_shoulder
1105
+ swap: left_shoulder
1106
+ type: upper
1107
+ 7:
1108
+ color:
1109
+ - 0
1110
+ - 255
1111
+ - 0
1112
+ id: 7
1113
+ name: left_elbow
1114
+ swap: right_elbow
1115
+ type: upper
1116
+ 8:
1117
+ color:
1118
+ - 255
1119
+ - 128
1120
+ - 0
1121
+ id: 8
1122
+ name: right_elbow
1123
+ swap: left_elbow
1124
+ type: upper
1125
+ 9:
1126
+ color:
1127
+ - 0
1128
+ - 255
1129
+ - 0
1130
+ id: 9
1131
+ name: left_wrist
1132
+ swap: right_wrist
1133
+ type: upper
1134
+ 10:
1135
+ color:
1136
+ - 255
1137
+ - 128
1138
+ - 0
1139
+ id: 10
1140
+ name: right_wrist
1141
+ swap: left_wrist
1142
+ type: upper
1143
+ 11:
1144
+ color:
1145
+ - 0
1146
+ - 255
1147
+ - 0
1148
+ id: 11
1149
+ name: left_hip
1150
+ swap: right_hip
1151
+ type: lower
1152
+ 12:
1153
+ color:
1154
+ - 255
1155
+ - 128
1156
+ - 0
1157
+ id: 12
1158
+ name: right_hip
1159
+ swap: left_hip
1160
+ type: lower
1161
+ 13:
1162
+ color:
1163
+ - 0
1164
+ - 255
1165
+ - 0
1166
+ id: 13
1167
+ name: left_knee
1168
+ swap: right_knee
1169
+ type: lower
1170
+ 14:
1171
+ color:
1172
+ - 255
1173
+ - 128
1174
+ - 0
1175
+ id: 14
1176
+ name: right_knee
1177
+ swap: left_knee
1178
+ type: lower
1179
+ 15:
1180
+ color:
1181
+ - 0
1182
+ - 255
1183
+ - 0
1184
+ id: 15
1185
+ name: left_ankle
1186
+ swap: right_ankle
1187
+ type: lower
1188
+ 16:
1189
+ color:
1190
+ - 255
1191
+ - 128
1192
+ - 0
1193
+ id: 16
1194
+ name: right_ankle
1195
+ swap: left_ankle
1196
+ type: lower
1197
+ paper_info:
1198
+ author: Lin, Tsung-Yi and Maire, Michael and Belongie, Serge and Hays, James
1199
+ and Perona, Pietro and Ramanan, Deva and Doll{\'a}r, Piotr and Zitnick,
1200
+ C Lawrence
1201
+ container: European conference on computer vision
1202
+ homepage: http://cocodataset.org/
1203
+ title: 'Microsoft coco: Common objects in context'
1204
+ year: '2014'
1205
+ sigmas:
1206
+ - 0.026
1207
+ - 0.025
1208
+ - 0.025
1209
+ - 0.035
1210
+ - 0.035
1211
+ - 0.079
1212
+ - 0.079
1213
+ - 0.072
1214
+ - 0.072
1215
+ - 0.062
1216
+ - 0.062
1217
+ - 0.107
1218
+ - 0.107
1219
+ - 0.087
1220
+ - 0.087
1221
+ - 0.089
1222
+ - 0.089
1223
+ skeleton_info:
1224
+ 0:
1225
+ color:
1226
+ - 0
1227
+ - 255
1228
+ - 0
1229
+ id: 0
1230
+ link: !!python/tuple
1231
+ - left_ankle
1232
+ - left_knee
1233
+ 1:
1234
+ color:
1235
+ - 0
1236
+ - 255
1237
+ - 0
1238
+ id: 1
1239
+ link: !!python/tuple
1240
+ - left_knee
1241
+ - left_hip
1242
+ 2:
1243
+ color:
1244
+ - 255
1245
+ - 128
1246
+ - 0
1247
+ id: 2
1248
+ link: !!python/tuple
1249
+ - right_ankle
1250
+ - right_knee
1251
+ 3:
1252
+ color:
1253
+ - 255
1254
+ - 128
1255
+ - 0
1256
+ id: 3
1257
+ link: !!python/tuple
1258
+ - right_knee
1259
+ - right_hip
1260
+ 4:
1261
+ color:
1262
+ - 51
1263
+ - 153
1264
+ - 255
1265
+ id: 4
1266
+ link: !!python/tuple
1267
+ - left_hip
1268
+ - right_hip
1269
+ 5:
1270
+ color:
1271
+ - 51
1272
+ - 153
1273
+ - 255
1274
+ id: 5
1275
+ link: !!python/tuple
1276
+ - left_shoulder
1277
+ - left_hip
1278
+ 6:
1279
+ color:
1280
+ - 51
1281
+ - 153
1282
+ - 255
1283
+ id: 6
1284
+ link: !!python/tuple
1285
+ - right_shoulder
1286
+ - right_hip
1287
+ 7:
1288
+ color:
1289
+ - 51
1290
+ - 153
1291
+ - 255
1292
+ id: 7
1293
+ link: !!python/tuple
1294
+ - left_shoulder
1295
+ - right_shoulder
1296
+ 8:
1297
+ color:
1298
+ - 0
1299
+ - 255
1300
+ - 0
1301
+ id: 8
1302
+ link: !!python/tuple
1303
+ - left_shoulder
1304
+ - left_elbow
1305
+ 9:
1306
+ color:
1307
+ - 255
1308
+ - 128
1309
+ - 0
1310
+ id: 9
1311
+ link: !!python/tuple
1312
+ - right_shoulder
1313
+ - right_elbow
1314
+ 10:
1315
+ color:
1316
+ - 0
1317
+ - 255
1318
+ - 0
1319
+ id: 10
1320
+ link: !!python/tuple
1321
+ - left_elbow
1322
+ - left_wrist
1323
+ 11:
1324
+ color:
1325
+ - 255
1326
+ - 128
1327
+ - 0
1328
+ id: 11
1329
+ link: !!python/tuple
1330
+ - right_elbow
1331
+ - right_wrist
1332
+ 12:
1333
+ color:
1334
+ - 51
1335
+ - 153
1336
+ - 255
1337
+ id: 12
1338
+ link: !!python/tuple
1339
+ - left_eye
1340
+ - right_eye
1341
+ 13:
1342
+ color:
1343
+ - 51
1344
+ - 153
1345
+ - 255
1346
+ id: 13
1347
+ link: !!python/tuple
1348
+ - nose
1349
+ - left_eye
1350
+ 14:
1351
+ color:
1352
+ - 51
1353
+ - 153
1354
+ - 255
1355
+ id: 14
1356
+ link: !!python/tuple
1357
+ - nose
1358
+ - right_eye
1359
+ 15:
1360
+ color:
1361
+ - 51
1362
+ - 153
1363
+ - 255
1364
+ id: 15
1365
+ link: !!python/tuple
1366
+ - left_eye
1367
+ - left_ear
1368
+ 16:
1369
+ color:
1370
+ - 51
1371
+ - 153
1372
+ - 255
1373
+ id: 16
1374
+ link: !!python/tuple
1375
+ - right_eye
1376
+ - right_ear
1377
+ 17:
1378
+ color:
1379
+ - 51
1380
+ - 153
1381
+ - 255
1382
+ id: 17
1383
+ link: !!python/tuple
1384
+ - left_ear
1385
+ - left_shoulder
1386
+ 18:
1387
+ color:
1388
+ - 51
1389
+ - 153
1390
+ - 255
1391
+ id: 18
1392
+ link: !!python/tuple
1393
+ - right_ear
1394
+ - right_shoulder
1395
+ img_prefix: data/coco/val2017/
1396
+ pipeline:
1397
+ - type: LoadImageFromFile
1398
+ - padding: 1.25
1399
+ type: TopDownGetBboxCenterScale
1400
+ - type: TopDownAffine
1401
+ - type: ToTensor
1402
+ - mean:
1403
+ - 0.485
1404
+ - 0.456
1405
+ - 0.406
1406
+ std:
1407
+ - 0.229
1408
+ - 0.224
1409
+ - 0.225
1410
+ type: NormalizeTensor
1411
+ - keys:
1412
+ - img
1413
+ meta_keys:
1414
+ - image_file
1415
+ - center
1416
+ - scale
1417
+ - rotation
1418
+ - bbox_score
1419
+ - flip_pairs
1420
+ type: Collect
1421
+ type: TopDownCocoDataset
1422
+ val_dataloader:
1423
+ samples_per_gpu: 32
1424
+ workers_per_gpu: 2
1425
+ data_cfg:
1426
+ bbox_file: data/coco/person_detection_results/COCO_val2017_detections_AP_H_56_person.json
1427
+ dataset_channel:
1428
+ - - 0
1429
+ - 1
1430
+ - 2
1431
+ - 3
1432
+ - 4
1433
+ - 5
1434
+ - 6
1435
+ - 7
1436
+ - 8
1437
+ - 9
1438
+ - 10
1439
+ - 11
1440
+ - 12
1441
+ - 13
1442
+ - 14
1443
+ - 15
1444
+ - 16
1445
+ det_bbox_thr: 0.0
1446
+ heatmap_size:
1447
+ - 48
1448
+ - 64
1449
+ image_size:
1450
+ - 192
1451
+ - 256
1452
+ inference_channel:
1453
+ - 0
1454
+ - 1
1455
+ - 2
1456
+ - 3
1457
+ - 4
1458
+ - 5
1459
+ - 6
1460
+ - 7
1461
+ - 8
1462
+ - 9
1463
+ - 10
1464
+ - 11
1465
+ - 12
1466
+ - 13
1467
+ - 14
1468
+ - 15
1469
+ - 16
1470
+ nms_thr: 1.0
1471
+ num_joints: 17
1472
+ num_output_channels: 17
1473
+ oks_thr: 0.9
1474
+ soft_nms: false
1475
+ use_gt_bbox: false
1476
+ vis_thr: 0.2
1477
+ data_root: data/coco
1478
+ dataset_info:
1479
+ dataset_name: coco
1480
+ joint_weights:
1481
+ - 1.0
1482
+ - 1.0
1483
+ - 1.0
1484
+ - 1.0
1485
+ - 1.0
1486
+ - 1.0
1487
+ - 1.0
1488
+ - 1.2
1489
+ - 1.2
1490
+ - 1.5
1491
+ - 1.5
1492
+ - 1.0
1493
+ - 1.0
1494
+ - 1.2
1495
+ - 1.2
1496
+ - 1.5
1497
+ - 1.5
1498
+ keypoint_info:
1499
+ 0:
1500
+ color:
1501
+ - 51
1502
+ - 153
1503
+ - 255
1504
+ id: 0
1505
+ name: nose
1506
+ swap: ''
1507
+ type: upper
1508
+ 1:
1509
+ color:
1510
+ - 51
1511
+ - 153
1512
+ - 255
1513
+ id: 1
1514
+ name: left_eye
1515
+ swap: right_eye
1516
+ type: upper
1517
+ 2:
1518
+ color:
1519
+ - 51
1520
+ - 153
1521
+ - 255
1522
+ id: 2
1523
+ name: right_eye
1524
+ swap: left_eye
1525
+ type: upper
1526
+ 3:
1527
+ color:
1528
+ - 51
1529
+ - 153
1530
+ - 255
1531
+ id: 3
1532
+ name: left_ear
1533
+ swap: right_ear
1534
+ type: upper
1535
+ 4:
1536
+ color:
1537
+ - 51
1538
+ - 153
1539
+ - 255
1540
+ id: 4
1541
+ name: right_ear
1542
+ swap: left_ear
1543
+ type: upper
1544
+ 5:
1545
+ color:
1546
+ - 0
1547
+ - 255
1548
+ - 0
1549
+ id: 5
1550
+ name: left_shoulder
1551
+ swap: right_shoulder
1552
+ type: upper
1553
+ 6:
1554
+ color:
1555
+ - 255
1556
+ - 128
1557
+ - 0
1558
+ id: 6
1559
+ name: right_shoulder
1560
+ swap: left_shoulder
1561
+ type: upper
1562
+ 7:
1563
+ color:
1564
+ - 0
1565
+ - 255
1566
+ - 0
1567
+ id: 7
1568
+ name: left_elbow
1569
+ swap: right_elbow
1570
+ type: upper
1571
+ 8:
1572
+ color:
1573
+ - 255
1574
+ - 128
1575
+ - 0
1576
+ id: 8
1577
+ name: right_elbow
1578
+ swap: left_elbow
1579
+ type: upper
1580
+ 9:
1581
+ color:
1582
+ - 0
1583
+ - 255
1584
+ - 0
1585
+ id: 9
1586
+ name: left_wrist
1587
+ swap: right_wrist
1588
+ type: upper
1589
+ 10:
1590
+ color:
1591
+ - 255
1592
+ - 128
1593
+ - 0
1594
+ id: 10
1595
+ name: right_wrist
1596
+ swap: left_wrist
1597
+ type: upper
1598
+ 11:
1599
+ color:
1600
+ - 0
1601
+ - 255
1602
+ - 0
1603
+ id: 11
1604
+ name: left_hip
1605
+ swap: right_hip
1606
+ type: lower
1607
+ 12:
1608
+ color:
1609
+ - 255
1610
+ - 128
1611
+ - 0
1612
+ id: 12
1613
+ name: right_hip
1614
+ swap: left_hip
1615
+ type: lower
1616
+ 13:
1617
+ color:
1618
+ - 0
1619
+ - 255
1620
+ - 0
1621
+ id: 13
1622
+ name: left_knee
1623
+ swap: right_knee
1624
+ type: lower
1625
+ 14:
1626
+ color:
1627
+ - 255
1628
+ - 128
1629
+ - 0
1630
+ id: 14
1631
+ name: right_knee
1632
+ swap: left_knee
1633
+ type: lower
1634
+ 15:
1635
+ color:
1636
+ - 0
1637
+ - 255
1638
+ - 0
1639
+ id: 15
1640
+ name: left_ankle
1641
+ swap: right_ankle
1642
+ type: lower
1643
+ 16:
1644
+ color:
1645
+ - 255
1646
+ - 128
1647
+ - 0
1648
+ id: 16
1649
+ name: right_ankle
1650
+ swap: left_ankle
1651
+ type: lower
1652
+ paper_info:
1653
+ author: Lin, Tsung-Yi and Maire, Michael and Belongie, Serge and Hays, James and
1654
+ Perona, Pietro and Ramanan, Deva and Doll{\'a}r, Piotr and Zitnick, C Lawrence
1655
+ container: European conference on computer vision
1656
+ homepage: http://cocodataset.org/
1657
+ title: 'Microsoft coco: Common objects in context'
1658
+ year: '2014'
1659
+ sigmas:
1660
+ - 0.026
1661
+ - 0.025
1662
+ - 0.025
1663
+ - 0.035
1664
+ - 0.035
1665
+ - 0.079
1666
+ - 0.079
1667
+ - 0.072
1668
+ - 0.072
1669
+ - 0.062
1670
+ - 0.062
1671
+ - 0.107
1672
+ - 0.107
1673
+ - 0.087
1674
+ - 0.087
1675
+ - 0.089
1676
+ - 0.089
1677
+ skeleton_info:
1678
+ 0:
1679
+ color:
1680
+ - 0
1681
+ - 255
1682
+ - 0
1683
+ id: 0
1684
+ link: !!python/tuple
1685
+ - left_ankle
1686
+ - left_knee
1687
+ 1:
1688
+ color:
1689
+ - 0
1690
+ - 255
1691
+ - 0
1692
+ id: 1
1693
+ link: !!python/tuple
1694
+ - left_knee
1695
+ - left_hip
1696
+ 2:
1697
+ color:
1698
+ - 255
1699
+ - 128
1700
+ - 0
1701
+ id: 2
1702
+ link: !!python/tuple
1703
+ - right_ankle
1704
+ - right_knee
1705
+ 3:
1706
+ color:
1707
+ - 255
1708
+ - 128
1709
+ - 0
1710
+ id: 3
1711
+ link: !!python/tuple
1712
+ - right_knee
1713
+ - right_hip
1714
+ 4:
1715
+ color:
1716
+ - 51
1717
+ - 153
1718
+ - 255
1719
+ id: 4
1720
+ link: !!python/tuple
1721
+ - left_hip
1722
+ - right_hip
1723
+ 5:
1724
+ color:
1725
+ - 51
1726
+ - 153
1727
+ - 255
1728
+ id: 5
1729
+ link: !!python/tuple
1730
+ - left_shoulder
1731
+ - left_hip
1732
+ 6:
1733
+ color:
1734
+ - 51
1735
+ - 153
1736
+ - 255
1737
+ id: 6
1738
+ link: !!python/tuple
1739
+ - right_shoulder
1740
+ - right_hip
1741
+ 7:
1742
+ color:
1743
+ - 51
1744
+ - 153
1745
+ - 255
1746
+ id: 7
1747
+ link: !!python/tuple
1748
+ - left_shoulder
1749
+ - right_shoulder
1750
+ 8:
1751
+ color:
1752
+ - 0
1753
+ - 255
1754
+ - 0
1755
+ id: 8
1756
+ link: !!python/tuple
1757
+ - left_shoulder
1758
+ - left_elbow
1759
+ 9:
1760
+ color:
1761
+ - 255
1762
+ - 128
1763
+ - 0
1764
+ id: 9
1765
+ link: !!python/tuple
1766
+ - right_shoulder
1767
+ - right_elbow
1768
+ 10:
1769
+ color:
1770
+ - 0
1771
+ - 255
1772
+ - 0
1773
+ id: 10
1774
+ link: !!python/tuple
1775
+ - left_elbow
1776
+ - left_wrist
1777
+ 11:
1778
+ color:
1779
+ - 255
1780
+ - 128
1781
+ - 0
1782
+ id: 11
1783
+ link: !!python/tuple
1784
+ - right_elbow
1785
+ - right_wrist
1786
+ 12:
1787
+ color:
1788
+ - 51
1789
+ - 153
1790
+ - 255
1791
+ id: 12
1792
+ link: !!python/tuple
1793
+ - left_eye
1794
+ - right_eye
1795
+ 13:
1796
+ color:
1797
+ - 51
1798
+ - 153
1799
+ - 255
1800
+ id: 13
1801
+ link: !!python/tuple
1802
+ - nose
1803
+ - left_eye
1804
+ 14:
1805
+ color:
1806
+ - 51
1807
+ - 153
1808
+ - 255
1809
+ id: 14
1810
+ link: !!python/tuple
1811
+ - nose
1812
+ - right_eye
1813
+ 15:
1814
+ color:
1815
+ - 51
1816
+ - 153
1817
+ - 255
1818
+ id: 15
1819
+ link: !!python/tuple
1820
+ - left_eye
1821
+ - left_ear
1822
+ 16:
1823
+ color:
1824
+ - 51
1825
+ - 153
1826
+ - 255
1827
+ id: 16
1828
+ link: !!python/tuple
1829
+ - right_eye
1830
+ - right_ear
1831
+ 17:
1832
+ color:
1833
+ - 51
1834
+ - 153
1835
+ - 255
1836
+ id: 17
1837
+ link: !!python/tuple
1838
+ - left_ear
1839
+ - left_shoulder
1840
+ 18:
1841
+ color:
1842
+ - 51
1843
+ - 153
1844
+ - 255
1845
+ id: 18
1846
+ link: !!python/tuple
1847
+ - right_ear
1848
+ - right_shoulder
1849
+ dist_params:
1850
+ backend: nccl
1851
+ evaluation:
1852
+ interval: 10
1853
+ metric: mAP
1854
+ save_best: AP
1855
+ load_from: null
1856
+ log_config:
1857
+ hooks:
1858
+ - type: TextLoggerHook
1859
+ interval: 50
1860
+ log_level: INFO
1861
+ lr_config:
1862
+ policy: step
1863
+ step:
1864
+ - 170
1865
+ - 200
1866
+ warmup: linear
1867
+ warmup_iters: 500
1868
+ warmup_ratio: 0.001
1869
+ model:
1870
+ backbone:
1871
+ extra:
1872
+ stage1:
1873
+ block: BOTTLENECK
1874
+ num_blocks: !!python/tuple
1875
+ - 4
1876
+ num_branches: 1
1877
+ num_channels: !!python/tuple
1878
+ - 64
1879
+ num_modules: 1
1880
+ stage2:
1881
+ block: BASIC
1882
+ num_blocks: !!python/tuple
1883
+ - 4
1884
+ - 4
1885
+ num_branches: 2
1886
+ num_channels: !!python/tuple
1887
+ - 48
1888
+ - 96
1889
+ num_modules: 1
1890
+ stage3:
1891
+ block: BASIC
1892
+ num_blocks: !!python/tuple
1893
+ - 4
1894
+ - 4
1895
+ - 4
1896
+ num_branches: 3
1897
+ num_channels: !!python/tuple
1898
+ - 48
1899
+ - 96
1900
+ - 192
1901
+ num_modules: 4
1902
+ stage4:
1903
+ block: BASIC
1904
+ num_blocks: !!python/tuple
1905
+ - 4
1906
+ - 4
1907
+ - 4
1908
+ - 4
1909
+ num_branches: 4
1910
+ num_channels: !!python/tuple
1911
+ - 48
1912
+ - 96
1913
+ - 192
1914
+ - 384
1915
+ num_modules: 3
1916
+ in_channels: 3
1917
+ type: HRNet
1918
+ keypoint_head:
1919
+ extra:
1920
+ final_conv_kernel: 1
1921
+ in_channels: 48
1922
+ loss_keypoint:
1923
+ type: JointsMSELoss
1924
+ use_target_weight: true
1925
+ num_deconv_layers: 0
1926
+ out_channels: 17
1927
+ type: TopdownHeatmapSimpleHead
1928
+ pretrained: https://download.openmmlab.com/mmpose/pretrain_models/hrnet_w48-8ef0771d.pth
1929
+ test_cfg:
1930
+ flip_test: true
1931
+ modulate_kernel: 11
1932
+ post_process: default
1933
+ shift_heatmap: true
1934
+ train_cfg: {}
1935
+ type: TopDown
1936
+ mp_start_method: fork
1937
+ opencv_num_threads: 0
1938
+ optimizer:
1939
+ lr: 0.0005
1940
+ type: Adam
1941
+ optimizer_config:
1942
+ grad_clip: null
1943
+ resume_from: null
1944
+ test_pipeline:
1945
+ - type: LoadImageFromFile
1946
+ - padding: 1.25
1947
+ type: TopDownGetBboxCenterScale
1948
+ - type: TopDownAffine
1949
+ - type: ToTensor
1950
+ - mean:
1951
+ - 0.485
1952
+ - 0.456
1953
+ - 0.406
1954
+ std:
1955
+ - 0.229
1956
+ - 0.224
1957
+ - 0.225
1958
+ type: NormalizeTensor
1959
+ - keys:
1960
+ - img
1961
+ meta_keys:
1962
+ - image_file
1963
+ - center
1964
+ - scale
1965
+ - rotation
1966
+ - bbox_score
1967
+ - flip_pairs
1968
+ type: Collect
1969
+ total_epochs: 210
1970
+ train_pipeline:
1971
+ - type: LoadImageFromFile
1972
+ - padding: 1.25
1973
+ type: TopDownGetBboxCenterScale
1974
+ - prob: 0.3
1975
+ shift_factor: 0.16
1976
+ type: TopDownRandomShiftBboxCenter
1977
+ - flip_prob: 0.5
1978
+ type: TopDownRandomFlip
1979
+ - num_joints_half_body: 8
1980
+ prob_half_body: 0.3
1981
+ type: TopDownHalfBodyTransform
1982
+ - rot_factor: 40
1983
+ scale_factor: 0.5
1984
+ type: TopDownGetRandomScaleRotation
1985
+ - type: TopDownAffine
1986
+ - type: ToTensor
1987
+ - mean:
1988
+ - 0.485
1989
+ - 0.456
1990
+ - 0.406
1991
+ std:
1992
+ - 0.229
1993
+ - 0.224
1994
+ - 0.225
1995
+ type: NormalizeTensor
1996
+ - sigma: 2
1997
+ type: TopDownGenerateTarget
1998
+ - keys:
1999
+ - img
2000
+ - target
2001
+ - target_weight
2002
+ meta_keys:
2003
+ - image_file
2004
+ - joints_3d
2005
+ - joints_3d_visible
2006
+ - center
2007
+ - scale
2008
+ - rotation
2009
+ - bbox_score
2010
+ - flip_pairs
2011
+ type: Collect
2012
+ val_pipeline:
2013
+ - type: LoadImageFromFile
2014
+ - padding: 1.25
2015
+ type: TopDownGetBboxCenterScale
2016
+ - type: TopDownAffine
2017
+ - type: ToTensor
2018
+ - mean:
2019
+ - 0.485
2020
+ - 0.456
2021
+ - 0.406
2022
+ std:
2023
+ - 0.229
2024
+ - 0.224
2025
+ - 0.225
2026
+ type: NormalizeTensor
2027
+ - keys:
2028
+ - img
2029
+ meta_keys:
2030
+ - image_file
2031
+ - center
2032
+ - scale
2033
+ - rotation
2034
+ - bbox_score
2035
+ - flip_pairs
2036
+ type: Collect
2037
+ workflow:
2038
+ - !!python/tuple
2039
+ - train
2040
+ - 1
keypose_mmpose_hrnet_384/README.md ADDED
@@ -0,0 +1 @@
 
 
1
+ Originally from https://github.com/open-mmlab/mmpose, converted to safetensors, used under Apache-2.0
keypose_mmpose_hrnet_384/hrnet_w48_coco_384x288-314c8528_20200708.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ecbf76bc185c98ce8d76b782ee818e051a0247b08de9397f2686e5fc4918d176
3
+ size 254911106
keypose_mmpose_hrnet_384/hrnet_w48_coco_384x288.yaml ADDED
@@ -0,0 +1,2040 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ channel_cfg:
2
+ dataset_channel:
3
+ - - 0
4
+ - 1
5
+ - 2
6
+ - 3
7
+ - 4
8
+ - 5
9
+ - 6
10
+ - 7
11
+ - 8
12
+ - 9
13
+ - 10
14
+ - 11
15
+ - 12
16
+ - 13
17
+ - 14
18
+ - 15
19
+ - 16
20
+ dataset_joints: 17
21
+ inference_channel:
22
+ - 0
23
+ - 1
24
+ - 2
25
+ - 3
26
+ - 4
27
+ - 5
28
+ - 6
29
+ - 7
30
+ - 8
31
+ - 9
32
+ - 10
33
+ - 11
34
+ - 12
35
+ - 13
36
+ - 14
37
+ - 15
38
+ - 16
39
+ num_output_channels: 17
40
+ checkpoint_config:
41
+ interval: 10
42
+ data:
43
+ samples_per_gpu: 32
44
+ test:
45
+ ann_file: data/coco/annotations/person_keypoints_val2017.json
46
+ data_cfg:
47
+ bbox_file: data/coco/person_detection_results/COCO_val2017_detections_AP_H_56_person.json
48
+ dataset_channel:
49
+ - - 0
50
+ - 1
51
+ - 2
52
+ - 3
53
+ - 4
54
+ - 5
55
+ - 6
56
+ - 7
57
+ - 8
58
+ - 9
59
+ - 10
60
+ - 11
61
+ - 12
62
+ - 13
63
+ - 14
64
+ - 15
65
+ - 16
66
+ det_bbox_thr: 0.0
67
+ heatmap_size:
68
+ - 72
69
+ - 96
70
+ image_size:
71
+ - 288
72
+ - 384
73
+ inference_channel:
74
+ - 0
75
+ - 1
76
+ - 2
77
+ - 3
78
+ - 4
79
+ - 5
80
+ - 6
81
+ - 7
82
+ - 8
83
+ - 9
84
+ - 10
85
+ - 11
86
+ - 12
87
+ - 13
88
+ - 14
89
+ - 15
90
+ - 16
91
+ nms_thr: 1.0
92
+ num_joints: 17
93
+ num_output_channels: 17
94
+ oks_thr: 0.9
95
+ soft_nms: false
96
+ use_gt_bbox: false
97
+ vis_thr: 0.2
98
+ dataset_info:
99
+ dataset_name: coco
100
+ joint_weights:
101
+ - 1.0
102
+ - 1.0
103
+ - 1.0
104
+ - 1.0
105
+ - 1.0
106
+ - 1.0
107
+ - 1.0
108
+ - 1.2
109
+ - 1.2
110
+ - 1.5
111
+ - 1.5
112
+ - 1.0
113
+ - 1.0
114
+ - 1.2
115
+ - 1.2
116
+ - 1.5
117
+ - 1.5
118
+ keypoint_info:
119
+ 0:
120
+ color:
121
+ - 51
122
+ - 153
123
+ - 255
124
+ id: 0
125
+ name: nose
126
+ swap: ''
127
+ type: upper
128
+ 1:
129
+ color:
130
+ - 51
131
+ - 153
132
+ - 255
133
+ id: 1
134
+ name: left_eye
135
+ swap: right_eye
136
+ type: upper
137
+ 2:
138
+ color:
139
+ - 51
140
+ - 153
141
+ - 255
142
+ id: 2
143
+ name: right_eye
144
+ swap: left_eye
145
+ type: upper
146
+ 3:
147
+ color:
148
+ - 51
149
+ - 153
150
+ - 255
151
+ id: 3
152
+ name: left_ear
153
+ swap: right_ear
154
+ type: upper
155
+ 4:
156
+ color:
157
+ - 51
158
+ - 153
159
+ - 255
160
+ id: 4
161
+ name: right_ear
162
+ swap: left_ear
163
+ type: upper
164
+ 5:
165
+ color:
166
+ - 0
167
+ - 255
168
+ - 0
169
+ id: 5
170
+ name: left_shoulder
171
+ swap: right_shoulder
172
+ type: upper
173
+ 6:
174
+ color:
175
+ - 255
176
+ - 128
177
+ - 0
178
+ id: 6
179
+ name: right_shoulder
180
+ swap: left_shoulder
181
+ type: upper
182
+ 7:
183
+ color:
184
+ - 0
185
+ - 255
186
+ - 0
187
+ id: 7
188
+ name: left_elbow
189
+ swap: right_elbow
190
+ type: upper
191
+ 8:
192
+ color:
193
+ - 255
194
+ - 128
195
+ - 0
196
+ id: 8
197
+ name: right_elbow
198
+ swap: left_elbow
199
+ type: upper
200
+ 9:
201
+ color:
202
+ - 0
203
+ - 255
204
+ - 0
205
+ id: 9
206
+ name: left_wrist
207
+ swap: right_wrist
208
+ type: upper
209
+ 10:
210
+ color:
211
+ - 255
212
+ - 128
213
+ - 0
214
+ id: 10
215
+ name: right_wrist
216
+ swap: left_wrist
217
+ type: upper
218
+ 11:
219
+ color:
220
+ - 0
221
+ - 255
222
+ - 0
223
+ id: 11
224
+ name: left_hip
225
+ swap: right_hip
226
+ type: lower
227
+ 12:
228
+ color:
229
+ - 255
230
+ - 128
231
+ - 0
232
+ id: 12
233
+ name: right_hip
234
+ swap: left_hip
235
+ type: lower
236
+ 13:
237
+ color:
238
+ - 0
239
+ - 255
240
+ - 0
241
+ id: 13
242
+ name: left_knee
243
+ swap: right_knee
244
+ type: lower
245
+ 14:
246
+ color:
247
+ - 255
248
+ - 128
249
+ - 0
250
+ id: 14
251
+ name: right_knee
252
+ swap: left_knee
253
+ type: lower
254
+ 15:
255
+ color:
256
+ - 0
257
+ - 255
258
+ - 0
259
+ id: 15
260
+ name: left_ankle
261
+ swap: right_ankle
262
+ type: lower
263
+ 16:
264
+ color:
265
+ - 255
266
+ - 128
267
+ - 0
268
+ id: 16
269
+ name: right_ankle
270
+ swap: left_ankle
271
+ type: lower
272
+ paper_info:
273
+ author: Lin, Tsung-Yi and Maire, Michael and Belongie, Serge and Hays, James
274
+ and Perona, Pietro and Ramanan, Deva and Doll{\'a}r, Piotr and Zitnick,
275
+ C Lawrence
276
+ container: European conference on computer vision
277
+ homepage: http://cocodataset.org/
278
+ title: 'Microsoft coco: Common objects in context'
279
+ year: '2014'
280
+ sigmas:
281
+ - 0.026
282
+ - 0.025
283
+ - 0.025
284
+ - 0.035
285
+ - 0.035
286
+ - 0.079
287
+ - 0.079
288
+ - 0.072
289
+ - 0.072
290
+ - 0.062
291
+ - 0.062
292
+ - 0.107
293
+ - 0.107
294
+ - 0.087
295
+ - 0.087
296
+ - 0.089
297
+ - 0.089
298
+ skeleton_info:
299
+ 0:
300
+ color:
301
+ - 0
302
+ - 255
303
+ - 0
304
+ id: 0
305
+ link: !!python/tuple
306
+ - left_ankle
307
+ - left_knee
308
+ 1:
309
+ color:
310
+ - 0
311
+ - 255
312
+ - 0
313
+ id: 1
314
+ link: !!python/tuple
315
+ - left_knee
316
+ - left_hip
317
+ 2:
318
+ color:
319
+ - 255
320
+ - 128
321
+ - 0
322
+ id: 2
323
+ link: !!python/tuple
324
+ - right_ankle
325
+ - right_knee
326
+ 3:
327
+ color:
328
+ - 255
329
+ - 128
330
+ - 0
331
+ id: 3
332
+ link: !!python/tuple
333
+ - right_knee
334
+ - right_hip
335
+ 4:
336
+ color:
337
+ - 51
338
+ - 153
339
+ - 255
340
+ id: 4
341
+ link: !!python/tuple
342
+ - left_hip
343
+ - right_hip
344
+ 5:
345
+ color:
346
+ - 51
347
+ - 153
348
+ - 255
349
+ id: 5
350
+ link: !!python/tuple
351
+ - left_shoulder
352
+ - left_hip
353
+ 6:
354
+ color:
355
+ - 51
356
+ - 153
357
+ - 255
358
+ id: 6
359
+ link: !!python/tuple
360
+ - right_shoulder
361
+ - right_hip
362
+ 7:
363
+ color:
364
+ - 51
365
+ - 153
366
+ - 255
367
+ id: 7
368
+ link: !!python/tuple
369
+ - left_shoulder
370
+ - right_shoulder
371
+ 8:
372
+ color:
373
+ - 0
374
+ - 255
375
+ - 0
376
+ id: 8
377
+ link: !!python/tuple
378
+ - left_shoulder
379
+ - left_elbow
380
+ 9:
381
+ color:
382
+ - 255
383
+ - 128
384
+ - 0
385
+ id: 9
386
+ link: !!python/tuple
387
+ - right_shoulder
388
+ - right_elbow
389
+ 10:
390
+ color:
391
+ - 0
392
+ - 255
393
+ - 0
394
+ id: 10
395
+ link: !!python/tuple
396
+ - left_elbow
397
+ - left_wrist
398
+ 11:
399
+ color:
400
+ - 255
401
+ - 128
402
+ - 0
403
+ id: 11
404
+ link: !!python/tuple
405
+ - right_elbow
406
+ - right_wrist
407
+ 12:
408
+ color:
409
+ - 51
410
+ - 153
411
+ - 255
412
+ id: 12
413
+ link: !!python/tuple
414
+ - left_eye
415
+ - right_eye
416
+ 13:
417
+ color:
418
+ - 51
419
+ - 153
420
+ - 255
421
+ id: 13
422
+ link: !!python/tuple
423
+ - nose
424
+ - left_eye
425
+ 14:
426
+ color:
427
+ - 51
428
+ - 153
429
+ - 255
430
+ id: 14
431
+ link: !!python/tuple
432
+ - nose
433
+ - right_eye
434
+ 15:
435
+ color:
436
+ - 51
437
+ - 153
438
+ - 255
439
+ id: 15
440
+ link: !!python/tuple
441
+ - left_eye
442
+ - left_ear
443
+ 16:
444
+ color:
445
+ - 51
446
+ - 153
447
+ - 255
448
+ id: 16
449
+ link: !!python/tuple
450
+ - right_eye
451
+ - right_ear
452
+ 17:
453
+ color:
454
+ - 51
455
+ - 153
456
+ - 255
457
+ id: 17
458
+ link: !!python/tuple
459
+ - left_ear
460
+ - left_shoulder
461
+ 18:
462
+ color:
463
+ - 51
464
+ - 153
465
+ - 255
466
+ id: 18
467
+ link: !!python/tuple
468
+ - right_ear
469
+ - right_shoulder
470
+ img_prefix: data/coco/val2017/
471
+ pipeline:
472
+ - type: LoadImageFromFile
473
+ - padding: 1.25
474
+ type: TopDownGetBboxCenterScale
475
+ - type: TopDownAffine
476
+ - type: ToTensor
477
+ - mean:
478
+ - 0.485
479
+ - 0.456
480
+ - 0.406
481
+ std:
482
+ - 0.229
483
+ - 0.224
484
+ - 0.225
485
+ type: NormalizeTensor
486
+ - keys:
487
+ - img
488
+ meta_keys:
489
+ - image_file
490
+ - center
491
+ - scale
492
+ - rotation
493
+ - bbox_score
494
+ - flip_pairs
495
+ type: Collect
496
+ type: TopDownCocoDataset
497
+ test_dataloader:
498
+ samples_per_gpu: 32
499
+ train:
500
+ ann_file: data/coco/annotations/person_keypoints_train2017.json
501
+ data_cfg:
502
+ bbox_file: data/coco/person_detection_results/COCO_val2017_detections_AP_H_56_person.json
503
+ dataset_channel:
504
+ - - 0
505
+ - 1
506
+ - 2
507
+ - 3
508
+ - 4
509
+ - 5
510
+ - 6
511
+ - 7
512
+ - 8
513
+ - 9
514
+ - 10
515
+ - 11
516
+ - 12
517
+ - 13
518
+ - 14
519
+ - 15
520
+ - 16
521
+ det_bbox_thr: 0.0
522
+ heatmap_size:
523
+ - 72
524
+ - 96
525
+ image_size:
526
+ - 288
527
+ - 384
528
+ inference_channel:
529
+ - 0
530
+ - 1
531
+ - 2
532
+ - 3
533
+ - 4
534
+ - 5
535
+ - 6
536
+ - 7
537
+ - 8
538
+ - 9
539
+ - 10
540
+ - 11
541
+ - 12
542
+ - 13
543
+ - 14
544
+ - 15
545
+ - 16
546
+ nms_thr: 1.0
547
+ num_joints: 17
548
+ num_output_channels: 17
549
+ oks_thr: 0.9
550
+ soft_nms: false
551
+ use_gt_bbox: false
552
+ vis_thr: 0.2
553
+ dataset_info:
554
+ dataset_name: coco
555
+ joint_weights:
556
+ - 1.0
557
+ - 1.0
558
+ - 1.0
559
+ - 1.0
560
+ - 1.0
561
+ - 1.0
562
+ - 1.0
563
+ - 1.2
564
+ - 1.2
565
+ - 1.5
566
+ - 1.5
567
+ - 1.0
568
+ - 1.0
569
+ - 1.2
570
+ - 1.2
571
+ - 1.5
572
+ - 1.5
573
+ keypoint_info:
574
+ 0:
575
+ color:
576
+ - 51
577
+ - 153
578
+ - 255
579
+ id: 0
580
+ name: nose
581
+ swap: ''
582
+ type: upper
583
+ 1:
584
+ color:
585
+ - 51
586
+ - 153
587
+ - 255
588
+ id: 1
589
+ name: left_eye
590
+ swap: right_eye
591
+ type: upper
592
+ 2:
593
+ color:
594
+ - 51
595
+ - 153
596
+ - 255
597
+ id: 2
598
+ name: right_eye
599
+ swap: left_eye
600
+ type: upper
601
+ 3:
602
+ color:
603
+ - 51
604
+ - 153
605
+ - 255
606
+ id: 3
607
+ name: left_ear
608
+ swap: right_ear
609
+ type: upper
610
+ 4:
611
+ color:
612
+ - 51
613
+ - 153
614
+ - 255
615
+ id: 4
616
+ name: right_ear
617
+ swap: left_ear
618
+ type: upper
619
+ 5:
620
+ color:
621
+ - 0
622
+ - 255
623
+ - 0
624
+ id: 5
625
+ name: left_shoulder
626
+ swap: right_shoulder
627
+ type: upper
628
+ 6:
629
+ color:
630
+ - 255
631
+ - 128
632
+ - 0
633
+ id: 6
634
+ name: right_shoulder
635
+ swap: left_shoulder
636
+ type: upper
637
+ 7:
638
+ color:
639
+ - 0
640
+ - 255
641
+ - 0
642
+ id: 7
643
+ name: left_elbow
644
+ swap: right_elbow
645
+ type: upper
646
+ 8:
647
+ color:
648
+ - 255
649
+ - 128
650
+ - 0
651
+ id: 8
652
+ name: right_elbow
653
+ swap: left_elbow
654
+ type: upper
655
+ 9:
656
+ color:
657
+ - 0
658
+ - 255
659
+ - 0
660
+ id: 9
661
+ name: left_wrist
662
+ swap: right_wrist
663
+ type: upper
664
+ 10:
665
+ color:
666
+ - 255
667
+ - 128
668
+ - 0
669
+ id: 10
670
+ name: right_wrist
671
+ swap: left_wrist
672
+ type: upper
673
+ 11:
674
+ color:
675
+ - 0
676
+ - 255
677
+ - 0
678
+ id: 11
679
+ name: left_hip
680
+ swap: right_hip
681
+ type: lower
682
+ 12:
683
+ color:
684
+ - 255
685
+ - 128
686
+ - 0
687
+ id: 12
688
+ name: right_hip
689
+ swap: left_hip
690
+ type: lower
691
+ 13:
692
+ color:
693
+ - 0
694
+ - 255
695
+ - 0
696
+ id: 13
697
+ name: left_knee
698
+ swap: right_knee
699
+ type: lower
700
+ 14:
701
+ color:
702
+ - 255
703
+ - 128
704
+ - 0
705
+ id: 14
706
+ name: right_knee
707
+ swap: left_knee
708
+ type: lower
709
+ 15:
710
+ color:
711
+ - 0
712
+ - 255
713
+ - 0
714
+ id: 15
715
+ name: left_ankle
716
+ swap: right_ankle
717
+ type: lower
718
+ 16:
719
+ color:
720
+ - 255
721
+ - 128
722
+ - 0
723
+ id: 16
724
+ name: right_ankle
725
+ swap: left_ankle
726
+ type: lower
727
+ paper_info:
728
+ author: Lin, Tsung-Yi and Maire, Michael and Belongie, Serge and Hays, James
729
+ and Perona, Pietro and Ramanan, Deva and Doll{\'a}r, Piotr and Zitnick,
730
+ C Lawrence
731
+ container: European conference on computer vision
732
+ homepage: http://cocodataset.org/
733
+ title: 'Microsoft coco: Common objects in context'
734
+ year: '2014'
735
+ sigmas:
736
+ - 0.026
737
+ - 0.025
738
+ - 0.025
739
+ - 0.035
740
+ - 0.035
741
+ - 0.079
742
+ - 0.079
743
+ - 0.072
744
+ - 0.072
745
+ - 0.062
746
+ - 0.062
747
+ - 0.107
748
+ - 0.107
749
+ - 0.087
750
+ - 0.087
751
+ - 0.089
752
+ - 0.089
753
+ skeleton_info:
754
+ 0:
755
+ color:
756
+ - 0
757
+ - 255
758
+ - 0
759
+ id: 0
760
+ link: !!python/tuple
761
+ - left_ankle
762
+ - left_knee
763
+ 1:
764
+ color:
765
+ - 0
766
+ - 255
767
+ - 0
768
+ id: 1
769
+ link: !!python/tuple
770
+ - left_knee
771
+ - left_hip
772
+ 2:
773
+ color:
774
+ - 255
775
+ - 128
776
+ - 0
777
+ id: 2
778
+ link: !!python/tuple
779
+ - right_ankle
780
+ - right_knee
781
+ 3:
782
+ color:
783
+ - 255
784
+ - 128
785
+ - 0
786
+ id: 3
787
+ link: !!python/tuple
788
+ - right_knee
789
+ - right_hip
790
+ 4:
791
+ color:
792
+ - 51
793
+ - 153
794
+ - 255
795
+ id: 4
796
+ link: !!python/tuple
797
+ - left_hip
798
+ - right_hip
799
+ 5:
800
+ color:
801
+ - 51
802
+ - 153
803
+ - 255
804
+ id: 5
805
+ link: !!python/tuple
806
+ - left_shoulder
807
+ - left_hip
808
+ 6:
809
+ color:
810
+ - 51
811
+ - 153
812
+ - 255
813
+ id: 6
814
+ link: !!python/tuple
815
+ - right_shoulder
816
+ - right_hip
817
+ 7:
818
+ color:
819
+ - 51
820
+ - 153
821
+ - 255
822
+ id: 7
823
+ link: !!python/tuple
824
+ - left_shoulder
825
+ - right_shoulder
826
+ 8:
827
+ color:
828
+ - 0
829
+ - 255
830
+ - 0
831
+ id: 8
832
+ link: !!python/tuple
833
+ - left_shoulder
834
+ - left_elbow
835
+ 9:
836
+ color:
837
+ - 255
838
+ - 128
839
+ - 0
840
+ id: 9
841
+ link: !!python/tuple
842
+ - right_shoulder
843
+ - right_elbow
844
+ 10:
845
+ color:
846
+ - 0
847
+ - 255
848
+ - 0
849
+ id: 10
850
+ link: !!python/tuple
851
+ - left_elbow
852
+ - left_wrist
853
+ 11:
854
+ color:
855
+ - 255
856
+ - 128
857
+ - 0
858
+ id: 11
859
+ link: !!python/tuple
860
+ - right_elbow
861
+ - right_wrist
862
+ 12:
863
+ color:
864
+ - 51
865
+ - 153
866
+ - 255
867
+ id: 12
868
+ link: !!python/tuple
869
+ - left_eye
870
+ - right_eye
871
+ 13:
872
+ color:
873
+ - 51
874
+ - 153
875
+ - 255
876
+ id: 13
877
+ link: !!python/tuple
878
+ - nose
879
+ - left_eye
880
+ 14:
881
+ color:
882
+ - 51
883
+ - 153
884
+ - 255
885
+ id: 14
886
+ link: !!python/tuple
887
+ - nose
888
+ - right_eye
889
+ 15:
890
+ color:
891
+ - 51
892
+ - 153
893
+ - 255
894
+ id: 15
895
+ link: !!python/tuple
896
+ - left_eye
897
+ - left_ear
898
+ 16:
899
+ color:
900
+ - 51
901
+ - 153
902
+ - 255
903
+ id: 16
904
+ link: !!python/tuple
905
+ - right_eye
906
+ - right_ear
907
+ 17:
908
+ color:
909
+ - 51
910
+ - 153
911
+ - 255
912
+ id: 17
913
+ link: !!python/tuple
914
+ - left_ear
915
+ - left_shoulder
916
+ 18:
917
+ color:
918
+ - 51
919
+ - 153
920
+ - 255
921
+ id: 18
922
+ link: !!python/tuple
923
+ - right_ear
924
+ - right_shoulder
925
+ img_prefix: data/coco/train2017/
926
+ pipeline:
927
+ - type: LoadImageFromFile
928
+ - padding: 1.25
929
+ type: TopDownGetBboxCenterScale
930
+ - prob: 0.3
931
+ shift_factor: 0.16
932
+ type: TopDownRandomShiftBboxCenter
933
+ - flip_prob: 0.5
934
+ type: TopDownRandomFlip
935
+ - num_joints_half_body: 8
936
+ prob_half_body: 0.3
937
+ type: TopDownHalfBodyTransform
938
+ - rot_factor: 40
939
+ scale_factor: 0.5
940
+ type: TopDownGetRandomScaleRotation
941
+ - type: TopDownAffine
942
+ - type: ToTensor
943
+ - mean:
944
+ - 0.485
945
+ - 0.456
946
+ - 0.406
947
+ std:
948
+ - 0.229
949
+ - 0.224
950
+ - 0.225
951
+ type: NormalizeTensor
952
+ - sigma: 3
953
+ type: TopDownGenerateTarget
954
+ - keys:
955
+ - img
956
+ - target
957
+ - target_weight
958
+ meta_keys:
959
+ - image_file
960
+ - joints_3d
961
+ - joints_3d_visible
962
+ - center
963
+ - scale
964
+ - rotation
965
+ - bbox_score
966
+ - flip_pairs
967
+ type: Collect
968
+ type: TopDownCocoDataset
969
+ val:
970
+ ann_file: data/coco/annotations/person_keypoints_val2017.json
971
+ data_cfg:
972
+ bbox_file: data/coco/person_detection_results/COCO_val2017_detections_AP_H_56_person.json
973
+ dataset_channel:
974
+ - - 0
975
+ - 1
976
+ - 2
977
+ - 3
978
+ - 4
979
+ - 5
980
+ - 6
981
+ - 7
982
+ - 8
983
+ - 9
984
+ - 10
985
+ - 11
986
+ - 12
987
+ - 13
988
+ - 14
989
+ - 15
990
+ - 16
991
+ det_bbox_thr: 0.0
992
+ heatmap_size:
993
+ - 72
994
+ - 96
995
+ image_size:
996
+ - 288
997
+ - 384
998
+ inference_channel:
999
+ - 0
1000
+ - 1
1001
+ - 2
1002
+ - 3
1003
+ - 4
1004
+ - 5
1005
+ - 6
1006
+ - 7
1007
+ - 8
1008
+ - 9
1009
+ - 10
1010
+ - 11
1011
+ - 12
1012
+ - 13
1013
+ - 14
1014
+ - 15
1015
+ - 16
1016
+ nms_thr: 1.0
1017
+ num_joints: 17
1018
+ num_output_channels: 17
1019
+ oks_thr: 0.9
1020
+ soft_nms: false
1021
+ use_gt_bbox: false
1022
+ vis_thr: 0.2
1023
+ dataset_info:
1024
+ dataset_name: coco
1025
+ joint_weights:
1026
+ - 1.0
1027
+ - 1.0
1028
+ - 1.0
1029
+ - 1.0
1030
+ - 1.0
1031
+ - 1.0
1032
+ - 1.0
1033
+ - 1.2
1034
+ - 1.2
1035
+ - 1.5
1036
+ - 1.5
1037
+ - 1.0
1038
+ - 1.0
1039
+ - 1.2
1040
+ - 1.2
1041
+ - 1.5
1042
+ - 1.5
1043
+ keypoint_info:
1044
+ 0:
1045
+ color:
1046
+ - 51
1047
+ - 153
1048
+ - 255
1049
+ id: 0
1050
+ name: nose
1051
+ swap: ''
1052
+ type: upper
1053
+ 1:
1054
+ color:
1055
+ - 51
1056
+ - 153
1057
+ - 255
1058
+ id: 1
1059
+ name: left_eye
1060
+ swap: right_eye
1061
+ type: upper
1062
+ 2:
1063
+ color:
1064
+ - 51
1065
+ - 153
1066
+ - 255
1067
+ id: 2
1068
+ name: right_eye
1069
+ swap: left_eye
1070
+ type: upper
1071
+ 3:
1072
+ color:
1073
+ - 51
1074
+ - 153
1075
+ - 255
1076
+ id: 3
1077
+ name: left_ear
1078
+ swap: right_ear
1079
+ type: upper
1080
+ 4:
1081
+ color:
1082
+ - 51
1083
+ - 153
1084
+ - 255
1085
+ id: 4
1086
+ name: right_ear
1087
+ swap: left_ear
1088
+ type: upper
1089
+ 5:
1090
+ color:
1091
+ - 0
1092
+ - 255
1093
+ - 0
1094
+ id: 5
1095
+ name: left_shoulder
1096
+ swap: right_shoulder
1097
+ type: upper
1098
+ 6:
1099
+ color:
1100
+ - 255
1101
+ - 128
1102
+ - 0
1103
+ id: 6
1104
+ name: right_shoulder
1105
+ swap: left_shoulder
1106
+ type: upper
1107
+ 7:
1108
+ color:
1109
+ - 0
1110
+ - 255
1111
+ - 0
1112
+ id: 7
1113
+ name: left_elbow
1114
+ swap: right_elbow
1115
+ type: upper
1116
+ 8:
1117
+ color:
1118
+ - 255
1119
+ - 128
1120
+ - 0
1121
+ id: 8
1122
+ name: right_elbow
1123
+ swap: left_elbow
1124
+ type: upper
1125
+ 9:
1126
+ color:
1127
+ - 0
1128
+ - 255
1129
+ - 0
1130
+ id: 9
1131
+ name: left_wrist
1132
+ swap: right_wrist
1133
+ type: upper
1134
+ 10:
1135
+ color:
1136
+ - 255
1137
+ - 128
1138
+ - 0
1139
+ id: 10
1140
+ name: right_wrist
1141
+ swap: left_wrist
1142
+ type: upper
1143
+ 11:
1144
+ color:
1145
+ - 0
1146
+ - 255
1147
+ - 0
1148
+ id: 11
1149
+ name: left_hip
1150
+ swap: right_hip
1151
+ type: lower
1152
+ 12:
1153
+ color:
1154
+ - 255
1155
+ - 128
1156
+ - 0
1157
+ id: 12
1158
+ name: right_hip
1159
+ swap: left_hip
1160
+ type: lower
1161
+ 13:
1162
+ color:
1163
+ - 0
1164
+ - 255
1165
+ - 0
1166
+ id: 13
1167
+ name: left_knee
1168
+ swap: right_knee
1169
+ type: lower
1170
+ 14:
1171
+ color:
1172
+ - 255
1173
+ - 128
1174
+ - 0
1175
+ id: 14
1176
+ name: right_knee
1177
+ swap: left_knee
1178
+ type: lower
1179
+ 15:
1180
+ color:
1181
+ - 0
1182
+ - 255
1183
+ - 0
1184
+ id: 15
1185
+ name: left_ankle
1186
+ swap: right_ankle
1187
+ type: lower
1188
+ 16:
1189
+ color:
1190
+ - 255
1191
+ - 128
1192
+ - 0
1193
+ id: 16
1194
+ name: right_ankle
1195
+ swap: left_ankle
1196
+ type: lower
1197
+ paper_info:
1198
+ author: Lin, Tsung-Yi and Maire, Michael and Belongie, Serge and Hays, James
1199
+ and Perona, Pietro and Ramanan, Deva and Doll{\'a}r, Piotr and Zitnick,
1200
+ C Lawrence
1201
+ container: European conference on computer vision
1202
+ homepage: http://cocodataset.org/
1203
+ title: 'Microsoft coco: Common objects in context'
1204
+ year: '2014'
1205
+ sigmas:
1206
+ - 0.026
1207
+ - 0.025
1208
+ - 0.025
1209
+ - 0.035
1210
+ - 0.035
1211
+ - 0.079
1212
+ - 0.079
1213
+ - 0.072
1214
+ - 0.072
1215
+ - 0.062
1216
+ - 0.062
1217
+ - 0.107
1218
+ - 0.107
1219
+ - 0.087
1220
+ - 0.087
1221
+ - 0.089
1222
+ - 0.089
1223
+ skeleton_info:
1224
+ 0:
1225
+ color:
1226
+ - 0
1227
+ - 255
1228
+ - 0
1229
+ id: 0
1230
+ link: !!python/tuple
1231
+ - left_ankle
1232
+ - left_knee
1233
+ 1:
1234
+ color:
1235
+ - 0
1236
+ - 255
1237
+ - 0
1238
+ id: 1
1239
+ link: !!python/tuple
1240
+ - left_knee
1241
+ - left_hip
1242
+ 2:
1243
+ color:
1244
+ - 255
1245
+ - 128
1246
+ - 0
1247
+ id: 2
1248
+ link: !!python/tuple
1249
+ - right_ankle
1250
+ - right_knee
1251
+ 3:
1252
+ color:
1253
+ - 255
1254
+ - 128
1255
+ - 0
1256
+ id: 3
1257
+ link: !!python/tuple
1258
+ - right_knee
1259
+ - right_hip
1260
+ 4:
1261
+ color:
1262
+ - 51
1263
+ - 153
1264
+ - 255
1265
+ id: 4
1266
+ link: !!python/tuple
1267
+ - left_hip
1268
+ - right_hip
1269
+ 5:
1270
+ color:
1271
+ - 51
1272
+ - 153
1273
+ - 255
1274
+ id: 5
1275
+ link: !!python/tuple
1276
+ - left_shoulder
1277
+ - left_hip
1278
+ 6:
1279
+ color:
1280
+ - 51
1281
+ - 153
1282
+ - 255
1283
+ id: 6
1284
+ link: !!python/tuple
1285
+ - right_shoulder
1286
+ - right_hip
1287
+ 7:
1288
+ color:
1289
+ - 51
1290
+ - 153
1291
+ - 255
1292
+ id: 7
1293
+ link: !!python/tuple
1294
+ - left_shoulder
1295
+ - right_shoulder
1296
+ 8:
1297
+ color:
1298
+ - 0
1299
+ - 255
1300
+ - 0
1301
+ id: 8
1302
+ link: !!python/tuple
1303
+ - left_shoulder
1304
+ - left_elbow
1305
+ 9:
1306
+ color:
1307
+ - 255
1308
+ - 128
1309
+ - 0
1310
+ id: 9
1311
+ link: !!python/tuple
1312
+ - right_shoulder
1313
+ - right_elbow
1314
+ 10:
1315
+ color:
1316
+ - 0
1317
+ - 255
1318
+ - 0
1319
+ id: 10
1320
+ link: !!python/tuple
1321
+ - left_elbow
1322
+ - left_wrist
1323
+ 11:
1324
+ color:
1325
+ - 255
1326
+ - 128
1327
+ - 0
1328
+ id: 11
1329
+ link: !!python/tuple
1330
+ - right_elbow
1331
+ - right_wrist
1332
+ 12:
1333
+ color:
1334
+ - 51
1335
+ - 153
1336
+ - 255
1337
+ id: 12
1338
+ link: !!python/tuple
1339
+ - left_eye
1340
+ - right_eye
1341
+ 13:
1342
+ color:
1343
+ - 51
1344
+ - 153
1345
+ - 255
1346
+ id: 13
1347
+ link: !!python/tuple
1348
+ - nose
1349
+ - left_eye
1350
+ 14:
1351
+ color:
1352
+ - 51
1353
+ - 153
1354
+ - 255
1355
+ id: 14
1356
+ link: !!python/tuple
1357
+ - nose
1358
+ - right_eye
1359
+ 15:
1360
+ color:
1361
+ - 51
1362
+ - 153
1363
+ - 255
1364
+ id: 15
1365
+ link: !!python/tuple
1366
+ - left_eye
1367
+ - left_ear
1368
+ 16:
1369
+ color:
1370
+ - 51
1371
+ - 153
1372
+ - 255
1373
+ id: 16
1374
+ link: !!python/tuple
1375
+ - right_eye
1376
+ - right_ear
1377
+ 17:
1378
+ color:
1379
+ - 51
1380
+ - 153
1381
+ - 255
1382
+ id: 17
1383
+ link: !!python/tuple
1384
+ - left_ear
1385
+ - left_shoulder
1386
+ 18:
1387
+ color:
1388
+ - 51
1389
+ - 153
1390
+ - 255
1391
+ id: 18
1392
+ link: !!python/tuple
1393
+ - right_ear
1394
+ - right_shoulder
1395
+ img_prefix: data/coco/val2017/
1396
+ pipeline:
1397
+ - type: LoadImageFromFile
1398
+ - padding: 1.25
1399
+ type: TopDownGetBboxCenterScale
1400
+ - type: TopDownAffine
1401
+ - type: ToTensor
1402
+ - mean:
1403
+ - 0.485
1404
+ - 0.456
1405
+ - 0.406
1406
+ std:
1407
+ - 0.229
1408
+ - 0.224
1409
+ - 0.225
1410
+ type: NormalizeTensor
1411
+ - keys:
1412
+ - img
1413
+ meta_keys:
1414
+ - image_file
1415
+ - center
1416
+ - scale
1417
+ - rotation
1418
+ - bbox_score
1419
+ - flip_pairs
1420
+ type: Collect
1421
+ type: TopDownCocoDataset
1422
+ val_dataloader:
1423
+ samples_per_gpu: 32
1424
+ workers_per_gpu: 2
1425
+ data_cfg:
1426
+ bbox_file: data/coco/person_detection_results/COCO_val2017_detections_AP_H_56_person.json
1427
+ dataset_channel:
1428
+ - - 0
1429
+ - 1
1430
+ - 2
1431
+ - 3
1432
+ - 4
1433
+ - 5
1434
+ - 6
1435
+ - 7
1436
+ - 8
1437
+ - 9
1438
+ - 10
1439
+ - 11
1440
+ - 12
1441
+ - 13
1442
+ - 14
1443
+ - 15
1444
+ - 16
1445
+ det_bbox_thr: 0.0
1446
+ heatmap_size:
1447
+ - 72
1448
+ - 96
1449
+ image_size:
1450
+ - 288
1451
+ - 384
1452
+ inference_channel:
1453
+ - 0
1454
+ - 1
1455
+ - 2
1456
+ - 3
1457
+ - 4
1458
+ - 5
1459
+ - 6
1460
+ - 7
1461
+ - 8
1462
+ - 9
1463
+ - 10
1464
+ - 11
1465
+ - 12
1466
+ - 13
1467
+ - 14
1468
+ - 15
1469
+ - 16
1470
+ nms_thr: 1.0
1471
+ num_joints: 17
1472
+ num_output_channels: 17
1473
+ oks_thr: 0.9
1474
+ soft_nms: false
1475
+ use_gt_bbox: false
1476
+ vis_thr: 0.2
1477
+ data_root: data/coco
1478
+ dataset_info:
1479
+ dataset_name: coco
1480
+ joint_weights:
1481
+ - 1.0
1482
+ - 1.0
1483
+ - 1.0
1484
+ - 1.0
1485
+ - 1.0
1486
+ - 1.0
1487
+ - 1.0
1488
+ - 1.2
1489
+ - 1.2
1490
+ - 1.5
1491
+ - 1.5
1492
+ - 1.0
1493
+ - 1.0
1494
+ - 1.2
1495
+ - 1.2
1496
+ - 1.5
1497
+ - 1.5
1498
+ keypoint_info:
1499
+ 0:
1500
+ color:
1501
+ - 51
1502
+ - 153
1503
+ - 255
1504
+ id: 0
1505
+ name: nose
1506
+ swap: ''
1507
+ type: upper
1508
+ 1:
1509
+ color:
1510
+ - 51
1511
+ - 153
1512
+ - 255
1513
+ id: 1
1514
+ name: left_eye
1515
+ swap: right_eye
1516
+ type: upper
1517
+ 2:
1518
+ color:
1519
+ - 51
1520
+ - 153
1521
+ - 255
1522
+ id: 2
1523
+ name: right_eye
1524
+ swap: left_eye
1525
+ type: upper
1526
+ 3:
1527
+ color:
1528
+ - 51
1529
+ - 153
1530
+ - 255
1531
+ id: 3
1532
+ name: left_ear
1533
+ swap: right_ear
1534
+ type: upper
1535
+ 4:
1536
+ color:
1537
+ - 51
1538
+ - 153
1539
+ - 255
1540
+ id: 4
1541
+ name: right_ear
1542
+ swap: left_ear
1543
+ type: upper
1544
+ 5:
1545
+ color:
1546
+ - 0
1547
+ - 255
1548
+ - 0
1549
+ id: 5
1550
+ name: left_shoulder
1551
+ swap: right_shoulder
1552
+ type: upper
1553
+ 6:
1554
+ color:
1555
+ - 255
1556
+ - 128
1557
+ - 0
1558
+ id: 6
1559
+ name: right_shoulder
1560
+ swap: left_shoulder
1561
+ type: upper
1562
+ 7:
1563
+ color:
1564
+ - 0
1565
+ - 255
1566
+ - 0
1567
+ id: 7
1568
+ name: left_elbow
1569
+ swap: right_elbow
1570
+ type: upper
1571
+ 8:
1572
+ color:
1573
+ - 255
1574
+ - 128
1575
+ - 0
1576
+ id: 8
1577
+ name: right_elbow
1578
+ swap: left_elbow
1579
+ type: upper
1580
+ 9:
1581
+ color:
1582
+ - 0
1583
+ - 255
1584
+ - 0
1585
+ id: 9
1586
+ name: left_wrist
1587
+ swap: right_wrist
1588
+ type: upper
1589
+ 10:
1590
+ color:
1591
+ - 255
1592
+ - 128
1593
+ - 0
1594
+ id: 10
1595
+ name: right_wrist
1596
+ swap: left_wrist
1597
+ type: upper
1598
+ 11:
1599
+ color:
1600
+ - 0
1601
+ - 255
1602
+ - 0
1603
+ id: 11
1604
+ name: left_hip
1605
+ swap: right_hip
1606
+ type: lower
1607
+ 12:
1608
+ color:
1609
+ - 255
1610
+ - 128
1611
+ - 0
1612
+ id: 12
1613
+ name: right_hip
1614
+ swap: left_hip
1615
+ type: lower
1616
+ 13:
1617
+ color:
1618
+ - 0
1619
+ - 255
1620
+ - 0
1621
+ id: 13
1622
+ name: left_knee
1623
+ swap: right_knee
1624
+ type: lower
1625
+ 14:
1626
+ color:
1627
+ - 255
1628
+ - 128
1629
+ - 0
1630
+ id: 14
1631
+ name: right_knee
1632
+ swap: left_knee
1633
+ type: lower
1634
+ 15:
1635
+ color:
1636
+ - 0
1637
+ - 255
1638
+ - 0
1639
+ id: 15
1640
+ name: left_ankle
1641
+ swap: right_ankle
1642
+ type: lower
1643
+ 16:
1644
+ color:
1645
+ - 255
1646
+ - 128
1647
+ - 0
1648
+ id: 16
1649
+ name: right_ankle
1650
+ swap: left_ankle
1651
+ type: lower
1652
+ paper_info:
1653
+ author: Lin, Tsung-Yi and Maire, Michael and Belongie, Serge and Hays, James and
1654
+ Perona, Pietro and Ramanan, Deva and Doll{\'a}r, Piotr and Zitnick, C Lawrence
1655
+ container: European conference on computer vision
1656
+ homepage: http://cocodataset.org/
1657
+ title: 'Microsoft coco: Common objects in context'
1658
+ year: '2014'
1659
+ sigmas:
1660
+ - 0.026
1661
+ - 0.025
1662
+ - 0.025
1663
+ - 0.035
1664
+ - 0.035
1665
+ - 0.079
1666
+ - 0.079
1667
+ - 0.072
1668
+ - 0.072
1669
+ - 0.062
1670
+ - 0.062
1671
+ - 0.107
1672
+ - 0.107
1673
+ - 0.087
1674
+ - 0.087
1675
+ - 0.089
1676
+ - 0.089
1677
+ skeleton_info:
1678
+ 0:
1679
+ color:
1680
+ - 0
1681
+ - 255
1682
+ - 0
1683
+ id: 0
1684
+ link: !!python/tuple
1685
+ - left_ankle
1686
+ - left_knee
1687
+ 1:
1688
+ color:
1689
+ - 0
1690
+ - 255
1691
+ - 0
1692
+ id: 1
1693
+ link: !!python/tuple
1694
+ - left_knee
1695
+ - left_hip
1696
+ 2:
1697
+ color:
1698
+ - 255
1699
+ - 128
1700
+ - 0
1701
+ id: 2
1702
+ link: !!python/tuple
1703
+ - right_ankle
1704
+ - right_knee
1705
+ 3:
1706
+ color:
1707
+ - 255
1708
+ - 128
1709
+ - 0
1710
+ id: 3
1711
+ link: !!python/tuple
1712
+ - right_knee
1713
+ - right_hip
1714
+ 4:
1715
+ color:
1716
+ - 51
1717
+ - 153
1718
+ - 255
1719
+ id: 4
1720
+ link: !!python/tuple
1721
+ - left_hip
1722
+ - right_hip
1723
+ 5:
1724
+ color:
1725
+ - 51
1726
+ - 153
1727
+ - 255
1728
+ id: 5
1729
+ link: !!python/tuple
1730
+ - left_shoulder
1731
+ - left_hip
1732
+ 6:
1733
+ color:
1734
+ - 51
1735
+ - 153
1736
+ - 255
1737
+ id: 6
1738
+ link: !!python/tuple
1739
+ - right_shoulder
1740
+ - right_hip
1741
+ 7:
1742
+ color:
1743
+ - 51
1744
+ - 153
1745
+ - 255
1746
+ id: 7
1747
+ link: !!python/tuple
1748
+ - left_shoulder
1749
+ - right_shoulder
1750
+ 8:
1751
+ color:
1752
+ - 0
1753
+ - 255
1754
+ - 0
1755
+ id: 8
1756
+ link: !!python/tuple
1757
+ - left_shoulder
1758
+ - left_elbow
1759
+ 9:
1760
+ color:
1761
+ - 255
1762
+ - 128
1763
+ - 0
1764
+ id: 9
1765
+ link: !!python/tuple
1766
+ - right_shoulder
1767
+ - right_elbow
1768
+ 10:
1769
+ color:
1770
+ - 0
1771
+ - 255
1772
+ - 0
1773
+ id: 10
1774
+ link: !!python/tuple
1775
+ - left_elbow
1776
+ - left_wrist
1777
+ 11:
1778
+ color:
1779
+ - 255
1780
+ - 128
1781
+ - 0
1782
+ id: 11
1783
+ link: !!python/tuple
1784
+ - right_elbow
1785
+ - right_wrist
1786
+ 12:
1787
+ color:
1788
+ - 51
1789
+ - 153
1790
+ - 255
1791
+ id: 12
1792
+ link: !!python/tuple
1793
+ - left_eye
1794
+ - right_eye
1795
+ 13:
1796
+ color:
1797
+ - 51
1798
+ - 153
1799
+ - 255
1800
+ id: 13
1801
+ link: !!python/tuple
1802
+ - nose
1803
+ - left_eye
1804
+ 14:
1805
+ color:
1806
+ - 51
1807
+ - 153
1808
+ - 255
1809
+ id: 14
1810
+ link: !!python/tuple
1811
+ - nose
1812
+ - right_eye
1813
+ 15:
1814
+ color:
1815
+ - 51
1816
+ - 153
1817
+ - 255
1818
+ id: 15
1819
+ link: !!python/tuple
1820
+ - left_eye
1821
+ - left_ear
1822
+ 16:
1823
+ color:
1824
+ - 51
1825
+ - 153
1826
+ - 255
1827
+ id: 16
1828
+ link: !!python/tuple
1829
+ - right_eye
1830
+ - right_ear
1831
+ 17:
1832
+ color:
1833
+ - 51
1834
+ - 153
1835
+ - 255
1836
+ id: 17
1837
+ link: !!python/tuple
1838
+ - left_ear
1839
+ - left_shoulder
1840
+ 18:
1841
+ color:
1842
+ - 51
1843
+ - 153
1844
+ - 255
1845
+ id: 18
1846
+ link: !!python/tuple
1847
+ - right_ear
1848
+ - right_shoulder
1849
+ dist_params:
1850
+ backend: nccl
1851
+ evaluation:
1852
+ interval: 10
1853
+ metric: mAP
1854
+ save_best: AP
1855
+ load_from: null
1856
+ log_config:
1857
+ hooks:
1858
+ - type: TextLoggerHook
1859
+ interval: 50
1860
+ log_level: INFO
1861
+ lr_config:
1862
+ policy: step
1863
+ step:
1864
+ - 170
1865
+ - 200
1866
+ warmup: linear
1867
+ warmup_iters: 500
1868
+ warmup_ratio: 0.001
1869
+ model:
1870
+ backbone:
1871
+ extra:
1872
+ stage1:
1873
+ block: BOTTLENECK
1874
+ num_blocks: !!python/tuple
1875
+ - 4
1876
+ num_branches: 1
1877
+ num_channels: !!python/tuple
1878
+ - 64
1879
+ num_modules: 1
1880
+ stage2:
1881
+ block: BASIC
1882
+ num_blocks: !!python/tuple
1883
+ - 4
1884
+ - 4
1885
+ num_branches: 2
1886
+ num_channels: !!python/tuple
1887
+ - 48
1888
+ - 96
1889
+ num_modules: 1
1890
+ stage3:
1891
+ block: BASIC
1892
+ num_blocks: !!python/tuple
1893
+ - 4
1894
+ - 4
1895
+ - 4
1896
+ num_branches: 3
1897
+ num_channels: !!python/tuple
1898
+ - 48
1899
+ - 96
1900
+ - 192
1901
+ num_modules: 4
1902
+ stage4:
1903
+ block: BASIC
1904
+ num_blocks: !!python/tuple
1905
+ - 4
1906
+ - 4
1907
+ - 4
1908
+ - 4
1909
+ num_branches: 4
1910
+ num_channels: !!python/tuple
1911
+ - 48
1912
+ - 96
1913
+ - 192
1914
+ - 384
1915
+ num_modules: 3
1916
+ in_channels: 3
1917
+ type: HRNet
1918
+ keypoint_head:
1919
+ extra:
1920
+ final_conv_kernel: 1
1921
+ in_channels: 48
1922
+ loss_keypoint:
1923
+ type: JointsMSELoss
1924
+ use_target_weight: true
1925
+ num_deconv_layers: 0
1926
+ out_channels: 17
1927
+ type: TopdownHeatmapSimpleHead
1928
+ pretrained: https://download.openmmlab.com/mmpose/pretrain_models/hrnet_w48-8ef0771d.pth
1929
+ test_cfg:
1930
+ flip_test: true
1931
+ modulate_kernel: 11
1932
+ post_process: default
1933
+ shift_heatmap: true
1934
+ train_cfg: {}
1935
+ type: TopDown
1936
+ mp_start_method: fork
1937
+ opencv_num_threads: 0
1938
+ optimizer:
1939
+ lr: 0.0005
1940
+ type: Adam
1941
+ optimizer_config:
1942
+ grad_clip: null
1943
+ resume_from: null
1944
+ test_pipeline:
1945
+ - type: LoadImageFromFile
1946
+ - padding: 1.25
1947
+ type: TopDownGetBboxCenterScale
1948
+ - type: TopDownAffine
1949
+ - type: ToTensor
1950
+ - mean:
1951
+ - 0.485
1952
+ - 0.456
1953
+ - 0.406
1954
+ std:
1955
+ - 0.229
1956
+ - 0.224
1957
+ - 0.225
1958
+ type: NormalizeTensor
1959
+ - keys:
1960
+ - img
1961
+ meta_keys:
1962
+ - image_file
1963
+ - center
1964
+ - scale
1965
+ - rotation
1966
+ - bbox_score
1967
+ - flip_pairs
1968
+ type: Collect
1969
+ total_epochs: 210
1970
+ train_pipeline:
1971
+ - type: LoadImageFromFile
1972
+ - padding: 1.25
1973
+ type: TopDownGetBboxCenterScale
1974
+ - prob: 0.3
1975
+ shift_factor: 0.16
1976
+ type: TopDownRandomShiftBboxCenter
1977
+ - flip_prob: 0.5
1978
+ type: TopDownRandomFlip
1979
+ - num_joints_half_body: 8
1980
+ prob_half_body: 0.3
1981
+ type: TopDownHalfBodyTransform
1982
+ - rot_factor: 40
1983
+ scale_factor: 0.5
1984
+ type: TopDownGetRandomScaleRotation
1985
+ - type: TopDownAffine
1986
+ - type: ToTensor
1987
+ - mean:
1988
+ - 0.485
1989
+ - 0.456
1990
+ - 0.406
1991
+ std:
1992
+ - 0.229
1993
+ - 0.224
1994
+ - 0.225
1995
+ type: NormalizeTensor
1996
+ - sigma: 3
1997
+ type: TopDownGenerateTarget
1998
+ - keys:
1999
+ - img
2000
+ - target
2001
+ - target_weight
2002
+ meta_keys:
2003
+ - image_file
2004
+ - joints_3d
2005
+ - joints_3d_visible
2006
+ - center
2007
+ - scale
2008
+ - rotation
2009
+ - bbox_score
2010
+ - flip_pairs
2011
+ type: Collect
2012
+ val_pipeline:
2013
+ - type: LoadImageFromFile
2014
+ - padding: 1.25
2015
+ type: TopDownGetBboxCenterScale
2016
+ - type: TopDownAffine
2017
+ - type: ToTensor
2018
+ - mean:
2019
+ - 0.485
2020
+ - 0.456
2021
+ - 0.406
2022
+ std:
2023
+ - 0.229
2024
+ - 0.224
2025
+ - 0.225
2026
+ type: NormalizeTensor
2027
+ - keys:
2028
+ - img
2029
+ meta_keys:
2030
+ - image_file
2031
+ - center
2032
+ - scale
2033
+ - rotation
2034
+ - bbox_score
2035
+ - flip_pairs
2036
+ type: Collect
2037
+ workflow:
2038
+ - !!python/tuple
2039
+ - train
2040
+ - 1