Upload 11 files
Browse files- emage_vq/.DS_Store +0 -0
- emage_vq/face/config.json +28 -0
- emage_vq/face/pytorch_model.bin +3 -0
- emage_vq/global/config.json +28 -0
- emage_vq/global/pytorch_model.bin +3 -0
- emage_vq/hands/config.json +28 -0
- emage_vq/hands/pytorch_model.bin +3 -0
- emage_vq/lower/config.json +28 -0
- emage_vq/lower/pytorch_model.bin +3 -0
- emage_vq/upper/config.json +28 -0
- emage_vq/upper/pytorch_model.bin +3 -0
emage_vq/.DS_Store
ADDED
|
Binary file (8.2 kB). View file
|
|
|
emage_vq/face/config.json
ADDED
|
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"EmageVQVAEConv"
|
| 4 |
+
],
|
| 5 |
+
"class_name": "EmageVQVAEConv",
|
| 6 |
+
"model_type": "emage_vqvaeconv",
|
| 7 |
+
"name_pyfile": "models.emage_audio.modeling_emage_audio",
|
| 8 |
+
"pose_dims": 106,
|
| 9 |
+
"pose_fps": 30,
|
| 10 |
+
"pose_length": 64,
|
| 11 |
+
"stride": 20,
|
| 12 |
+
"torch_dtype": "float32",
|
| 13 |
+
"transformers_version": "4.30.2",
|
| 14 |
+
"vae_codebook_size": 256,
|
| 15 |
+
"vae_grow": [
|
| 16 |
+
1,
|
| 17 |
+
1,
|
| 18 |
+
2,
|
| 19 |
+
1
|
| 20 |
+
],
|
| 21 |
+
"vae_layer": 2,
|
| 22 |
+
"vae_length": 256,
|
| 23 |
+
"vae_quantizer_lambda": 1.0,
|
| 24 |
+
"vae_test_dim": 106,
|
| 25 |
+
"vae_test_len": 64,
|
| 26 |
+
"vae_test_stride": 20,
|
| 27 |
+
"variational": false
|
| 28 |
+
}
|
emage_vq/face/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1ddee4ad1c66a7c21d4f44a4d5ee020b970d9e6532713ec9896de1720f1531c3
|
| 3 |
+
size 8932994
|
emage_vq/global/config.json
ADDED
|
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"EmageVAEConv"
|
| 4 |
+
],
|
| 5 |
+
"class_name": "EmageVAEConv",
|
| 6 |
+
"model_type": "emage_vaeconv",
|
| 7 |
+
"name_pyfile": "models.emage_audio.modeling_emage_audio",
|
| 8 |
+
"pose_dims": 61,
|
| 9 |
+
"pose_fps": 30,
|
| 10 |
+
"pose_length": 64,
|
| 11 |
+
"stride": 20,
|
| 12 |
+
"torch_dtype": "float32",
|
| 13 |
+
"transformers_version": "4.30.2",
|
| 14 |
+
"vae_codebook_size": 256,
|
| 15 |
+
"vae_grow": [
|
| 16 |
+
1,
|
| 17 |
+
1,
|
| 18 |
+
2,
|
| 19 |
+
1
|
| 20 |
+
],
|
| 21 |
+
"vae_layer": 4,
|
| 22 |
+
"vae_length": 256,
|
| 23 |
+
"vae_quantizer_lambda": 1.0,
|
| 24 |
+
"vae_test_dim": 61,
|
| 25 |
+
"vae_test_len": 64,
|
| 26 |
+
"vae_test_stride": 20,
|
| 27 |
+
"variational": false
|
| 28 |
+
}
|
emage_vq/global/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:42d4d883ff611bc82926a246b6e21791ae0eae73efef86248b2221a19c2c5a9f
|
| 3 |
+
size 14608303
|
emage_vq/hands/config.json
ADDED
|
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"EmageVQVAEConv"
|
| 4 |
+
],
|
| 5 |
+
"class_name": "EmageVQVAEConv",
|
| 6 |
+
"model_type": "emage_vqvaeconv",
|
| 7 |
+
"name_pyfile": "models.emage_audio.modeling_emage_audio",
|
| 8 |
+
"pose_dims": 180,
|
| 9 |
+
"pose_fps": 30,
|
| 10 |
+
"pose_length": 64,
|
| 11 |
+
"stride": 20,
|
| 12 |
+
"torch_dtype": "float32",
|
| 13 |
+
"transformers_version": "4.30.2",
|
| 14 |
+
"vae_codebook_size": 256,
|
| 15 |
+
"vae_grow": [
|
| 16 |
+
1,
|
| 17 |
+
1,
|
| 18 |
+
2,
|
| 19 |
+
1
|
| 20 |
+
],
|
| 21 |
+
"vae_layer": 2,
|
| 22 |
+
"vae_length": 256,
|
| 23 |
+
"vae_quantizer_lambda": 1.0,
|
| 24 |
+
"vae_test_dim": 180,
|
| 25 |
+
"vae_test_len": 64,
|
| 26 |
+
"vae_test_stride": 20,
|
| 27 |
+
"variational": false
|
| 28 |
+
}
|
emage_vq/hands/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:72872387623e13c45dea2307d61b029506fd08f5b0eb0824600fc138ad0290fb
|
| 3 |
+
size 9642306
|
emage_vq/lower/config.json
ADDED
|
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"EmageVQVAEConv"
|
| 4 |
+
],
|
| 5 |
+
"class_name": "EmageVQVAEConv",
|
| 6 |
+
"model_type": "emage_vqvaeconv",
|
| 7 |
+
"name_pyfile": "models.emage_audio.modeling_emage_audio",
|
| 8 |
+
"pose_dims": 61,
|
| 9 |
+
"pose_fps": 30,
|
| 10 |
+
"pose_length": 64,
|
| 11 |
+
"stride": 20,
|
| 12 |
+
"torch_dtype": "float32",
|
| 13 |
+
"transformers_version": "4.30.2",
|
| 14 |
+
"vae_codebook_size": 256,
|
| 15 |
+
"vae_grow": [
|
| 16 |
+
1,
|
| 17 |
+
1,
|
| 18 |
+
2,
|
| 19 |
+
1
|
| 20 |
+
],
|
| 21 |
+
"vae_layer": 4,
|
| 22 |
+
"vae_length": 256,
|
| 23 |
+
"vae_quantizer_lambda": 1.0,
|
| 24 |
+
"vae_test_dim": 61,
|
| 25 |
+
"vae_test_len": 64,
|
| 26 |
+
"vae_test_stride": 20,
|
| 27 |
+
"variational": false
|
| 28 |
+
}
|
emage_vq/lower/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eafc580ae68c43ada6f87c83ae96924094ab7e88befbc62bd952655a576f3027
|
| 3 |
+
size 14870770
|
emage_vq/upper/config.json
ADDED
|
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"EmageVQVAEConv"
|
| 4 |
+
],
|
| 5 |
+
"class_name": "EmageVQVAEConv",
|
| 6 |
+
"model_type": "emage_vqvaeconv",
|
| 7 |
+
"name_pyfile": "models.emage_audio.modeling_emage_audio",
|
| 8 |
+
"pose_dims": 78,
|
| 9 |
+
"pose_fps": 30,
|
| 10 |
+
"pose_length": 64,
|
| 11 |
+
"stride": 20,
|
| 12 |
+
"torch_dtype": "float32",
|
| 13 |
+
"transformers_version": "4.30.2",
|
| 14 |
+
"vae_codebook_size": 256,
|
| 15 |
+
"vae_grow": [
|
| 16 |
+
1,
|
| 17 |
+
1,
|
| 18 |
+
2,
|
| 19 |
+
1
|
| 20 |
+
],
|
| 21 |
+
"vae_layer": 2,
|
| 22 |
+
"vae_length": 256,
|
| 23 |
+
"vae_quantizer_lambda": 1.0,
|
| 24 |
+
"vae_test_dim": 78,
|
| 25 |
+
"vae_test_len": 64,
|
| 26 |
+
"vae_test_stride": 20,
|
| 27 |
+
"variational": false
|
| 28 |
+
}
|
emage_vq/upper/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fe2f67701f9fc63c7f1bb0857d37dec3a6cdb082e2fe92b4556d1f9cc70afb88
|
| 3 |
+
size 8698882
|