Uploading MVTCAE in asenella/MVTCAE_mmnist_python311
Browse files- README.md +13 -0
- decoders.pkl +3 -0
- encoders.pkl +3 -0
- environment.json +1 -0
- model.pt +3 -0
- model_config.json +1 -0
README.md
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
language: en
|
| 3 |
+
tags:
|
| 4 |
+
- multivae
|
| 5 |
+
license: apache-2.0
|
| 6 |
+
---
|
| 7 |
+
|
| 8 |
+
### Downloading this model from the Hub
|
| 9 |
+
This model was trained with multivae. It can be downloaded or reloaded using the method `load_from_hf_hub`
|
| 10 |
+
```python
|
| 11 |
+
>>> from multivae.models import AutoModel
|
| 12 |
+
>>> model = AutoModel.load_from_hf_hub(hf_hub_path="your_hf_username/repo_name")
|
| 13 |
+
```
|
decoders.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c64bfff078b22380be5f9b1fbf6a2a89320d0551f735dd5ef70c97177bfa5e57
|
| 3 |
+
size 22922973
|
encoders.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bc732d05d941dae2a5b9a8aa6145050507f5811f235a99806befb69a7e5986d7
|
| 3 |
+
size 43880819
|
environment.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"name": "EnvironmentConfig", "python_version": "3.11"}
|
model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:82bad3e07ff7342720a89a9459b7f50b808d5b2418f8c1c395ab8ec2bf689f43
|
| 3 |
+
size 66730837
|
model_config.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"name": "MVTCAEConfig", "n_modalities": 5, "latent_dim": 512, "input_dims": {"m0": [3, 28, 28], "m1": [3, 28, 28], "m2": [3, 28, 28], "m3": [3, 28, 28], "m4": [3, 28, 28]}, "uses_likelihood_rescaling": false, "rescale_factors": null, "decoders_dist": {"m0": "laplace", "m1": "laplace", "m2": "laplace", "m3": "laplace", "m4": "laplace"}, "decoder_dist_params": {"m0": {"scale": 0.75}, "m1": {"scale": 0.75}, "m2": {"scale": 0.75}, "m3": {"scale": 0.75}, "m4": {"scale": 0.75}}, "custom_architectures": ["encoders", "decoders"], "alpha": 0.8333333333333334, "beta": 2.5}
|