thanks to kabachuha ❤
Browse files- README.md +23 -0
- VQGAN_autoencoder.pth +3 -0
- configuration.json +35 -0
- open_clip_pytorch_model.bin +3 -0
- text2video_pytorch_model.pth +3 -0
README.md
ADDED
|
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: cc-by-nc-4.0
|
| 3 |
+
task_categories:
|
| 4 |
+
- text-to-video
|
| 5 |
+
language:
|
| 6 |
+
- en
|
| 7 |
+
tags:
|
| 8 |
+
- anime
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
This is https://huggingface.co/datasets/strangeman3107/animov-0.1 model by strangeman3107 that was converted by [me](https://github.com/kabachuha) into the ModelScope original format using this script https://github.com/ExponentialML/Text-To-Video-Finetuning/pull/52.
|
| 12 |
+
|
| 13 |
+
Ready to use in Auto1111 webui with this extension https://github.com/deforum-art/sd-webui-text2video
|
| 14 |
+
|
| 15 |
+
---
|
| 16 |
+
|
| 17 |
+
Now, copyting info from the original page
|
| 18 |
+
|
| 19 |
+
This is a text2video model for diffusers, fine-tuned with a [modelscope](https://huggingface.co/damo-vilab/text-to-video-ms-1.7b) to have an anime-style appearance.
|
| 20 |
+
It was trained at 384x384 resolution.
|
| 21 |
+
It still generates unstable content often. The usage is the same as with the original modelscope model.
|
| 22 |
+
|
| 23 |
+
example images are [here](https://imgur.com/a/sCwmKG1).
|
VQGAN_autoencoder.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:930e9865584beae2405d29bc06a05db3bb6a5b34eedd40a7db29b9156ed7d098
|
| 3 |
+
size 2607657443
|
configuration.json
ADDED
|
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{ "framework": "pytorch",
|
| 2 |
+
"task": "text-to-video-synthesis",
|
| 3 |
+
"model": {
|
| 4 |
+
"type": "latent-text-to-video-synthesis",
|
| 5 |
+
"model_args": {
|
| 6 |
+
"ckpt_clip": "open_clip_pytorch_model.bin",
|
| 7 |
+
"ckpt_unet": "text2video_pytorch_model.pth",
|
| 8 |
+
"ckpt_autoencoder": "VQGAN_autoencoder.pth",
|
| 9 |
+
"max_frames": 16,
|
| 10 |
+
"tiny_gpu": 1
|
| 11 |
+
},
|
| 12 |
+
"model_cfg": {
|
| 13 |
+
"unet_in_dim": 4,
|
| 14 |
+
"unet_dim": 320,
|
| 15 |
+
"unet_y_dim": 768,
|
| 16 |
+
"unet_context_dim": 1024,
|
| 17 |
+
"unet_out_dim": 4,
|
| 18 |
+
"unet_dim_mult": [1, 2, 4, 4],
|
| 19 |
+
"unet_num_heads": 8,
|
| 20 |
+
"unet_head_dim": 64,
|
| 21 |
+
"unet_res_blocks": 2,
|
| 22 |
+
"unet_attn_scales": [1, 0.5, 0.25],
|
| 23 |
+
"unet_dropout": 0.1,
|
| 24 |
+
"temporal_attention": "True",
|
| 25 |
+
"num_timesteps": 1000,
|
| 26 |
+
"mean_type": "eps",
|
| 27 |
+
"var_type": "fixed_small",
|
| 28 |
+
"loss_type": "mse"
|
| 29 |
+
}
|
| 30 |
+
},
|
| 31 |
+
"pipeline": {
|
| 32 |
+
"type": "latent-text-to-video-synthesis"
|
| 33 |
+
}
|
| 34 |
+
}
|
| 35 |
+
|
open_clip_pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:73c32c62eebf1112b0693ff9e3ecfa0573ba02cd279420ea4da4af1cbfb39e3b
|
| 3 |
+
size 1972451989
|
text2video_pytorch_model.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8b710dd73d49c598b339ea9c76c78750fa38e6477793f6373be18087dbe9740c
|
| 3 |
+
size 2822972283
|