vidfom commited on
Commit
618f472
·
verified ·
1 Parent(s): d1e67e8

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitignore +47 -0
  2. Custom Resolutions Instructions.txt +16 -0
  3. Dockerfile +92 -0
  4. LICENSE.txt +46 -0
  5. README.md +256 -0
  6. defaults/ReadMe.txt +13 -0
  7. defaults/alpha.json +19 -0
  8. defaults/alpha_sf.json +17 -0
  9. defaults/animate.json +17 -0
  10. defaults/chatterbox.json +18 -0
  11. defaults/fantasy.json +11 -0
  12. defaults/flf2v_720p.json +16 -0
  13. defaults/flux.json +15 -0
  14. defaults/flux_chroma.json +17 -0
  15. defaults/flux_dev_kontext.json +16 -0
  16. defaults/flux_dev_kontext_dreamomni2.json +19 -0
  17. defaults/flux_dev_umo.json +23 -0
  18. defaults/flux_dev_uso.json +16 -0
  19. defaults/flux_krea.json +15 -0
  20. defaults/flux_schnell.json +16 -0
  21. defaults/flux_srpo.json +14 -0
  22. defaults/flux_srpo_uso.json +16 -0
  23. defaults/fun_inp.json +13 -0
  24. defaults/fun_inp_1.3B.json +11 -0
  25. defaults/hunyuan.json +12 -0
  26. defaults/hunyuan_avatar.json +12 -0
  27. defaults/hunyuan_custom.json +12 -0
  28. defaults/hunyuan_custom_audio.json +12 -0
  29. defaults/hunyuan_custom_edit.json +12 -0
  30. defaults/hunyuan_i2v.json +12 -0
  31. defaults/hunyuan_t2v_accvideo.json +30 -0
  32. defaults/hunyuan_t2v_fast.json +32 -0
  33. defaults/i2v.json +13 -0
  34. defaults/i2v_2_2.json +25 -0
  35. defaults/i2v_2_2_multitalk.json +18 -0
  36. defaults/i2v_720p.json +14 -0
  37. defaults/i2v_fusionix.json +11 -0
  38. defaults/infinitetalk.json +16 -0
  39. defaults/infinitetalk_multi.json +16 -0
  40. defaults/ltxv_13B.json +19 -0
  41. defaults/ltxv_distilled.json +15 -0
  42. defaults/lucy_edit.json +20 -0
  43. defaults/lucy_edit_fastwan.json +17 -0
  44. defaults/lynx.json +18 -0
  45. defaults/moviigen.json +16 -0
  46. defaults/multitalk.json +15 -0
  47. defaults/multitalk_720p.json +13 -0
  48. defaults/ovi.json +18 -0
  49. defaults/ovi_fastwan.json +17 -0
  50. defaults/phantom_1.3B.json +11 -0
.gitignore ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ .*
2
+ *.py[cod]
3
+ # *.jpg
4
+ *.jpeg
5
+ # *.png
6
+ *.gif
7
+ *.bmp
8
+ *.mp4
9
+ *.mov
10
+ *.mkv
11
+ *.log
12
+ *.zip
13
+ *.pt
14
+ *.pth
15
+ *.ckpt
16
+ *.safetensors
17
+ #*.json
18
+ # *.txt
19
+ *.backup
20
+ *.pkl
21
+ *.html
22
+ *.pdf
23
+ *.whl
24
+ *.exe
25
+ cache
26
+ __pycache__/
27
+ storage/
28
+ samples/
29
+ !.gitignore
30
+ !requirements.txt
31
+ .DS_Store
32
+ *DS_Store
33
+ google/
34
+ Wan2.1-T2V-14B/
35
+ Wan2.1-T2V-1.3B/
36
+ Wan2.1-I2V-14B-480P/
37
+ Wan2.1-I2V-14B-720P/
38
+ outputs/
39
+ outputs2/
40
+ gradio_outputs/
41
+ ckpts/
42
+ loras/
43
+ loras_i2v/
44
+
45
+ settings/
46
+
47
+ wgp_config.json
Custom Resolutions Instructions.txt ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ You can override the choice of Resolutions offered by WanGP, if you create a file "resolutions.json" in the main WanGP folder.
2
+ This file is composed of a list of 2 elements sublists. Each 2 elements sublist should have the format ["Label", "WxH"] where W, H are respectively the Width and Height of the resolution. Please make sure that W and H are multiples of 16. The letter "x" should be placed inbetween these two dimensions.
3
+
4
+ Here is below a sample "resolutions.json" file :
5
+
6
+ [
7
+ ["1280x720 (16:9, 720p)", "1280x720"],
8
+ ["720x1280 (9:16, 720p)", "720x1280"],
9
+ ["1024x1024 (1:1, 720p)", "1024x1024"],
10
+ ["1280x544 (21:9, 720p)", "1280x544"],
11
+ ["544x1280 (9:21, 720p)", "544x1280"],
12
+ ["1104x832 (4:3, 720p)", "1104x832"],
13
+ ["832x1104 (3:4, 720p)", "832x1104"],
14
+ ["960x960 (1:1, 720p)", "960x960"],
15
+ ["832x480 (16:9, 480p)", "832x480"]
16
+ ]
Dockerfile ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ FROM nvidia/cuda:12.4.1-cudnn-devel-ubuntu22.04
2
+
3
+ # Build arg for GPU architectures - specify which CUDA compute capabilities to compile for
4
+ # Common values:
5
+ # 7.0 - Tesla V100
6
+ # 7.5 - RTX 2060, 2070, 2080, Titan RTX
7
+ # 8.0 - A100, A800 (Ampere data center)
8
+ # 8.6 - RTX 3060, 3070, 3080, 3090 (Ampere consumer)
9
+ # 8.9 - RTX 4070, 4080, 4090 (Ada Lovelace)
10
+ # 9.0 - H100, H800 (Hopper data center)
11
+ # 12.0 - RTX 5070, 5080, 5090 (Blackwell) - Note: sm_120 architecture
12
+ #
13
+ # Examples:
14
+ # RTX 3060: --build-arg CUDA_ARCHITECTURES="8.6"
15
+ # RTX 4090: --build-arg CUDA_ARCHITECTURES="8.9"
16
+ # Multiple: --build-arg CUDA_ARCHITECTURES="8.0;8.6;8.9"
17
+ #
18
+ # Note: Including 8.9 or 9.0 may cause compilation issues on some setups
19
+ # Default includes 8.0 and 8.6 for broad Ampere compatibility
20
+ ARG CUDA_ARCHITECTURES="8.0;8.6"
21
+
22
+ ENV DEBIAN_FRONTEND=noninteractive
23
+
24
+ # Install system dependencies
25
+ RUN apt update && \
26
+ apt install -y \
27
+ python3 python3-pip git wget curl cmake ninja-build \
28
+ libgl1 libglib2.0-0 ffmpeg && \
29
+ apt clean
30
+
31
+ WORKDIR /workspace
32
+
33
+ COPY requirements.txt .
34
+
35
+ # Upgrade pip first
36
+ RUN pip install --upgrade pip setuptools wheel
37
+
38
+ # Install requirements if exists
39
+ RUN pip install -r requirements.txt
40
+
41
+ # Install PyTorch with CUDA support
42
+ RUN pip install --extra-index-url https://download.pytorch.org/whl/cu124 \
43
+ torch==2.6.0+cu124 torchvision==0.21.0+cu124
44
+
45
+ # Install SageAttention from git (patch GPU detection)
46
+ ENV TORCH_CUDA_ARCH_LIST="${CUDA_ARCHITECTURES}"
47
+ ENV FORCE_CUDA="1"
48
+ ENV MAX_JOBS="1"
49
+
50
+ COPY <<EOF /tmp/patch_setup.py
51
+ import os
52
+ with open('setup.py', 'r') as f:
53
+ content = f.read()
54
+
55
+ # Get architectures from environment variable
56
+ arch_list = os.environ.get('TORCH_CUDA_ARCH_LIST')
57
+ arch_set = '{' + ', '.join([f'"{arch}"' for arch in arch_list.split(';')]) + '}'
58
+
59
+ # Replace the GPU detection section
60
+ old_section = '''compute_capabilities = set()
61
+ device_count = torch.cuda.device_count()
62
+ for i in range(device_count):
63
+ major, minor = torch.cuda.get_device_capability(i)
64
+ if major < 8:
65
+ warnings.warn(f"skipping GPU {i} with compute capability {major}.{minor}")
66
+ continue
67
+ compute_capabilities.add(f"{major}.{minor}")'''
68
+
69
+ new_section = 'compute_capabilities = ' + arch_set + '''
70
+ print(f"Manually set compute capabilities: {compute_capabilities}")'''
71
+
72
+ content = content.replace(old_section, new_section)
73
+
74
+ with open('setup.py', 'w') as f:
75
+ f.write(content)
76
+ EOF
77
+
78
+ RUN git clone https://github.com/thu-ml/SageAttention.git /tmp/sageattention && \
79
+ cd /tmp/sageattention && \
80
+ python3 /tmp/patch_setup.py && \
81
+ pip install --no-build-isolation .
82
+
83
+ RUN useradd -u 1000 -ms /bin/bash user
84
+
85
+ RUN chown -R user:user /workspace
86
+
87
+ RUN mkdir /home/user/.cache && \
88
+ chown -R user:user /home/user/.cache
89
+
90
+ COPY entrypoint.sh /workspace/entrypoint.sh
91
+
92
+ ENTRYPOINT ["/workspace/entrypoint.sh"]
LICENSE.txt ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ WanGP NON-COMMERCIAL EVALUATION LICENSE 1.0
2
+
3
+ Definitions
4
+ 1.1 “Software” means the source code, binaries, libraries, utilities and UI released under this license.
5
+ 1.2 “Output” means images, videos or other media produced by running the Software.
6
+ 1.3 “Commercial Use” means:
7
+ a) selling, sublicensing, renting, leasing, or otherwise distributing the Software, in whole or in part, for a fee or other consideration; or
8
+ b) offering the Software (or any derivative) as part of a paid product or hosted service; or
9
+ c) using the Software (or any derivative) to provide cloud-based or backend services, where end users access or pay for those services.
10
+
11
+ License Grant
12
+ Subject to Section 3:
13
+ a) You are granted a worldwide, non-exclusive, royalty-free, revocable license to use, reproduce, modify and distribute the Software for non-commercial purposes only.
14
+ b) You are granted a worldwide, non-exclusive, royalty-free, irrevocable license to use, reproduce, modify and distribute the Output for any purpose, including commercial sale, provided that any commercial distribution of the Output includes a clear notice that the Output was produced (in whole or in part) using WanGP, along with a hyperlink to the WanGP application’s About tab or repository.
15
+
16
+ Restrictions
17
+ 3.1 You MAY NOT distribute, sublicense or otherwise make available the Software (or any derivative) for Commercial Use.
18
+ 3.2 You MAY sell, license or otherwise commercially exploit the Output without restriction.
19
+ 3.3 If you wish to use the Software for Commercial Use, you must obtain a separate commercial license from the Licensor.
20
+
21
+ Third-Party Components 4.1 The Software includes components licensed under various open-source licenses (e.g., Apache 2.0, MIT, BSD). 4.2 You must comply with all applicable terms of those third-party licenses, including preservation of copyright notices, inclusion of required license texts, and patent-grant provisions. 4.3 You can find the full text of each third-party license via the “About” tab in the WanGP application, which provides links to their original GitHub repositories.
22
+
23
+ Attribution
24
+ 5.1 You must give appropriate credit by including:
25
+ • a copy of this license (or a link to it), and
26
+ • a notice that your use is based on “WanGP”.
27
+ 5.2 You may do so in any reasonable manner, but not in any way that suggests the Licensor endorses you or your use.
28
+
29
+ Disclaimer of Warranty & Liability
30
+ THE SOFTWARE IS PROVIDED “AS IS”, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED.
31
+ IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE.
32
+
33
+ Commercial Licensing The Licensor may offer commercial licenses for the Software, which grant rights to use the Software for Commercial Use. Please contact [deepbeepmeep@yahoo.com] for terms and pricing.
34
+
35
+ Effective Date & Previous Versions
36
+ 8.1 This license is effective as of the date the LICENSE file is updated in the WanGP repository.
37
+ 8.2 Any copies of the Software obtained under prior license terms before this Effective Date remain governed by those prior terms; such granted rights are irrevocable.
38
+ 8.3 Use of the Software after the release of any subsequent version by the Licensor is subject to the terms of the then-current license, unless a separate agreement is in place.
39
+
40
+ Acceptable Use / Moral Clause
41
+ 9.1 You MAY NOT use the Software or the Output to facilitate or produce content that is illegal, harmful, violent, harassing, defamatory, fraudulent, or otherwise violates applicable laws or fundamental human rights.
42
+ 9.2 You MAY NOT deploy the Software or Output in contexts that promote hate speech, extremist ideology, human rights abuses, or other actions that could foreseeably cause significant harm to individuals or groups.
43
+ 9.3 The Licensor reserves the right to terminate the rights granted under this license if a licensee materially breaches this Acceptable Use clause.
44
+
45
+ END OF LICENSE
46
+
README.md ADDED
@@ -0,0 +1,256 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # WanGP
2
+
3
+ -----
4
+ <p align="center">
5
+ <b>WanGP by DeepBeepMeep : The best Open Source Video Generative Models Accessible to the GPU Poor</b>
6
+ </p>
7
+
8
+ WanGP supports the Wan (and derived models), Hunyuan Video and LTV Video models with:
9
+ - Low VRAM requirements (as low as 6 GB of VRAM is sufficient for certain models)
10
+ - Support for old Nvidia GPUs (RTX 10XX, 20xx, ...)
11
+ - Support for AMD GPUs Radeon RX 76XX, 77XX, 78XX & 79XX, instructions in the Installation Section Below.
12
+ - Very Fast on the latest GPUs
13
+ - Easy to use Full Web based interface
14
+ - Auto download of the required model adapted to your specific architecture
15
+ - Tools integrated to facilitate Video Generation : Mask Editor, Prompt Enhancer, Temporal and Spatial Generation, MMAudio, Video Browser, Pose / Depth / Flow extractor
16
+ - Loras Support to customize each model
17
+ - Queuing system : make your shopping list of videos to generate and come back later
18
+
19
+ **Discord Server to get Help from Other Users and show your Best Videos:** https://discord.gg/g7efUW9jGV
20
+
21
+ **Follow DeepBeepMeep on Twitter/X to get the Latest News**: https://x.com/deepbeepmeep
22
+
23
+ -----
24
+
25
+ ### You have your choice of Dark or Light Theme
26
+
27
+
28
+ <img width="1895" height="1023" alt="Screenshot 2025-10-23 210313" src="https://github.com/user-attachments/assets/3778ae4e-6a95-4752-ba47-bb160c653310" />
29
+
30
+ -----
31
+ <img width="1899" height="1020" alt="Screenshot 2025-10-23 210500" src="https://github.com/user-attachments/assets/5e524260-ad24-4203-acf2-6622676a83bb" />
32
+
33
+ -----
34
+ ![Screen Recording 2025-10-23 210625 - frame at 0m9s](https://github.com/user-attachments/assets/c65a815e-09fa-41a7-bc49-5f879b0b8ece)
35
+
36
+ -----
37
+
38
+ ## 🔥 Latest Updates :
39
+ ### October 26 2025: WanGP v9.21, Why isn't all my VRAM used ?
40
+
41
+
42
+ *WanGP exclusive*: VRAM requirements have never been that low !
43
+
44
+ **Wan 2.2 Ovi 10 GB** for all the GPU Poors of the World: *only 6 GB of VRAM to generate 121 frames at 720p*. With 16 GB of VRAM, you may even be able to load all the model in VRAM with *Memory Profile 3*
45
+
46
+ To get the x10 speed effect just apply the FastWan Lora Accelerator that comes prepackaged with Ovi (acccessible in the dropdown box Settings at the top)
47
+
48
+ *update v9.21*: Got FastWan to work with Ovi: it is now 10 times faster ! (not including the VAE)
49
+
50
+ ### October 24 2025: WanGP v9.10, What else will you ever need after this one ?
51
+
52
+ With WanGP v9 you will have enough features to go to a desert island with no internet connection and comes back with a full Hollywood movie.
53
+
54
+ First here are the new models supported:
55
+ - **Wan 2.1 Alpha** : a very requested model that can generate videos with *semi transparent background* (as it is very lora picky it supports only the *Self Forcing / lightning* loras accelerators)
56
+ - **Chatterbox Multilingual**: the first *Voice Generator* in WanGP. Let's say you have a flu and lost your voice (somehow I can't think of another usecase), the world will still be able to hear you as *Chatterbox* can generate up to 15s clips of your voice using a recorded voice sample. Chatterbox works with numerous languages out the box.
57
+ - **Flux DreamOmni2** : another wannabe *Nano Banana* image Editor / image composer. The *Edit Mode* ("Conditional Image is first Main Subject ...") seems to work better than the *Gen Mode* (Conditional Images are People / Objects ..."). If you have at least 16 GB of VRAM it is recommended to force profile 3 for this model (it uses an autoregressive model for the prompt encoding and the start may be slow).
58
+ - **Ditto** (new with *WanGP 9.1* !): a powerful Video 2 Video model, can change for instance the style or the material visible in the video. Be aware it is an instruct based model, so the prompt should contain intructions.
59
+
60
+ Upgraded Features:
61
+ - A new **Audio Gallery** to store your Chatterbox generations and import your audio assets. *Metadata support* (stored gen settings) for *Wav files* generated with WanGP available from day one.
62
+ - **Matanyone** improvements: you can now use it during a video gen, it will *suspend gracefully the Gen in progress*. *Input Video / Images* can be resized for faster processing & lower VRAM. Image version can now generate *Green screens* (not used by WanGP but I did it because someone asked for it and I am nice) and *Alpha masks*.
63
+ - **Images Stored in Metadata**: Video Gen *Settings Metadata* that are stored in the Generated Videos can now contain the Start Image, Image Refs used to generate the Video. Many thanks to **Gunther-Schulz** for this contribution
64
+ - **Three Levels of Hierarchy** to browse the models / finetunes: you can collect as many finetunes as you want now and they will no longer encumber the UI.
65
+ - Added **Loras Accelerators** for *Wan 2.1 1.3B*, *Wan 2.2 i2v*, *Flux* and the latest *Wan 2.2 Lightning*
66
+ - Finetunes now support **Custom Text Encoders** : you will need to use the "text_encoder_URLs" key. Please check the finetunes doc.
67
+ - Sometime Less is More: removed the palingenesis finetunes that were controversial
68
+
69
+ Huge Kudos & Thanks to **Tophness** that has outdone himself with these Great Features:
70
+ - **Multicolors Queue** items with **Drag & Drop** to reorder them
71
+ - **Edit a Gen Request** that is already in the queue
72
+ - Added **Plugin support** to WanGP : found that features are missing in WanGP, you can now add tabs at the top in WanGP. Each tab may contain a full embedded App that can share data with the Video Generator of WanGP. Please check the Plugin guide written by Tophness and don't hesitate to contact him or me on the Discord if you have a plugin you want to share. I have added a new Plugins channels to discuss idea of plugins and help each other developing plugins. *Idea for a PlugIn that may end up popular*: a screen where you view the hard drive space used per model and that will let you remove unused models weights
73
+ - Two Plugins ready to use designed & developped by **Tophness**: an **Extended Gallery** and a **Lora multipliers Wizard**
74
+
75
+ WanGP v9 is now targetting Pytorch 2.8 although it should still work with 2.7, don't forget to upgrade by doing:
76
+ ```bash
77
+ pip install torch==2.8.0 torchvision torchaudio --index-url https://download.pytorch.org/whl/test/cu128
78
+ ```
79
+ You will need to upgrade Sage Attention or Flash (check the installation guide)
80
+
81
+ *Update info: you might have some git error message while upgrading to v9 if WanGP is already installed.*
82
+ Sorry about that if that's the case, you will need to reinstall WanGP.
83
+ There are two different ways to fix this issue while still preserving your data:
84
+ 1) **Command Line**
85
+ If you have access to a terminal window :
86
+ ```
87
+ cd installation_path_of_wangp
88
+ git fetch origin && git reset --hard origin/main
89
+ pip install -r requirements.txt
90
+ ```
91
+
92
+ 2) **Generic Method**
93
+ a) move outside the installation WanGP folder the folders **ckpts**, **settings**, **outputs** and all the **loras** folders and the file **wgp_config.json**
94
+ b) delete the WanGP folder and reinstall
95
+ c) move back what you moved in a)
96
+
97
+ ## 🔥 Latest Updates :
98
+ ### October 6 2025: WanGP v8.999 - A few last things before the Big Unknown ...
99
+
100
+ This new version hasn't any new model...
101
+
102
+ ...but temptation to upgrade will be high as it contains a few Loras related features that may change your Life:
103
+ - **Ready to use Loras Accelerators Profiles** per type of model that you can apply on your current *Generation Settings*. Next time I will recommend a *Lora Accelerator*, it will be only one click away. And best of all of the required Loras will be downloaded automatically. When you apply an *Accelerator Profile*, input fields like the *Number of Denoising Steps* *Activated Loras*, *Loras Multipliers* (such as "1;0 0;1" ...) will be automatically filled. However your video specific fields will be preserved, so it will be easy to switch between Profiles to experiment. With *WanGP 8.993*, the *Accelerator Loras* are now merged with *Non Accelerator Loras". Things are getting too easy...
104
+
105
+ - **Embedded Loras URL** : WanGP will now try to remember every Lora URLs it sees. For instance if someone sends you some settings that contain Loras URLs or you extract the Settings of Video generated by a friend with Loras URLs, these URLs will be automatically added to *WanGP URL Cache*. Conversely everything you will share (Videos, Settings, Lset files) will contain the download URLs if they are known. You can also download directly a Lora in WanGP by using the *Download Lora* button a the bottom. The Lora will be immediatly available and added to WanGP lora URL cache. This will work with *Hugging Face* as a repository. Support for CivitAi will come as soon as someone will nice enough to post a GitHub PR ...
106
+
107
+ - **.lset file** supports embedded Loras URLs. It has never been easier to share a Lora with a friend. As a reminder a .lset file can be created directly from *WanGP Web Interface* and it contains a list of Loras and their multipliers, a Prompt and Instructions how to use these loras (like the Lora's *Trigger*). So with embedded Loras URL, you can send an .lset file by email or share it on discord: it is just a 1 KB tiny text, but with it other people will be able to use Gigabytes Loras as these will be automatically downloaded.
108
+
109
+ I have created the new Discord Channel **share-your-settings** where you can post your *Settings* or *Lset files*. I will be pleased to add new Loras Accelerators in the list of WanGP *Accelerators Profiles if you post some good ones there.
110
+
111
+ *With the 8.993 update*, I have added support for **Scaled FP8 format**. As a sample case, I have created finetunes for the **Wan 2.2 PalinGenesis** Finetune which is quite popular recently. You will find it in 3 flavors : *t2v*, *i2v* and *Lightning Accelerated for t2v*.
112
+
113
+ The *Scaled FP8 format* is widely used as it the format used by ... *ComfyUI*. So I except a flood of Finetunes in the *share-your-finetune* channel. If not it means this feature was useless and I will remove it &#x1F608;&#x1F608;&#x1F608;
114
+
115
+ Not enough Space left on your SSD to download more models ? Would like to reuse Scaled FP8 files in your ComfyUI Folder without duplicating them ? Here comes *WanGP 8.994* **Multiple Checkpoints Folders** : you just need to move the files into different folders / hard drives or reuse existing folders and let know WanGP about it in the *Config Tab* and WanGP will be able to put all the parts together.
116
+
117
+ Last but not least the Lora's documentation has been updated.
118
+
119
+ *update 8.991*: full power of *Vace Lynx* unleashed with new combinations such as Landscape + Face / Clothes + Face / Injectd Frame (Start/End frames/...) + Face
120
+ *update 8.992*: optimized gen with Lora, should be 10% faster if many loras
121
+ *update 8.993*: Support for *Scaled FP8* format and samples *Paligenesis* finetunes, merged Loras Accelerators and Non Accelerators
122
+ *update 8.994*: Added custom checkpoints folders
123
+ *update 8.999*: fixed a lora + fp8 bug and version sync for the jump to the unknown
124
+
125
+ ### September 30 2025: WanGP v8.9 - Combinatorics
126
+
127
+ This new version of WanGP introduces **Wan 2.1 Lynx** the best Control Net so far to transfer *Facial Identity*. You will be amazed to recognize your friends even with a completely different hair style. Congrats to the *Byte Dance team* for this achievement. Lynx works quite with well *Fusionix t2v* 10 steps.
128
+
129
+ *WanGP 8.9* also illustrate how existing WanGP features can be easily combined with new models. For instance with *Lynx* you will get out of the box *Video to Video* and *Image/Text to Image*.
130
+
131
+ Another fun combination is *Vace* + *Lynx*, which works much better than *Vace StandIn*. I have added sliders to change the weight of Vace & Lynx to allow you to tune the effects.
132
+
133
+
134
+
135
+ See full changelog: **[Changelog](docs/CHANGELOG.md)**
136
+
137
+ ## 📋 Table of Contents
138
+
139
+ - [🚀 Quick Start](#-quick-start)
140
+ - [📦 Installation](#-installation)
141
+ - [🎯 Usage](#-usage)
142
+ - [📚 Documentation](#-documentation)
143
+ - [🔗 Related Projects](#-related-projects)
144
+
145
+ ## 🚀 Quick Start
146
+
147
+ **One-click installation:**
148
+ - Get started instantly with [Pinokio App](https://pinokio.computer/)
149
+ - Use Redtash1 [One Click Install with Sage](https://github.com/Redtash1/Wan2GP-Windows-One-Click-Install-With-Sage)
150
+
151
+ **Manual installation:**
152
+ ```bash
153
+ git clone https://github.com/deepbeepmeep/Wan2GP.git
154
+ cd Wan2GP
155
+ conda create -n wan2gp python=3.10.9
156
+ conda activate wan2gp
157
+ pip install torch==2.8.0 torchvision torchaudio --index-url https://download.pytorch.org/whl/test/cu128
158
+ pip install -r requirements.txt
159
+ ```
160
+
161
+ **Run the application:**
162
+ ```bash
163
+ python wgp.py
164
+ ```
165
+
166
+ **Update the application:**
167
+ If using Pinokio use Pinokio to update otherwise:
168
+ Get in the directory where WanGP is installed and:
169
+ ```bash
170
+ git pull
171
+ conda activate wan2gp
172
+ pip install -r requirements.txt
173
+ ```
174
+
175
+ if you get some error messages related to git, you may try the following (beware this will overwrite local changes made to the source code of WanGP):
176
+ ```bash
177
+ git fetch origin && git reset --hard origin/main
178
+ conda activate wan2gp
179
+ pip install -r requirements.txt
180
+ ```
181
+
182
+ ## 🐳 Docker:
183
+
184
+ **For Debian-based systems (Ubuntu, Debian, etc.):**
185
+
186
+ ```bash
187
+ ./run-docker-cuda-deb.sh
188
+ ```
189
+
190
+ This automated script will:
191
+
192
+ - Detect your GPU model and VRAM automatically
193
+ - Select optimal CUDA architecture for your GPU
194
+ - Install NVIDIA Docker runtime if needed
195
+ - Build a Docker image with all dependencies
196
+ - Run WanGP with optimal settings for your hardware
197
+
198
+ **Docker environment includes:**
199
+
200
+ - NVIDIA CUDA 12.4.1 with cuDNN support
201
+ - PyTorch 2.6.0 with CUDA 12.4 support
202
+ - SageAttention compiled for your specific GPU architecture
203
+ - Optimized environment variables for performance (TF32, threading, etc.)
204
+ - Automatic cache directory mounting for faster subsequent runs
205
+ - Current directory mounted in container - all downloaded models, loras, generated videos and files are saved locally
206
+
207
+ **Supported GPUs:** RTX 40XX, RTX 30XX, RTX 20XX, GTX 16XX, GTX 10XX, Tesla V100, A100, H100, and more.
208
+
209
+ ## 📦 Installation
210
+
211
+ ### Nvidia
212
+ For detailed installation instructions for different GPU generations:
213
+ - **[Installation Guide](docs/INSTALLATION.md)** - Complete setup instructions for RTX 10XX to RTX 50XX
214
+
215
+ ### AMD
216
+ For detailed installation instructions for different GPU generations:
217
+ - **[Installation Guide](docs/AMD-INSTALLATION.md)** - Complete setup instructions for Radeon RX 76XX, 77XX, 78XX & 79XX
218
+
219
+ ## 🎯 Usage
220
+
221
+ ### Basic Usage
222
+ - **[Getting Started Guide](docs/GETTING_STARTED.md)** - First steps and basic usage
223
+ - **[Models Overview](docs/MODELS.md)** - Available models and their capabilities
224
+
225
+ ### Advanced Features
226
+ - **[Loras Guide](docs/LORAS.md)** - Using and managing Loras for customization
227
+ - **[Finetunes](docs/FINETUNES.md)** - Add manually new models to WanGP
228
+ - **[VACE ControlNet](docs/VACE.md)** - Advanced video control and manipulation
229
+ - **[Command Line Reference](docs/CLI.md)** - All available command line options
230
+
231
+ ## 📚 Documentation
232
+
233
+ - **[Changelog](docs/CHANGELOG.md)** - Latest updates and version history
234
+ - **[Troubleshooting](docs/TROUBLESHOOTING.md)** - Common issues and solutions
235
+
236
+ ## 📚 Video Guides
237
+ - Nice Video that explain how to use Vace:\
238
+ https://www.youtube.com/watch?v=FMo9oN2EAvE
239
+ - Another Vace guide:\
240
+ https://www.youtube.com/watch?v=T5jNiEhf9xk
241
+
242
+ ## 🔗 Related Projects
243
+
244
+ ### Other Models for the GPU Poor
245
+ - **[HuanyuanVideoGP](https://github.com/deepbeepmeep/HunyuanVideoGP)** - One of the best open source Text to Video generators
246
+ - **[Hunyuan3D-2GP](https://github.com/deepbeepmeep/Hunyuan3D-2GP)** - Image to 3D and text to 3D tool
247
+ - **[FluxFillGP](https://github.com/deepbeepmeep/FluxFillGP)** - Inpainting/outpainting tools based on Flux
248
+ - **[Cosmos1GP](https://github.com/deepbeepmeep/Cosmos1GP)** - Text to world generator and image/video to world
249
+ - **[OminiControlGP](https://github.com/deepbeepmeep/OminiControlGP)** - Flux-derived application for object transfer
250
+ - **[YuE GP](https://github.com/deepbeepmeep/YuEGP)** - Song generator with instruments and singer's voice
251
+
252
+ ---
253
+
254
+ <p align="center">
255
+ Made with ❤️ by DeepBeepMeep
256
+ </p>
defaults/ReadMe.txt ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Please dot not modify any file in this Folder.
2
+
3
+ If you want to change a property of a default model, copy the corrresponding model file in the ./finetunes folder and modify the properties you want to change in the new file.
4
+ If a property is not in the new file, it will be inherited automatically from the default file that matches the same name file.
5
+
6
+ For instance to hide a model:
7
+
8
+ {
9
+ "model":
10
+ {
11
+ "visible": false
12
+ }
13
+ }
defaults/alpha.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Wan2.1 Alpha 14B",
5
+ "architecture" : "alpha",
6
+ "description": "This model successfully generates various scenes with accurate and clearly rendered transparency. Notably, it can synthesize diverse semi-transparent objects, glowing effects, and fine-grained details such as hair. For each video generated you will find a Zip file with the same name that will contain the corresponding RGBA images.",
7
+ "URLs": "t2v",
8
+ "preload_URLs": [
9
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan_alpha_2.1_vae_rgb_channel.safetensors",
10
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan_alpha_2.1_vae_alpha_channel.safetensors"
11
+ ],
12
+ "loras": [
13
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan_alpha_2.1_dora.safetensors"
14
+ ],
15
+ "loras_multipliers": [ 1 ]
16
+ },
17
+ "prompt": "A large orange octopus is seen resting. The background of the video is transparent."
18
+
19
+ }
defaults/alpha_sf.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Wan2.1 Alpha Lightning 14B",
5
+ "architecture" : "alpha",
6
+ "description": "This model is accelerated by the Lightning / SelfForcing process. It successfully generates various scenes with accurate and clearly rendered transparency. Notably, it can synthesize diverse semi-transparent objects, glowing effects, and fine-grained details such as hair. For each video generated you will find a Zip file with the same name that will contain the corresponding RGBA images.",
7
+ "URLs": "t2v_sf",
8
+ "preload_URLs": "alpha",
9
+ "loras": "alpha",
10
+ "loras_multipliers": [ 1 ],
11
+ "profiles_dir" : [""]
12
+ },
13
+ "prompt": "A large orange octopus is seen resting. The background of the video is transparent.",
14
+ "num_inference_steps": 4,
15
+ "guidance_scale": 1,
16
+ "flow_shift": 3
17
+ }
defaults/animate.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Wan2.2 Animate 14B",
4
+ "architecture": "animate",
5
+ "description": "Wan-Animate takes a video and a character image as input, and generates a video in either 'Animation' or 'Replacement' mode. Sliding Window of 81 frames at least are recommeded to obtain the best Style continuity.",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_animate_14B_bf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_animate_14B_quanto_fp16_int8.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_animate_14B_quanto_bf16_int8.safetensors"
10
+ ],
11
+ "preload_URLs" :
12
+ [
13
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_animate_relighting_lora.safetensors"
14
+ ],
15
+ "group": "wan2_2"
16
+ }
17
+ }
defaults/chatterbox.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "TTS Chatterbox Multilingual",
4
+ "architecture": "chatterbox",
5
+ "description": "Resemble AI's open multilingual TTS with language selection via model mode.",
6
+ "URLs": []
7
+ },
8
+ "prompt": "Welcome to Chatterbox !",
9
+ "negative_prompt": "",
10
+ "audio_prompt_type": "A",
11
+ "model_mode": "en",
12
+ "repeat_generation": 1,
13
+ "video_length": 0,
14
+ "num_inference_steps": 0,
15
+ "pace": 0.5,
16
+ "exaggeration": 0.5,
17
+ "temperature": 0.8
18
+ }
defaults/fantasy.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Fantasy Talking 720p 14B",
5
+ "architecture" : "fantasy",
6
+ "modules": [ ["https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_fantasy_speaking_14B_bf16.safetensors"]],
7
+ "description": "The Fantasy Talking model corresponds to the original Wan image 2 video model combined with the Fantasy Speaking module to process an audio Input.",
8
+ "URLs": "i2v_720p"
9
+ },
10
+ "resolution": "1280x720"
11
+ }
defaults/flf2v_720p.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "First Last Frame to Video 720p (FLF2V) 14B",
5
+ "architecture" : "flf2v_720p",
6
+ "visible" : true,
7
+ "description": "The First Last Frame 2 Video model is the official model Image 2 Video model that supports Start and End frames.",
8
+ "URLs": [
9
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_FLF2V_720p_14B_mbf16.safetensors",
10
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_FLF2V_720p_14B_quanto_mbf16_int8.safetensors",
11
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_FLF2V_720p_14B_quanto_mfp16_int8.safetensors"
12
+ ],
13
+ "auto_quantize": true
14
+ },
15
+ "resolution": "1280x720"
16
+ }
defaults/flux.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 Dev 12B",
4
+ "architecture": "flux",
5
+ "description": "FLUX.1 Dev is a 12 billion parameter rectified flow transformer capable of generating images from text descriptions.",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-dev_bf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-dev_quanto_bf16_int8.safetensors"
9
+ ],
10
+ "image_outputs": true
11
+ },
12
+ "prompt": "draw a hat",
13
+ "resolution": "1280x720",
14
+ "batch_size": 1
15
+ }
defaults/flux_chroma.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 Chroma 1 HD 8.9B",
4
+ "architecture": "flux_chroma",
5
+ "description": "FLUX.1 Chroma is a 8.9 billion parameters model. As a base model, Chroma1 is intentionally designed to be an excellent starting point for finetuning. It provides a strong, neutral foundation for developers, researchers, and artists to create specialized models..",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-chroma_hd_bf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-chroma_hd_quanto_bf16_int8.safetensors"
9
+ ],
10
+ "image_outputs": true
11
+ },
12
+ "prompt": "draw a hat",
13
+ "resolution": "1280x720",
14
+ "guidance_scale": 3.0,
15
+ "num_inference_steps": 20,
16
+ "batch_size": 1
17
+ }
defaults/flux_dev_kontext.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 Dev Kontext 12B",
4
+ "architecture": "flux_dev_kontext",
5
+ "description": "FLUX.1 Kontext is a 12 billion parameter rectified flow transformer capable of editing images based on instructions stored in the Prompt. Please be aware that Flux Kontext is picky on the resolution of the input image and the output dimensions may not match the dimensions of the input image.",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1_kontext_dev_bf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1_kontext_dev_quanto_bf16_int8.safetensors"
9
+ ]
10
+ },
11
+ "prompt": "add a hat",
12
+ "resolution": "1280x720",
13
+ "batch_size": 1
14
+ }
15
+
16
+
defaults/flux_dev_kontext_dreamomni2.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 DreamOmni2 12B",
4
+ "architecture": "flux_dev_kontext_dreamomni2",
5
+ "description": "DreamOmni2 is a Multimodal Instruction-based Editing and Generation Model",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1_kontext_dev_bf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1_kontext_dev_quanto_bf16_int8.safetensors"
9
+ ],
10
+ "preload_URLs": [ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux_dreamomni2_edit_lora.safetensors",
11
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux_dreamomni2_gen_lora.safetensors"
12
+ ]
13
+ },
14
+ "prompt": "In the scene, the character from the first image stands on the left, and the character from the second image stands on the right. They are shaking hands against the backdrop of a spaceship interior.",
15
+ "resolution": "1280x720",
16
+ "batch_size": 1
17
+ }
18
+
19
+
defaults/flux_dev_umo.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 UMO Dev 12B",
4
+ "architecture": "flux_dev_umo",
5
+ "description": "FLUX.1 UMO Dev is a model that can Edit Images with a specialization in combining multiple image references (resized internally at 512x512 max) to produce an Image output. Best Image preservation at 768x768 Resolution Output.",
6
+ "URLs": "flux",
7
+ "loras": ["https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-dev-UMO_dit_lora_bf16.safetensors"],
8
+ "resolutions": [ ["1024x1024 (1:1)", "1024x1024"],
9
+ ["768x1024 (3:4)", "768x1024"],
10
+ ["1024x768 (4:3)", "1024x768"],
11
+ ["512x1024 (1:2)", "512x1024"],
12
+ ["1024x512 (2:1)", "1024x512"],
13
+ ["768x768 (1:1)", "768x768"],
14
+ ["768x512 (3:2)", "768x512"],
15
+ ["512x768 (2:3)", "512x768"]]
16
+ },
17
+ "prompt": "the man is wearing a hat",
18
+ "embedded_guidance_scale": 4,
19
+ "resolution": "768x768",
20
+ "batch_size": 1
21
+ }
22
+
23
+
defaults/flux_dev_uso.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 USO Dev 12B",
4
+ "architecture": "flux_dev_uso",
5
+ "description": "FLUX.1 USO Dev is a model that can Edit Images with a specialization in Style Transfers (up to two).",
6
+ "modules": [ ["https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-dev-USO_projector_bf16.safetensors"]],
7
+ "URLs": "flux",
8
+ "loras": ["https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-dev-USO_dit_lora_bf16.safetensors"]
9
+ },
10
+ "prompt": "the man is wearing a hat",
11
+ "embedded_guidance_scale": 4,
12
+ "resolution": "1024x1024",
13
+ "batch_size": 1
14
+ }
15
+
16
+
defaults/flux_krea.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 Dev Krea 12B",
4
+ "architecture": "flux",
5
+ "description": "Cutting-edge output quality, with a focus on aesthetic photography..",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-krea-dev_bf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-krea-dev_quanto_bf16_int8.safetensors"
9
+ ],
10
+ "image_outputs": true
11
+ },
12
+ "prompt": "draw a hat",
13
+ "resolution": "1280x720",
14
+ "batch_size": 1
15
+ }
defaults/flux_schnell.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 Schnell 12B",
4
+ "architecture": "flux_schnell",
5
+ "description": "FLUX.1 Schnell is a 12 billion parameter rectified flow transformer capable of generating images from text descriptions. As a distilled model it requires fewer denoising steps.",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-schnell_bf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-schnell_quanto_bf16_int8.safetensors"
9
+ ],
10
+ "image_outputs": true
11
+ },
12
+ "prompt": "draw a hat",
13
+ "resolution": "1280x720",
14
+ "num_inference_steps": 10,
15
+ "batch_size": 1
16
+ }
defaults/flux_srpo.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 Dev SRPO 12B",
4
+ "architecture": "flux",
5
+ "description": "By fine-tuning the FLUX.1.dev model with optimized denoising and online reward adjustment, SRPO improves its human-evaluated realism and aesthetic quality by over 3x.",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-srpo-dev_bf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-srpo-dev_quanto_bf16_int8.safetensors"
9
+ ]
10
+ },
11
+ "prompt": "draw a hat",
12
+ "resolution": "1024x1024",
13
+ "batch_size": 1
14
+ }
defaults/flux_srpo_uso.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 USO SRPO 12B",
4
+ "architecture": "flux_dev_uso",
5
+ "description": "FLUX.1 USO SRPO is a model that can Edit Images with a specialization in Style Transfers (up to two). It leverages the improved Image quality brought by the SRPO process",
6
+ "modules": [ "flux_dev_uso"],
7
+ "URLs": "flux_srpo",
8
+ "loras": "flux_dev_uso"
9
+ },
10
+ "prompt": "the man is wearing a hat",
11
+ "embedded_guidance_scale": 4,
12
+ "resolution": "1024x1024",
13
+ "batch_size": 1
14
+ }
15
+
16
+
defaults/fun_inp.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Fun InP image2video 14B",
5
+ "architecture" : "fun_inp",
6
+ "description": "The Fun model is an alternative image 2 video that supports out the box End Image fixing (contrary to the original Wan image 2 video model).",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_Fun_InP_14B_bf16.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_Fun_InP_14B_quanto_int8.safetensors",
10
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_Fun_InP_14B_quanto_fp16_int8.safetensors"
11
+ ]
12
+ }
13
+ }
defaults/fun_inp_1.3B.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Fun InP image2video 1.3B",
5
+ "architecture" : "fun_inp_1.3B",
6
+ "description": "The Fun model is an alternative image 2 video that supports out the box End Image fixing (contrary to the original Wan image 2 video model). The 1.3B adds also image 2 to video capability to the 1.3B model.",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_Fun_InP_1.3B_bf16.safetensors"
9
+ ]
10
+ }
11
+ }
defaults/hunyuan.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Hunyuan Video Text2video 720p 13B",
5
+ "architecture" : "hunyuan",
6
+ "description": "Probably the best text 2 video model available.",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/HunyuanVideo/resolve/main/hunyuan_video_720_bf16.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/HunyuanVideo/resolve/main/hunyuan_video_720_quanto_int8.safetensors"
10
+ ]
11
+ }
12
+ }
defaults/hunyuan_avatar.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Hunyuan Video Avatar 720p 13B",
5
+ "architecture" : "hunyuan_avatar",
6
+ "description": "With the Hunyuan Video Avatar model you can animate a person based on the content of an audio input. Please note that the video generator works by processing 128 frames segment at a time (even if you ask less). The good news is that it will concatenate multiple segments for long video generation (max 3 segments recommended as the quality will get worse).",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/HunyuanVideo/resolve/main/hunyuan_video_avatar_720_bf16.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/HunyuanVideo/resolve/main/hunyuan_video_avatar_720_quanto_bf16_int8.safetensors"
10
+ ]
11
+ }
12
+ }
defaults/hunyuan_custom.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Hunyuan Video Custom 720p 13B",
5
+ "architecture" : "hunyuan_custom",
6
+ "description": "The Hunyuan Video Custom model is probably the best model to transfer people (only people for the moment) as it is quite good to keep their identity. However it is slow as to get good results, you need to generate 720p videos with 30 steps.",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/HunyuanVideo/resolve/main/hunyuan_video_custom_720_bf16.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/HunyuanVideo/resolve/main/hunyuan_video_custom_720_quanto_bf16_int8.safetensors"
10
+ ]
11
+ }
12
+ }
defaults/hunyuan_custom_audio.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Hunyuan Video Custom Audio 720p 13B",
5
+ "architecture" : "hunyuan_custom_audio",
6
+ "description": "The Hunyuan Video Custom Audio model can be used to generate scenes of a person speaking given a Reference Image and a Recorded Voice or Song. The reference image is not a start image and therefore one can represent the person in a different context.The video length can be anything up to 10s. It is also quite good to generate no sound Video based on a person.",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/HunyuanVideo/resolve/main/hunyuan_video_custom_audio_720_bf16.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/HunyuanVideo/resolve/main/hunyuan_video_custom_audio_720_quanto_bf16_int8.safetensors"
10
+ ]
11
+ }
12
+ }
defaults/hunyuan_custom_edit.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Hunyuan Video Custom Edit 720p 13B",
5
+ "architecture" : "hunyuan_custom_edit",
6
+ "description": "The Hunyuan Video Custom Edit model can be used to do Video inpainting on a person (add accessories or completely replace the person). You will need in any case to define a Video Mask which will indicate which area of the Video should be edited.",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/HunyuanVideo/resolve/main/hunyuan_video_custom_edit_720_bf16.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/HunyuanVideo/resolve/main/hunyuan_video_custom_edit_720_quanto_bf16_int8.safetensors"
10
+ ]
11
+ }
12
+ }
defaults/hunyuan_i2v.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Hunyuan Video Image2video 720p 13B",
5
+ "architecture" : "hunyuan_i2v",
6
+ "description": "A good looking image 2 video model, but not so good in prompt adherence.",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/HunyuanVideo/resolve/main/hunyuan_video_i2v_720_bf16v2.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/HunyuanVideo/resolve/main/hunyuan_video_i2v_720_quanto_int8v2.safetensors"
10
+ ]
11
+ }
12
+ }
defaults/hunyuan_t2v_accvideo.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Hunyuan Video Text2video 720p AccVideo 13B",
4
+ "architecture": "hunyuan",
5
+ "description": " AccVideo is a novel efficient distillation method to accelerate video diffusion models with synthetic datset. Our method is 8.5x faster than HunyuanVideo.",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/HunyuanVideo/resolve/main/accvideo_hunyuan_video_720_quanto_int8.safetensors"
8
+ ],
9
+ "preload_URLs": [
10
+ ],
11
+ "auto_quantize": true
12
+ },
13
+ "negative_prompt": "",
14
+ "resolution": "832x480",
15
+ "video_length": 81,
16
+ "seed": 42,
17
+ "num_inference_steps": 5,
18
+ "flow_shift": 7,
19
+ "embedded_guidance_scale": 6,
20
+ "repeat_generation": 1,
21
+ "loras_multipliers": "",
22
+ "temporal_upsampling": "",
23
+ "spatial_upsampling": "",
24
+ "RIFLEx_setting": 0,
25
+ "slg_start_perc": 10,
26
+ "slg_end_perc": 90,
27
+ "prompt_enhancer": "",
28
+ "activated_loras": [
29
+ ]
30
+ }
defaults/hunyuan_t2v_fast.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Hunyuan Video Text2video 720p FastHunyuan 13B",
4
+ "architecture": "hunyuan",
5
+ "description": "Fast Hunyuan is an accelerated HunyuanVideo model. It can sample high quality videos with 6 diffusion steps.",
6
+ "settings_dir": [ "" ],
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/HunyuanVideo/resolve/main/fast_hunyuan_video_720_quanto_int8.safetensors"
9
+ ],
10
+ "preload_URLs": [
11
+ "https://huggingface.co/DeepBeepMeep/HunyuanVideo/resolve/main/fast_hunyuan_video_720_quanto_int8_map.json"
12
+ ],
13
+ "auto_quantize": true
14
+ },
15
+ "negative_prompt": "",
16
+ "resolution": "832x480",
17
+ "video_length": 81,
18
+ "seed": 42,
19
+ "num_inference_steps": 6,
20
+ "flow_shift": 17,
21
+ "embedded_guidance_scale": 6,
22
+ "repeat_generation": 1,
23
+ "loras_multipliers": "",
24
+ "temporal_upsampling": "",
25
+ "spatial_upsampling": "",
26
+ "RIFLEx_setting": 0,
27
+ "slg_start_perc": 10,
28
+ "slg_end_perc": 90,
29
+ "prompt_enhancer": "",
30
+ "activated_loras": [
31
+ ]
32
+ }
defaults/i2v.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Wan2.1 Image2video 480p 14B",
5
+ "architecture" : "i2v",
6
+ "description": "The standard Wan Image 2 Video specialized to generate 480p images. It also offers Start and End Image support (End Image is not supported in the original model but seems to work well)",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_image2video_480p_14B_mbf16.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_image2video_480p_14B_quanto_mbf16_int8.safetensors",
10
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_image2video_480p_14B_quanto_mfp16_int8.safetensors"
11
+ ]
12
+ }
13
+ }
defaults/i2v_2_2.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Wan2.2 Image2video 14B",
5
+ "architecture" : "i2v_2_2",
6
+ "description": "Wan 2.2 Image 2 Video model. Contrary to the Wan Image2video 2.1 this model is structurally close to the t2v model. You will need consequently to store Loras for this model in the t2v Lora Folder.",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_image2video_14B_high_mbf16.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_image2video_14B_high_quanto_mbf16_int8.safetensors",
10
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_image2video_14B_high_quanto_mfp16_int8.safetensors"
11
+ ],
12
+ "URLs2": [
13
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_image2video_14B_low_mbf16.safetensors",
14
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_image2video_14B_low_quanto_mbf16_int8.safetensors",
15
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_image2video_14B_low_quanto_mfp16_int8.safetensors"
16
+ ],
17
+ "group": "wan2_2"
18
+ },
19
+ "guidance_phases": 2,
20
+ "switch_threshold" : 900,
21
+ "guidance_scale" : 3.5,
22
+ "guidance2_scale" : 3.5,
23
+ "flow_shift" : 5
24
+
25
+ }
defaults/i2v_2_2_multitalk.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Wan2.2 Multitalk 14B",
5
+ "architecture" : "i2v_2_2_multitalk",
6
+ "description": "The Multitalk module of Wan 2.1 has been combined with the Wan 2.2 image 2 video. It lets you have up to two people have a conversation.",
7
+ "modules": ["multitalk"],
8
+ "URLs": "i2v_2_2",
9
+ "URLs2": "i2v_2_2",
10
+ "group": "wan2_2",
11
+ "visible": false
12
+ },
13
+ "switch_threshold" : 900,
14
+ "guidance_scale" : 3.5,
15
+ "guidance2_scale" : 3.5,
16
+ "flow_shift" : 5
17
+
18
+ }
defaults/i2v_720p.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Wan2.1 Image2video 720p 14B",
5
+ "architecture" : "i2v",
6
+ "description": "The standard Wan Image 2 Video specialized to generate 720p images. It also offers Start and End Image support (End Image is not supported in the original model but seems to work well).",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_image2video_720p_14B_mbf16.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_image2video_720p_14B_quanto_mbf16_int8.safetensors",
10
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_image2video_720p_14B_quanto_mfp16_int8.safetensors"
11
+ ]
12
+ },
13
+ "resolution": "1280x720"
14
+ }
defaults/i2v_fusionix.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Wan2.1 Image2video 480p FusioniX 14B",
5
+ "architecture" : "i2v",
6
+ "description": "A powerful merged image-to-video model based on the original WAN 2.1 I2V model, enhanced using multiple open-source components and LoRAs to boost motion realism, temporal consistency, and expressive detail.",
7
+ "URLs": "i2v",
8
+ "settings_dir": [ "" ],
9
+ "loras": ["https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/loras_accelerators/Wan2.1_I2V_14B_FusionX_LoRA.safetensors"]
10
+ }
11
+ }
defaults/infinitetalk.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Infinitetalk Single Speaker 480p 14B",
4
+ "architecture": "infinitetalk",
5
+ "modules": [
6
+ [
7
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_infinitetalk_single_14B_mbf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_infinitetalk_single_14B_quanto_mbf16_int8.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_infinitetalk_single_14B_quanto_mfp16_int8.safetensors"
10
+ ]
11
+ ],
12
+ "description": "The Infinitetalk model is an improved version of Multitalk that supports very long videos. This is the single speaker version. Sliding Window size must be 81 frames to get smooth transitions between shots.",
13
+ "one_speaker_only": true,
14
+ "URLs": "i2v"
15
+ }
16
+ }
defaults/infinitetalk_multi.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Infinitetalk Multi Speakers 480p 14B",
4
+ "architecture": "infinitetalk",
5
+ "modules": [
6
+ [
7
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_infinitetalk_multi_14B_mbf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_infinitetalk_multi_14B_quanto_mfp16_int8.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_infinitetalk_multi_14B_quanto_mbf16_int8.safetensors"
10
+ ]
11
+ ],
12
+ "description": "The Infinitetalk model is an improved version of Multitalk that supports very long videos. This is the multi speakers version.Sliding Window size must be 81 frames to get smooth transitions between shots",
13
+ "multi_speakers_only": true,
14
+ "URLs": "i2v"
15
+ }
16
+ }
defaults/ltxv_13B.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "LTX Video 0.9.8 13B",
5
+ "architecture" : "ltxv_13B",
6
+ "description": "LTX Video is a fast model that can be used to generate very very long videos (up to 1800 frames !).It is recommended to keep the number of steps to 30 or you will need to update the file 'ltxv_video/configs/ltxv-13b-0.9.8-dev.yaml'.The LTX Video model expects very long prompts, so don't hesitate to use the Prompt Enhancer.",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/LTX_Video/resolve/main/ltxv_0.9.8_13B_dev_bf16.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/LTX_Video/resolve/main/ltxv_0.9.8_13B_dev_quanto_bf16_int8.safetensors"
10
+ ],
11
+ "preload_URLs" : [
12
+ "https://huggingface.co/DeepBeepMeep/LTX_Video/resolve/main/ltxv-097-ic-lora-pose-control-diffusers.safetensors",
13
+ "https://huggingface.co/DeepBeepMeep/LTX_Video/resolve/main/ltxv-097-ic-lora-depth-control-diffusers.safetensors",
14
+ "https://huggingface.co/DeepBeepMeep/LTX_Video/resolve/main/ltxv-097-ic-lora-canny-control-diffusers.safetensors"
15
+ ],
16
+ "LTXV_config": "models/ltx_video/configs/ltxv-13b-0.9.8-dev.yaml"
17
+ },
18
+ "num_inference_steps": 30
19
+ }
defaults/ltxv_distilled.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "LTX Video 0.9.8 Distilled 13B",
5
+ "architecture" : "ltxv_13B",
6
+ "description": "LTX Video is a fast model that can be used to generate very long videos (up to 1800 frames !).This distilled version is a very fast version and retains a high level of quality. The LTX Video model expects very long prompts, so don't hesitate to use the Prompt Enhancer.",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/LTX_Video/resolve/main/ltxv_0.9.8_13B_distilled_bf16.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/LTX_Video/resolve/main/ltxv_0.9.8_13B_distilled_quanto_bf16_int8.safetensors"
10
+ ],
11
+ "preload_URLs" : "ltxv_13B",
12
+ "LTXV_config": "models/ltx_video/configs/ltxv-13b-0.9.8-distilled.yaml"
13
+ },
14
+ "num_inference_steps": 6
15
+ }
defaults/lucy_edit.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Wan2.2 Lucy Edit 5B",
4
+ "architecture": "lucy_edit",
5
+ "description": "Lucy Edit is a video editing model that performs instruction-guided edits on videos using free-text prompts. It supports a variety of edits, such as clothing & accessory changes, character changes, object insertions, and scene replacements while preserving the motion and composition perfectly.",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_lucy_edit_mbf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_lucy_edit_quanto_mbf16_int8.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_lucy_edit_quanto_mfp16_int8.safetensors"
10
+ ],
11
+ "settings_dir": "ti2v_2_2",
12
+ "group": "wan2_2"
13
+ },
14
+ "prompt": "change the clothes to red",
15
+ "video_length": 81,
16
+ "guidance_scale": 5,
17
+ "flow_shift": 5,
18
+ "num_inference_steps": 30,
19
+ "resolution": "1280x720"
20
+ }
defaults/lucy_edit_fastwan.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Wan2.2 Lucy Edit FastWan 5B",
4
+ "architecture": "lucy_edit",
5
+ "description": "Lucy Edit is a video editing model that performs instruction-guided edits on videos using free-text prompts. It supports a variety of edits, such as clothing & accessory changes, character changes, object insertions, and scene replacements while preserving the motion and composition perfectly. This is the FastWan version for faster generation.",
6
+ "URLs": "lucy_edit",
7
+ "group": "wan2_2",
8
+ "settings_dir": [ "" ],
9
+ "loras": "ti2v_2_2_fastwan"
10
+ },
11
+ "prompt": "change the clothes to red",
12
+ "video_length": 81,
13
+ "guidance_scale": 1,
14
+ "flow_shift": 3,
15
+ "num_inference_steps": 5,
16
+ "resolution": "1280x720"
17
+ }
defaults/lynx.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Wan2.1 Lynx 14B",
4
+ "modules": [
5
+ [
6
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_lynx_full_module_14B_bf16.safetensors",
7
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_lynx_full_module_14B_quanto_bf16_int8.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_lynx_full_module_14B_quanto_fp16_int8.safetensors"
9
+ ]
10
+ ],
11
+ "architecture": "lynx",
12
+ "description": "The Lynx ControlNet offers State of the Art Identity Preservation. You need to provide a Reference Image which is a close up of a person face to transfer this person in the Video.",
13
+ "URLs": "t2v",
14
+ "preload_URLs": [
15
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_lynx_full_arc_resampler.safetensors"
16
+ ]
17
+ }
18
+ }
defaults/moviigen.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "MoviiGen 1080p 14B",
5
+ "architecture" : "t2v",
6
+ "description": "MoviiGen 1.1, a cutting-edge video generation model that excels in cinematic aesthetics and visual quality. Use it to generate videos in 720p or 1080p in the 21:9 ratio.",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_moviigen1.1_14B_mbf16.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_moviigen1.1_14B_quanto_mbf16_int8.safetensors",
10
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_moviigen1.1_14B_quanto_mfp16_int8.safetensors"
11
+ ],
12
+ "auto_quantize": true
13
+ },
14
+ "resolution": "1280x720",
15
+ "video_length": 81
16
+ }
defaults/multitalk.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Multitalk 480p 14B",
5
+ "architecture" : "multitalk",
6
+ "modules": [
7
+ ["https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_multitalk_14B_mbf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_multitalk_14B_quanto_mbf16_int8.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_multitalk_14B_quanto_mfp16_int8.safetensors"]
10
+ ],
11
+ "description": "The Multitalk model corresponds to the original Wan image 2 video model combined with the Multitalk module. It lets you have up to two people have a conversation.",
12
+ "URLs": "i2v",
13
+ "teacache_coefficients" : [-3.02331670e+02, 2.23948934e+02, -5.25463970e+01, 5.87348440e+00, -2.01973289e-01]
14
+ }
15
+ }
defaults/multitalk_720p.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Multitalk 720p 14B",
5
+ "architecture" : "multitalk",
6
+ "modules": ["multitalk"],
7
+ "description": "The Multitalk model corresponds to the original Wan image 2 video 720p model combined with the Multitalk module. It lets you have up to two people have a conversation.",
8
+ "URLs": "i2v_720p",
9
+ "teacache_coefficients" : [-114.36346466, 65.26524496, -18.82220707, 4.91518089, -0.23412683],
10
+ "auto_quantize": true
11
+ },
12
+ "resolution": "1280x720"
13
+ }
defaults/ovi.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Wan2.2 Ovi 10B",
4
+ "architecture": "ovi",
5
+ "description": "Ovi will generate an Audio soundtrack with the Video. It is specialized in speaking characters. Use the tags &lt;S&gt; and &lt;E&gt; to delimit the speaker words and &lt;AUDCAP&gt; and &lt;ENDAUDCAP&gt; to set the background noise.",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_ovi_video_10B_bf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_ovi_video_10B_quanto_bf16_int8.safetensors"
9
+ ],
10
+ "URLs2": [
11
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_ovi_audio_10B_bf16.safetensors",
12
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_ovi_audio_10B_quanto_bf16_int8.safetensors"
13
+ ]
14
+ },
15
+ "num_inference_steps": 30,
16
+ "prompt" : "A singer in a glittering jacket grips the microphone, sweat shining on his brow, and shouts, <S>The end is night<E>. The crowd roars in response, fists in the air. Behind him, a guitarist steps to the mic and adds to say <S>We must all find a bunker where to hide.<E>. The energy peaks as the lights flare brighter.. <AUDCAP>Electric guitar riffs, cheering crowd, shouted male voices.<ENDAUDCAP>"
17
+
18
+ }
defaults/ovi_fastwan.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Wan2.2 Ovi FastWan 10B",
4
+ "architecture": "ovi",
5
+ "description": "Ovi will generate an Audio soundtrack with the Video. It is specialized in speaking characters. Use the tags &lt;S&gt; and &lt;E&gt; to delimit the speaker words and &lt;AUDCAP&gt; and &lt;ENDAUDCAP&gt; to set the background noise. This version of Ovi is accelerated by FastWan",
6
+ "URLs": "ovi",
7
+ "URLs2": "ovi",
8
+ "loras": ["https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/loras_accelerators/Wan2_2_5B_FastWanFullAttn_lora_rank_128_bf16.safetensors"],
9
+ "loras_multipliers": [1]
10
+ },
11
+ "num_inference_steps": 6,
12
+ "prompt" : "A singer in a glittering jacket grips the microphone, sweat shining on his brow, and shouts, <S>The end is night<E>. The crowd roars in response, fists in the air. Behind him, a guitarist steps to the mic and adds to say <S>We must all find a bunker where to hide.<E>. The energy peaks as the lights flare brighter.. <AUDCAP>Electric guitar riffs, cheering crowd, shouted male voices.<ENDAUDCAP>",
13
+ "guidance_scale": 1,
14
+ "audio_guidance_scale": 1,
15
+ "flow_shift": 6,
16
+ "num_inference_steps": 6
17
+ }
defaults/phantom_1.3B.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Phantom 1.3B",
5
+ "architecture" : "phantom_1.3B",
6
+ "description": "The Phantom model is specialized in transferring people or objects of your choice into a generated Video. It produces very nice results when used at 720p.",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2_1_phantom_1.3B_mbf16.safetensors"
9
+ ]
10
+ }
11
+ }