spikedrop commited on
Commit
4cae7a7
·
verified ·
1 Parent(s): 8c74f4b

Delete ComfyUI

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. ComfyUI/custom_nodes_snapshot.yaml +0 -44
  2. ComfyUI/models/checkpoints/DreamShaper8_LCM.metadata.json +0 -18
  3. ComfyUI/models/checkpoints/DreamShaper8_LCM.safetensors +0 -3
  4. ComfyUI/models/checkpoints/Juggernaut-XI-byRunDiffusion.metadata.json +0 -18
  5. ComfyUI/models/checkpoints/endlessreality_v7.metadata.json +0 -18
  6. ComfyUI/models/checkpoints/endlessreality_v7.safetensors +0 -3
  7. ComfyUI/models/checkpoints/epiCPhotoGasmVAE.metadata.json +0 -18
  8. ComfyUI/models/checkpoints/epiCPhotoGasmVAE.safetensors +0 -3
  9. ComfyUI/models/checkpoints/ltxv-2b-0.9.6-distilled-04-25.metadata.json +0 -18
  10. ComfyUI/models/checkpoints/put_checkpoints_here +0 -0
  11. ComfyUI/models/checkpoints/realisticVisionV60B1_v60B1VAE.metadata.json +0 -18
  12. ComfyUI/models/checkpoints/realisticVisionV60B1_v60B1VAE.safetensors +0 -3
  13. ComfyUI/models/controlnet/FLUX.1-dev-ControlNet-Union-Pro-2.0/diffusion_pytorch_model.safetensors +0 -3
  14. ComfyUI/models/controlnet/FLUX.1/FLUX.1-dev-ControlNet-Union-Pro-2.0/diffusion_pytorch_model.safetensors +0 -3
  15. ComfyUI/models/controlnet/control_v11p_sd15_scribble_fp16.safetensors +0 -3
  16. ComfyUI/models/controlnet/put_controlnets_and_t2i_here +0 -0
  17. ComfyUI/models/loras/30sTechnicolorMovieV2SD1.metadata.json +0 -367
  18. ComfyUI/models/loras/30sTechnicolorMovieV2SD1.safetensors +0 -3
  19. ComfyUI/models/loras/30sTechnicolorMovieV2SD1.webp +0 -0
  20. ComfyUI/models/loras/AnimateLCM_sd15_t2v_lora.metadata.json +0 -18
  21. ComfyUI/models/loras/AnimateLCM_sd15_t2v_lora.safetensors +0 -3
  22. ComfyUI/models/loras/SDXL/LastWords_SDXL_rank16_fp16.metadata.json +0 -18
  23. ComfyUI/models/loras/SDXL/LastWords_SDXL_rank16_fp16.safetensors +0 -3
  24. ComfyUI/models/loras/SDXL/aker.metadata.json +0 -18
  25. ComfyUI/models/loras/SDXL/aker.safetensors +0 -3
  26. ComfyUI/models/loras/SDXL/boreal-v2.metadata.json +0 -18
  27. ComfyUI/models/loras/SDXL/boreal-v2.safetensors +0 -3
  28. ComfyUI/models/loras/boreal-v2.metadata.json +0 -18
  29. ComfyUI/models/loras/boreal-v2.safetensors +0 -3
  30. ComfyUI/models/loras/flux1-canny-dev-lora.metadata.json +0 -18
  31. ComfyUI/models/loras/flux1-canny-dev-lora.safetensors +0 -3
  32. ComfyUI/models/loras/flux1-depth-dev-lora.metadata.json +0 -18
  33. ComfyUI/models/loras/flux1-depth-dev-lora.safetensors +0 -3
  34. ComfyUI/models/loras/lora_manager_stats.json +0 -13
  35. ComfyUI/models/loras/put_loras_here +0 -0
  36. ComfyUI/models/loras/pytorch_lora_weights.metadata.json +0 -18
  37. ComfyUI/models/loras/pytorch_lora_weights.safetensors +0 -3
  38. ComfyUI/models/loras/studio_ghibli_wan14b_t2v_v01.metadata.json +0 -369
  39. ComfyUI/models/loras/studio_ghibli_wan14b_t2v_v01.mp4 +0 -3
  40. ComfyUI/models/loras/studio_ghibli_wan14b_t2v_v01.safetensors +0 -3
  41. ComfyUI/models/sonic/.gitattributes +0 -35
  42. ComfyUI/models/sonic/RIFE/flownet.pkl +0 -3
  43. ComfyUI/models/sonic/audio2bucket.pth +0 -3
  44. ComfyUI/models/sonic/audio2token.pth +0 -3
  45. ComfyUI/models/sonic/whisper-tiny/config.json +0 -144
  46. ComfyUI/models/sonic/whisper-tiny/model.safetensors +0 -3
  47. ComfyUI/models/sonic/whisper-tiny/preprocessor_config.json +0 -0
  48. ComfyUI/models/sonic/yoloface_v5m.pt +0 -3
  49. ComfyUI/models/upscale_models/4x_NMKD-Siax_200k.pth +0 -3
  50. ComfyUI/models/upscale_models/4x_foolhardy_Remacri.pth +0 -3
ComfyUI/custom_nodes_snapshot.yaml DELETED
@@ -1,44 +0,0 @@
1
- comfyui: f935d42d8ee399e57028d33e0142730d0c163a91
2
- git_custom_nodes:
3
- https://github.com/Fannovel16/comfyui_controlnet_aux.git:
4
- hash: 83463c2e4b04e729268e57f638b4212e0da4badc
5
- disabled: false
6
- https://github.com/city96/ComfyUI-GGUF.git:
7
- hash: 54a4854e0c006cf61494d29644ed5f4a20ad02c3
8
- disabled: false
9
- https://github.com/kaibioinfo/ComfyUI_AdvancedRefluxControl:
10
- hash: 2b95c2c866399ca1914b4da486fe52808f7a9c60
11
- disabled: false
12
- https://codeberg.org/Gourieff/comfyui-reactor-node.git:
13
- hash: 976a8c51bab55fe5443392566ab4e6a9315eed6a
14
- disabled: true
15
- https://github.com/kijai/ComfyUI-KJNodes.git:
16
- hash: c3dc82108a2a86c17094107ead61d63f8c76200e
17
- disabled: false
18
- https://github.com/willmiao/ComfyUI-Lora-Manager.git:
19
- hash: b4e22cd375ed1ed1d27c69209aca6303a873f88a
20
- disabled: false
21
- https://github.com/crystian/ComfyUI-Crystools.git:
22
- hash: 0820a7560bcc405ef6d0a7c5c53a83cc02ae7db2
23
- disabled: false
24
- https://github.com/Fannovel16/ComfyUI-Frame-Interpolation.git:
25
- hash: 7578205bd810cf96257b2e32d534b4513892af37
26
- disabled: false
27
- https://github.com/jnxmx/ComfyUI_HuggingFace_Downloader.git:
28
- hash: 6071167641e2c8bdd506774dad05d07435cb3fbd
29
- disabled: false
30
- https://github.com/Kosinkadink/ComfyUI-VideoHelperSuite.git:
31
- hash: 598e181e97d50174cc18677dfcdaa7b9b6ec0cb2
32
- disabled: false
33
- https://github.com/ltdrdata/ComfyUI-Manager.git:
34
- hash: a66bada8a34e74905530dad33aecd3b1665efe1d
35
- disabled: false
36
- cnr_custom_nodes:
37
- comfyui_ultimatesdupscale: 1.1.2
38
- comfyui-propost: 1.1.3
39
- comfyui_ipadapter_plus: 2.0.0
40
- comfyui_pulid_flux_ll: 1.1.4
41
- file_custom_nodes:
42
- - filename: websocket_image_save.py
43
- disabled: false
44
- pips: {}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/checkpoints/DreamShaper8_LCM.metadata.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "file_name": "DreamShaper8_LCM",
3
- "model_name": "DreamShaper8_LCM",
4
- "file_path": "/workspace/ComfyUI/models/checkpoints/DreamShaper8_LCM.safetensors",
5
- "size": 2133804992,
6
- "modified": 1745264041.924619,
7
- "sha256": "a4f3e1526c5dc4fcbe342f5c410d83ae202c7a415fcefcbb92e0f93fcd0a87c3",
8
- "base_model": "Unknown",
9
- "preview_url": "",
10
- "preview_nsfw_level": 0,
11
- "notes": "",
12
- "from_civitai": true,
13
- "civitai": null,
14
- "tags": [],
15
- "modelDescription": "",
16
- "civitai_deleted": false,
17
- "model_type": "checkpoint"
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/checkpoints/DreamShaper8_LCM.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a4f3e1526c5dc4fcbe342f5c410d83ae202c7a415fcefcbb92e0f93fcd0a87c3
3
- size 2133804992
 
 
 
 
ComfyUI/models/checkpoints/Juggernaut-XI-byRunDiffusion.metadata.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "file_name": "Juggernaut-XI-byRunDiffusion",
3
- "model_name": "Juggernaut-XI-byRunDiffusion",
4
- "file_path": "/workspace/ComfyUI/models/checkpoints/Juggernaut-XI-byRunDiffusion.safetensors",
5
- "size": 7105350536,
6
- "modified": 1745540492.4777641,
7
- "sha256": "33e58e86686f6b386c526682b5da9228ead4f91d994abd4b053442dc5b42719e",
8
- "base_model": "Unknown",
9
- "preview_url": "",
10
- "preview_nsfw_level": 0,
11
- "notes": "",
12
- "from_civitai": true,
13
- "civitai": null,
14
- "tags": [],
15
- "modelDescription": "",
16
- "civitai_deleted": false,
17
- "model_type": "checkpoint"
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/checkpoints/endlessreality_v7.metadata.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "file_name": "endlessreality_v7",
3
- "model_name": "endlessreality_v7",
4
- "file_path": "/workspace/ComfyUI/models/checkpoints/endlessreality_v7.safetensors",
5
- "size": 2132625894,
6
- "modified": 1745259190.8363886,
7
- "sha256": "89efaf43e5ce529a761a149883b5a22c409d9ed6dbf5c4d22a457c0a69fae2b0",
8
- "base_model": "Unknown",
9
- "preview_url": "",
10
- "preview_nsfw_level": 0,
11
- "notes": "",
12
- "from_civitai": true,
13
- "civitai": null,
14
- "tags": [],
15
- "modelDescription": "",
16
- "civitai_deleted": false,
17
- "model_type": "checkpoint"
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/checkpoints/endlessreality_v7.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:89efaf43e5ce529a761a149883b5a22c409d9ed6dbf5c4d22a457c0a69fae2b0
3
- size 2132625894
 
 
 
 
ComfyUI/models/checkpoints/epiCPhotoGasmVAE.metadata.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "file_name": "epiCPhotoGasmVAE",
3
- "model_name": "epiCPhotoGasmVAE",
4
- "file_path": "/workspace/ComfyUI/models/checkpoints/epiCPhotoGasmVAE.safetensors",
5
- "size": 2299934092,
6
- "modified": 1745358275.775316,
7
- "sha256": "2e5cddfb80f48396b5cac32864b511c25cdf5be5752b7300448838e2121faca0",
8
- "base_model": "Unknown",
9
- "preview_url": "",
10
- "preview_nsfw_level": 0,
11
- "notes": "",
12
- "from_civitai": true,
13
- "civitai": null,
14
- "tags": [],
15
- "modelDescription": "",
16
- "civitai_deleted": false,
17
- "model_type": "checkpoint"
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/checkpoints/epiCPhotoGasmVAE.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2e5cddfb80f48396b5cac32864b511c25cdf5be5752b7300448838e2121faca0
3
- size 2299934092
 
 
 
 
ComfyUI/models/checkpoints/ltxv-2b-0.9.6-distilled-04-25.metadata.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "file_name": "ltxv-2b-0.9.6-distilled-04-25",
3
- "model_name": "ltxv-2b-0.9.6-distilled-04-25",
4
- "file_path": "/workspace/ComfyUI/models/checkpoints/ltxv-2b-0.9.6-distilled-04-25.safetensors",
5
- "size": 6340744028,
6
- "modified": 1745540968.2027185,
7
- "sha256": "94891bd4bd08de30d484befbfc54fdcffe6d1596a131baad700b9baa5e1de86b",
8
- "base_model": "Unknown",
9
- "preview_url": "",
10
- "preview_nsfw_level": 0,
11
- "notes": "",
12
- "from_civitai": true,
13
- "civitai": null,
14
- "tags": [],
15
- "modelDescription": "",
16
- "civitai_deleted": false,
17
- "model_type": "checkpoint"
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/checkpoints/put_checkpoints_here DELETED
File without changes
ComfyUI/models/checkpoints/realisticVisionV60B1_v60B1VAE.metadata.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "file_name": "realisticVisionV60B1_v60B1VAE",
3
- "model_name": "realisticVisionV60B1_v60B1VAE",
4
- "file_path": "/workspace/ComfyUI/models/checkpoints/realisticVisionV60B1_v60B1VAE.safetensors",
5
- "size": 2132625894,
6
- "modified": 1745323771.7214153,
7
- "sha256": "fe7578cb5ee0be63aa15baa894ab5d1751ff9b5b25ef611d5fafb2186d930c30",
8
- "base_model": "Unknown",
9
- "preview_url": "",
10
- "preview_nsfw_level": 0,
11
- "notes": "",
12
- "from_civitai": true,
13
- "civitai": null,
14
- "tags": [],
15
- "modelDescription": "",
16
- "civitai_deleted": false,
17
- "model_type": "checkpoint"
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/checkpoints/realisticVisionV60B1_v60B1VAE.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe7578cb5ee0be63aa15baa894ab5d1751ff9b5b25ef611d5fafb2186d930c30
3
- size 2132625894
 
 
 
 
ComfyUI/models/controlnet/FLUX.1-dev-ControlNet-Union-Pro-2.0/diffusion_pytorch_model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9d03f63f36206bab2f36aed5cfedc8693c2881397534e9d5f9ae9a0a41362517
3
- size 4281779224
 
 
 
 
ComfyUI/models/controlnet/FLUX.1/FLUX.1-dev-ControlNet-Union-Pro-2.0/diffusion_pytorch_model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9d03f63f36206bab2f36aed5cfedc8693c2881397534e9d5f9ae9a0a41362517
3
- size 4281779224
 
 
 
 
ComfyUI/models/controlnet/control_v11p_sd15_scribble_fp16.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:99edfd25b54c18c0ab19fba8c5618f741aac1f8c3101e7fa62cce925ad87ae68
3
- size 722601100
 
 
 
 
ComfyUI/models/controlnet/put_controlnets_and_t2i_here DELETED
File without changes
ComfyUI/models/loras/30sTechnicolorMovieV2SD1.metadata.json DELETED
@@ -1,367 +0,0 @@
1
- {
2
- "file_name": "30sTechnicolorMovieV2SD1",
3
- "model_name": "30s Technicolor Movie (SD1, SDXL, Pony, Flux)",
4
- "file_path": "/workspace/ComfyUI/models/loras/30sTechnicolorMovieV2SD1.safetensors",
5
- "size": 38035920,
6
- "modified": 1745273225.1968696,
7
- "sha256": "b534b340b9726444539367134c9235fb2ed181a4e99eac07ef1e231868ea2c10",
8
- "base_model": "SD 1.5",
9
- "preview_url": "/workspace/ComfyUI/models/loras/30sTechnicolorMovieV2SD1.webp",
10
- "preview_nsfw_level": 1,
11
- "notes": "",
12
- "from_civitai": true,
13
- "civitai": {
14
- "id": 1128217,
15
- "modelId": 886686,
16
- "name": "SD1.5V2",
17
- "createdAt": "2024-12-04T10:04:03.952Z",
18
- "updatedAt": "2024-12-07T07:57:46.031Z",
19
- "status": "Published",
20
- "publishedAt": "2024-12-08T18:54:41.751Z",
21
- "trainedWords": [
22
- "ArsMovieStill, movie still from a 1930s technicolor movie"
23
- ],
24
- "trainingStatus": null,
25
- "trainingDetails": null,
26
- "baseModel": "SD 1.5",
27
- "baseModelType": null,
28
- "earlyAccessEndsAt": null,
29
- "earlyAccessConfig": {
30
- "timeframe": 0,
31
- "donationGoal": 1000,
32
- "downloadPrice": 100,
33
- "donationGoalId": 4365,
34
- "generationPrice": 50,
35
- "chargeForDownload": true,
36
- "originalTimeframe": "12",
37
- "chargeForGeneration": true,
38
- "donationGoalEnabled": true,
39
- "originalPublishedAt": "2024-12-07T07:57:46.023",
40
- "generationTrialLimit": 10
41
- },
42
- "description": null,
43
- "uploadType": "Created",
44
- "usageControl": "Download",
45
- "air": "urn:air:sd1:lora:civitai:886686@1128217",
46
- "stats": {
47
- "downloadCount": 337,
48
- "ratingCount": 0,
49
- "rating": 0,
50
- "thumbsUpCount": 31
51
- },
52
- "model": {
53
- "name": "30s Technicolor Movie (SD1, SDXL, Pony, Flux)",
54
- "type": "LORA",
55
- "nsfw": false,
56
- "poi": false
57
- },
58
- "files": [
59
- {
60
- "id": 1032875,
61
- "sizeKB": 37144.453125,
62
- "name": "30sTechnicolorMovieV2SD1.safetensors",
63
- "type": "Model",
64
- "pickleScanResult": "Success",
65
- "pickleScanMessage": "No Pickle imports",
66
- "virusScanResult": "Success",
67
- "virusScanMessage": null,
68
- "scannedAt": "2024-12-04T10:06:07.832Z",
69
- "metadata": {
70
- "format": "SafeTensor",
71
- "size": null,
72
- "fp": null
73
- },
74
- "hashes": {
75
- "AutoV1": "F65DADDA",
76
- "AutoV2": "B534B340B9",
77
- "SHA256": "B534B340B9726444539367134C9235FB2ED181A4E99EAC07EF1E231868EA2C10",
78
- "CRC32": "70DDB1AF",
79
- "BLAKE3": "344EC4C7AFE970EA9E7AB750FC4994518C653DA3D9D769923974090DD061066D",
80
- "AutoV3": "68C6D3B142BD"
81
- },
82
- "primary": true,
83
- "downloadUrl": "https://civitai.com/api/download/models/1128217"
84
- }
85
- ],
86
- "images": [
87
- {
88
- "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/392c4f78-f3c4-4356-a569-ac9042e50cba/width=1024/43619891.jpeg",
89
- "nsfwLevel": 1,
90
- "width": 1024,
91
- "height": 1024,
92
- "hash": "U55}~v.QK5yW*JoeaJyCS~MdVEt6?ZaLMeoI",
93
- "type": "image",
94
- "metadata": {
95
- "hash": "U55}~v.QK5yW*JoeaJyCS~MdVEt6?ZaLMeoI",
96
- "size": 1354543,
97
- "width": 1024,
98
- "height": 1024
99
- },
100
- "meta": {
101
- "Size": "512x512",
102
- "seed": 2872236929,
103
- "Model": "epicrealism_pureEvolutionV5",
104
- "steps": 40,
105
- "hashes": {
106
- "model": "76be5be1b2",
107
- "lora:30sTechnicolorMovieV2SD1": "68c6d3b142bd",
108
- "embed:verybadimagenegative_v1.3": "d70463f870",
109
- "embed:epiCPhotoGasm-softPhoto-neg": "68452a2cab",
110
- "embed:epiCPhotoGasm-colorfulPhoto-neg": "cea9c24d4c"
111
- },
112
- "prompt": "<lora:30sTechnicolorMovieV2SD1:1> ArsMovieStill, movie still from a 1930s technicolor movie, The image shows a woman in a black dress with a necklace and earrings standing in front of a crescent moon., 1girl, jewelry, closed eyes, solo, necklace, moon, black hair, makeup, earrings, lipstick",
113
- "Version": "v1.9.4",
114
- "sampler": "DPM++ 2M",
115
- "cfgScale": 7,
116
- "TI hashes": {
117
- "verybadimagenegative_v1.3": "d70463f87042",
118
- "epiCPhotoGasm-softPhoto-neg": "68452a2cab70",
119
- "epiCPhotoGasm-colorfulPhoto-neg": "cea9c24d4cbb"
120
- },
121
- "resources": [
122
- {
123
- "hash": "68c6d3b142bd",
124
- "name": "30sTechnicolorMovieV2SD1",
125
- "type": "lora",
126
- "weight": 1
127
- },
128
- {
129
- "hash": "76be5be1b2",
130
- "name": "epicrealism_pureEvolutionV5",
131
- "type": "model"
132
- }
133
- ],
134
- "Model hash": "76be5be1b2",
135
- "Hires upscale": "2",
136
- "Schedule type": "Karras",
137
- "Hires upscaler": "4x-UltraSharp",
138
- "negativePrompt": "cartoon, CGI , render, illustration, (painting) , drawing, epiCPhotoGasm-colorfulPhoto-neg, epiCPhotoGasm-softPhoto-neg, verybadimagenegative_v1.3,",
139
- "ADetailer model": "face_yolov8n.pt",
140
- "ADetailer version": "24.11.1",
141
- "Denoising strength": "0.7",
142
- "ADetailer mask blur": "4",
143
- "ADetailer confidence": "0.3",
144
- "ADetailer dilate erode": "4",
145
- "Downcast alphas_cumprod": "True",
146
- "ADetailer inpaint padding": "32",
147
- "ADetailer denoising strength": "0.4",
148
- "ADetailer inpaint only masked": "True"
149
- },
150
- "availability": "Public",
151
- "hasMeta": true,
152
- "hasPositivePrompt": true,
153
- "onSite": false,
154
- "remixOfId": null
155
- },
156
- {
157
- "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/5ff81773-e076-4ec1-ab27-12f93c559a93/width=1024/43619897.jpeg",
158
- "nsfwLevel": 1,
159
- "width": 1024,
160
- "height": 1024,
161
- "hash": "U86+e8.lIoV[*J%fxtX8OExY-mbHw]j]oysl",
162
- "type": "image",
163
- "metadata": {
164
- "hash": "U86+e8.lIoV[*J%fxtX8OExY-mbHw]j]oysl",
165
- "size": 1627396,
166
- "width": 1024,
167
- "height": 1024
168
- },
169
- "meta": {
170
- "Size": "512x512",
171
- "seed": 2614348895,
172
- "Model": "epicrealism_pureEvolutionV5",
173
- "steps": 40,
174
- "hashes": {
175
- "model": "76be5be1b2",
176
- "lora:30sTechnicolorMovieV2SD1": "68c6d3b142bd",
177
- "embed:verybadimagenegative_v1.3": "d70463f870",
178
- "embed:epiCPhotoGasm-softPhoto-neg": "68452a2cab",
179
- "embed:epiCPhotoGasm-colorfulPhoto-neg": "cea9c24d4c"
180
- },
181
- "prompt": "<lora:30sTechnicolorMovieV2SD1:1> ArsMovieStill, movie still from a 1930s technicolor movie, The image shows a fawn standing on top of a rock in the woods surrounded by trees and illuminated by a full moon in the night sky., no humans, night, moon, star (sky), sky, outdoors, tree, starry sky, night sky, nature",
182
- "Version": "v1.9.4",
183
- "sampler": "DPM++ 2M",
184
- "cfgScale": 7,
185
- "resources": [
186
- {
187
- "hash": "68c6d3b142bd",
188
- "name": "30sTechnicolorMovieV2SD1",
189
- "type": "lora",
190
- "weight": 1
191
- },
192
- {
193
- "hash": "76be5be1b2",
194
- "name": "epicrealism_pureEvolutionV5",
195
- "type": "model"
196
- }
197
- ],
198
- "Model hash": "76be5be1b2",
199
- "Hires upscale": "2",
200
- "Schedule type": "Karras",
201
- "Hires upscaler": "4x-UltraSharp",
202
- "negativePrompt": "cartoon, CGI , render, illustration, (painting) , drawing, epiCPhotoGasm-colorfulPhoto-neg, epiCPhotoGasm-softPhoto-neg, verybadimagenegative_v1.3,",
203
- "ADetailer model": "face_yolov8n.pt",
204
- "ADetailer version": "24.11.1",
205
- "Denoising strength": "0.7",
206
- "ADetailer mask blur": "4",
207
- "ADetailer confidence": "0.3",
208
- "ADetailer dilate erode": "4",
209
- "Downcast alphas_cumprod": "True",
210
- "ADetailer inpaint padding": "32",
211
- "ADetailer denoising strength": "0.4",
212
- "ADetailer inpaint only masked": "True"
213
- },
214
- "availability": "Public",
215
- "hasMeta": true,
216
- "hasPositivePrompt": true,
217
- "onSite": false,
218
- "remixOfId": null
219
- },
220
- {
221
- "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e74ee3e9-f7bc-4a71-bb5e-11ccdb41d9f8/width=1024/43619892.jpeg",
222
- "nsfwLevel": 1,
223
- "width": 1024,
224
- "height": 1024,
225
- "hash": "UhD_84xvxaR*.TxWt7W=tSs:axR*R%oMRjba",
226
- "type": "image",
227
- "metadata": {
228
- "hash": "UhD_84xvxaR*.TxWt7W=tSs:axR*R%oMRjba",
229
- "size": 1849131,
230
- "width": 1024,
231
- "height": 1024
232
- },
233
- "meta": {
234
- "Size": "512x512",
235
- "seed": 388896898,
236
- "Model": "epicrealism_pureEvolutionV5",
237
- "steps": 40,
238
- "hashes": {
239
- "model": "76be5be1b2",
240
- "lora:30sTechnicolorMovieV2SD1": "68c6d3b142bd",
241
- "embed:verybadimagenegative_v1.3": "d70463f870",
242
- "embed:epiCPhotoGasm-softPhoto-neg": "68452a2cab",
243
- "embed:epiCPhotoGasm-colorfulPhoto-neg": "cea9c24d4c"
244
- },
245
- "prompt": "<lora:30sTechnicolorMovieV2SD1:1> ArsMovieStill, movie still from a 1930s technicolor movie, The image shows a painting of a castle perched atop a hill surrounded by a river a bridge a group of trees plants grass the hills and a cloudy sky., no humans, tree, scenery, outdoors, sky, day, castle, water, blue sky, grass",
246
- "Version": "v1.9.4",
247
- "sampler": "DPM++ 2M",
248
- "cfgScale": 7,
249
- "resources": [
250
- {
251
- "hash": "68c6d3b142bd",
252
- "name": "30sTechnicolorMovieV2SD1",
253
- "type": "lora",
254
- "weight": 1
255
- },
256
- {
257
- "hash": "76be5be1b2",
258
- "name": "epicrealism_pureEvolutionV5",
259
- "type": "model"
260
- }
261
- ],
262
- "Model hash": "76be5be1b2",
263
- "Hires upscale": "2",
264
- "Schedule type": "Karras",
265
- "Hires upscaler": "4x-UltraSharp",
266
- "negativePrompt": "cartoon, CGI , render, illustration, (painting) , drawing, epiCPhotoGasm-colorfulPhoto-neg, epiCPhotoGasm-softPhoto-neg, verybadimagenegative_v1.3,",
267
- "ADetailer model": "face_yolov8n.pt",
268
- "ADetailer version": "24.11.1",
269
- "Denoising strength": "0.7",
270
- "ADetailer mask blur": "4",
271
- "ADetailer confidence": "0.3",
272
- "ADetailer dilate erode": "4",
273
- "Downcast alphas_cumprod": "True",
274
- "ADetailer inpaint padding": "32",
275
- "ADetailer denoising strength": "0.4",
276
- "ADetailer inpaint only masked": "True"
277
- },
278
- "availability": "Public",
279
- "hasMeta": true,
280
- "hasPositivePrompt": true,
281
- "onSite": false,
282
- "remixOfId": null
283
- },
284
- {
285
- "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/4e16fa9c-4b82-479b-94e1-33137a5ee230/width=1024/43619894.jpeg",
286
- "nsfwLevel": 1,
287
- "width": 1024,
288
- "height": 1024,
289
- "hash": "U8BCDZ*02^XSHW.8%ho|PB%g~VtR-:%Mx]oy",
290
- "type": "image",
291
- "metadata": {
292
- "hash": "U8BCDZ*02^XSHW.8%ho|PB%g~VtR-:%Mx]oy",
293
- "size": 1461354,
294
- "width": 1024,
295
- "height": 1024
296
- },
297
- "meta": {
298
- "Size": "512x512",
299
- "seed": 4111700315,
300
- "Model": "epicrealism_pureEvolutionV5",
301
- "steps": 40,
302
- "hashes": {
303
- "model": "76be5be1b2",
304
- "lora:30sTechnicolorMovieV2SD1": "68c6d3b142bd",
305
- "embed:verybadimagenegative_v1.3": "d70463f870",
306
- "embed:epiCPhotoGasm-softPhoto-neg": "68452a2cab",
307
- "embed:epiCPhotoGasm-colorfulPhoto-neg": "cea9c24d4c"
308
- },
309
- "prompt": "<lora:30sTechnicolorMovieV2SD1:1> ArsMovieStill, movie still from a 1930s technicolor movie, The image shows a woman wearing a red dress and a red hooded cloak with a blurred background. Her face is clearly visible with her eyes looking directly at the viewer. Her hair is pulled back in a neat bun and her lips are painted a deep red. She has a mysterious and captivating expression on her face., 1girl, solo, red lips, blue eyes, makeup, black hair, looking at viewer, lipstick, bangs, portrait",
310
- "Version": "v1.9.4",
311
- "sampler": "DPM++ 2M",
312
- "cfgScale": 7,
313
- "TI hashes": {
314
- "verybadimagenegative_v1.3": "d70463f87042",
315
- "epiCPhotoGasm-softPhoto-neg": "68452a2cab70",
316
- "epiCPhotoGasm-colorfulPhoto-neg": "cea9c24d4cbb"
317
- },
318
- "resources": [
319
- {
320
- "hash": "68c6d3b142bd",
321
- "name": "30sTechnicolorMovieV2SD1",
322
- "type": "lora",
323
- "weight": 1
324
- },
325
- {
326
- "hash": "76be5be1b2",
327
- "name": "epicrealism_pureEvolutionV5",
328
- "type": "model"
329
- }
330
- ],
331
- "Model hash": "76be5be1b2",
332
- "Hires upscale": "2",
333
- "Schedule type": "Karras",
334
- "Hires upscaler": "4x-UltraSharp",
335
- "negativePrompt": "cartoon, CGI , render, illustration, (painting) , drawing, epiCPhotoGasm-colorfulPhoto-neg, epiCPhotoGasm-softPhoto-neg, verybadimagenegative_v1.3,",
336
- "ADetailer model": "face_yolov8n.pt",
337
- "ADetailer version": "24.11.1",
338
- "Denoising strength": "0.7",
339
- "ADetailer mask blur": "4",
340
- "ADetailer confidence": "0.3",
341
- "ADetailer dilate erode": "4",
342
- "Downcast alphas_cumprod": "True",
343
- "ADetailer inpaint padding": "32",
344
- "ADetailer denoising strength": "0.4",
345
- "ADetailer inpaint only masked": "True"
346
- },
347
- "availability": "Public",
348
- "hasMeta": true,
349
- "hasPositivePrompt": true,
350
- "onSite": false,
351
- "remixOfId": null
352
- }
353
- ],
354
- "downloadUrl": "https://civitai.com/api/download/models/1128217"
355
- },
356
- "tags": [
357
- "movie",
358
- "midjourney",
359
- "style",
360
- "still",
361
- "movie still",
362
- "midjourney v6"
363
- ],
364
- "modelDescription": "<p>This LoRA aims to capture the look and feel of the early <strong>technicolor movies</strong> in the <strong>30s</strong></p><p>The trigger words are <strong>ArsMovieStill, movie still from a 1930s technicolor movie</strong></p><p></p><p><strong>30s Technicolor V2 is here:</strong></p><ul><li><p>Dataset expanded by 400% (150 → 600 images) for greater diversity.</p></li><li><p>Sub-themes include all major movie genres of the 1930s.</p></li><li><p>Folklore, Fairytales, and various fantasy themes added on top.</p></li><li><p>Enhanced background variety while maintaining a strong focus on characters.<br /><br /></p></li></ul>",
365
- "civitai_deleted": false,
366
- "usage_tips": "{}"
367
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/loras/30sTechnicolorMovieV2SD1.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b534b340b9726444539367134c9235fb2ed181a4e99eac07ef1e231868ea2c10
3
- size 38035920
 
 
 
 
ComfyUI/models/loras/30sTechnicolorMovieV2SD1.webp DELETED
Binary file (15.4 kB)
 
ComfyUI/models/loras/AnimateLCM_sd15_t2v_lora.metadata.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "file_name": "AnimateLCM_sd15_t2v_lora",
3
- "model_name": "AnimateLCM_sd15_t2v_lora",
4
- "file_path": "/workspace/ComfyUI/models/loras/AnimateLCM_sd15_t2v_lora.safetensors",
5
- "size": 134621556,
6
- "modified": 1745358251.8548393,
7
- "sha256": "8f90d840e075ff588a58e22c6586e2ae9a6f7922996ee6649a7f01072333afe4",
8
- "base_model": "Unknown",
9
- "preview_url": "",
10
- "preview_nsfw_level": 0,
11
- "notes": "",
12
- "from_civitai": true,
13
- "civitai": null,
14
- "tags": [],
15
- "modelDescription": "",
16
- "civitai_deleted": false,
17
- "usage_tips": "{}"
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/loras/AnimateLCM_sd15_t2v_lora.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f90d840e075ff588a58e22c6586e2ae9a6f7922996ee6649a7f01072333afe4
3
- size 134621556
 
 
 
 
ComfyUI/models/loras/SDXL/LastWords_SDXL_rank16_fp16.metadata.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "file_name": "LastWords_SDXL_rank16_fp16",
3
- "model_name": "LastWords_SDXL_rank16_fp16",
4
- "file_path": "/workspace/ComfyUI/models/loras/SDXL/LastWords_SDXL_rank16_fp16.safetensors",
5
- "size": 85459428,
6
- "modified": 1745538742.734948,
7
- "sha256": "800836f926b7ba65c8263e9c8f3243c42d62db2e07b71def79a20d9be1424972",
8
- "base_model": "SDXL 1.0",
9
- "preview_url": "",
10
- "preview_nsfw_level": 0,
11
- "notes": "",
12
- "from_civitai": true,
13
- "civitai": null,
14
- "tags": [],
15
- "modelDescription": "",
16
- "civitai_deleted": false,
17
- "usage_tips": "{}"
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/loras/SDXL/LastWords_SDXL_rank16_fp16.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:800836f926b7ba65c8263e9c8f3243c42d62db2e07b71def79a20d9be1424972
3
- size 85459428
 
 
 
 
ComfyUI/models/loras/SDXL/aker.metadata.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "file_name": "aker",
3
- "model_name": "aker",
4
- "file_path": "/workspace/ComfyUI/models/loras/SDXL/aker.safetensors",
5
- "size": 21591876,
6
- "modified": 1745538742.978951,
7
- "sha256": "f4da28a2dd547c27e63006e2b203ceddb573c1f82365e33196b7f353f6cf8de1",
8
- "base_model": "SDXL 1.0",
9
- "preview_url": "",
10
- "preview_nsfw_level": 0,
11
- "notes": "",
12
- "from_civitai": true,
13
- "civitai": null,
14
- "tags": [],
15
- "modelDescription": "",
16
- "civitai_deleted": false,
17
- "usage_tips": "{}"
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/loras/SDXL/aker.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f4da28a2dd547c27e63006e2b203ceddb573c1f82365e33196b7f353f6cf8de1
3
- size 21591876
 
 
 
 
ComfyUI/models/loras/SDXL/boreal-v2.metadata.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "file_name": "boreal-v2",
3
- "model_name": "boreal-v2",
4
- "file_path": "/workspace/ComfyUI/models/loras/SDXL/boreal-v2.safetensors",
5
- "size": 171969478,
6
- "modified": 1745538744.1529658,
7
- "sha256": "566c00ed0a38aa80911338c1f8b48b1843faa7be1e98a14383ef5317d1f4fa94",
8
- "base_model": "SD 1.5",
9
- "preview_url": "",
10
- "preview_nsfw_level": 0,
11
- "notes": "",
12
- "from_civitai": true,
13
- "civitai": null,
14
- "tags": [],
15
- "modelDescription": "",
16
- "civitai_deleted": false,
17
- "usage_tips": "{}"
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/loras/SDXL/boreal-v2.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:566c00ed0a38aa80911338c1f8b48b1843faa7be1e98a14383ef5317d1f4fa94
3
- size 171969478
 
 
 
 
ComfyUI/models/loras/boreal-v2.metadata.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "file_name": "boreal-v2",
3
- "model_name": "boreal-v2",
4
- "file_path": "/workspace/ComfyUI/models/loras/boreal-v2.safetensors",
5
- "size": 171969478,
6
- "modified": 1745538745.1099777,
7
- "sha256": "566c00ed0a38aa80911338c1f8b48b1843faa7be1e98a14383ef5317d1f4fa94",
8
- "base_model": "SD 1.5",
9
- "preview_url": "",
10
- "preview_nsfw_level": 0,
11
- "notes": "",
12
- "from_civitai": true,
13
- "civitai": null,
14
- "tags": [],
15
- "modelDescription": "",
16
- "civitai_deleted": false,
17
- "usage_tips": "{}"
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/loras/boreal-v2.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:566c00ed0a38aa80911338c1f8b48b1843faa7be1e98a14383ef5317d1f4fa94
3
- size 171969478
 
 
 
 
ComfyUI/models/loras/flux1-canny-dev-lora.metadata.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "file_name": "flux1-canny-dev-lora",
3
- "model_name": "flux1-canny-dev-lora",
4
- "file_path": "/workspace/ComfyUI/models/loras/flux1-canny-dev-lora.safetensors",
5
- "size": 1244443944,
6
- "modified": 1745577033.5130565,
7
- "sha256": "8eaa21b9c43d5e7242844deb64b8cf22ae9010f813f955ca8c05f240b8a98f7e",
8
- "base_model": "Unknown",
9
- "preview_url": "",
10
- "preview_nsfw_level": 0,
11
- "notes": "",
12
- "from_civitai": true,
13
- "civitai": null,
14
- "tags": [],
15
- "modelDescription": "",
16
- "civitai_deleted": false,
17
- "usage_tips": "{}"
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/loras/flux1-canny-dev-lora.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8eaa21b9c43d5e7242844deb64b8cf22ae9010f813f955ca8c05f240b8a98f7e
3
- size 1244443944
 
 
 
 
ComfyUI/models/loras/flux1-depth-dev-lora.metadata.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "file_name": "flux1-depth-dev-lora",
3
- "model_name": "flux1-depth-dev-lora",
4
- "file_path": "/workspace/ComfyUI/models/loras/flux1-depth-dev-lora.safetensors",
5
- "size": 1244440512,
6
- "modified": 1745249510.1192756,
7
- "sha256": "1938b38ea0fdd98080fa3e48beb2bedfbc7ad102d8b65e6614de704a46d8b907",
8
- "base_model": "Unknown",
9
- "preview_url": "",
10
- "preview_nsfw_level": 0,
11
- "notes": "",
12
- "from_civitai": true,
13
- "civitai": null,
14
- "tags": [],
15
- "modelDescription": "",
16
- "civitai_deleted": false,
17
- "usage_tips": "{}"
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/loras/flux1-depth-dev-lora.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1938b38ea0fdd98080fa3e48beb2bedfbc7ad102d8b65e6614de704a46d8b907
3
- size 1244440512
 
 
 
 
ComfyUI/models/loras/lora_manager_stats.json DELETED
@@ -1,13 +0,0 @@
1
- {
2
- "checkpoints": {
3
- "a4f3e1526c5dc4fcbe342f5c410d83ae202c7a415fcefcbb92e0f93fcd0a87c3": 2,
4
- "89efaf43e5ce529a761a149883b5a22c409d9ed6dbf5c4d22a457c0a69fae2b0": 1,
5
- "1be961341be8f5307ef26c787199f80bf4e0de3c1c0b4617095aa6ee5550dfce": 87
6
- },
7
- "loras": {
8
- "800836f926b7ba65c8263e9c8f3243c42d62db2e07b71def79a20d9be1424972": 13,
9
- "f4da28a2dd547c27e63006e2b203ceddb573c1f82365e33196b7f353f6cf8de1": 12
10
- },
11
- "total_executions": 157,
12
- "last_save_time": 1745592184.0134323
13
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/loras/put_loras_here DELETED
File without changes
ComfyUI/models/loras/pytorch_lora_weights.metadata.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "file_name": "pytorch_lora_weights",
3
- "model_name": "pytorch_lora_weights",
4
- "file_path": "/workspace/ComfyUI/models/loras/pytorch_lora_weights.safetensors",
5
- "size": 134621556,
6
- "modified": 1745260658.767056,
7
- "sha256": "8f90d840e075ff588a58e22c6586e2ae9a6f7922996ee6649a7f01072333afe4",
8
- "base_model": "Unknown",
9
- "preview_url": "",
10
- "preview_nsfw_level": 0,
11
- "notes": "",
12
- "from_civitai": true,
13
- "civitai": null,
14
- "tags": [],
15
- "modelDescription": "",
16
- "civitai_deleted": false,
17
- "usage_tips": "{}"
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/loras/pytorch_lora_weights.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f90d840e075ff588a58e22c6586e2ae9a6f7922996ee6649a7f01072333afe4
3
- size 134621556
 
 
 
 
ComfyUI/models/loras/studio_ghibli_wan14b_t2v_v01.metadata.json DELETED
@@ -1,369 +0,0 @@
1
- {
2
- "file_name": "studio_ghibli_wan14b_t2v_v01",
3
- "model_name": "Studio Ghibli 🎥 Wan2.1-T2V-14B",
4
- "file_path": "/workspace/ComfyUI/models/loras/studio_ghibli_wan14b_t2v_v01.safetensors",
5
- "size": 306849176,
6
- "modified": 1745253935.7079911,
7
- "sha256": "dd2fe1258db4288133f866103638f8a635cee065e50aef818b0f0569335e9eab",
8
- "base_model": "Wan Video",
9
- "preview_url": "/workspace/ComfyUI/models/loras/studio_ghibli_wan14b_t2v_v01.mp4",
10
- "preview_nsfw_level": 1,
11
- "notes": "",
12
- "from_civitai": true,
13
- "civitai": {
14
- "id": 1587891,
15
- "modelId": 1404755,
16
- "name": "v1.0",
17
- "createdAt": "2025-03-27T17:34:08.255Z",
18
- "updatedAt": "2025-04-16T15:44:42.855Z",
19
- "status": "Published",
20
- "publishedAt": "2025-03-28T08:47:19.427Z",
21
- "trainedWords": [
22
- "Studio Ghibli style"
23
- ],
24
- "trainingStatus": null,
25
- "trainingDetails": null,
26
- "baseModel": "Wan Video",
27
- "baseModelType": null,
28
- "earlyAccessEndsAt": null,
29
- "earlyAccessConfig": null,
30
- "description": null,
31
- "uploadType": "Created",
32
- "usageControl": "Download",
33
- "air": "urn:air:wanvideo:lora:civitai:1404755@1587891",
34
- "stats": {
35
- "downloadCount": 3351,
36
- "ratingCount": 0,
37
- "rating": 0,
38
- "thumbsUpCount": 414
39
- },
40
- "model": {
41
- "name": "Studio Ghibli 🎥 Wan2.1-T2V-14B",
42
- "type": "LORA",
43
- "nsfw": false,
44
- "poi": false
45
- },
46
- "files": [
47
- {
48
- "id": 1496871,
49
- "sizeKB": 293425.7109375,
50
- "name": "studio_ghibli_wan14b_t2v_v01_dataset.zip",
51
- "type": "Training Data",
52
- "pickleScanResult": "Success",
53
- "pickleScanMessage": "No Pickle imports",
54
- "virusScanResult": "Success",
55
- "virusScanMessage": null,
56
- "scannedAt": "2025-03-29T21:36:03.133Z",
57
- "metadata": {
58
- "format": "Other",
59
- "size": null,
60
- "fp": null
61
- },
62
- "hashes": {
63
- "AutoV1": "2A04055F",
64
- "AutoV2": "2954231B76",
65
- "SHA256": "2954231B76654E901BE31481DDBE862D34EF831122FE46C54B5146DB21E94BAA",
66
- "CRC32": "D27ECE6A",
67
- "BLAKE3": "40BE499D8ACA9C4D62EB4FFE6A340A0B71EF23935AEA689A820046BB71755F2C",
68
- "AutoV3": "7ACCC23B1700"
69
- },
70
- "primary": false,
71
- "downloadUrl": "https://civitai.com/api/download/models/1587891?type=Training%20Data"
72
- },
73
- {
74
- "id": 1487655,
75
- "sizeKB": 299657.3984375,
76
- "name": "studio_ghibli_wan14b_t2v_v01.safetensors",
77
- "type": "Model",
78
- "pickleScanResult": "Success",
79
- "pickleScanMessage": "No Pickle imports",
80
- "virusScanResult": "Success",
81
- "virusScanMessage": null,
82
- "scannedAt": "2025-03-27T17:41:47.311Z",
83
- "metadata": {
84
- "format": "SafeTensor",
85
- "size": null,
86
- "fp": null
87
- },
88
- "hashes": {
89
- "AutoV1": "31B3A416",
90
- "AutoV2": "DD2FE1258D",
91
- "SHA256": "DD2FE1258DB4288133F866103638F8A635CEE065E50AEF818B0F0569335E9EAB",
92
- "CRC32": "EED20A08",
93
- "BLAKE3": "1AE0474A8CC913027ADCE751FBDC374773B34C0341C7131ACDBD9893E533BB85",
94
- "AutoV3": "F4528E54C253"
95
- },
96
- "primary": true,
97
- "downloadUrl": "https://civitai.com/api/download/models/1587891"
98
- }
99
- ],
100
- "images": [
101
- {
102
- "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/61878df7-e96c-454c-8ec5-d8ab75ca83bd/width=640/66380373.mp4",
103
- "nsfwLevel": 1,
104
- "width": 640,
105
- "height": 480,
106
- "hash": "UMHCcdE-1jw^uhoz^%9bxyRPxFELEAM{VskB",
107
- "type": "video",
108
- "metadata": {
109
- "hash": "UMHCcdE-1jw^uhoz^%9bxyRPxFELEAM{VskB",
110
- "size": 1280407,
111
- "audio": false,
112
- "width": 640,
113
- "height": 480,
114
- "duration": 5.063
115
- },
116
- "meta": {
117
- "prompt": "Studio Ghibli style. A girl with a playful smile leans forward on the ferry railing, holding onto her sunhat as the wind threatens to carry it away. She looks directly at the camera with bright, excited eyes, her laughter almost visible in her expression. The sun reflects off the sparkling water, and behind her, the coastline of a charming island town approaches. The camera slowly zooms out, capturing the anticipation of arrival.",
118
- "negativePrompt": "色调艳丽,过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多,倒着走, 3D, MMD, MikuMikuDance, SFM, Source Filmmaker, Blender, Unity, Unreal, CGI, bad quality"
119
- },
120
- "availability": "Public",
121
- "hasMeta": true,
122
- "hasPositivePrompt": true,
123
- "onSite": false,
124
- "remixOfId": null
125
- },
126
- {
127
- "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e8407d4a-8ec5-4828-8e65-57de2ec28e38/width=640/66380371.mp4",
128
- "nsfwLevel": 1,
129
- "width": 640,
130
- "height": 480,
131
- "hash": "UCCiN+~A0~5n0MM}I:S08{IUWCj]~BxY%fo}",
132
- "type": "video",
133
- "metadata": {
134
- "hash": "UCCiN+~A0~5n0MM}I:S08{IUWCj]~BxY%fo}",
135
- "size": 338205,
136
- "audio": false,
137
- "width": 640,
138
- "height": 480,
139
- "duration": 5.063
140
- },
141
- "meta": {
142
- "prompt": "Studio Ghibli style. A young girl with short brown hair, adorned with a red bow, is seated at a wooden table, appearing surprised or startled. She is wearing a dark blue dress and is holding a fork, seemingly in the middle of eating a piece of food. The table is set with a basket of bread rolls, a plate with a piece of toast and a slice of tomato, a mug with a black cat design, and a bottle of milk. The lighting is soft and warm, creating a cozy atmosphere. The background features dark wooden walls, adding to the intimate and homely setting. The scene is captured in a medium shot, highlighting the girl's facial expression and the details of the breakfast table. The overall atmosphere is warm and inviting, with a sense of everyday life and simplicity.",
143
- "negativePrompt": "色调艳丽,过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多,倒着走, 3D, MMD, MikuMikuDance, SFM, Source Filmmaker, Blender, Unity, Unreal, CGI, bad quality"
144
- },
145
- "availability": "Public",
146
- "hasMeta": true,
147
- "hasPositivePrompt": true,
148
- "onSite": false,
149
- "remixOfId": null
150
- },
151
- {
152
- "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/854d9f0e-2025-40ea-ac64-02ba1e338e1b/width=640/66380309.mp4",
153
- "nsfwLevel": 1,
154
- "width": 640,
155
- "height": 480,
156
- "hash": "UFJIUd$+0j4r_K04-O$xTg-.RUN_7i-Os:xt",
157
- "type": "video",
158
- "metadata": {
159
- "hash": "UFJIUd$+0j4r_K04-O$xTg-.RUN_7i-Os:xt",
160
- "size": 558337,
161
- "audio": false,
162
- "width": 640,
163
- "height": 480,
164
- "duration": 5.063
165
- },
166
- "meta": {
167
- "prompt": "Studio Ghibli style. Camera zooms out. A young woman with short blonde hair and gray eyes stands on the deck of ship, watching directly at the camera with a calm smile. She wears a black tank top. She smiles and waves her hand. There is an amazing tropical beach visilbe behind her. The warm glow of the late afternoon sun reflects off the water. Clouds drift lazily in the vast sky as a cool gust of wind tousles her hair. In the distance, seabirds glide gracefully above the waves. The camera zooms out, revealing the ship’s full deck bathed in golden light, the vast ocean stretching endlessly beyond the horizon.",
168
- "negativePrompt": "色调艳丽,过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多,倒着走, 3D, MMD, MikuMikuDance, SFM, Source Filmmaker, Blender, Unity, Unreal, CGI, bad quality"
169
- },
170
- "availability": "Public",
171
- "hasMeta": true,
172
- "hasPositivePrompt": true,
173
- "onSite": false,
174
- "remixOfId": null
175
- },
176
- {
177
- "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/3aec5b1c-590a-4b7e-b984-62358eee465a/width=640/66380393.mp4",
178
- "nsfwLevel": 1,
179
- "width": 640,
180
- "height": 480,
181
- "hash": "UABov}IU00?a{ckDERba14xD^iE20eoy-:NH",
182
- "type": "video",
183
- "metadata": {
184
- "hash": "UABov}IU00?a{ckDERba14xD^iE20eoy-:NH",
185
- "size": 366484,
186
- "audio": false,
187
- "width": 640,
188
- "height": 480,
189
- "duration": 5.063
190
- },
191
- "meta": {
192
- "prompt": "Studio Ghibli style. A young boy with short, dark hair and glasses is standing in front of a large, dark background with red Japanese characters. He is wearing a blue uniform with a high collar and a small patch on the left side of his chest. The boy appears to be speaking or shouting, with his mouth wide open and his eyes wide. The background is dark, with the red characters standing out prominently. The lighting is soft, with a warm glow illuminating the boy's face. The overall atmosphere of the scene is tense and dramatic. The shot is a close-up, emphasizing the boy's facial expressions and the red characters in the background. The style is characteristic of Studio Ghibli animation, with detailed character design and a focus on emotion.",
193
- "negativePrompt": "色调艳丽,过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多,倒着走, 3D, MMD, MikuMikuDance, SFM, Source Filmmaker, Blender, Unity, Unreal, CGI, bad quality"
194
- },
195
- "availability": "Public",
196
- "hasMeta": true,
197
- "hasPositivePrompt": true,
198
- "onSite": false,
199
- "remixOfId": null
200
- },
201
- {
202
- "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/26ba24ac-f6e0-4ea2-8b0c-c984c7663e89/width=640/66380394.mp4",
203
- "nsfwLevel": 1,
204
- "width": 640,
205
- "height": 480,
206
- "hash": "U6Iz-L7S0Uzk014Y^M?:02%g?a#Q?OtwM2VN",
207
- "type": "video",
208
- "metadata": {
209
- "hash": "U6Iz-L7S0Uzk014Y^M?:02%g?a#Q?OtwM2VN",
210
- "size": 820890,
211
- "audio": false,
212
- "width": 640,
213
- "height": 480,
214
- "duration": 5.063
215
- },
216
- "meta": {
217
- "prompt": "Studio Ghibli style. Two children are joyfully laughing together, their faces lit up with delight. The boy, with short black hair, is wearing a yellow shirt with orange accents and red pants. The girl, with brown hair tied in pigtails, is dressed in a pink dress with a white collar. They are standing close to each other, leaning against a window sill, with the background showing a simple, muted interior setting. The lighting is soft and warm, creating a cozy and cheerful atmosphere. The scene is captured in a close-up shot, emphasizing the children's expressions and the bond between them. The overall style is vibrant and animated, characteristic of Studio Ghibli's artistry.",
218
- "negativePrompt": "色调艳丽,过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多,倒着走, 3D, MMD, MikuMikuDance, SFM, Source Filmmaker, Blender, Unity, Unreal, CGI, bad quality"
219
- },
220
- "availability": "Public",
221
- "hasMeta": true,
222
- "hasPositivePrompt": true,
223
- "onSite": false,
224
- "remixOfId": null
225
- },
226
- {
227
- "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/242e8482-7224-47eb-8fec-3b176328ff41/width=640/66380395.mp4",
228
- "nsfwLevel": 1,
229
- "width": 640,
230
- "height": 480,
231
- "hash": "UmIrj5~UkXJCtlMyxtt7tRM|jZoyt7ofWBM|",
232
- "type": "video",
233
- "metadata": {
234
- "hash": "UmIrj5~UkXJCtlMyxtt7tRM|jZoyt7ofWBM|",
235
- "size": 767592,
236
- "audio": false,
237
- "width": 640,
238
- "height": 480,
239
- "duration": 5.063
240
- },
241
- "meta": {
242
- "prompt": "Studio Ghibli style. Camera zooms out. A young woman with dark skin and dreadlocks tied back in a headscarf gazes into the camera with a soft smile. She wears a linen blouse tucked into comfortable trousers. The ocean stretches behind her, waves breaking against the sides of the ship as a distant lighthouse stands tall against the bright sky.",
243
- "negativePrompt": "色调艳丽,过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多,倒着走, 3D, MMD, MikuMikuDance, SFM, Source Filmmaker, Blender, Unity, Unreal, CGI, bad quality"
244
- },
245
- "availability": "Public",
246
- "hasMeta": true,
247
- "hasPositivePrompt": true,
248
- "onSite": false,
249
- "remixOfId": null
250
- },
251
- {
252
- "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/112fe459-827c-4e34-9f2d-838d3d08b2ea/width=640/66380397.mp4",
253
- "nsfwLevel": 1,
254
- "width": 640,
255
- "height": 480,
256
- "hash": "U9A^zO?ZD+x[^Us9o{j[-|.6-:RlE2N2RPW;",
257
- "type": "video",
258
- "metadata": {
259
- "hash": "U9A^zO?ZD+x[^Us9o{j[-|.6-:RlE2N2RPW;",
260
- "size": 1088972,
261
- "audio": false,
262
- "width": 640,
263
- "height": 480,
264
- "duration": 5.063
265
- },
266
- "meta": {
267
- "prompt": "Studio Ghibli style. A young blonde girl stands atop a grassy mountain ridge overlooking a vast, misty seashore. Her light blue dress clings to her frame as a gentle rain falls, droplets shimmering in the soft afternoon light. Wind tousles her golden hair, and she gazes down at the distant beach, where waves roll onto the shore in rhythmic motion. Dark clouds drift lazily across the sky, their edges tinged with warm hues from the hidden sun. The scene feels both melancholic and serene, with the soft patter of rain and the distant cry of seagulls blending into the moment. A slow, cinematic pan captures the movement of raindrops on leaves and the gentle swaying of wildflowers around her.",
268
- "negativePrompt": "色调艳丽,过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多,倒着走, 3D, MMD, MikuMikuDance, SFM, Source Filmmaker, Blender, Unity, Unreal, CGI, bad quality"
269
- },
270
- "availability": "Public",
271
- "hasMeta": true,
272
- "hasPositivePrompt": true,
273
- "onSite": false,
274
- "remixOfId": null
275
- },
276
- {
277
- "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/df1b62bf-1def-4d86-bb5b-d38c473c42dd/width=640/66380396.mp4",
278
- "nsfwLevel": 1,
279
- "width": 640,
280
- "height": 480,
281
- "hash": "UEIhKY5A?E~1u1s%9uESBCs+xtT0~UIq-;?Z",
282
- "type": "video",
283
- "metadata": {
284
- "hash": "UEIhKY5A?E~1u1s%9uESBCs+xtT0~UIq-;?Z",
285
- "size": 518413,
286
- "audio": false,
287
- "width": 640,
288
- "height": 480,
289
- "duration": 5.063
290
- },
291
- "meta": {
292
- "prompt": "Studio Ghibli style. A woman in a traditional blue kimono is gently brushing the hair of a young boy with a wooden comb. The boy, wearing a yellow shirt and orange suspenders, looks up at her with a smile. A young girl with brown hair, wearing a white dress with red ribbons, is sitting on the woman's lap, observing the scene. The background features a window with a view of greenery outside, suggesting a serene and cozy indoor setting. The lighting is soft and natural, creating a warm and intimate atmosphere. The scene is captured in a medium shot, highlighting the interaction between the characters and the gentle, peaceful mood of the moment.",
293
- "negativePrompt": "色调艳丽,过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多,倒着走, 3D, MMD, MikuMikuDance, SFM, Source Filmmaker, Blender, Unity, Unreal, CGI, bad quality"
294
- },
295
- "availability": "Public",
296
- "hasMeta": true,
297
- "hasPositivePrompt": true,
298
- "onSite": false,
299
- "remixOfId": null
300
- },
301
- {
302
- "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e1a3033b-8dab-4cec-9929-8f7725626c87/width=640/66380402.mp4",
303
- "nsfwLevel": 1,
304
- "width": 640,
305
- "height": 480,
306
- "hash": "U9INW~5+00}a0xN1=eE102?G}tI.;CX4KIe=",
307
- "type": "video",
308
- "metadata": {
309
- "hash": "U9INW~5+00}a0xN1=eE102?G}tI.;CX4KIe=",
310
- "size": 508111,
311
- "audio": false,
312
- "width": 640,
313
- "height": 480,
314
- "duration": 5.063
315
- },
316
- "meta": {
317
- "prompt": "Studio Ghibli style. A young girl with short brown hair is sitting at a desk, engrossed in writing in a notebook. She is wearing a pink sweater over a white collared shirt and has a pencil in her mouth. The desk is cluttered with books and papers, suggesting she is deeply focused on her work. Behind her, a calendar hangs on the wall, listing dates in a grid format. The room is warmly lit, creating a cozy atmosphere. The scene is captured in a close-up shot, emphasizing the girl's concentration and the details of her surroundings. The overall atmosphere is one of quiet determination and studiousness.",
318
- "negativePrompt": "色调艳丽,过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多,倒着走, 3D, MMD, MikuMikuDance, SFM, Source Filmmaker, Blender, Unity, Unreal, CGI, bad quality"
319
- },
320
- "availability": "Public",
321
- "hasMeta": true,
322
- "hasPositivePrompt": true,
323
- "onSite": false,
324
- "remixOfId": null
325
- },
326
- {
327
- "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/43f308de-941d-48da-b1e3-c1f6feb21874/width=640/66380401.mp4",
328
- "nsfwLevel": 2,
329
- "width": 640,
330
- "height": 480,
331
- "hash": "U6DTtm?h00%$1Uxo~mR84Ti:EptI8wRRjYIB",
332
- "type": "video",
333
- "metadata": {
334
- "hash": "U6DTtm?h00%$1Uxo~mR84Ti:EptI8wRRjYIB",
335
- "size": 662474,
336
- "audio": false,
337
- "width": 640,
338
- "height": 480,
339
- "duration": 5.063
340
- },
341
- "meta": {
342
- "prompt": "Studio Ghibli style. Young woman with short blonde hair and grey eyes, tank top, denim shorts, in a very detailed cluttered open restaraunt on a hill above amazing ocean beach.",
343
- "negativePrompt": "色调艳丽,过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多,倒着走, 3D, MMD, MikuMikuDance, SFM, Source Filmmaker, Blender, Unity, Unreal, CGI, bad quality"
344
- },
345
- "availability": "Public",
346
- "hasMeta": true,
347
- "hasPositivePrompt": true,
348
- "onSite": false,
349
- "remixOfId": null
350
- }
351
- ],
352
- "downloadUrl": "https://civitai.com/api/download/models/1587891"
353
- },
354
- "tags": [
355
- "anime",
356
- "animation",
357
- "studio ghibli",
358
- "ghibli",
359
- "style",
360
- "hayao miyazaki",
361
- "wan",
362
- "isao takahata",
363
- "wanvideo",
364
- "wan-14b"
365
- ],
366
- "modelDescription": "<h3 id=\"description-yysq8dhwq\"><strong>Description</strong></h3><p>I am very happy to share my <em>magnum opus</em> LoRA, which I've been working on for the past month since Wan came out. This is indeed the best LoRA <s>on Civitai</s> I have ever trained, and I have to say once again - <a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/Wan-Video/Wan2.1\">WanVideo </a>is an amazing model.</p><p></p><p>It was trained for ~90 hours on an RTX 3090 with <a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/kohya-ss/musubi-tuner\">musubi-tuner</a> using a mixed dataset of 240 clips and 120 images. This could have been done faster, but I was obsessed with pushing the limits to create a <em>state-of-the-art</em> style model. It’s up to you to judge if I succeeded.</p><p></p><h3 id=\"usage-np4eb9411\"><strong>Usage</strong></h3><p>The trigger phrase is <strong>Studio Ghibli style</strong> - all captions for training data were prefixed with these words.</p><p>All clips I publish in gallery are raw model outputs using a single LoRA, without post-processing, upscaling, or interpolation.</p><p></p><p>Compatibility with other LoRAs and with Wan-I2V models has not been tested.</p><p></p><p>Workflows are embedded with each clip. You can download example JSON workflow <a target=\"_blank\" rel=\"ugc\" href=\"https://files.catbox.moe/1nrkms.json\">here</a>. I use <a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/kijai/ComfyUI-WanVideoWrapper\">Kijai's wrapper</a> and enable a lot of optimizations in workflow (more information <a target=\"_blank\" rel=\"ugc\" href=\"https://civitai.com/models/1360623/ryuko-matoi-wan21-t2v-14b\">here</a>), including fp8_e5m2 checkpoints + torch.compile, SageAttention, TeaCache, Enhance-A-Video, Fp16_fast, SLG, and (sometimes) Zero-Star. Rendering a 640x480x81 clip takes about 5 minutes (RTX 3090).</p><p>WanVideo Sampler's parameters I use are the following:</p><pre><code>Sampler: unipc\nSteps: 20\nCfg: 6\nShift: 7</code></pre><p><span style=\"color:rgb(134, 142, 150)\">I believe that without optimizations and with an increase in steps, it is possible to achieve higher-quality clips, but I don't have the time or hardware resources to verify this.</span></p><p></p><p>To generate most prompts, I usually apply the following meta-prompt in ChatGPT (or Claude, or any other capable LLM), that helps to enhance \"raw\" descriptions. This prompt is based on official <a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/Wan-Video/Wan2.1/blob/main/wan/utils/prompt_extend.py\">prompt extension code</a> by Wan developers and looks like this:</p><pre><code>You are a prompt engineer, specializing in refining user inputs into high-quality prompts for video generation in the distinct Studio Ghibli style. You ensure that the output aligns with the original intent while enriching details for visual and motion clarity.\n\nTask Requirements:\n- If the user input is too brief, expand it with reasonable details to create a more vivid and complete scene without altering the core meaning.\n- Emphasize key features such as characters' appearances, expressions, clothing, postures, and spatial relationships.\n- Always maintain the Studio Ghibli visual aesthetic - soft watercolor-like backgrounds, expressive yet simple character designs, and a warm, nostalgic atmosphere.\n- Enhance descriptions of motion and camera movements for natural animation flow. Include gentle, organic movements that match Ghibli's storytelling style.\n- Preserve original text in quotes or titles while ensuring the prompt is clear, immersive, and 80-100 words long.\n- All prompts must begin with \"Studio Ghibli style.\" No other art styles should be used.\n\nExample Revised Prompts:\n\"Studio Ghibli style. A young girl with short brown hair and curious eyes stands on a sunlit grassy hill, wind gently rustling her simple white dress. She watches a group of birds soar across the golden sky, her bare feet sinking slightly into the soft earth. The scene is bathed in warm, nostalgic light, with lush trees swaying in the distance. A gentle breeze carries the sounds of nature. Medium shot, slightly low angle, with a slow cinematic pan capturing the serene movement.\"\n\"Studio Ghibli style. A small village at sunset, lanterns glowing softly under the eaves of wooden houses. A young boy in a blue yukata runs down a narrow stone path, his sandals tapping against the ground as he chases a firefly. His excited expression reflects in the shimmering river beside him. The atmosphere is rich with warm oranges and cool blues, evoking a peaceful summer evening. Medium shot with a smooth tracking movement following the boy's energetic steps.\"\n\"Studio Ghibli style. A mystical forest bathed in morning mist, where towering trees arch over a moss-covered path. A girl in a simple green cloak gently places her hand on the back of a massive, gentle-eyed creature resembling an ancient deer. Its fur shimmers faintly as sunlight pierces through the thick canopy, illuminating drifting pollen. The camera slowly zooms in, emphasizing their quiet connection. A soft gust of wind stirs the leaves, and tiny glowing spirits peek from behind the roots.\"\n\nInstructions:\nI will now provide a prompt for you to rewrite. Please expand and refine it in English while ensuring it adheres to the Studio Ghibli aesthetic. Even if the input is an instruction rather than a description, rewrite it into a complete, visually rich prompt without additional responses or quotation marks.\n\nThe prompt is: \"YOUR PROMPT HERE\".</code></pre><p>Replace YOUR PROMPT HERE with something like <em>Young blonde girl stands on the mountain near seashore beach under rain</em> or whatever<em>.</em></p><p>The negative prompt always includes the same base text (but may have additional words added depending on the specific prompt):</p><pre><code>色调艳丽,过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多,倒着走, 3D, MMD, MikuMikuDance, SFM, Source Filmmaker, Blender, Unity, Unreal, CGI, bad quality</code></pre><p></p><h3 id=\"dataset-jkw0xf4ku\"><strong>Dataset</strong></h3><p><span style=\"color:rgb(190, 75, 219)\">In this and the following sections, I'll be doing a bit of yapping :) Feel free to skip ahead and just read the </span><strong><span style=\"color:rgb(190, 75, 219)\">Conclusion</span></strong><span style=\"color:rgb(190, 75, 219)\">, but maybe someone will find some useful bits of information in this wall of text. So...</span></p><p></p><p>Dataset selection stage was the \"easiest\" part, I already have all the Ghibli films in highest possible quality and splitted into scenes - over 30,000 clips in 1920x1040 resolution and high bitrate. They're patiently waiting for the day I finally will make a full fine-tune some video model with them.</p><p></p><p>And I had already prepped around 300 clips for training v0.7 of <a target=\"_blank\" rel=\"ugc\" href=\"https://civitai.com/models/1084814/studio-ghibli-style-hunyuanvideo\">HV LoRA</a> (in fact, I was just about to start the training when Wan came out). These clips were in the range of 65-129 frames, which I consider optimal for training HV on videos, and they were all 24 fps. For Wan, though, I wanted them to be in a different frame range (not exceeding 81 frames, explanation see later in the \"Training\" section). I also needed them to be in 16 fps. <span style=\"color:rgb(134, 142, 150)\">I'm still </span><a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/kohya-ss/musubi-tuner/discussions/200\">not entirely sure</a><span style=\"color:rgb(134, 142, 150)\"> if strict 16 fps is necessary, but I had some issues with HV when clips were in 30 fps instead of HV’s native 24 fps, so I decided to stick with 16 fps.</span></p><p></p><p><em>I should mention, that for processing dataset, I usually make a lot of small \"one-time\" scripts (with the help of Claude, ChatGPT, and DeepSeek) - that includes mini-GUIs for manual selection of videos, one-liners for splitting frames, scripts for outputting various helper stats, dissecting clips by ranges, creating buckets in advance, etc. I don't publish these scripts because they're messy, full of hardcoded values, and designed for one-time use anyway. And nowadays anyone can easily create similar scripts by making requests to the aforementioned LLMs.</em></p><p></p><p>Converting all clips to 16 fps narrowed the range of frames in each video from 65-129 to around 45-88 frames, which messed up my meticulously planned, frame-perfect ranges for the frame buckets I had set up for training. Thankfully, it wasn't a big deal because I had some rules in place when selecting videos for training, specifically to handle situations like this.</p><p>First of all, the scene shouldn't have rapid transitions during its duration. I needed this because I couldn't predict the exact duration (in frames) of target frame buckets that trainer will establish for training - model size, VRAM, and other factors all affect this. Example: I might want to use a single 81-frame long clip for training, but I won't be able to do this, because I will get OOM on RTX 3090. So will have to choose some <a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/kohya-ss/musubi-tuner/blob/main/dataset/dataset_config.md#frame_extraction-options\">frame extraction strategy</a>, depending of which clip might be splitted onto several shorter parts (<a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/kohya-ss/musubi-tuner/issues/59#issuecomment-2613708935\">here is excellent breakdown</a> of various strategies). And its semantic coherence might be broken (like, on first fragment of the clip a girl might open her mouth , but from clipped first fragment it will become ambiguous whether she is gonna cry or laugh), and that kind of context incoherence may make Wan's <span style=\"color:rgb(37, 38, 43)\">UMT5 encoder feel sad.</span></p><p>Another thing to consider is that I wanted to reuse captions for any fragment of the original clip without dealing with recaptioning and recaching embeddings via the text encoder. Captioning videos takes quite a long time, but if a scene changes drastically throughout its range, the original caption might not fit all fragments, reducing training quality. By following rules \"clip should not contain rapid context transitions\" and \"clip should be self-contained, i.e. it should not feature events that may not be understood from within the clip itself\", even if a scene is to be split into subfragments, the captions would (with an acceptable margin of error) still apply to each fragment.</p><p></p><p>After conversion I looked through all clips and reduced total number of them to 240 (just took out some clips that did contained too much transitions or, vica-versa, were too static), which formed the first part of the dataset.</p><p></p><p>I decided to use a mixed dataset of videos and images. So second part of the dataset was formed by 120 images (at 768x768 resolution), taken from screencaps of various Ghibli movies.</p><p><span style=\"color:rgb(134, 142, 150)\">There's an alternative approach where you train on images first and then fine-tune on videos (it was successfully applied by the creator of </span><a target=\"_blank\" rel=\"ugc\" href=\"https://civitai.com/models/1332383/wan-lora-arcane-jinx-v1-wan-13b\">this LoRA</a><span style=\"color:rgb(134, 142, 150)\">), but I personally think it's not as good as mixing in a single batch (though I don't have hard numbers to back this up). To back up my assumptions, here is </span><a target=\"_blank\" rel=\"ugc\" href=\"https://civitai.com/models/1132089/flat-color-style?modelVersionId=1474944\">very good LoRA</a><span style=\"color:rgb(134, 142, 150)\"> that uses the same mixed approach to training (and btw it was also done on a 24 GB GPU, if I am not mistaken).</span></p><p></p><p>To properly enable effective video training on mixed dataset on consumer-level GPUs I had to find the right balance between resolution, duration, and training time, and I decided to do this by mixing low-res high-duration videos and high-res images - I will give more details about this in <strong>Training</strong> section.</p><p></p><p>Considering captioning: images for dataset were actually just reused from some of my HV datasets, and they were captioned earlier using my \"swiss army knife\" VLM for (SFW-only) dataset captioning, also known as <a target=\"_blank\" rel=\"ugc\" href=\"https://huggingface.co/Qwen/Qwen2-VL-7B-Instruct\"><strong>Qwen2-VL-7B-Instruct</strong></a>. I used the following captioning prompt:</p><pre><code>Create a very detailed description of this scene. Do not use numbered lists or line breaks. IMPORTANT: The output description MUST ALWAYS start with the unaltered phrase 'Studio Ghibli style. ', followed by your detailed description. The description should 1) describe the main content of the scene, 2) describe the environment and lighting details, 3) identify the type of shot (e.g., aerial shot, close-up, medium shot, long shot), and 4) include the atmosphere of the scene (e.g., cozy, tense, mysterious). Here's a template you MUST use: 'Studio Ghibli style. {Primary Subject Action/Description}. {Environment and Lighting Details}. {Style and Technical Specifications}'.</code></pre><p>I had some doubts about whether I should recaption them since the target caption structure was specifically designed for HunyuanVideo, and I worried that Wan might need a completely different approach. I left them as-is, and have no idea if this was the right decision, but, broadly speaking, modern text encoders are powerful enough to ignore such limitations. As we know, models like Flux and some others can even be trained without captions at all (although I believe training with captions is always better than without - but only if captions are relevant to the content).</p><p></p><p>For captioning videos I tested a bunch of local models that can natively caption video content:</p><ul><li><p><a target=\"_blank\" rel=\"ugc\" href=\"https://huggingface.co/THUDM/cogvlm2-video-llama3-chat\">CogVLM2-Video-Llama3-Chat</a> (usually this is my go-to option for clip captioning)</p></li><li><p><a target=\"_blank\" rel=\"ugc\" href=\"https://huggingface.co/openbmb/MiniCPM-V-2_6\">MiniCPM-V 2.6</a></p></li><li><p><a target=\"_blank\" rel=\"ugc\" href=\"https://huggingface.co/GoodiesHere/Apollo-LMMs-Apollo-7B-t32\">Apollo-LMMs-Apollo-7B-t32</a></p></li><li><p><a target=\"_blank\" rel=\"ugc\" href=\"https://huggingface.co/llava-hf/llava-onevision-qwen2-7b-ov-hf\">LLaVA-Onevision</a></p></li><li><p><a target=\"_blank\" rel=\"ugc\" href=\"https://huggingface.co/OpenGVLab/VideoChat-Flash-Qwen2_5-2B_res448\">VideoChat-Flash-2B</a></p></li><li><p><a target=\"_blank\" rel=\"ugc\" href=\"https://huggingface.co/DAMO-NLP-SG/VideoLLaMA3-7B\">VideoLLaMA 3</a></p></li><li><p><a target=\"_blank\" rel=\"ugc\" href=\"https://huggingface.co/AIDC-AI/Ovis2-16B\">Ovis2-16B </a>(this one seems really good! But I had already dataset captioned when I found it, so will use it in future LoRAs)</p></li></ul><p>There are more models out there, but these are the ones I tested. For this LoRA, I ended up using <strong><span style=\"color:rgb(37, 38, 43)\">Apollo-7B</span></strong>. I used this simple VLM prompt:</p><pre><code>Create a very detailed description of this video. IMPORTANT: The output description MUST ALWAYS start with the unaltered phrase 'Studio Ghibli style. ', followed by your detailed description.</code></pre><p>I’m attaching the full dataset I used as an addendum to the model. While it does kinda contain copyrighted material, I think this falls under fair use.</p><p></p><h3 id=\"training-yk6zn3ean\"><strong>Training</strong></h3><p>If anyone interested, here is list of trainers that I considered for training WanVideo:</p><ul><li><p><a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/tdrussell/diffusion-pipe\">diffusion-pipe</a> - OG of the HV training, but also allows memory-efficient Wan training; config-driven, has third-party GUI and runpod templates (read more <a target=\"_blank\" rel=\"ugc\" href=\"https://civitai.com/articles/12828/new-diffusion-pipe-interface-for-lora-training\">here </a>and <a target=\"_blank\" rel=\"ugc\" href=\"https://civitai.com/articles/12330/wansdxlflux-all-in-one-lora-training-diffusion-pipe-with-auto-captioning\">here</a>). For HV I used it exclusively. Requires WSL to run on Windows.</p></li><li><p><a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/kohya-ss/musubi-tuner\"><strong>Musubi Tuner</strong></a><strong> - Maintained by responsible and friendly developer. Config-driven, has cozy community, tons of options. Currently my choice for Wan training.</strong></p></li><li><p><a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/ostris/ai-toolkit\">AI Toolkit</a> - My favorite trainer for Flux recently got support for Wan. It's fast, easy-to-use, config-driven, also has first-party UI (which I do not use 🤷), but currently supports training 14B only without captions, which is the main reason I do not use it.</p></li><li><p><a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/modelscope/DiffSynth-Studio\">DiffSynth Studio</a> - I haven't had the time to test it yet and am unsure if it can train Wan models with 24 GB VRAM. However, it’s maintained by ModelScope, making it worth a closer look. I plan to test it soon.</p></li><li><p><a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/a-r-r-o-w/finetrainers\">finetrainers </a>- Has support for Wan training, but doesn't seem to work with 24 GB GPUs (yet)</p></li><li><p><a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/bghira/SimpleTuner\">SimpleTuner </a>- Gained support for Wan last week, so I haven't had a chance to try it yet. It definitely deserves attention since the main developer is a truly passionate and knowledgeable person.</p></li><li><p><a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/Bria-AI/Zero-to-Wan\">Zero-to-Wan</a> - Supports training only for 1.3B models.</p></li><li><p><a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/spacepxl/WanTraining\">WanTraining</a> - I have to mention this project, as it's supported by a developer who’s done impressive work with it, including <a target=\"_blank\" rel=\"ugc\" href=\"https://civitai.com/models/1337418/wan21-13b-cfg-distill-lora\">guidance-distilled LoRA</a> and <a target=\"_blank\" rel=\"ugc\" href=\"https://huggingface.co/spacepxl/Wan2.1-control-loras\">control LoRA<span style=\"color:rgb(54, 65, 83)\">.</span></a></p></li></ul><p></p><p>So, I used Musubi Tuner. <em><span style=\"color:rgb(134, 142, 150)\">For reference, here are my hardware params: i5-12600KF, RTX 3090, Windows 11, 64Gb RAM. </span></em>The commands and config files I used were the following.</p><p></p><ul><li><p>For <span style=\"color:rgb(34, 34, 34)\">caching VAE latents (nothing specific here, just default command)</span></p></li></ul><pre><code>python wan_cache_latents.py --dataset_config G:/samples/musubi-tuner/_studio_ghibli_wan14b_v01_dataset.toml --vae G:/samples/musubi-tuner/wan14b/vae/wan_2.1_vae.safetensors</code></pre><ul><li><p>For <span style=\"color:rgb(34, 34, 34)\">caching text encoder embeddings (default):</span></p></li></ul><pre><code>python wan_cache_text_encoder_outputs.py --dataset_config G:/samples/musubi-tuner/_studio_ghibli_wan14b_v01_dataset.toml --t5 G:/samples/musubi-tuner/wan14b/tenc/models_t5_umt5-xxl-enc-bf16.pth --batch_size 16 </code></pre><ul><li><p>For launching training:</p></li></ul><pre><code>accelerate launch --num_cpu_threads_per_process 1 --mixed_precision bf16 wan_train_network.py ^\n --task t2v-14B ^\n --dit G:/samples/musubi-tuner/wan14b/dit/wan2.1_t2v_14B_bf16.safetensors ^\n\t--vae G:/samples/musubi-tuner/wan14b/vae/wan_2.1_vae.safetensors ^\n\t--t5 G:/samples/musubi-tuner/wan14b/tenc/models_t5_umt5-xxl-enc-bf16.pth ^\n\t--sdpa ^\n\t--blocks_to_swap 10 ^\n\t--mixed_precision bf16 ^\n\t--fp8_base ^\n\t--fp8_scaled ^\n\t--fp8_t5 ^\n\t--dataset_config G:/samples/musubi-tuner/_studio_ghibli_wan14b_v01_dataset.toml ^\n --optimizer_type adamw8bit ^\n\t--learning_rate 5e-5 ^\n\t--gradient_checkpointing ^\n --max_data_loader_n_workers 2 ^\n\t--persistent_data_loader_workers ^\n --network_module networks.lora_wan ^\n\t--network_dim 32 ^\n\t--network_alpha 32 ^\n --timestep_sampling shift ^\n\t--discrete_flow_shift 3.0 ^\n\t--save_every_n_epochs 1 ^\n\t--seed 2025 ^\n --output_dir G:/samples/musubi-tuner/output ^\n\t--output_name studio_ghibli_wan14b_v01 ^\n\t--log_config ^\n\t--log_with tensorboard ^\n\t--logging_dir G:/samples/musubi-tuner/logs ^\n\t--sample_prompts G:/samples/musubi-tuner/_studio_ghibli_wan14b_v01_sampling.txt ^\n\t--save_state ^\n\t--max_train_epochs 50 ^\n\t--sample_every_n_epochs 1</code></pre><p>Again, nothing to see here, actually. I had to use <em>blocks_to_swap</em> parameter because otherwise, with my dataset config (see below), I confronted into 24 Gb VRAM constraints. Hyperparameters were mostly left on defaults. I didn't want to risk anything after a bad experience - 60 hours of HV training lost due to getting too ambitious with flow shift values and adaptive optimizers instead of good old <em>adamw</em>.</p><p></p><ul><li><p><span style=\"color:rgb(31, 35, 40)\">Prompt file for </span>sampling during training:</p></li></ul><pre><code># prompt 1\nStudio Ghibli style. Woman with blonde hair is walking on the beach, camera zoom out. --w 384 --h 384 --f 45 --d 7 --s 20\n\n# prompt 2\nStudio Ghibli style. Woman dancing in the bar. --w 384 --h 384 --f 45 --d 7 --s 20</code></pre><ul><li><p>Dataset configuration (the most important part; I'll explain the thoughts that led me to it afterward):</p></li></ul><pre><code>[general]\ncaption_extension = \".txt\"\nenable_bucket = true\nbucket_no_upscale = true\n\n[[datasets]]\nimage_directory = \"H:/datasets/studio_ghibli_wan_video_v01/images/768x768\"\ncache_directory = \"H:/datasets/studio_ghibli_wan_video_v01/images/768x768/cache\"\nresolution = [768, 768]\nbatch_size = 1\nnum_repeats = 1\n\n[[datasets]]\nvideo_directory = \"H:/datasets/studio_ghibli_wan_video_v01/videos/1920x1040\"\ncache_directory = \"H:/datasets/studio_ghibli_wan_video_v01/videos/1920x1040/cache_1\"\nresolution = [768, 416]\nbatch_size = 1\nnum_repeats = 1\nframe_extraction = \"head\"\ntarget_frames = [1, 21]\n\n[[datasets]]\nvideo_directory = \"H:/datasets/studio_ghibli_wan_video_v01/videos/1920x1040\"\ncache_directory = \"H:/datasets/studio_ghibli_wan_video_v01/videos/1920x1040/cache_2\"\nresolution = [384, 208]\nbatch_size = 1\nnum_repeats = 1\nframe_extraction = \"uniform\"\ntarget_frames = [45]\nframe_sample = 2</code></pre><p>My dataset setup consists of three parts.</p><p>I'll start with the <strong>last one</strong>, which includes the main data array - 240 clips in 1920x1040 resolution and duration that varies from 45 to 88 frames.</p><p>Obviously, training on full-resolution 1920x1040, full-duration clips on an RTX 3090 was out of the question. I needed to find the minimum resolution and frame duration that would avoid OOM errors while keeping the bucket fragments as long as possible. Longer fragments help the model learn motion, timing, and spatial patterns (like hair twitching, fabric swaying, liquid dynamics etc.) of the Ghibli style - something you can't achieve with still frames.</p><p>From training HV, I remembered a good starting point for estimation of available resolution range for 24 Gb GPU is 512x512x33. I decided on the \"uniform\" frame extraction pattern, ensuring all extracted fragments were no fewer than 45 frames. Since, as I wrote before, after conversion to 16fps, maxed out at 88 frames, this approach kept the clips from being divided into more than two spans, which would've made epochs too long. At the same time, timespan of 45 frames (~3s) should be enough for model to learn spatial flow of the style.</p><p>With the target fixed to 45 frames, I started testing different resolutions. I used a script to analyze all clips in a folder and suggest valid width-height combinations that maintained the original aspect ratio (1920/1040 ≈ 1.85) and were divisible by 16 (a <a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/kohya-ss/musubi-tuner/blob/c8fea7413ed5599ad1a75ff8e3e25870e78748d4/dataset/image_video_dataset.py#L301\">model requirement</a>).</p><p>Eventually, I found that using <strong>[384, 208]</strong> for the bucket size and setting <strong>--blocks_to_swap 10</strong> prevented OOM errors and pushing into shared memory (which eventually led to 160 s/it). The downside was that training speed dropped to around <strong>11-12 s/it</strong>. In hindsight, lowering the resolution to [368, 192] could have bumped the speed up to ~8 s/it, which would've been great (close to what I get when training Flux at 1024p in AI Toolkit). And that would've saved me around 20 hours of training over the full 90-hour run (~28000 steps), although I didn't expect it to go &gt; 20K steps back then.</p><p></p><p>And it needs to be noted, that I trained on Windows with my monitor connected to the GPU (and used my PC for coding at the same time 😼). On Linux (for example, with diffusion-pipe) and with using internal GPU for monitor output, it might be possible to use slightly higher spatiotemporal resolutions without hitting OOM or shared memory limits (something I think is Windows-specific).</p><p></p><p>Now about the <strong>first part</strong> (120 images in 768x768 resolution). Initially, I wanted to train on 1024p images, but I decided it'd be overkill and slow things down. My plan was to train on HD images and low-res videos simultaneously to ensure better generalization. The idea was that high-resolution images would compensate for the lower resolution of the clips. And joint video + image pretraining is how <a target=\"_blank\" rel=\"ugc\" href=\"https://cdn.blrimages.net/input/ab6373a7-c5a4-4f5a-a60d-cb1a8538b7d4.image/avif\">WAN was trained</a> anyway, so I figured this approach would favor \"upstream\" style learning as well.</p><p></p><p>Finally, the <strong>second part</strong>, which is also important for generalization (again, that is not as \"scientific\" assumption, but it seems reasonable). The idea was to reuse the same clips from the third section but now train only on the first frame and the first 21 frames. This approach, I hoped, would facilitate learning temporal style motion features. At the same time, it let me bump up the resolution for the second section to [768, 416].</p><p></p><p>As the result, I hoped to achieve \"cross-generalization\" between:</p><ul><li><p><strong>Section 1's</strong> high-res images (768x768)</p></li><li><p><strong>Section 2's</strong> medium-res single frames and 21-frame clips (768x416)</p></li><li><p><strong>Section 3's</strong> low-res 45-frame clips (384x208)</p><p></p></li></ul><p>Additionally, both the second and the larger part of the third sections shared the same starting frame, which I believed would benefit LoRA usage in I2V scenarios. All this seemed like the best way to fully utilize my dataset without hitting hardware limits.</p><p>Of course, I'm <a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/kohya-ss/musubi-tuner/issues/46#issuecomment-2599189500\">not the first</a> to come up with this approach, but it seems logical and reasonable, so I hope more creators realize you don’t need an A100 to train a video-based LoRA for Wan.</p><p></p><p><em><span style=\"color:rgb(134, 142, 150)\">Funny fact: I expected one epoch to consist of 1080 samples: 120 images (1st dataset section) + 240 single frames (2nd dataset section, \"head\" frame bucket=1) + 240 clips of 21 frames each (2nd dataset section, \"head\" frame bucket=21) + 480 clips of 45 frames each (2nd dataset section, \"uniform\" frame bucket=45, sampled 2 times). However, after I started training, I discovered it was actually 1078 samples. When I dug into it, I found that two of the clips reported by my scripts (which use the ffprobe command from ffmpeg to count the number of frames) were actually shorter than 45 frames, so there was an issue with rounding. This wasn't a big deal, so I just continued training without those two clips, but that was the reason the number of steps for the final LoRA seemed so off :)</span></em></p><p></p><p>The training itself went smoothly. I won't reveal loss graphs since I <s>am too shy</s> don't think they mean much. I mostly use them to check if the loss distribution starts looking too similar across epochs - that's my cue for potential overfitting.</p><p>I trained up to 28000 steps, then spent several days selecting the best checkpoint. Another thing I think I could have done better is taking checkpoints not just at the end of each epoch, but also in between. Since each epoch is 1078 steps long, it's possible that a checkpoint with even better results than the one I ended up with was lost somewhere in between.</p><p>I'm considering integrating validation loss estimation into my training pipeline (more on this <a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/spacepxl/demystifying-sd-finetuning\">here</a>), but I haven't done it yet.</p><p></p><p>Could this be simplified? Probably yes. In my next LoRA, I'll test whether the extra image dataset in section 1 was redundant. I could've just set up a separate dataset section and reused clips' first frame, but with high resolution. On the other hand, I wanted the dataset to be as varied as possible, so I used screencaps from different scenes than the clips, in this sense they were not redundant.</p><p>I'm not even sure if the second section was necessary. Since WAN itself (according to its <a target=\"_blank\" rel=\"ugc\" href=\"https://files.alicdn.com/tpsservice/5c9de1c74de03972b7aa657e5a54756b.pdf\">technical report</a>) was pretrained on 192px clips, training at around 352x192x45 should be effective and make the most of my hardware. Ideally, I'd use 5-second clips (16 fps * 5s + 1 = 81 frames), but that’s just not feasible on the RTX 3090 without aggressive block swapping.</p><p></p><h3 id=\"conclusion-7elayk28w\"><strong>Conclusion</strong></h3><p>Aside from the fun and the hundreds of insanely good clips, here are some insights I've gained from training this LoRA. I should mention that these practices are based on my personal experience and observations, I don't have any strictly analytical evidence to prove their effectiveness and I only tried style training so far. I plan to explore concept training very soon to test some of my other assumptions and see if they can be applied as well.</p><ul><li><p>You can train Wan-14B on consumer-level GPUs using videos. 368x192x45 seems like a solid starting point.</p></li><li><p>Compensate for motion-targeted style learning on low-res videos by using high-res images to ensure better generalization.</p></li><li><p>Combine various frame extraction methods on the same datasets to maximize effectiveness and hardware usage.</p><p></p></li></ul><p>A lot, if not all, of what I've learned to make this LoRA comes from reading countless <a target=\"_blank\" rel=\"ugc\" href=\"https://www.reddit.com/r/StableDiffusion/\">r/StableDiffusion</a> posts, 24/7 lurking on the awesome <a target=\"_blank\" rel=\"ugc\" href=\"https://banodoco.ai/\">Banodoco Discord</a>, reading comments <s>and opening every NSFW clip</s> to every single WanVideo model here on Civitai, and diving into every issue I could find in the <a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/kohya-ss/musubi-tuner\">musubi-tuner</a>, <a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/tdrussell/diffusion-pipe\">diffusion-pipe</a>, <a target=\"_blank\" rel=\"ugc\" href=\"https://github.com/Wan-Video/Wan2.1\">Wan2.1</a>, and other repositories. 😽</p>",
367
- "civitai_deleted": false,
368
- "usage_tips": "{}"
369
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/loras/studio_ghibli_wan14b_t2v_v01.mp4 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:cdb49dcb6f44c54170c8885e50fb8afc367a909bdf90ce2cebb3bc7076072b17
3
- size 1280407
 
 
 
 
ComfyUI/models/loras/studio_ghibli_wan14b_t2v_v01.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:dd2fe1258db4288133f866103638f8a635cee065e50aef818b0f0569335e9eab
3
- size 306849176
 
 
 
 
ComfyUI/models/sonic/.gitattributes DELETED
@@ -1,35 +0,0 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/sonic/RIFE/flownet.pkl DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe854fc8996547c953f732aaa3b78cae76cc0a12833ae856ea0749c4c570d7d8
3
- size 12186817
 
 
 
 
ComfyUI/models/sonic/audio2bucket.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:4af942fded37b70d0a0a7993b8c0f46b5b1e104bcf7b6316f7b509900db69d26
3
- size 87050998
 
 
 
 
ComfyUI/models/sonic/audio2token.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:68cf305813bf5e4682c4f6a80955233cfad6a5e194fd4a6963f27e4e8ee490d8
3
- size 217206084
 
 
 
 
ComfyUI/models/sonic/whisper-tiny/config.json DELETED
@@ -1,144 +0,0 @@
1
- {
2
- "_name_or_path": "openai/whisper-tiny",
3
- "activation_dropout": 0.0,
4
- "activation_function": "gelu",
5
- "architectures": [
6
- "WhisperForConditionalGeneration"
7
- ],
8
- "attention_dropout": 0.0,
9
- "begin_suppress_tokens": [
10
- 220,
11
- 50257
12
- ],
13
- "bos_token_id": 50257,
14
- "d_model": 384,
15
- "decoder_attention_heads": 6,
16
- "decoder_ffn_dim": 1536,
17
- "decoder_layerdrop": 0.0,
18
- "decoder_layers": 4,
19
- "decoder_start_token_id": 50258,
20
- "dropout": 0.0,
21
- "encoder_attention_heads": 6,
22
- "encoder_ffn_dim": 1536,
23
- "encoder_layerdrop": 0.0,
24
- "encoder_layers": 4,
25
- "eos_token_id": 50257,
26
- "forced_decoder_ids": [
27
- [
28
- 1,
29
- 50259
30
- ],
31
- [
32
- 2,
33
- 50359
34
- ],
35
- [
36
- 3,
37
- 50363
38
- ]
39
- ],
40
- "init_std": 0.02,
41
- "is_encoder_decoder": true,
42
- "max_length": 448,
43
- "max_source_positions": 1500,
44
- "max_target_positions": 448,
45
- "model_type": "whisper",
46
- "num_hidden_layers": 4,
47
- "num_mel_bins": 80,
48
- "pad_token_id": 50257,
49
- "scale_embedding": false,
50
- "suppress_tokens": [
51
- 1,
52
- 2,
53
- 7,
54
- 8,
55
- 9,
56
- 10,
57
- 14,
58
- 25,
59
- 26,
60
- 27,
61
- 28,
62
- 29,
63
- 31,
64
- 58,
65
- 59,
66
- 60,
67
- 61,
68
- 62,
69
- 63,
70
- 90,
71
- 91,
72
- 92,
73
- 93,
74
- 359,
75
- 503,
76
- 522,
77
- 542,
78
- 873,
79
- 893,
80
- 902,
81
- 918,
82
- 922,
83
- 931,
84
- 1350,
85
- 1853,
86
- 1982,
87
- 2460,
88
- 2627,
89
- 3246,
90
- 3253,
91
- 3268,
92
- 3536,
93
- 3846,
94
- 3961,
95
- 4183,
96
- 4667,
97
- 6585,
98
- 6647,
99
- 7273,
100
- 9061,
101
- 9383,
102
- 10428,
103
- 10929,
104
- 11938,
105
- 12033,
106
- 12331,
107
- 12562,
108
- 13793,
109
- 14157,
110
- 14635,
111
- 15265,
112
- 15618,
113
- 16553,
114
- 16604,
115
- 18362,
116
- 18956,
117
- 20075,
118
- 21675,
119
- 22520,
120
- 26130,
121
- 26161,
122
- 26435,
123
- 28279,
124
- 29464,
125
- 31650,
126
- 32302,
127
- 32470,
128
- 36865,
129
- 42863,
130
- 47425,
131
- 49870,
132
- 50254,
133
- 50258,
134
- 50358,
135
- 50359,
136
- 50360,
137
- 50361,
138
- 50362
139
- ],
140
- "torch_dtype": "float32",
141
- "transformers_version": "4.27.0.dev0",
142
- "use_cache": true,
143
- "vocab_size": 51865
144
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ComfyUI/models/sonic/whisper-tiny/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ebd0e69e78190ffe1438491fa05cc1f5c1aa3a4c4db3bc1723adbb551ea2395
3
- size 151061672
 
 
 
 
ComfyUI/models/sonic/whisper-tiny/preprocessor_config.json DELETED
The diff for this file is too large to render. See raw diff
 
ComfyUI/models/sonic/yoloface_v5m.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:5ef5928d2ee1350ea7050ad7524b26a2b55e5c69fee49cd499667bde6a215b17
3
- size 84574759
 
 
 
 
ComfyUI/models/upscale_models/4x_NMKD-Siax_200k.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:560424d9f68625713fc47e9e7289a98aabe1d744e1cd6a9ae5a35e9957fd127e
3
- size 66957746
 
 
 
 
ComfyUI/models/upscale_models/4x_foolhardy_Remacri.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e1a73bd89c2da1ae494774746398689048b5a892bd9653e146713f9df8bca86a
3
- size 67025055