Buckets:
| import{s as Pf,o as Hf,n as C}from"../chunks/scheduler.8c3d61f6.js";import{S as Vf,i as If,g as n,s as t,r as p,A as Ff,h as s,f as i,c as a,j as w,u as m,x as f,k as v,y as o,a as L,v as h,d as _,t as u,w as g}from"../chunks/index.da70eac4.js";import{T as P}from"../chunks/Tip.1d9b8c37.js";import{D as $}from"../chunks/Docstring.fa488882.js";import{C as Hs}from"../chunks/CodeBlock.a9c4becf.js";import{E as Ps}from"../chunks/ExampleCodeBlock.ec9feb8f.js";import{H as ne,E as Uf}from"../chunks/index.dfbaf638.js";function Xf(T){let r,b='To learn more about how to load LoRA weights, see the <a href="../../using-diffusers/loading_adapters#lora">LoRA</a> loading guide.';return{c(){r=n("p"),r.innerHTML=b},l(l){r=s(l,"P",{"data-svelte-h":!0}),f(r)!=="svelte-1fw6lx1"&&(r.innerHTML=b)},m(l,c){L(l,r,c)},p:C,d(l){l&&i(r)}}}function Ef(T){let r,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,M="This function is experimental and might change in the future.";return{c(){r=n("p"),r.textContent=b,l=t(),c=n("p"),c.textContent=M},l(d){r=s(d,"P",{"data-svelte-h":!0}),f(r)!=="svelte-15l1sdn"&&(r.textContent=b),l=a(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=M)},m(d,y){L(d,r,y),L(d,l,y),L(d,c,y)},p:C,d(d){d&&(i(r),i(l),i(c))}}}function Wf(T){let r,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,M="This function is experimental and might change in the future.";return{c(){r=n("p"),r.textContent=b,l=t(),c=n("p"),c.textContent=M},l(d){r=s(d,"P",{"data-svelte-h":!0}),f(r)!=="svelte-15l1sdn"&&(r.textContent=b),l=a(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=M)},m(d,y){L(d,r,y),L(d,l,y),L(d,c,y)},p:C,d(d){d&&(i(r),i(l),i(c))}}}function qf(T){let r,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,M="This function is experimental and might change in the future.";return{c(){r=n("p"),r.textContent=b,l=t(),c=n("p"),c.textContent=M},l(d){r=s(d,"P",{"data-svelte-h":!0}),f(r)!=="svelte-15l1sdn"&&(r.textContent=b),l=a(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=M)},m(d,y){L(d,r,y),L(d,l,y),L(d,c,y)},p:C,d(d){d&&(i(r),i(l),i(c))}}}function Nf(T){let r,b="This is an experimental API.";return{c(){r=n("p"),r.textContent=b},l(l){r=s(l,"P",{"data-svelte-h":!0}),f(r)!=="svelte-8w79b9"&&(r.textContent=b)},m(l,c){L(l,r,c)},p:C,d(l){l&&i(r)}}}function zf(T){let r,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,M="This function is experimental and might change in the future.";return{c(){r=n("p"),r.textContent=b,l=t(),c=n("p"),c.textContent=M},l(d){r=s(d,"P",{"data-svelte-h":!0}),f(r)!=="svelte-15l1sdn"&&(r.textContent=b),l=a(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=M)},m(d,y){L(d,r,y),L(d,l,y),L(d,c,y)},p:C,d(d){d&&(i(r),i(l),i(c))}}}function Bf(T){let r,b="This is an experimental API.";return{c(){r=n("p"),r.textContent=b},l(l){r=s(l,"P",{"data-svelte-h":!0}),f(r)!=="svelte-8w79b9"&&(r.textContent=b)},m(l,c){L(l,r,c)},p:C,d(l){l&&i(r)}}}function jf(T){let r,b="Examples:",l,c,M;return c=new Hs({props:{code:"JTIzJTIwQXNzdW1pbmclMjAlNjBwaXBlbGluZSU2MCUyMGlzJTIwYWxyZWFkeSUyMGxvYWRlZCUyMHdpdGglMjB0aGUlMjBMb1JBJTIwcGFyYW1ldGVycy4lMEFwaXBlbGluZS51bmxvYWRfbG9yYV93ZWlnaHRzKCklMEEuLi4=",highlighted:'<span class="hljs-meta">>>> </span><span class="hljs-comment"># Assuming `pipeline` is already loaded with the LoRA parameters.</span>\n<span class="hljs-meta">>>> </span>pipeline.unload_lora_weights()\n<span class="hljs-meta">>>> </span>...',wrap:!1}}),{c(){r=n("p"),r.textContent=b,l=t(),p(c.$$.fragment)},l(d){r=s(d,"P",{"data-svelte-h":!0}),f(r)!=="svelte-kvfsh7"&&(r.textContent=b),l=a(d),m(c.$$.fragment,d)},m(d,y){L(d,r,y),L(d,l,y),h(c,d,y),M=!0},p:C,i(d){M||(_(c.$$.fragment,d),M=!0)},o(d){u(c.$$.fragment,d),M=!1},d(d){d&&(i(r),i(l)),g(c,d)}}}function Gf(T){let r,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,M="This function is experimental and might change in the future.";return{c(){r=n("p"),r.textContent=b,l=t(),c=n("p"),c.textContent=M},l(d){r=s(d,"P",{"data-svelte-h":!0}),f(r)!=="svelte-15l1sdn"&&(r.textContent=b),l=a(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=M)},m(d,y){L(d,r,y),L(d,l,y),L(d,c,y)},p:C,d(d){d&&(i(r),i(l),i(c))}}}function Jf(T){let r,b="This is an experimental API.";return{c(){r=n("p"),r.textContent=b},l(l){r=s(l,"P",{"data-svelte-h":!0}),f(r)!=="svelte-8w79b9"&&(r.textContent=b)},m(l,c){L(l,r,c)},p:C,d(l){l&&i(r)}}}function Zf(T){let r,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,M="This function is experimental and might change in the future.";return{c(){r=n("p"),r.textContent=b,l=t(),c=n("p"),c.textContent=M},l(d){r=s(d,"P",{"data-svelte-h":!0}),f(r)!=="svelte-15l1sdn"&&(r.textContent=b),l=a(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=M)},m(d,y){L(d,r,y),L(d,l,y),L(d,c,y)},p:C,d(d){d&&(i(r),i(l),i(c))}}}function Yf(T){let r,b="This is an experimental API.";return{c(){r=n("p"),r.textContent=b},l(l){r=s(l,"P",{"data-svelte-h":!0}),f(r)!=="svelte-8w79b9"&&(r.textContent=b)},m(l,c){L(l,r,c)},p:C,d(l){l&&i(r)}}}function Of(T){let r,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,M="This function is experimental and might change in the future.";return{c(){r=n("p"),r.textContent=b,l=t(),c=n("p"),c.textContent=M},l(d){r=s(d,"P",{"data-svelte-h":!0}),f(r)!=="svelte-15l1sdn"&&(r.textContent=b),l=a(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=M)},m(d,y){L(d,r,y),L(d,l,y),L(d,c,y)},p:C,d(d){d&&(i(r),i(l),i(c))}}}function Qf(T){let r,b="This is an experimental API.";return{c(){r=n("p"),r.textContent=b},l(l){r=s(l,"P",{"data-svelte-h":!0}),f(r)!=="svelte-8w79b9"&&(r.textContent=b)},m(l,c){L(l,r,c)},p:C,d(l){l&&i(r)}}}function Kf(T){let r,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,M="This function is experimental and might change in the future.";return{c(){r=n("p"),r.textContent=b,l=t(),c=n("p"),c.textContent=M},l(d){r=s(d,"P",{"data-svelte-h":!0}),f(r)!=="svelte-15l1sdn"&&(r.textContent=b),l=a(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=M)},m(d,y){L(d,r,y),L(d,l,y),L(d,c,y)},p:C,d(d){d&&(i(r),i(l),i(c))}}}function ep(T){let r,b="This is an experimental API.";return{c(){r=n("p"),r.textContent=b},l(l){r=s(l,"P",{"data-svelte-h":!0}),f(r)!=="svelte-8w79b9"&&(r.textContent=b)},m(l,c){L(l,r,c)},p:C,d(l){l&&i(r)}}}function op(T){let r,b="We support loading original format HunyuanVideo LoRA checkpoints.",l,c,M="This function is experimental and might change in the future.";return{c(){r=n("p"),r.textContent=b,l=t(),c=n("p"),c.textContent=M},l(d){r=s(d,"P",{"data-svelte-h":!0}),f(r)!=="svelte-gyrs6h"&&(r.textContent=b),l=a(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=M)},m(d,y){L(d,r,y),L(d,l,y),L(d,c,y)},p:C,d(d){d&&(i(r),i(l),i(c))}}}function tp(T){let r,b="This is an experimental API.";return{c(){r=n("p"),r.textContent=b},l(l){r=s(l,"P",{"data-svelte-h":!0}),f(r)!=="svelte-8w79b9"&&(r.textContent=b)},m(l,c){L(l,r,c)},p:C,d(l){l&&i(r)}}}function ap(T){let r,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,M="This function is experimental and might change in the future.";return{c(){r=n("p"),r.textContent=b,l=t(),c=n("p"),c.textContent=M},l(d){r=s(d,"P",{"data-svelte-h":!0}),f(r)!=="svelte-15l1sdn"&&(r.textContent=b),l=a(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=M)},m(d,y){L(d,r,y),L(d,l,y),L(d,c,y)},p:C,d(d){d&&(i(r),i(l),i(c))}}}function rp(T){let r,b="This is an experimental API.";return{c(){r=n("p"),r.textContent=b},l(l){r=s(l,"P",{"data-svelte-h":!0}),f(r)!=="svelte-8w79b9"&&(r.textContent=b)},m(l,c){L(l,r,c)},p:C,d(l){l&&i(r)}}}function np(T){let r,b="This is an experimental API.";return{c(){r=n("p"),r.textContent=b},l(l){r=s(l,"P",{"data-svelte-h":!0}),f(r)!=="svelte-8w79b9"&&(r.textContent=b)},m(l,c){L(l,r,c)},p:C,d(l){l&&i(r)}}}function sp(T){let r,b="Example:",l,c,M;return c=new Hs({props:{code:"ZnJvbSUyMGRpZmZ1c2VycyUyMGltcG9ydCUyMERpZmZ1c2lvblBpcGVsaW5lJTBBaW1wb3J0JTIwdG9yY2glMEElMEFwaXBlbGluZSUyMCUzRCUyMERpZmZ1c2lvblBpcGVsaW5lLmZyb21fcHJldHJhaW5lZCglMEElMjAlMjAlMjAlMjAlMjJzdGFiaWxpdHlhaSUyRnN0YWJsZS1kaWZmdXNpb24teGwtYmFzZS0xLjAlMjIlMkMlMjB0b3JjaF9kdHlwZSUzRHRvcmNoLmZsb2F0MTYlMEEpLnRvKCUyMmN1ZGElMjIpJTBBcGlwZWxpbmUubG9hZF9sb3JhX3dlaWdodHMoJTIybmVyaWpzJTJGcGl4ZWwtYXJ0LXhsJTIyJTJDJTIwd2VpZ2h0X25hbWUlM0QlMjJwaXhlbC1hcnQteGwuc2FmZXRlbnNvcnMlMjIlMkMlMjBhZGFwdGVyX25hbWUlM0QlMjJwaXhlbCUyMiklMEFwaXBlbGluZS5mdXNlX2xvcmEobG9yYV9zY2FsZSUzRDAuNyk=",highlighted:`<span class="hljs-keyword">from</span> diffusers <span class="hljs-keyword">import</span> DiffusionPipeline | |
| <span class="hljs-keyword">import</span> torch | |
| pipeline = DiffusionPipeline.from_pretrained( | |
| <span class="hljs-string">"stabilityai/stable-diffusion-xl-base-1.0"</span>, torch_dtype=torch.float16 | |
| ).to(<span class="hljs-string">"cuda"</span>) | |
| pipeline.load_lora_weights(<span class="hljs-string">"nerijs/pixel-art-xl"</span>, weight_name=<span class="hljs-string">"pixel-art-xl.safetensors"</span>, adapter_name=<span class="hljs-string">"pixel"</span>) | |
| pipeline.fuse_lora(lora_scale=<span class="hljs-number">0.7</span>)`,wrap:!1}}),{c(){r=n("p"),r.textContent=b,l=t(),p(c.$$.fragment)},l(d){r=s(d,"P",{"data-svelte-h":!0}),f(r)!=="svelte-11lpom8"&&(r.textContent=b),l=a(d),m(c.$$.fragment,d)},m(d,y){L(d,r,y),L(d,l,y),h(c,d,y),M=!0},p:C,i(d){M||(_(c.$$.fragment,d),M=!0)},o(d){u(c.$$.fragment,d),M=!1},d(d){d&&(i(r),i(l)),g(c,d)}}}function ip(T){let r,b="Example:",l,c,M;return c=new Hs({props:{code:"ZnJvbSUyMGRpZmZ1c2VycyUyMGltcG9ydCUyMERpZmZ1c2lvblBpcGVsaW5lJTBBJTBBcGlwZWxpbmUlMjAlM0QlMjBEaWZmdXNpb25QaXBlbGluZS5mcm9tX3ByZXRyYWluZWQoJTBBJTIwJTIwJTIwJTIwJTIyc3RhYmlsaXR5YWklMkZzdGFibGUtZGlmZnVzaW9uLXhsLWJhc2UtMS4wJTIyJTJDJTBBKS50byglMjJjdWRhJTIyKSUwQXBpcGVsaW5lLmxvYWRfbG9yYV93ZWlnaHRzKCUyMkNpcm9OMjAyMiUyRnRveS1mYWNlJTIyJTJDJTIwd2VpZ2h0X25hbWUlM0QlMjJ0b3lfZmFjZV9zZHhsLnNhZmV0ZW5zb3JzJTIyJTJDJTIwYWRhcHRlcl9uYW1lJTNEJTIydG95JTIyKSUwQXBpcGVsaW5lLmdldF9hY3RpdmVfYWRhcHRlcnMoKQ==",highlighted:`<span class="hljs-keyword">from</span> diffusers <span class="hljs-keyword">import</span> DiffusionPipeline | |
| pipeline = DiffusionPipeline.from_pretrained( | |
| <span class="hljs-string">"stabilityai/stable-diffusion-xl-base-1.0"</span>, | |
| ).to(<span class="hljs-string">"cuda"</span>) | |
| pipeline.load_lora_weights(<span class="hljs-string">"CiroN2022/toy-face"</span>, weight_name=<span class="hljs-string">"toy_face_sdxl.safetensors"</span>, adapter_name=<span class="hljs-string">"toy"</span>) | |
| pipeline.get_active_adapters()`,wrap:!1}}),{c(){r=n("p"),r.textContent=b,l=t(),p(c.$$.fragment)},l(d){r=s(d,"P",{"data-svelte-h":!0}),f(r)!=="svelte-11lpom8"&&(r.textContent=b),l=a(d),m(c.$$.fragment,d)},m(d,y){L(d,r,y),L(d,l,y),h(c,d,y),M=!0},p:C,i(d){M||(_(c.$$.fragment,d),M=!0)},o(d){u(c.$$.fragment,d),M=!1},d(d){d&&(i(r),i(l)),g(c,d)}}}function dp(T){let r,b="This is an experimental API.";return{c(){r=n("p"),r.textContent=b},l(l){r=s(l,"P",{"data-svelte-h":!0}),f(r)!=="svelte-8w79b9"&&(r.textContent=b)},m(l,c){L(l,r,c)},p:C,d(l){l&&i(r)}}}function lp(T){let r,b="Examples:",l,c,M;return c=new Hs({props:{code:"JTIzJTIwQXNzdW1pbmclMjAlNjBwaXBlbGluZSU2MCUyMGlzJTIwYWxyZWFkeSUyMGxvYWRlZCUyMHdpdGglMjB0aGUlMjBMb1JBJTIwcGFyYW1ldGVycy4lMEFwaXBlbGluZS51bmxvYWRfbG9yYV93ZWlnaHRzKCklMEEuLi4=",highlighted:'<span class="hljs-meta">>>> </span><span class="hljs-comment"># Assuming `pipeline` is already loaded with the LoRA parameters.</span>\n<span class="hljs-meta">>>> </span>pipeline.unload_lora_weights()\n<span class="hljs-meta">>>> </span>...',wrap:!1}}),{c(){r=n("p"),r.textContent=b,l=t(),p(c.$$.fragment)},l(d){r=s(d,"P",{"data-svelte-h":!0}),f(r)!=="svelte-kvfsh7"&&(r.textContent=b),l=a(d),m(c.$$.fragment,d)},m(d,y){L(d,r,y),L(d,l,y),h(c,d,y),M=!0},p:C,i(d){M||(_(c.$$.fragment,d),M=!0)},o(d){u(c.$$.fragment,d),M=!1},d(d){d&&(i(r),i(l)),g(c,d)}}}function cp(T){let r,b,l,c,M,d,y,Jl='LoRA is a fast and lightweight training method that inserts and trains a significantly smaller number of parameters instead of all the model parameters. This produces a smaller file (~100 MBs) and makes it easier to quickly train a model to learn a new concept. LoRA weights are typically loaded into the denoiser, text encoder or both. The denoiser usually corresponds to a UNet (<a href="/docs/diffusers/pr_11105/en/api/models/unet2d-cond#diffusers.UNet2DConditionModel">UNet2DConditionModel</a>, for example) or a Transformer (<a href="/docs/diffusers/pr_11105/en/api/models/sd3_transformer2d#diffusers.SD3Transformer2DModel">SD3Transformer2DModel</a>, for example). There are several classes for loading LoRA weights:',In,qo,Zl='<li><code>StableDiffusionLoraLoaderMixin</code> provides functions for loading and unloading, fusing and unfusing, enabling and disabling, and more functions for managing LoRA weights. This class can be used with any model.</li> <li><code>StableDiffusionXLLoraLoaderMixin</code> is a <a href="../../api/pipelines/stable_diffusion/stable_diffusion_xl">Stable Diffusion (SDXL)</a> version of the <code>StableDiffusionLoraLoaderMixin</code> class for loading and saving LoRA weights. It can only be used with the SDXL model.</li> <li><code>SD3LoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/blog/sd3" rel="nofollow">Stable Diffusion 3</a>.</li> <li><code>FluxLoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/flux" rel="nofollow">Flux</a>.</li> <li><code>CogVideoXLoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/cogvideox" rel="nofollow">CogVideoX</a>.</li> <li><code>Mochi1LoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/mochi" rel="nofollow">Mochi</a>.</li> <li><code>LTXVideoLoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/ltx_video" rel="nofollow">LTX-Video</a>.</li> <li><code>SanaLoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/sana" rel="nofollow">Sana</a>.</li> <li><code>HunyuanVideoLoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/hunyuan_video" rel="nofollow">HunyuanVideo</a>.</li> <li><code>Lumina2LoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/lumina2" rel="nofollow">Lumina2</a>.</li> <li><code>AmusedLoraLoaderMixin</code> is for the <a href="/docs/diffusers/pr_11105/en/api/pipelines/amused#diffusers.AmusedPipeline">AmusedPipeline</a>.</li> <li><code>LoraBaseMixin</code> provides a base class with several utility methods to fuse, unfuse, unload, LoRAs and more.</li>',Fn,Ie,Un,No,Xn,V,zo,Vs,Ha,Yl=`Load LoRA layers into Stable Diffusion <a href="/docs/diffusers/pr_11105/en/api/models/unet2d-cond#diffusers.UNet2DConditionModel">UNet2DConditionModel</a> and | |
| <a href="https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModel" rel="nofollow"><code>CLIPTextModel</code></a>.`,Is,Fe,Bo,Fs,Va,Ol="This will load the LoRA layers specified in <code>state_dict</code> into <code>text_encoder</code>",Us,Ue,jo,Xs,Ia,Ql="This will load the LoRA layers specified in <code>state_dict</code> into <code>unet</code>.",Es,B,Go,Ws,Fa,Kl=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.unet</code> and | |
| <code>self.text_encoder</code>.`,qs,Ua,ec="All kwargs are forwarded to <code>self.lora_state_dict</code>.",Ns,Xa,oc=`See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is | |
| loaded.`,zs,Ea,tc=`See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet">load_lora_into_unet()</a> for more details on how the state dict is | |
| loaded into <code>self.unet</code>.`,Bs,Wa,ac=`See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder">load_lora_into_text_encoder()</a> for more details on how the state | |
| dict is loaded into <code>self.text_encoder</code>.`,js,ie,Jo,Gs,qa,rc="Return state dict for lora weights and the network alphas.",Js,Xe,Zs,Ee,Zo,Ys,Na,nc="Save the LoRA parameters corresponding to the UNet and text encoder.",En,Yo,Wn,I,Oo,Os,za,sc=`Load LoRA layers into Stable Diffusion XL <a href="/docs/diffusers/pr_11105/en/api/models/unet2d-cond#diffusers.UNet2DConditionModel">UNet2DConditionModel</a>, | |
| <a href="https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModel" rel="nofollow"><code>CLIPTextModel</code></a>, and | |
| <a href="https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModelWithProjection" rel="nofollow"><code>CLIPTextModelWithProjection</code></a>.`,Qs,We,Qo,Ks,Ba,ic="This will load the LoRA layers specified in <code>state_dict</code> into <code>text_encoder</code>",ei,qe,Ko,oi,ja,dc="This will load the LoRA layers specified in <code>state_dict</code> into <code>unet</code>.",ti,j,et,ai,Ga,lc=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.unet</code> and | |
| <code>self.text_encoder</code>.`,ri,Ja,cc="All kwargs are forwarded to <code>self.lora_state_dict</code>.",ni,Za,fc=`See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is | |
| loaded.`,si,Ya,pc=`See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet">load_lora_into_unet()</a> for more details on how the state dict is | |
| loaded into <code>self.unet</code>.`,ii,Oa,mc=`See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder">load_lora_into_text_encoder()</a> for more details on how the state | |
| dict is loaded into <code>self.text_encoder</code>.`,di,de,ot,li,Qa,hc="Return state dict for lora weights and the network alphas.",ci,Ne,fi,ze,tt,pi,Ka,_c="Save the LoRA parameters corresponding to the UNet and text encoder.",qn,at,Nn,k,rt,mi,er,uc=`Load LoRA layers into <a href="/docs/diffusers/pr_11105/en/api/models/sd3_transformer2d#diffusers.SD3Transformer2DModel">SD3Transformer2DModel</a>, | |
| <a href="https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModel" rel="nofollow"><code>CLIPTextModel</code></a>, and | |
| <a href="https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModelWithProjection" rel="nofollow"><code>CLIPTextModelWithProjection</code></a>.`,hi,or,gc='Specific to <a href="/docs/diffusers/pr_11105/en/api/pipelines/stable_diffusion/stable_diffusion_3#diffusers.StableDiffusion3Pipeline">StableDiffusion3Pipeline</a>.',_i,Be,nt,ui,tr,xc="This will load the LoRA layers specified in <code>state_dict</code> into <code>text_encoder</code>",gi,je,st,xi,ar,Lc="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",Li,Z,it,bi,rr,bc=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.unet</code> and | |
| <code>self.text_encoder</code>.`,wi,nr,wc="All kwargs are forwarded to <code>self.lora_state_dict</code>.",vi,sr,vc=`See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is | |
| loaded.`,$i,ir,$c=`See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,yi,le,dt,Mi,dr,yc="Return state dict for lora weights and the network alphas.",Ti,Ge,Di,Je,lt,Si,lr,Mc="Save the LoRA parameters corresponding to the UNet and text encoder.",Ci,ce,ct,ki,cr,Tc=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,Ai,Ze,zn,ft,Bn,D,pt,Ri,fr,Dc=`Load LoRA layers into <a href="/docs/diffusers/pr_11105/en/api/models/flux_transformer#diffusers.FluxTransformer2DModel">FluxTransformer2DModel</a>, | |
| <a href="https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModel" rel="nofollow"><code>CLIPTextModel</code></a>.`,Pi,pr,Sc='Specific to <a href="/docs/diffusers/pr_11105/en/api/pipelines/stable_diffusion/stable_diffusion_3#diffusers.StableDiffusion3Pipeline">StableDiffusion3Pipeline</a>.',Hi,Ye,mt,Vi,mr,Cc="This will load the LoRA layers specified in <code>state_dict</code> into <code>text_encoder</code>",Ii,Oe,ht,Fi,hr,kc="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",Ui,Y,_t,Xi,_r,Ac=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>.`,Ei,ur,Rc="All kwargs are forwarded to <code>self.lora_state_dict</code>.",Wi,gr,Pc=`See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is | |
| loaded.`,qi,xr,Hc=`See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,Ni,fe,ut,zi,Lr,Vc="Return state dict for lora weights and the network alphas.",Bi,Qe,ji,Ke,gt,Gi,br,Ic="Save the LoRA parameters corresponding to the UNet and text encoder.",Ji,pe,xt,Zi,wr,Fc=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,Yi,eo,Oi,me,Lt,Qi,vr,Uc="Unloads the LoRA parameters.",Ki,oo,jn,bt,Gn,F,wt,ed,$r,Xc='Load LoRA layers into <a href="/docs/diffusers/pr_11105/en/api/models/cogvideox_transformer3d#diffusers.CogVideoXTransformer3DModel">CogVideoXTransformer3DModel</a>. Specific to <a href="/docs/diffusers/pr_11105/en/api/pipelines/cogvideox#diffusers.CogVideoXPipeline">CogVideoXPipeline</a>.',od,to,vt,td,yr,Ec="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",ad,ao,$t,rd,Mr,Wc=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,nd,he,yt,sd,Tr,qc="Return state dict for lora weights and the network alphas.",id,ro,dd,no,Mt,ld,Dr,Nc="Save the LoRA parameters corresponding to the UNet and text encoder.",cd,_e,Tt,fd,Sr,zc=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,pd,so,Jn,Dt,Zn,U,St,md,Cr,Bc='Load LoRA layers into <a href="/docs/diffusers/pr_11105/en/api/models/mochi_transformer3d#diffusers.MochiTransformer3DModel">MochiTransformer3DModel</a>. Specific to <a href="/docs/diffusers/pr_11105/en/api/pipelines/mochi#diffusers.MochiPipeline">MochiPipeline</a>.',hd,io,Ct,_d,kr,jc="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",ud,lo,kt,gd,Ar,Gc=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,xd,ue,At,Ld,Rr,Jc="Return state dict for lora weights and the network alphas.",bd,co,wd,fo,Rt,vd,Pr,Zc="Save the LoRA parameters corresponding to the UNet and text encoder.",$d,ge,Pt,yd,Hr,Yc=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,Md,po,Yn,Ht,On,X,Vt,Td,Vr,Oc='Load LoRA layers into <a href="/docs/diffusers/pr_11105/en/api/models/ltx_video_transformer3d#diffusers.LTXVideoTransformer3DModel">LTXVideoTransformer3DModel</a>. Specific to <a href="/docs/diffusers/pr_11105/en/api/pipelines/ltx_video#diffusers.LTXPipeline">LTXPipeline</a>.',Dd,mo,It,Sd,Ir,Qc="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",Cd,ho,Ft,kd,Fr,Kc=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,Ad,xe,Ut,Rd,Ur,ef="Return state dict for lora weights and the network alphas.",Pd,_o,Hd,uo,Xt,Vd,Xr,of="Save the LoRA parameters corresponding to the UNet and text encoder.",Id,Le,Et,Fd,Er,tf=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,Ud,go,Qn,Wt,Kn,E,qt,Xd,Wr,af='Load LoRA layers into <a href="/docs/diffusers/pr_11105/en/api/models/sana_transformer2d#diffusers.SanaTransformer2DModel">SanaTransformer2DModel</a>. Specific to <a href="/docs/diffusers/pr_11105/en/api/pipelines/sana#diffusers.SanaPipeline">SanaPipeline</a>.',Ed,xo,Nt,Wd,qr,rf="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",qd,Lo,zt,Nd,Nr,nf=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,zd,be,Bt,Bd,zr,sf="Return state dict for lora weights and the network alphas.",jd,bo,Gd,wo,jt,Jd,Br,df="Save the LoRA parameters corresponding to the UNet and text encoder.",Zd,we,Gt,Yd,jr,lf=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,Od,vo,es,Jt,os,W,Zt,Qd,Gr,cf='Load LoRA layers into <a href="/docs/diffusers/pr_11105/en/api/models/hunyuan_video_transformer_3d#diffusers.HunyuanVideoTransformer3DModel">HunyuanVideoTransformer3DModel</a>. Specific to <a href="/docs/diffusers/pr_11105/en/api/pipelines/hunyuan_video#diffusers.HunyuanVideoPipeline">HunyuanVideoPipeline</a>.',Kd,$o,Yt,el,Jr,ff="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",ol,yo,Ot,tl,Zr,pf=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,al,ve,Qt,rl,Yr,mf="Return state dict for lora weights and the network alphas.",nl,Mo,sl,To,Kt,il,Or,hf="Save the LoRA parameters corresponding to the UNet and text encoder.",dl,$e,ea,ll,Qr,_f=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,cl,Do,ts,oa,as,q,ta,fl,Kr,uf='Load LoRA layers into <a href="/docs/diffusers/pr_11105/en/api/models/lumina2_transformer2d#diffusers.Lumina2Transformer2DModel">Lumina2Transformer2DModel</a>. Specific to <code>Lumina2Text2ImgPipeline</code>.',pl,So,aa,ml,en,gf="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",hl,Co,ra,_l,on,xf=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,ul,ye,na,gl,tn,Lf="Return state dict for lora weights and the network alphas.",xl,ko,Ll,Ao,sa,bl,an,bf="Save the LoRA parameters corresponding to the UNet and text encoder.",wl,Me,ia,vl,rn,wf=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,$l,Ro,rs,da,ns,ke,la,yl,Po,ca,Ml,nn,vf="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",ss,fa,is,S,pa,Tl,sn,$f="Utility class for handling LoRAs.",Dl,dn,ma,Sl,Te,ha,Cl,ln,yf="Enables the possibility to hotswap LoRA adapters.",kl,cn,Mf=`Calling this method is only required when hotswapping adapters and if the model is compiled or if the ranks of | |
| the loaded adapters differ.`,Al,se,_a,Rl,fn,Tf="Fuses the LoRA parameters into the original parameters of the corresponding blocks.",Pl,Ho,Hl,Vo,Vl,De,ua,Il,pn,Df="Gets the list of the current active adapters.",Fl,Io,Ul,Fo,ga,Xl,mn,Sf="Gets the current list of all available adapters in the pipeline.",El,Uo,xa,Wl,hn,Cf=`Moves the LoRAs listed in <code>adapter_names</code> to a target device. Useful for offloading the LoRA to the CPU in case | |
| you want to load multiple adapters and free some GPU memory.`,ql,Se,La,Nl,_n,kf=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,zl,Xo,Bl,Ce,ba,jl,un,Af="Unloads the LoRA parameters.",Gl,Eo,ds,wa,ls,Vn,cs;return M=new ne({props:{title:"LoRA",local:"lora",headingTag:"h1"}}),Ie=new P({props:{$$slots:{default:[Xf]},$$scope:{ctx:T}}}),No=new ne({props:{title:"StableDiffusionLoraLoaderMixin",local:"diffusers.loaders.StableDiffusionLoraLoaderMixin",headingTag:"h2"}}),zo=new $({props:{name:"class diffusers.loaders.StableDiffusionLoraLoaderMixin",anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L71"}}),Bo=new $({props:{name:"load_lora_into_text_encoder",anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder",parameters:[{name:"state_dict",val:""},{name:"network_alphas",val:""},{name:"text_encoder",val:""},{name:"prefix",val:" = None"},{name:"lora_scale",val:" = 1.0"},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The key should be prefixed with an | |
| additional <code>text_encoder</code> to distinguish between unet lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder.network_alphas",description:`<strong>network_alphas</strong> (<code>Dict[str, float]</code>) — | |
| The value of the network alpha used for stable learning and preventing underflow. This value has the | |
| same meaning as the <code>--network_alpha</code> option in the kohya-ss trainer script. Refer to <a href="https://github.com/darkstorm2150/sd-scripts/blob/main/docs/train_network_README-en.md#execute-learning" rel="nofollow">this | |
| link</a>.`,name:"network_alphas"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder.text_encoder",description:`<strong>text_encoder</strong> (<code>CLIPTextModel</code>) — | |
| The text encoder model to load the LoRA layers into.`,name:"text_encoder"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder.prefix",description:`<strong>prefix</strong> (<code>str</code>) — | |
| Expected prefix of the <code>text_encoder</code> in the <code>state_dict</code>.`,name:"prefix"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder.lora_scale",description:`<strong>lora_scale</strong> (<code>float</code>) — | |
| How much to scale the output of the lora linear layer before it is added with the output of the regular | |
| lora layer.`,name:"lora_scale"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L371"}}),jo=new $({props:{name:"load_lora_into_unet",anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet",parameters:[{name:"state_dict",val:""},{name:"network_alphas",val:""},{name:"unet",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet.network_alphas",description:`<strong>network_alphas</strong> (<code>Dict[str, float]</code>) — | |
| The value of the network alpha used for stable learning and preventing underflow. This value has the | |
| same meaning as the <code>--network_alpha</code> option in the kohya-ss trainer script. Refer to <a href="https://github.com/darkstorm2150/sd-scripts/blob/main/docs/train_network_README-en.md#execute-learning" rel="nofollow">this | |
| link</a>.`,name:"network_alphas"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet.unet",description:`<strong>unet</strong> (<code>UNet2DConditionModel</code>) — | |
| The UNet model to load the LoRA layers into.`,name:"unet"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L294"}}),Go=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:" = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L81"}}),Jo=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11105/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.weight_name",description:`<strong>weight_name</strong> (<code>str</code>, <em>optional</em>, defaults to None) — | |
| Name of the serialized state dict file.`,name:"weight_name"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L179"}}),Xe=new P({props:{warning:!0,$$slots:{default:[Ef]},$$scope:{ctx:T}}}),Zo=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"unet_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"text_encoder_lora_layers",val:": typing.Dict[str, torch.nn.modules.module.Module] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.save_lora_weights.unet_lora_layers",description:`<strong>unet_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>unet</code>.`,name:"unet_lora_layers"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.save_lora_weights.text_encoder_lora_layers",description:`<strong>text_encoder_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>text_encoder</code>. Must explicitly pass the text | |
| encoder LoRA state dict because it comes from 🤗 Transformers.`,name:"text_encoder_lora_layers"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L445"}}),Yo=new ne({props:{title:"StableDiffusionXLLoraLoaderMixin",local:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin",headingTag:"h2"}}),Oo=new $({props:{name:"class diffusers.loaders.StableDiffusionXLLoraLoaderMixin",anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L567"}}),Qo=new $({props:{name:"load_lora_into_text_encoder",anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder",parameters:[{name:"state_dict",val:""},{name:"network_alphas",val:""},{name:"text_encoder",val:""},{name:"prefix",val:" = None"},{name:"lora_scale",val:" = 1.0"},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The key should be prefixed with an | |
| additional <code>text_encoder</code> to distinguish between unet lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder.network_alphas",description:`<strong>network_alphas</strong> (<code>Dict[str, float]</code>) — | |
| The value of the network alpha used for stable learning and preventing underflow. This value has the | |
| same meaning as the <code>--network_alpha</code> option in the kohya-ss trainer script. Refer to <a href="https://github.com/darkstorm2150/sd-scripts/blob/main/docs/train_network_README-en.md#execute-learning" rel="nofollow">this | |
| link</a>.`,name:"network_alphas"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder.text_encoder",description:`<strong>text_encoder</strong> (<code>CLIPTextModel</code>) — | |
| The text encoder model to load the LoRA layers into.`,name:"text_encoder"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder.prefix",description:`<strong>prefix</strong> (<code>str</code>) — | |
| Expected prefix of the <code>text_encoder</code> in the <code>state_dict</code>.`,name:"prefix"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder.lora_scale",description:`<strong>lora_scale</strong> (<code>float</code>) — | |
| How much to scale the output of the lora linear layer before it is added with the output of the regular | |
| lora layer.`,name:"lora_scale"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L862"}}),Ko=new $({props:{name:"load_lora_into_unet",anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_unet",parameters:[{name:"state_dict",val:""},{name:"network_alphas",val:""},{name:"unet",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_unet.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_unet.network_alphas",description:`<strong>network_alphas</strong> (<code>Dict[str, float]</code>) — | |
| The value of the network alpha used for stable learning and preventing underflow. This value has the | |
| same meaning as the <code>--network_alpha</code> option in the kohya-ss trainer script. Refer to <a href="https://github.com/darkstorm2150/sd-scripts/blob/main/docs/train_network_README-en.md#execute-learning" rel="nofollow">this | |
| link</a>.`,name:"network_alphas"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_unet.unet",description:`<strong>unet</strong> (<code>UNet2DConditionModel</code>) — | |
| The UNet model to load the LoRA layers into.`,name:"unet"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_unet.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_unet.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_unet.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L784"}}),et=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:": typing.Optional[str] = None"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L578"}}),ot=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11105/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.weight_name",description:`<strong>weight_name</strong> (<code>str</code>, <em>optional</em>, defaults to None) — | |
| Name of the serialized state dict file.`,name:"weight_name"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L668"}}),Ne=new P({props:{warning:!0,$$slots:{default:[Wf]},$$scope:{ctx:T}}}),tt=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"unet_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"text_encoder_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"text_encoder_2_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights.unet_lora_layers",description:`<strong>unet_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>unet</code>.`,name:"unet_lora_layers"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights.text_encoder_lora_layers",description:`<strong>text_encoder_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>text_encoder</code>. Must explicitly pass the text | |
| encoder LoRA state dict because it comes from 🤗 Transformers.`,name:"text_encoder_lora_layers"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights.text_encoder_2_lora_layers",description:`<strong>text_encoder_2_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>text_encoder_2</code>. Must explicitly pass the text | |
| encoder LoRA state dict because it comes from 🤗 Transformers.`,name:"text_encoder_2_lora_layers"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L937"}}),at=new ne({props:{title:"SD3LoraLoaderMixin",local:"diffusers.loaders.SD3LoraLoaderMixin",headingTag:"h2"}}),rt=new $({props:{name:"class diffusers.loaders.SD3LoraLoaderMixin",anchor:"diffusers.loaders.SD3LoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L1067"}}),nt=new $({props:{name:"load_lora_into_text_encoder",anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder",parameters:[{name:"state_dict",val:""},{name:"network_alphas",val:""},{name:"text_encoder",val:""},{name:"prefix",val:" = None"},{name:"lora_scale",val:" = 1.0"},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The key should be prefixed with an | |
| additional <code>text_encoder</code> to distinguish between unet lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder.network_alphas",description:`<strong>network_alphas</strong> (<code>Dict[str, float]</code>) — | |
| The value of the network alpha used for stable learning and preventing underflow. This value has the | |
| same meaning as the <code>--network_alpha</code> option in the kohya-ss trainer script. Refer to <a href="https://github.com/darkstorm2150/sd-scripts/blob/main/docs/train_network_README-en.md#execute-learning" rel="nofollow">this | |
| link</a>.`,name:"network_alphas"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder.text_encoder",description:`<strong>text_encoder</strong> (<code>CLIPTextModel</code>) — | |
| The text encoder model to load the LoRA layers into.`,name:"text_encoder"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder.prefix",description:`<strong>prefix</strong> (<code>str</code>) — | |
| Expected prefix of the <code>text_encoder</code> in the <code>state_dict</code>.`,name:"prefix"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder.lora_scale",description:`<strong>lora_scale</strong> (<code>float</code>) — | |
| How much to scale the output of the lora linear layer before it is added with the output of the regular | |
| lora layer.`,name:"lora_scale"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L1342"}}),st=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>SD3Transformer2DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L1282"}}),it=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:" = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L1177"}}),dt=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11105/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L1080"}}),Ge=new P({props:{warning:!0,$$slots:{default:[qf]},$$scope:{ctx:T}}}),lt=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.SD3LoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"text_encoder_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"text_encoder_2_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.SD3LoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.save_lora_weights.text_encoder_lora_layers",description:`<strong>text_encoder_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>text_encoder</code>. Must explicitly pass the text | |
| encoder LoRA state dict because it comes from 🤗 Transformers.`,name:"text_encoder_lora_layers"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.save_lora_weights.text_encoder_2_lora_layers",description:`<strong>text_encoder_2_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>text_encoder_2</code>. Must explicitly pass the text | |
| encoder LoRA state dict because it comes from 🤗 Transformers.`,name:"text_encoder_2_lora_layers"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L1417"}}),ct=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.SD3LoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer', 'text_encoder', 'text_encoder_2']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.SD3LoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.unfuse_lora.unfuse_text_encoder",description:`<strong>unfuse_text_encoder</strong> (<code>bool</code>, defaults to <code>True</code>) — | |
| Whether to unfuse the text encoder LoRA parameters. If the text encoder wasn’t monkey-patched with the | |
| LoRA parameters then it won’t have any effect.`,name:"unfuse_text_encoder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L1529"}}),Ze=new P({props:{warning:!0,$$slots:{default:[Nf]},$$scope:{ctx:T}}}),ft=new ne({props:{title:"FluxLoraLoaderMixin",local:"diffusers.loaders.FluxLoraLoaderMixin",headingTag:"h2"}}),pt=new $({props:{name:"class diffusers.loaders.FluxLoraLoaderMixin",anchor:"diffusers.loaders.FluxLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L1550"}}),mt=new $({props:{name:"load_lora_into_text_encoder",anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder",parameters:[{name:"state_dict",val:""},{name:"network_alphas",val:""},{name:"text_encoder",val:""},{name:"prefix",val:" = None"},{name:"lora_scale",val:" = 1.0"},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The key should be prefixed with an | |
| additional <code>text_encoder</code> to distinguish between unet lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder.network_alphas",description:`<strong>network_alphas</strong> (<code>Dict[str, float]</code>) — | |
| The value of the network alpha used for stable learning and preventing underflow. This value has the | |
| same meaning as the <code>--network_alpha</code> option in the kohya-ss trainer script. Refer to <a href="https://github.com/darkstorm2150/sd-scripts/blob/main/docs/train_network_README-en.md#execute-learning" rel="nofollow">this | |
| link</a>.`,name:"network_alphas"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder.text_encoder",description:`<strong>text_encoder</strong> (<code>CLIPTextModel</code>) — | |
| The text encoder model to load the LoRA layers into.`,name:"text_encoder"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder.prefix",description:`<strong>prefix</strong> (<code>str</code>) — | |
| Expected prefix of the <code>text_encoder</code> in the <code>state_dict</code>.`,name:"prefix"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder.lora_scale",description:`<strong>lora_scale</strong> (<code>float</code>) — | |
| How much to scale the output of the lora linear layer before it is added with the output of the regular | |
| lora layer.`,name:"lora_scale"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L1962"}}),ht=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"network_alphas",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_transformer.network_alphas",description:`<strong>network_alphas</strong> (<code>Dict[str, float]</code>) — | |
| The value of the network alpha used for stable learning and preventing underflow. This value has the | |
| same meaning as the <code>--network_alpha</code> option in the kohya-ss trainer script. Refer to <a href="https://github.com/darkstorm2150/sd-scripts/blob/main/docs/train_network_README-en.md#execute-learning" rel="nofollow">this | |
| link</a>.`,name:"network_alphas"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>FluxTransformer2DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L1837"}}),_t=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:" = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| \`Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter. If the new | |
| adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need to call an | |
| additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L1697"}}),ut=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"return_alphas",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11105/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L1563"}}),Qe=new P({props:{warning:!0,$$slots:{default:[zf]},$$scope:{ctx:T}}}),gt=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.FluxLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"text_encoder_lora_layers",val:": typing.Dict[str, torch.nn.modules.module.Module] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.FluxLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.save_lora_weights.text_encoder_lora_layers",description:`<strong>text_encoder_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>text_encoder</code>. Must explicitly pass the text | |
| encoder LoRA state dict because it comes from 🤗 Transformers.`,name:"text_encoder_lora_layers"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L2037"}}),xt=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.FluxLoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer', 'text_encoder']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.FluxLoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L2152"}}),eo=new P({props:{warning:!0,$$slots:{default:[Bf]},$$scope:{ctx:T}}}),Lt=new $({props:{name:"unload_lora_weights",anchor:"diffusers.loaders.FluxLoraLoaderMixin.unload_lora_weights",parameters:[{name:"reset_to_overwritten_params",val:" = False"}],parametersDescription:[{anchor:"diffusers.loaders.FluxLoraLoaderMixin.unload_lora_weights.reset_to_overwritten_params",description:`<strong>reset_to_overwritten_params</strong> (<code>bool</code>, defaults to <code>False</code>) — Whether to reset the LoRA-loaded modules | |
| to their original params. Refer to the <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/flux" rel="nofollow">Flux | |
| documentation</a> to learn more.`,name:"reset_to_overwritten_params"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L2173"}}),oo=new Ps({props:{anchor:"diffusers.loaders.FluxLoraLoaderMixin.unload_lora_weights.example",$$slots:{default:[jf]},$$scope:{ctx:T}}}),bt=new ne({props:{title:"CogVideoXLoraLoaderMixin",local:"diffusers.loaders.CogVideoXLoraLoaderMixin",headingTag:"h2"}}),wt=new $({props:{name:"class diffusers.loaders.CogVideoXLoraLoaderMixin",anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L2644"}}),vt=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>CogVideoXTransformer3DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L2800"}}),$t=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:" = None"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L2750"}}),yt=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11105/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L2652"}}),ro=new P({props:{warning:!0,$$slots:{default:[Gf]},$$scope:{ctx:T}}}),Mt=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L2861"}}),Tt=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L2956"}}),so=new P({props:{warning:!0,$$slots:{default:[Jf]},$$scope:{ctx:T}}}),Dt=new ne({props:{title:"Mochi1LoraLoaderMixin",local:"diffusers.loaders.Mochi1LoraLoaderMixin",headingTag:"h2"}}),St=new $({props:{name:"class diffusers.loaders.Mochi1LoraLoaderMixin",anchor:"diffusers.loaders.Mochi1LoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L2974"}}),Ct=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>MochiTransformer3DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3131"}}),kt=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:" = None"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3081"}}),At=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11105/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L2982"}}),co=new P({props:{warning:!0,$$slots:{default:[Zf]},$$scope:{ctx:T}}}),Rt=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3192"}}),Pt=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3289"}}),po=new P({props:{warning:!0,$$slots:{default:[Yf]},$$scope:{ctx:T}}}),Ht=new ne({props:{title:"LTXVideoLoraLoaderMixin",local:"diffusers.loaders.LTXVideoLoraLoaderMixin",headingTag:"h2"}}),Vt=new $({props:{name:"class diffusers.loaders.LTXVideoLoraLoaderMixin",anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3307"}}),It=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>LTXVideoTransformer3DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3464"}}),Ft=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:" = None"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3414"}}),Ut=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11105/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3315"}}),_o=new P({props:{warning:!0,$$slots:{default:[Of]},$$scope:{ctx:T}}}),Xt=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3525"}}),Et=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3622"}}),go=new P({props:{warning:!0,$$slots:{default:[Qf]},$$scope:{ctx:T}}}),Wt=new ne({props:{title:"SanaLoraLoaderMixin",local:"diffusers.loaders.SanaLoraLoaderMixin",headingTag:"h2"}}),qt=new $({props:{name:"class diffusers.loaders.SanaLoraLoaderMixin",anchor:"diffusers.loaders.SanaLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3640"}}),Nt=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>SanaTransformer2DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3797"}}),zt=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:" = None"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3747"}}),Bt=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11105/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3648"}}),bo=new P({props:{warning:!0,$$slots:{default:[Kf]},$$scope:{ctx:T}}}),jt=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.SanaLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.SanaLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3858"}}),Gt=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.SanaLoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.SanaLoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3955"}}),vo=new P({props:{warning:!0,$$slots:{default:[ep]},$$scope:{ctx:T}}}),Jt=new ne({props:{title:"HunyuanVideoLoraLoaderMixin",local:"diffusers.loaders.HunyuanVideoLoraLoaderMixin",headingTag:"h2"}}),Zt=new $({props:{name:"class diffusers.loaders.HunyuanVideoLoraLoaderMixin",anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3973"}}),Yt=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>HunyuanVideoTransformer3DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L4133"}}),Ot=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:" = None"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L4083"}}),Qt=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11105/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L3981"}}),Mo=new P({props:{warning:!0,$$slots:{default:[op]},$$scope:{ctx:T}}}),Kt=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L4194"}}),ea=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L4291"}}),Do=new P({props:{warning:!0,$$slots:{default:[tp]},$$scope:{ctx:T}}}),oa=new ne({props:{title:"Lumina2LoraLoaderMixin",local:"diffusers.loaders.Lumina2LoraLoaderMixin",headingTag:"h2"}}),ta=new $({props:{name:"class diffusers.loaders.Lumina2LoraLoaderMixin",anchor:"diffusers.loaders.Lumina2LoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L4309"}}),aa=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>Lumina2Transformer2DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L4470"}}),ra=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:" = None"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11105/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L4420"}}),na=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11105/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L4317"}}),ko=new P({props:{warning:!0,$$slots:{default:[ap]},$$scope:{ctx:T}}}),sa=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L4531"}}),ia=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L4628"}}),Ro=new P({props:{warning:!0,$$slots:{default:[rp]},$$scope:{ctx:T}}}),da=new ne({props:{title:"AmusedLoraLoaderMixin",local:"diffusers.loaders.AmusedLoraLoaderMixin",headingTag:"h2"}}),la=new $({props:{name:"class diffusers.loaders.AmusedLoraLoaderMixin",anchor:"diffusers.loaders.AmusedLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L2437"}}),ca=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.AmusedLoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"network_alphas",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.AmusedLoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.AmusedLoraLoaderMixin.load_lora_into_transformer.network_alphas",description:`<strong>network_alphas</strong> (<code>Dict[str, float]</code>) — | |
| The value of the network alpha used for stable learning and preventing underflow. This value has the | |
| same meaning as the <code>--network_alpha</code> option in the kohya-ss trainer script. Refer to <a href="https://github.com/darkstorm2150/sd-scripts/blob/main/docs/train_network_README-en.md#execute-learning" rel="nofollow">this | |
| link</a>.`,name:"network_alphas"},{anchor:"diffusers.loaders.AmusedLoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>UVit2DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.AmusedLoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.AmusedLoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.AmusedLoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> — (<code>bool</code>, <em>optional</em>) | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_pipeline.py#L2442"}}),fa=new ne({props:{title:"LoraBaseMixin",local:"diffusers.loaders.lora_base.LoraBaseMixin",headingTag:"h2"}}),pa=new $({props:{name:"class diffusers.loaders.lora_base.LoraBaseMixin",anchor:"diffusers.loaders.lora_base.LoraBaseMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_base.py#L464"}}),ma=new $({props:{name:"delete_adapters",anchor:"diffusers.loaders.lora_base.LoraBaseMixin.delete_adapters",parameters:[{name:"adapter_names",val:": typing.Union[typing.List[str], str]"}],parametersDescription:[{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.delete_adapters.Deletes",description:`<strong>Deletes</strong> the LoRA layers of <code>adapter_name</code> for the unet and text-encoder(s). — | |
| adapter_names (<code>Union[List[str], str]</code>): | |
| The names of the adapter to delete. Can be a single string or a list of strings`,name:"Deletes"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_base.py#L765"}}),ha=new $({props:{name:"enable_lora_hotswap",anchor:"diffusers.loaders.lora_base.LoraBaseMixin.enable_lora_hotswap",parameters:[{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.enable_lora_hotswap.target_rank",description:`<strong>target_rank</strong> (<code>int</code>) — | |
| The highest rank among all the adapters that will be loaded.`,name:"target_rank"},{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.enable_lora_hotswap.check_compiled",description:`<strong>check_compiled</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"error"</code>) — | |
| How to handle the case when the model is already compiled, which should generally be avoided. The | |
| options are:<ul> | |
| <li>“error” (default): raise an error</li> | |
| <li>“warn”: issue a warning</li> | |
| <li>“ignore”: do nothing</li> | |
| </ul>`,name:"check_compiled"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_base.py#L917"}}),_a=new $({props:{name:"fuse_lora",anchor:"diffusers.loaders.lora_base.LoraBaseMixin.fuse_lora",parameters:[{name:"components",val:": typing.List[str] = []"},{name:"lora_scale",val:": float = 1.0"},{name:"safe_fusing",val:": bool = False"},{name:"adapter_names",val:": typing.Optional[typing.List[str]] = None"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.fuse_lora.components",description:"<strong>components</strong> — (<code>List[str]</code>): List of LoRA-injectable components to fuse the LoRAs into.",name:"components"},{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.fuse_lora.lora_scale",description:`<strong>lora_scale</strong> (<code>float</code>, defaults to 1.0) — | |
| Controls how much to influence the outputs with the LoRA parameters.`,name:"lora_scale"},{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.fuse_lora.safe_fusing",description:`<strong>safe_fusing</strong> (<code>bool</code>, defaults to <code>False</code>) — | |
| Whether to check fused weights for NaN values before fusing and if values are NaN not fusing them.`,name:"safe_fusing"},{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.fuse_lora.adapter_names",description:`<strong>adapter_names</strong> (<code>List[str]</code>, <em>optional</em>) — | |
| Adapter names to be used for fusing. If nothing is passed, all active adapters will be fused.`,name:"adapter_names"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_base.py#L531"}}),Ho=new P({props:{warning:!0,$$slots:{default:[np]},$$scope:{ctx:T}}}),Vo=new Ps({props:{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.fuse_lora.example",$$slots:{default:[sp]},$$scope:{ctx:T}}}),ua=new $({props:{name:"get_active_adapters",anchor:"diffusers.loaders.lora_base.LoraBaseMixin.get_active_adapters",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_base.py#L787"}}),Io=new Ps({props:{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.get_active_adapters.example",$$slots:{default:[ip]},$$scope:{ctx:T}}}),ga=new $({props:{name:"get_list_adapters",anchor:"diffusers.loaders.lora_base.LoraBaseMixin.get_list_adapters",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_base.py#L820"}}),xa=new $({props:{name:"set_lora_device",anchor:"diffusers.loaders.lora_base.LoraBaseMixin.set_lora_device",parameters:[{name:"adapter_names",val:": typing.List[str]"},{name:"device",val:": typing.Union[torch.device, str, int]"}],parametersDescription:[{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.set_lora_device.adapter_names",description:`<strong>adapter_names</strong> (<code>List[str]</code>) — | |
| List of adapters to send device to.`,name:"adapter_names"},{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.set_lora_device.device",description:`<strong>device</strong> (<code>Union[torch.device, str, int]</code>) — | |
| Device to send the adapters to. Can be either a torch device, a str or an integer.`,name:"device"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_base.py#L842"}}),La=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.lora_base.LoraBaseMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = []"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.unfuse_lora.unfuse_unet",description:"<strong>unfuse_unet</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_unet"},{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.unfuse_lora.unfuse_text_encoder",description:`<strong>unfuse_text_encoder</strong> (<code>bool</code>, defaults to <code>True</code>) — | |
| Whether to unfuse the text encoder LoRA parameters. If the text encoder wasn’t monkey-patched with the | |
| LoRA parameters then it won’t have any effect.`,name:"unfuse_text_encoder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_base.py#L612"}}),Xo=new P({props:{warning:!0,$$slots:{default:[dp]},$$scope:{ctx:T}}}),ba=new $({props:{name:"unload_lora_weights",anchor:"diffusers.loaders.lora_base.LoraBaseMixin.unload_lora_weights",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11105/src/diffusers/loaders/lora_base.py#L508"}}),Eo=new Ps({props:{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.unload_lora_weights.example",$$slots:{default:[lp]},$$scope:{ctx:T}}}),wa=new Uf({props:{source:"https://github.com/huggingface/diffusers/blob/main/docs/source/en/api/loaders/lora.md"}}),{c(){r=n("meta"),b=t(),l=n("p"),c=t(),p(M.$$.fragment),d=t(),y=n("p"),y.innerHTML=Jl,In=t(),qo=n("ul"),qo.innerHTML=Zl,Fn=t(),p(Ie.$$.fragment),Un=t(),p(No.$$.fragment),Xn=t(),V=n("div"),p(zo.$$.fragment),Vs=t(),Ha=n("p"),Ha.innerHTML=Yl,Is=t(),Fe=n("div"),p(Bo.$$.fragment),Fs=t(),Va=n("p"),Va.innerHTML=Ol,Us=t(),Ue=n("div"),p(jo.$$.fragment),Xs=t(),Ia=n("p"),Ia.innerHTML=Ql,Es=t(),B=n("div"),p(Go.$$.fragment),Ws=t(),Fa=n("p"),Fa.innerHTML=Kl,qs=t(),Ua=n("p"),Ua.innerHTML=ec,Ns=t(),Xa=n("p"),Xa.innerHTML=oc,zs=t(),Ea=n("p"),Ea.innerHTML=tc,Bs=t(),Wa=n("p"),Wa.innerHTML=ac,js=t(),ie=n("div"),p(Jo.$$.fragment),Gs=t(),qa=n("p"),qa.textContent=rc,Js=t(),p(Xe.$$.fragment),Zs=t(),Ee=n("div"),p(Zo.$$.fragment),Ys=t(),Na=n("p"),Na.textContent=nc,En=t(),p(Yo.$$.fragment),Wn=t(),I=n("div"),p(Oo.$$.fragment),Os=t(),za=n("p"),za.innerHTML=sc,Qs=t(),We=n("div"),p(Qo.$$.fragment),Ks=t(),Ba=n("p"),Ba.innerHTML=ic,ei=t(),qe=n("div"),p(Ko.$$.fragment),oi=t(),ja=n("p"),ja.innerHTML=dc,ti=t(),j=n("div"),p(et.$$.fragment),ai=t(),Ga=n("p"),Ga.innerHTML=lc,ri=t(),Ja=n("p"),Ja.innerHTML=cc,ni=t(),Za=n("p"),Za.innerHTML=fc,si=t(),Ya=n("p"),Ya.innerHTML=pc,ii=t(),Oa=n("p"),Oa.innerHTML=mc,di=t(),de=n("div"),p(ot.$$.fragment),li=t(),Qa=n("p"),Qa.textContent=hc,ci=t(),p(Ne.$$.fragment),fi=t(),ze=n("div"),p(tt.$$.fragment),pi=t(),Ka=n("p"),Ka.textContent=_c,qn=t(),p(at.$$.fragment),Nn=t(),k=n("div"),p(rt.$$.fragment),mi=t(),er=n("p"),er.innerHTML=uc,hi=t(),or=n("p"),or.innerHTML=gc,_i=t(),Be=n("div"),p(nt.$$.fragment),ui=t(),tr=n("p"),tr.innerHTML=xc,gi=t(),je=n("div"),p(st.$$.fragment),xi=t(),ar=n("p"),ar.innerHTML=Lc,Li=t(),Z=n("div"),p(it.$$.fragment),bi=t(),rr=n("p"),rr.innerHTML=bc,wi=t(),nr=n("p"),nr.innerHTML=wc,vi=t(),sr=n("p"),sr.innerHTML=vc,$i=t(),ir=n("p"),ir.innerHTML=$c,yi=t(),le=n("div"),p(dt.$$.fragment),Mi=t(),dr=n("p"),dr.textContent=yc,Ti=t(),p(Ge.$$.fragment),Di=t(),Je=n("div"),p(lt.$$.fragment),Si=t(),lr=n("p"),lr.textContent=Mc,Ci=t(),ce=n("div"),p(ct.$$.fragment),ki=t(),cr=n("p"),cr.innerHTML=Tc,Ai=t(),p(Ze.$$.fragment),zn=t(),p(ft.$$.fragment),Bn=t(),D=n("div"),p(pt.$$.fragment),Ri=t(),fr=n("p"),fr.innerHTML=Dc,Pi=t(),pr=n("p"),pr.innerHTML=Sc,Hi=t(),Ye=n("div"),p(mt.$$.fragment),Vi=t(),mr=n("p"),mr.innerHTML=Cc,Ii=t(),Oe=n("div"),p(ht.$$.fragment),Fi=t(),hr=n("p"),hr.innerHTML=kc,Ui=t(),Y=n("div"),p(_t.$$.fragment),Xi=t(),_r=n("p"),_r.innerHTML=Ac,Ei=t(),ur=n("p"),ur.innerHTML=Rc,Wi=t(),gr=n("p"),gr.innerHTML=Pc,qi=t(),xr=n("p"),xr.innerHTML=Hc,Ni=t(),fe=n("div"),p(ut.$$.fragment),zi=t(),Lr=n("p"),Lr.textContent=Vc,Bi=t(),p(Qe.$$.fragment),ji=t(),Ke=n("div"),p(gt.$$.fragment),Gi=t(),br=n("p"),br.textContent=Ic,Ji=t(),pe=n("div"),p(xt.$$.fragment),Zi=t(),wr=n("p"),wr.innerHTML=Fc,Yi=t(),p(eo.$$.fragment),Oi=t(),me=n("div"),p(Lt.$$.fragment),Qi=t(),vr=n("p"),vr.textContent=Uc,Ki=t(),p(oo.$$.fragment),jn=t(),p(bt.$$.fragment),Gn=t(),F=n("div"),p(wt.$$.fragment),ed=t(),$r=n("p"),$r.innerHTML=Xc,od=t(),to=n("div"),p(vt.$$.fragment),td=t(),yr=n("p"),yr.innerHTML=Ec,ad=t(),ao=n("div"),p($t.$$.fragment),rd=t(),Mr=n("p"),Mr.innerHTML=Wc,nd=t(),he=n("div"),p(yt.$$.fragment),sd=t(),Tr=n("p"),Tr.textContent=qc,id=t(),p(ro.$$.fragment),dd=t(),no=n("div"),p(Mt.$$.fragment),ld=t(),Dr=n("p"),Dr.textContent=Nc,cd=t(),_e=n("div"),p(Tt.$$.fragment),fd=t(),Sr=n("p"),Sr.innerHTML=zc,pd=t(),p(so.$$.fragment),Jn=t(),p(Dt.$$.fragment),Zn=t(),U=n("div"),p(St.$$.fragment),md=t(),Cr=n("p"),Cr.innerHTML=Bc,hd=t(),io=n("div"),p(Ct.$$.fragment),_d=t(),kr=n("p"),kr.innerHTML=jc,ud=t(),lo=n("div"),p(kt.$$.fragment),gd=t(),Ar=n("p"),Ar.innerHTML=Gc,xd=t(),ue=n("div"),p(At.$$.fragment),Ld=t(),Rr=n("p"),Rr.textContent=Jc,bd=t(),p(co.$$.fragment),wd=t(),fo=n("div"),p(Rt.$$.fragment),vd=t(),Pr=n("p"),Pr.textContent=Zc,$d=t(),ge=n("div"),p(Pt.$$.fragment),yd=t(),Hr=n("p"),Hr.innerHTML=Yc,Md=t(),p(po.$$.fragment),Yn=t(),p(Ht.$$.fragment),On=t(),X=n("div"),p(Vt.$$.fragment),Td=t(),Vr=n("p"),Vr.innerHTML=Oc,Dd=t(),mo=n("div"),p(It.$$.fragment),Sd=t(),Ir=n("p"),Ir.innerHTML=Qc,Cd=t(),ho=n("div"),p(Ft.$$.fragment),kd=t(),Fr=n("p"),Fr.innerHTML=Kc,Ad=t(),xe=n("div"),p(Ut.$$.fragment),Rd=t(),Ur=n("p"),Ur.textContent=ef,Pd=t(),p(_o.$$.fragment),Hd=t(),uo=n("div"),p(Xt.$$.fragment),Vd=t(),Xr=n("p"),Xr.textContent=of,Id=t(),Le=n("div"),p(Et.$$.fragment),Fd=t(),Er=n("p"),Er.innerHTML=tf,Ud=t(),p(go.$$.fragment),Qn=t(),p(Wt.$$.fragment),Kn=t(),E=n("div"),p(qt.$$.fragment),Xd=t(),Wr=n("p"),Wr.innerHTML=af,Ed=t(),xo=n("div"),p(Nt.$$.fragment),Wd=t(),qr=n("p"),qr.innerHTML=rf,qd=t(),Lo=n("div"),p(zt.$$.fragment),Nd=t(),Nr=n("p"),Nr.innerHTML=nf,zd=t(),be=n("div"),p(Bt.$$.fragment),Bd=t(),zr=n("p"),zr.textContent=sf,jd=t(),p(bo.$$.fragment),Gd=t(),wo=n("div"),p(jt.$$.fragment),Jd=t(),Br=n("p"),Br.textContent=df,Zd=t(),we=n("div"),p(Gt.$$.fragment),Yd=t(),jr=n("p"),jr.innerHTML=lf,Od=t(),p(vo.$$.fragment),es=t(),p(Jt.$$.fragment),os=t(),W=n("div"),p(Zt.$$.fragment),Qd=t(),Gr=n("p"),Gr.innerHTML=cf,Kd=t(),$o=n("div"),p(Yt.$$.fragment),el=t(),Jr=n("p"),Jr.innerHTML=ff,ol=t(),yo=n("div"),p(Ot.$$.fragment),tl=t(),Zr=n("p"),Zr.innerHTML=pf,al=t(),ve=n("div"),p(Qt.$$.fragment),rl=t(),Yr=n("p"),Yr.textContent=mf,nl=t(),p(Mo.$$.fragment),sl=t(),To=n("div"),p(Kt.$$.fragment),il=t(),Or=n("p"),Or.textContent=hf,dl=t(),$e=n("div"),p(ea.$$.fragment),ll=t(),Qr=n("p"),Qr.innerHTML=_f,cl=t(),p(Do.$$.fragment),ts=t(),p(oa.$$.fragment),as=t(),q=n("div"),p(ta.$$.fragment),fl=t(),Kr=n("p"),Kr.innerHTML=uf,pl=t(),So=n("div"),p(aa.$$.fragment),ml=t(),en=n("p"),en.innerHTML=gf,hl=t(),Co=n("div"),p(ra.$$.fragment),_l=t(),on=n("p"),on.innerHTML=xf,ul=t(),ye=n("div"),p(na.$$.fragment),gl=t(),tn=n("p"),tn.textContent=Lf,xl=t(),p(ko.$$.fragment),Ll=t(),Ao=n("div"),p(sa.$$.fragment),bl=t(),an=n("p"),an.textContent=bf,wl=t(),Me=n("div"),p(ia.$$.fragment),vl=t(),rn=n("p"),rn.innerHTML=wf,$l=t(),p(Ro.$$.fragment),rs=t(),p(da.$$.fragment),ns=t(),ke=n("div"),p(la.$$.fragment),yl=t(),Po=n("div"),p(ca.$$.fragment),Ml=t(),nn=n("p"),nn.innerHTML=vf,ss=t(),p(fa.$$.fragment),is=t(),S=n("div"),p(pa.$$.fragment),Tl=t(),sn=n("p"),sn.textContent=$f,Dl=t(),dn=n("div"),p(ma.$$.fragment),Sl=t(),Te=n("div"),p(ha.$$.fragment),Cl=t(),ln=n("p"),ln.textContent=yf,kl=t(),cn=n("p"),cn.textContent=Mf,Al=t(),se=n("div"),p(_a.$$.fragment),Rl=t(),fn=n("p"),fn.textContent=Tf,Pl=t(),p(Ho.$$.fragment),Hl=t(),p(Vo.$$.fragment),Vl=t(),De=n("div"),p(ua.$$.fragment),Il=t(),pn=n("p"),pn.textContent=Df,Fl=t(),p(Io.$$.fragment),Ul=t(),Fo=n("div"),p(ga.$$.fragment),Xl=t(),mn=n("p"),mn.textContent=Sf,El=t(),Uo=n("div"),p(xa.$$.fragment),Wl=t(),hn=n("p"),hn.innerHTML=Cf,ql=t(),Se=n("div"),p(La.$$.fragment),Nl=t(),_n=n("p"),_n.innerHTML=kf,zl=t(),p(Xo.$$.fragment),Bl=t(),Ce=n("div"),p(ba.$$.fragment),jl=t(),un=n("p"),un.textContent=Af,Gl=t(),p(Eo.$$.fragment),ds=t(),p(wa.$$.fragment),ls=t(),Vn=n("p"),this.h()},l(e){const x=Ff("svelte-u9bgzb",document.head);r=s(x,"META",{name:!0,content:!0}),x.forEach(i),b=a(e),l=s(e,"P",{}),w(l).forEach(i),c=a(e),m(M.$$.fragment,e),d=a(e),y=s(e,"P",{"data-svelte-h":!0}),f(y)!=="svelte-1aaqe11"&&(y.innerHTML=Jl),In=a(e),qo=s(e,"UL",{"data-svelte-h":!0}),f(qo)!=="svelte-hi235f"&&(qo.innerHTML=Zl),Fn=a(e),m(Ie.$$.fragment,e),Un=a(e),m(No.$$.fragment,e),Xn=a(e),V=s(e,"DIV",{class:!0});var N=w(V);m(zo.$$.fragment,N),Vs=a(N),Ha=s(N,"P",{"data-svelte-h":!0}),f(Ha)!=="svelte-1rz3bfv"&&(Ha.innerHTML=Yl),Is=a(N),Fe=s(N,"DIV",{class:!0});var va=w(Fe);m(Bo.$$.fragment,va),Fs=a(va),Va=s(va,"P",{"data-svelte-h":!0}),f(Va)!=="svelte-1062ci4"&&(Va.innerHTML=Ol),va.forEach(i),Us=a(N),Ue=s(N,"DIV",{class:!0});var $a=w(Ue);m(jo.$$.fragment,$a),Xs=a($a),Ia=s($a,"P",{"data-svelte-h":!0}),f(Ia)!=="svelte-u3q4so"&&(Ia.innerHTML=Ql),$a.forEach(i),Es=a(N),B=s(N,"DIV",{class:!0});var G=w(B);m(Go.$$.fragment,G),Ws=a(G),Fa=s(G,"P",{"data-svelte-h":!0}),f(Fa)!=="svelte-vs7s0z"&&(Fa.innerHTML=Kl),qs=a(G),Ua=s(G,"P",{"data-svelte-h":!0}),f(Ua)!=="svelte-15b960v"&&(Ua.innerHTML=ec),Ns=a(G),Xa=s(G,"P",{"data-svelte-h":!0}),f(Xa)!=="svelte-qff88w"&&(Xa.innerHTML=oc),zs=a(G),Ea=s(G,"P",{"data-svelte-h":!0}),f(Ea)!=="svelte-g6s5gf"&&(Ea.innerHTML=tc),Bs=a(G),Wa=s(G,"P",{"data-svelte-h":!0}),f(Wa)!=="svelte-ong6t9"&&(Wa.innerHTML=ac),G.forEach(i),js=a(N),ie=s(N,"DIV",{class:!0});var Ae=w(ie);m(Jo.$$.fragment,Ae),Gs=a(Ae),qa=s(Ae,"P",{"data-svelte-h":!0}),f(qa)!=="svelte-flusvq"&&(qa.textContent=rc),Js=a(Ae),m(Xe.$$.fragment,Ae),Ae.forEach(i),Zs=a(N),Ee=s(N,"DIV",{class:!0});var ya=w(Ee);m(Zo.$$.fragment,ya),Ys=a(ya),Na=s(ya,"P",{"data-svelte-h":!0}),f(Na)!=="svelte-1ufq5ot"&&(Na.textContent=nc),ya.forEach(i),N.forEach(i),En=a(e),m(Yo.$$.fragment,e),Wn=a(e),I=s(e,"DIV",{class:!0});var z=w(I);m(Oo.$$.fragment,z),Os=a(z),za=s(z,"P",{"data-svelte-h":!0}),f(za)!=="svelte-1mmsq14"&&(za.innerHTML=sc),Qs=a(z),We=s(z,"DIV",{class:!0});var Ma=w(We);m(Qo.$$.fragment,Ma),Ks=a(Ma),Ba=s(Ma,"P",{"data-svelte-h":!0}),f(Ba)!=="svelte-1062ci4"&&(Ba.innerHTML=ic),Ma.forEach(i),ei=a(z),qe=s(z,"DIV",{class:!0});var Ta=w(qe);m(Ko.$$.fragment,Ta),oi=a(Ta),ja=s(Ta,"P",{"data-svelte-h":!0}),f(ja)!=="svelte-u3q4so"&&(ja.innerHTML=dc),Ta.forEach(i),ti=a(z),j=s(z,"DIV",{class:!0});var J=w(j);m(et.$$.fragment,J),ai=a(J),Ga=s(J,"P",{"data-svelte-h":!0}),f(Ga)!=="svelte-vs7s0z"&&(Ga.innerHTML=lc),ri=a(J),Ja=s(J,"P",{"data-svelte-h":!0}),f(Ja)!=="svelte-15b960v"&&(Ja.innerHTML=cc),ni=a(J),Za=s(J,"P",{"data-svelte-h":!0}),f(Za)!=="svelte-qff88w"&&(Za.innerHTML=fc),si=a(J),Ya=s(J,"P",{"data-svelte-h":!0}),f(Ya)!=="svelte-g6s5gf"&&(Ya.innerHTML=pc),ii=a(J),Oa=s(J,"P",{"data-svelte-h":!0}),f(Oa)!=="svelte-ong6t9"&&(Oa.innerHTML=mc),J.forEach(i),di=a(z),de=s(z,"DIV",{class:!0});var Re=w(de);m(ot.$$.fragment,Re),li=a(Re),Qa=s(Re,"P",{"data-svelte-h":!0}),f(Qa)!=="svelte-flusvq"&&(Qa.textContent=hc),ci=a(Re),m(Ne.$$.fragment,Re),Re.forEach(i),fi=a(z),ze=s(z,"DIV",{class:!0});var Da=w(ze);m(tt.$$.fragment,Da),pi=a(Da),Ka=s(Da,"P",{"data-svelte-h":!0}),f(Ka)!=="svelte-1ufq5ot"&&(Ka.textContent=_c),Da.forEach(i),z.forEach(i),qn=a(e),m(at.$$.fragment,e),Nn=a(e),k=s(e,"DIV",{class:!0});var R=w(k);m(rt.$$.fragment,R),mi=a(R),er=s(R,"P",{"data-svelte-h":!0}),f(er)!=="svelte-1ribxih"&&(er.innerHTML=uc),hi=a(R),or=s(R,"P",{"data-svelte-h":!0}),f(or)!=="svelte-b8z8fv"&&(or.innerHTML=gc),_i=a(R),Be=s(R,"DIV",{class:!0});var Sa=w(Be);m(nt.$$.fragment,Sa),ui=a(Sa),tr=s(Sa,"P",{"data-svelte-h":!0}),f(tr)!=="svelte-1062ci4"&&(tr.innerHTML=xc),Sa.forEach(i),gi=a(R),je=s(R,"DIV",{class:!0});var Ca=w(je);m(st.$$.fragment,Ca),xi=a(Ca),ar=s(Ca,"P",{"data-svelte-h":!0}),f(ar)!=="svelte-1lgbsz7"&&(ar.innerHTML=Lc),Ca.forEach(i),Li=a(R),Z=s(R,"DIV",{class:!0});var ae=w(Z);m(it.$$.fragment,ae),bi=a(ae),rr=s(ae,"P",{"data-svelte-h":!0}),f(rr)!=="svelte-vs7s0z"&&(rr.innerHTML=bc),wi=a(ae),nr=s(ae,"P",{"data-svelte-h":!0}),f(nr)!=="svelte-15b960v"&&(nr.innerHTML=wc),vi=a(ae),sr=s(ae,"P",{"data-svelte-h":!0}),f(sr)!=="svelte-qff88w"&&(sr.innerHTML=vc),$i=a(ae),ir=s(ae,"P",{"data-svelte-h":!0}),f(ir)!=="svelte-1ukghd4"&&(ir.innerHTML=$c),ae.forEach(i),yi=a(R),le=s(R,"DIV",{class:!0});var Pe=w(le);m(dt.$$.fragment,Pe),Mi=a(Pe),dr=s(Pe,"P",{"data-svelte-h":!0}),f(dr)!=="svelte-flusvq"&&(dr.textContent=yc),Ti=a(Pe),m(Ge.$$.fragment,Pe),Pe.forEach(i),Di=a(R),Je=s(R,"DIV",{class:!0});var ka=w(Je);m(lt.$$.fragment,ka),Si=a(ka),lr=s(ka,"P",{"data-svelte-h":!0}),f(lr)!=="svelte-1ufq5ot"&&(lr.textContent=Mc),ka.forEach(i),Ci=a(R),ce=s(R,"DIV",{class:!0});var He=w(ce);m(ct.$$.fragment,He),ki=a(He),cr=s(He,"P",{"data-svelte-h":!0}),f(cr)!=="svelte-ioswce"&&(cr.innerHTML=Tc),Ai=a(He),m(Ze.$$.fragment,He),He.forEach(i),R.forEach(i),zn=a(e),m(ft.$$.fragment,e),Bn=a(e),D=s(e,"DIV",{class:!0});var A=w(D);m(pt.$$.fragment,A),Ri=a(A),fr=s(A,"P",{"data-svelte-h":!0}),f(fr)!=="svelte-dbp88o"&&(fr.innerHTML=Dc),Pi=a(A),pr=s(A,"P",{"data-svelte-h":!0}),f(pr)!=="svelte-b8z8fv"&&(pr.innerHTML=Sc),Hi=a(A),Ye=s(A,"DIV",{class:!0});var Aa=w(Ye);m(mt.$$.fragment,Aa),Vi=a(Aa),mr=s(Aa,"P",{"data-svelte-h":!0}),f(mr)!=="svelte-1062ci4"&&(mr.innerHTML=Cc),Aa.forEach(i),Ii=a(A),Oe=s(A,"DIV",{class:!0});var Ra=w(Oe);m(ht.$$.fragment,Ra),Fi=a(Ra),hr=s(Ra,"P",{"data-svelte-h":!0}),f(hr)!=="svelte-1lgbsz7"&&(hr.innerHTML=kc),Ra.forEach(i),Ui=a(A),Y=s(A,"DIV",{class:!0});var re=w(Y);m(_t.$$.fragment,re),Xi=a(re),_r=s(re,"P",{"data-svelte-h":!0}),f(_r)!=="svelte-178gcly"&&(_r.innerHTML=Ac),Ei=a(re),ur=s(re,"P",{"data-svelte-h":!0}),f(ur)!=="svelte-15b960v"&&(ur.innerHTML=Rc),Wi=a(re),gr=s(re,"P",{"data-svelte-h":!0}),f(gr)!=="svelte-qff88w"&&(gr.innerHTML=Pc),qi=a(re),xr=s(re,"P",{"data-svelte-h":!0}),f(xr)!=="svelte-1ukghd4"&&(xr.innerHTML=Hc),re.forEach(i),Ni=a(A),fe=s(A,"DIV",{class:!0});var Ve=w(fe);m(ut.$$.fragment,Ve),zi=a(Ve),Lr=s(Ve,"P",{"data-svelte-h":!0}),f(Lr)!=="svelte-flusvq"&&(Lr.textContent=Vc),Bi=a(Ve),m(Qe.$$.fragment,Ve),Ve.forEach(i),ji=a(A),Ke=s(A,"DIV",{class:!0});var Pa=w(Ke);m(gt.$$.fragment,Pa),Gi=a(Pa),br=s(Pa,"P",{"data-svelte-h":!0}),f(br)!=="svelte-1ufq5ot"&&(br.textContent=Ic),Pa.forEach(i),Ji=a(A),pe=s(A,"DIV",{class:!0});var gn=w(pe);m(xt.$$.fragment,gn),Zi=a(gn),wr=s(gn,"P",{"data-svelte-h":!0}),f(wr)!=="svelte-ioswce"&&(wr.innerHTML=Fc),Yi=a(gn),m(eo.$$.fragment,gn),gn.forEach(i),Oi=a(A),me=s(A,"DIV",{class:!0});var xn=w(me);m(Lt.$$.fragment,xn),Qi=a(xn),vr=s(xn,"P",{"data-svelte-h":!0}),f(vr)!=="svelte-119cgd9"&&(vr.textContent=Uc),Ki=a(xn),m(oo.$$.fragment,xn),xn.forEach(i),A.forEach(i),jn=a(e),m(bt.$$.fragment,e),Gn=a(e),F=s(e,"DIV",{class:!0});var O=w(F);m(wt.$$.fragment,O),ed=a(O),$r=s(O,"P",{"data-svelte-h":!0}),f($r)!=="svelte-1sg5qir"&&($r.innerHTML=Xc),od=a(O),to=s(O,"DIV",{class:!0});var fs=w(to);m(vt.$$.fragment,fs),td=a(fs),yr=s(fs,"P",{"data-svelte-h":!0}),f(yr)!=="svelte-1lgbsz7"&&(yr.innerHTML=Ec),fs.forEach(i),ad=a(O),ao=s(O,"DIV",{class:!0});var ps=w(ao);m($t.$$.fragment,ps),rd=a(ps),Mr=s(ps,"P",{"data-svelte-h":!0}),f(Mr)!=="svelte-t8m9w5"&&(Mr.innerHTML=Wc),ps.forEach(i),nd=a(O),he=s(O,"DIV",{class:!0});var Ln=w(he);m(yt.$$.fragment,Ln),sd=a(Ln),Tr=s(Ln,"P",{"data-svelte-h":!0}),f(Tr)!=="svelte-flusvq"&&(Tr.textContent=qc),id=a(Ln),m(ro.$$.fragment,Ln),Ln.forEach(i),dd=a(O),no=s(O,"DIV",{class:!0});var ms=w(no);m(Mt.$$.fragment,ms),ld=a(ms),Dr=s(ms,"P",{"data-svelte-h":!0}),f(Dr)!=="svelte-1ufq5ot"&&(Dr.textContent=Nc),ms.forEach(i),cd=a(O),_e=s(O,"DIV",{class:!0});var bn=w(_e);m(Tt.$$.fragment,bn),fd=a(bn),Sr=s(bn,"P",{"data-svelte-h":!0}),f(Sr)!=="svelte-ioswce"&&(Sr.innerHTML=zc),pd=a(bn),m(so.$$.fragment,bn),bn.forEach(i),O.forEach(i),Jn=a(e),m(Dt.$$.fragment,e),Zn=a(e),U=s(e,"DIV",{class:!0});var Q=w(U);m(St.$$.fragment,Q),md=a(Q),Cr=s(Q,"P",{"data-svelte-h":!0}),f(Cr)!=="svelte-74s0ub"&&(Cr.innerHTML=Bc),hd=a(Q),io=s(Q,"DIV",{class:!0});var hs=w(io);m(Ct.$$.fragment,hs),_d=a(hs),kr=s(hs,"P",{"data-svelte-h":!0}),f(kr)!=="svelte-1lgbsz7"&&(kr.innerHTML=jc),hs.forEach(i),ud=a(Q),lo=s(Q,"DIV",{class:!0});var _s=w(lo);m(kt.$$.fragment,_s),gd=a(_s),Ar=s(_s,"P",{"data-svelte-h":!0}),f(Ar)!=="svelte-t8m9w5"&&(Ar.innerHTML=Gc),_s.forEach(i),xd=a(Q),ue=s(Q,"DIV",{class:!0});var wn=w(ue);m(At.$$.fragment,wn),Ld=a(wn),Rr=s(wn,"P",{"data-svelte-h":!0}),f(Rr)!=="svelte-flusvq"&&(Rr.textContent=Jc),bd=a(wn),m(co.$$.fragment,wn),wn.forEach(i),wd=a(Q),fo=s(Q,"DIV",{class:!0});var us=w(fo);m(Rt.$$.fragment,us),vd=a(us),Pr=s(us,"P",{"data-svelte-h":!0}),f(Pr)!=="svelte-1ufq5ot"&&(Pr.textContent=Zc),us.forEach(i),$d=a(Q),ge=s(Q,"DIV",{class:!0});var vn=w(ge);m(Pt.$$.fragment,vn),yd=a(vn),Hr=s(vn,"P",{"data-svelte-h":!0}),f(Hr)!=="svelte-ioswce"&&(Hr.innerHTML=Yc),Md=a(vn),m(po.$$.fragment,vn),vn.forEach(i),Q.forEach(i),Yn=a(e),m(Ht.$$.fragment,e),On=a(e),X=s(e,"DIV",{class:!0});var K=w(X);m(Vt.$$.fragment,K),Td=a(K),Vr=s(K,"P",{"data-svelte-h":!0}),f(Vr)!=="svelte-1bvq5vj"&&(Vr.innerHTML=Oc),Dd=a(K),mo=s(K,"DIV",{class:!0});var gs=w(mo);m(It.$$.fragment,gs),Sd=a(gs),Ir=s(gs,"P",{"data-svelte-h":!0}),f(Ir)!=="svelte-1lgbsz7"&&(Ir.innerHTML=Qc),gs.forEach(i),Cd=a(K),ho=s(K,"DIV",{class:!0});var xs=w(ho);m(Ft.$$.fragment,xs),kd=a(xs),Fr=s(xs,"P",{"data-svelte-h":!0}),f(Fr)!=="svelte-t8m9w5"&&(Fr.innerHTML=Kc),xs.forEach(i),Ad=a(K),xe=s(K,"DIV",{class:!0});var $n=w(xe);m(Ut.$$.fragment,$n),Rd=a($n),Ur=s($n,"P",{"data-svelte-h":!0}),f(Ur)!=="svelte-flusvq"&&(Ur.textContent=ef),Pd=a($n),m(_o.$$.fragment,$n),$n.forEach(i),Hd=a(K),uo=s(K,"DIV",{class:!0});var Ls=w(uo);m(Xt.$$.fragment,Ls),Vd=a(Ls),Xr=s(Ls,"P",{"data-svelte-h":!0}),f(Xr)!=="svelte-1ufq5ot"&&(Xr.textContent=of),Ls.forEach(i),Id=a(K),Le=s(K,"DIV",{class:!0});var yn=w(Le);m(Et.$$.fragment,yn),Fd=a(yn),Er=s(yn,"P",{"data-svelte-h":!0}),f(Er)!=="svelte-ioswce"&&(Er.innerHTML=tf),Ud=a(yn),m(go.$$.fragment,yn),yn.forEach(i),K.forEach(i),Qn=a(e),m(Wt.$$.fragment,e),Kn=a(e),E=s(e,"DIV",{class:!0});var ee=w(E);m(qt.$$.fragment,ee),Xd=a(ee),Wr=s(ee,"P",{"data-svelte-h":!0}),f(Wr)!=="svelte-26iv7o"&&(Wr.innerHTML=af),Ed=a(ee),xo=s(ee,"DIV",{class:!0});var bs=w(xo);m(Nt.$$.fragment,bs),Wd=a(bs),qr=s(bs,"P",{"data-svelte-h":!0}),f(qr)!=="svelte-1lgbsz7"&&(qr.innerHTML=rf),bs.forEach(i),qd=a(ee),Lo=s(ee,"DIV",{class:!0});var ws=w(Lo);m(zt.$$.fragment,ws),Nd=a(ws),Nr=s(ws,"P",{"data-svelte-h":!0}),f(Nr)!=="svelte-t8m9w5"&&(Nr.innerHTML=nf),ws.forEach(i),zd=a(ee),be=s(ee,"DIV",{class:!0});var Mn=w(be);m(Bt.$$.fragment,Mn),Bd=a(Mn),zr=s(Mn,"P",{"data-svelte-h":!0}),f(zr)!=="svelte-flusvq"&&(zr.textContent=sf),jd=a(Mn),m(bo.$$.fragment,Mn),Mn.forEach(i),Gd=a(ee),wo=s(ee,"DIV",{class:!0});var vs=w(wo);m(jt.$$.fragment,vs),Jd=a(vs),Br=s(vs,"P",{"data-svelte-h":!0}),f(Br)!=="svelte-1ufq5ot"&&(Br.textContent=df),vs.forEach(i),Zd=a(ee),we=s(ee,"DIV",{class:!0});var Tn=w(we);m(Gt.$$.fragment,Tn),Yd=a(Tn),jr=s(Tn,"P",{"data-svelte-h":!0}),f(jr)!=="svelte-ioswce"&&(jr.innerHTML=lf),Od=a(Tn),m(vo.$$.fragment,Tn),Tn.forEach(i),ee.forEach(i),es=a(e),m(Jt.$$.fragment,e),os=a(e),W=s(e,"DIV",{class:!0});var oe=w(W);m(Zt.$$.fragment,oe),Qd=a(oe),Gr=s(oe,"P",{"data-svelte-h":!0}),f(Gr)!=="svelte-h45s88"&&(Gr.innerHTML=cf),Kd=a(oe),$o=s(oe,"DIV",{class:!0});var $s=w($o);m(Yt.$$.fragment,$s),el=a($s),Jr=s($s,"P",{"data-svelte-h":!0}),f(Jr)!=="svelte-1lgbsz7"&&(Jr.innerHTML=ff),$s.forEach(i),ol=a(oe),yo=s(oe,"DIV",{class:!0});var ys=w(yo);m(Ot.$$.fragment,ys),tl=a(ys),Zr=s(ys,"P",{"data-svelte-h":!0}),f(Zr)!=="svelte-t8m9w5"&&(Zr.innerHTML=pf),ys.forEach(i),al=a(oe),ve=s(oe,"DIV",{class:!0});var Dn=w(ve);m(Qt.$$.fragment,Dn),rl=a(Dn),Yr=s(Dn,"P",{"data-svelte-h":!0}),f(Yr)!=="svelte-flusvq"&&(Yr.textContent=mf),nl=a(Dn),m(Mo.$$.fragment,Dn),Dn.forEach(i),sl=a(oe),To=s(oe,"DIV",{class:!0});var Ms=w(To);m(Kt.$$.fragment,Ms),il=a(Ms),Or=s(Ms,"P",{"data-svelte-h":!0}),f(Or)!=="svelte-1ufq5ot"&&(Or.textContent=hf),Ms.forEach(i),dl=a(oe),$e=s(oe,"DIV",{class:!0});var Sn=w($e);m(ea.$$.fragment,Sn),ll=a(Sn),Qr=s(Sn,"P",{"data-svelte-h":!0}),f(Qr)!=="svelte-ioswce"&&(Qr.innerHTML=_f),cl=a(Sn),m(Do.$$.fragment,Sn),Sn.forEach(i),oe.forEach(i),ts=a(e),m(oa.$$.fragment,e),as=a(e),q=s(e,"DIV",{class:!0});var te=w(q);m(ta.$$.fragment,te),fl=a(te),Kr=s(te,"P",{"data-svelte-h":!0}),f(Kr)!=="svelte-u52146"&&(Kr.innerHTML=uf),pl=a(te),So=s(te,"DIV",{class:!0});var Ts=w(So);m(aa.$$.fragment,Ts),ml=a(Ts),en=s(Ts,"P",{"data-svelte-h":!0}),f(en)!=="svelte-1lgbsz7"&&(en.innerHTML=gf),Ts.forEach(i),hl=a(te),Co=s(te,"DIV",{class:!0});var Ds=w(Co);m(ra.$$.fragment,Ds),_l=a(Ds),on=s(Ds,"P",{"data-svelte-h":!0}),f(on)!=="svelte-t8m9w5"&&(on.innerHTML=xf),Ds.forEach(i),ul=a(te),ye=s(te,"DIV",{class:!0});var Cn=w(ye);m(na.$$.fragment,Cn),gl=a(Cn),tn=s(Cn,"P",{"data-svelte-h":!0}),f(tn)!=="svelte-flusvq"&&(tn.textContent=Lf),xl=a(Cn),m(ko.$$.fragment,Cn),Cn.forEach(i),Ll=a(te),Ao=s(te,"DIV",{class:!0});var Ss=w(Ao);m(sa.$$.fragment,Ss),bl=a(Ss),an=s(Ss,"P",{"data-svelte-h":!0}),f(an)!=="svelte-1ufq5ot"&&(an.textContent=bf),Ss.forEach(i),wl=a(te),Me=s(te,"DIV",{class:!0});var kn=w(Me);m(ia.$$.fragment,kn),vl=a(kn),rn=s(kn,"P",{"data-svelte-h":!0}),f(rn)!=="svelte-ioswce"&&(rn.innerHTML=wf),$l=a(kn),m(Ro.$$.fragment,kn),kn.forEach(i),te.forEach(i),rs=a(e),m(da.$$.fragment,e),ns=a(e),ke=s(e,"DIV",{class:!0});var Cs=w(ke);m(la.$$.fragment,Cs),yl=a(Cs),Po=s(Cs,"DIV",{class:!0});var ks=w(Po);m(ca.$$.fragment,ks),Ml=a(ks),nn=s(ks,"P",{"data-svelte-h":!0}),f(nn)!=="svelte-1lgbsz7"&&(nn.innerHTML=vf),ks.forEach(i),Cs.forEach(i),ss=a(e),m(fa.$$.fragment,e),is=a(e),S=s(e,"DIV",{class:!0});var H=w(S);m(pa.$$.fragment,H),Tl=a(H),sn=s(H,"P",{"data-svelte-h":!0}),f(sn)!=="svelte-1q4bbx"&&(sn.textContent=$f),Dl=a(H),dn=s(H,"DIV",{class:!0});var Rf=w(dn);m(ma.$$.fragment,Rf),Rf.forEach(i),Sl=a(H),Te=s(H,"DIV",{class:!0});var An=w(Te);m(ha.$$.fragment,An),Cl=a(An),ln=s(An,"P",{"data-svelte-h":!0}),f(ln)!=="svelte-w22v11"&&(ln.textContent=yf),kl=a(An),cn=s(An,"P",{"data-svelte-h":!0}),f(cn)!=="svelte-zdadya"&&(cn.textContent=Mf),An.forEach(i),Al=a(H),se=s(H,"DIV",{class:!0});var Wo=w(se);m(_a.$$.fragment,Wo),Rl=a(Wo),fn=s(Wo,"P",{"data-svelte-h":!0}),f(fn)!=="svelte-1nr2dy0"&&(fn.textContent=Tf),Pl=a(Wo),m(Ho.$$.fragment,Wo),Hl=a(Wo),m(Vo.$$.fragment,Wo),Wo.forEach(i),Vl=a(H),De=s(H,"DIV",{class:!0});var Rn=w(De);m(ua.$$.fragment,Rn),Il=a(Rn),pn=s(Rn,"P",{"data-svelte-h":!0}),f(pn)!=="svelte-h0os0v"&&(pn.textContent=Df),Fl=a(Rn),m(Io.$$.fragment,Rn),Rn.forEach(i),Ul=a(H),Fo=s(H,"DIV",{class:!0});var As=w(Fo);m(ga.$$.fragment,As),Xl=a(As),mn=s(As,"P",{"data-svelte-h":!0}),f(mn)!=="svelte-1825k9e"&&(mn.textContent=Sf),As.forEach(i),El=a(H),Uo=s(H,"DIV",{class:!0});var Rs=w(Uo);m(xa.$$.fragment,Rs),Wl=a(Rs),hn=s(Rs,"P",{"data-svelte-h":!0}),f(hn)!=="svelte-rvubqa"&&(hn.innerHTML=Cf),Rs.forEach(i),ql=a(H),Se=s(H,"DIV",{class:!0});var Pn=w(Se);m(La.$$.fragment,Pn),Nl=a(Pn),_n=s(Pn,"P",{"data-svelte-h":!0}),f(_n)!=="svelte-ioswce"&&(_n.innerHTML=kf),zl=a(Pn),m(Xo.$$.fragment,Pn),Pn.forEach(i),Bl=a(H),Ce=s(H,"DIV",{class:!0});var Hn=w(Ce);m(ba.$$.fragment,Hn),jl=a(Hn),un=s(Hn,"P",{"data-svelte-h":!0}),f(un)!=="svelte-119cgd9"&&(un.textContent=Af),Gl=a(Hn),m(Eo.$$.fragment,Hn),Hn.forEach(i),H.forEach(i),ds=a(e),m(wa.$$.fragment,e),ls=a(e),Vn=s(e,"P",{}),w(Vn).forEach(i),this.h()},h(){v(r,"name","hf:doc:metadata"),v(r,"content",fp),v(Fe,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Ue,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(B,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(ie,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Ee,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(V,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(We,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(qe,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(j,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(de,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(ze,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(I,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Be,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(je,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Z,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(le,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Je,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(ce,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(k,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Ye,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Oe,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Y,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(fe,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Ke,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(pe,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(me,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(D,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(to,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(ao,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(he,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(no,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(_e,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(F,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(io,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(lo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(ue,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(fo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(ge,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(U,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(mo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(ho,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(xe,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(uo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Le,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(X,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(xo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Lo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(be,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(wo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(we,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(E,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v($o,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(yo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(ve,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(To,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v($e,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(W,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(So,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Co,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(ye,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Ao,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Me,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(q,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Po,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(ke,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(dn,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Te,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(se,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(De,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Fo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Uo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Se,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(Ce,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),v(S,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8")},m(e,x){o(document.head,r),L(e,b,x),L(e,l,x),L(e,c,x),h(M,e,x),L(e,d,x),L(e,y,x),L(e,In,x),L(e,qo,x),L(e,Fn,x),h(Ie,e,x),L(e,Un,x),h(No,e,x),L(e,Xn,x),L(e,V,x),h(zo,V,null),o(V,Vs),o(V,Ha),o(V,Is),o(V,Fe),h(Bo,Fe,null),o(Fe,Fs),o(Fe,Va),o(V,Us),o(V,Ue),h(jo,Ue,null),o(Ue,Xs),o(Ue,Ia),o(V,Es),o(V,B),h(Go,B,null),o(B,Ws),o(B,Fa),o(B,qs),o(B,Ua),o(B,Ns),o(B,Xa),o(B,zs),o(B,Ea),o(B,Bs),o(B,Wa),o(V,js),o(V,ie),h(Jo,ie,null),o(ie,Gs),o(ie,qa),o(ie,Js),h(Xe,ie,null),o(V,Zs),o(V,Ee),h(Zo,Ee,null),o(Ee,Ys),o(Ee,Na),L(e,En,x),h(Yo,e,x),L(e,Wn,x),L(e,I,x),h(Oo,I,null),o(I,Os),o(I,za),o(I,Qs),o(I,We),h(Qo,We,null),o(We,Ks),o(We,Ba),o(I,ei),o(I,qe),h(Ko,qe,null),o(qe,oi),o(qe,ja),o(I,ti),o(I,j),h(et,j,null),o(j,ai),o(j,Ga),o(j,ri),o(j,Ja),o(j,ni),o(j,Za),o(j,si),o(j,Ya),o(j,ii),o(j,Oa),o(I,di),o(I,de),h(ot,de,null),o(de,li),o(de,Qa),o(de,ci),h(Ne,de,null),o(I,fi),o(I,ze),h(tt,ze,null),o(ze,pi),o(ze,Ka),L(e,qn,x),h(at,e,x),L(e,Nn,x),L(e,k,x),h(rt,k,null),o(k,mi),o(k,er),o(k,hi),o(k,or),o(k,_i),o(k,Be),h(nt,Be,null),o(Be,ui),o(Be,tr),o(k,gi),o(k,je),h(st,je,null),o(je,xi),o(je,ar),o(k,Li),o(k,Z),h(it,Z,null),o(Z,bi),o(Z,rr),o(Z,wi),o(Z,nr),o(Z,vi),o(Z,sr),o(Z,$i),o(Z,ir),o(k,yi),o(k,le),h(dt,le,null),o(le,Mi),o(le,dr),o(le,Ti),h(Ge,le,null),o(k,Di),o(k,Je),h(lt,Je,null),o(Je,Si),o(Je,lr),o(k,Ci),o(k,ce),h(ct,ce,null),o(ce,ki),o(ce,cr),o(ce,Ai),h(Ze,ce,null),L(e,zn,x),h(ft,e,x),L(e,Bn,x),L(e,D,x),h(pt,D,null),o(D,Ri),o(D,fr),o(D,Pi),o(D,pr),o(D,Hi),o(D,Ye),h(mt,Ye,null),o(Ye,Vi),o(Ye,mr),o(D,Ii),o(D,Oe),h(ht,Oe,null),o(Oe,Fi),o(Oe,hr),o(D,Ui),o(D,Y),h(_t,Y,null),o(Y,Xi),o(Y,_r),o(Y,Ei),o(Y,ur),o(Y,Wi),o(Y,gr),o(Y,qi),o(Y,xr),o(D,Ni),o(D,fe),h(ut,fe,null),o(fe,zi),o(fe,Lr),o(fe,Bi),h(Qe,fe,null),o(D,ji),o(D,Ke),h(gt,Ke,null),o(Ke,Gi),o(Ke,br),o(D,Ji),o(D,pe),h(xt,pe,null),o(pe,Zi),o(pe,wr),o(pe,Yi),h(eo,pe,null),o(D,Oi),o(D,me),h(Lt,me,null),o(me,Qi),o(me,vr),o(me,Ki),h(oo,me,null),L(e,jn,x),h(bt,e,x),L(e,Gn,x),L(e,F,x),h(wt,F,null),o(F,ed),o(F,$r),o(F,od),o(F,to),h(vt,to,null),o(to,td),o(to,yr),o(F,ad),o(F,ao),h($t,ao,null),o(ao,rd),o(ao,Mr),o(F,nd),o(F,he),h(yt,he,null),o(he,sd),o(he,Tr),o(he,id),h(ro,he,null),o(F,dd),o(F,no),h(Mt,no,null),o(no,ld),o(no,Dr),o(F,cd),o(F,_e),h(Tt,_e,null),o(_e,fd),o(_e,Sr),o(_e,pd),h(so,_e,null),L(e,Jn,x),h(Dt,e,x),L(e,Zn,x),L(e,U,x),h(St,U,null),o(U,md),o(U,Cr),o(U,hd),o(U,io),h(Ct,io,null),o(io,_d),o(io,kr),o(U,ud),o(U,lo),h(kt,lo,null),o(lo,gd),o(lo,Ar),o(U,xd),o(U,ue),h(At,ue,null),o(ue,Ld),o(ue,Rr),o(ue,bd),h(co,ue,null),o(U,wd),o(U,fo),h(Rt,fo,null),o(fo,vd),o(fo,Pr),o(U,$d),o(U,ge),h(Pt,ge,null),o(ge,yd),o(ge,Hr),o(ge,Md),h(po,ge,null),L(e,Yn,x),h(Ht,e,x),L(e,On,x),L(e,X,x),h(Vt,X,null),o(X,Td),o(X,Vr),o(X,Dd),o(X,mo),h(It,mo,null),o(mo,Sd),o(mo,Ir),o(X,Cd),o(X,ho),h(Ft,ho,null),o(ho,kd),o(ho,Fr),o(X,Ad),o(X,xe),h(Ut,xe,null),o(xe,Rd),o(xe,Ur),o(xe,Pd),h(_o,xe,null),o(X,Hd),o(X,uo),h(Xt,uo,null),o(uo,Vd),o(uo,Xr),o(X,Id),o(X,Le),h(Et,Le,null),o(Le,Fd),o(Le,Er),o(Le,Ud),h(go,Le,null),L(e,Qn,x),h(Wt,e,x),L(e,Kn,x),L(e,E,x),h(qt,E,null),o(E,Xd),o(E,Wr),o(E,Ed),o(E,xo),h(Nt,xo,null),o(xo,Wd),o(xo,qr),o(E,qd),o(E,Lo),h(zt,Lo,null),o(Lo,Nd),o(Lo,Nr),o(E,zd),o(E,be),h(Bt,be,null),o(be,Bd),o(be,zr),o(be,jd),h(bo,be,null),o(E,Gd),o(E,wo),h(jt,wo,null),o(wo,Jd),o(wo,Br),o(E,Zd),o(E,we),h(Gt,we,null),o(we,Yd),o(we,jr),o(we,Od),h(vo,we,null),L(e,es,x),h(Jt,e,x),L(e,os,x),L(e,W,x),h(Zt,W,null),o(W,Qd),o(W,Gr),o(W,Kd),o(W,$o),h(Yt,$o,null),o($o,el),o($o,Jr),o(W,ol),o(W,yo),h(Ot,yo,null),o(yo,tl),o(yo,Zr),o(W,al),o(W,ve),h(Qt,ve,null),o(ve,rl),o(ve,Yr),o(ve,nl),h(Mo,ve,null),o(W,sl),o(W,To),h(Kt,To,null),o(To,il),o(To,Or),o(W,dl),o(W,$e),h(ea,$e,null),o($e,ll),o($e,Qr),o($e,cl),h(Do,$e,null),L(e,ts,x),h(oa,e,x),L(e,as,x),L(e,q,x),h(ta,q,null),o(q,fl),o(q,Kr),o(q,pl),o(q,So),h(aa,So,null),o(So,ml),o(So,en),o(q,hl),o(q,Co),h(ra,Co,null),o(Co,_l),o(Co,on),o(q,ul),o(q,ye),h(na,ye,null),o(ye,gl),o(ye,tn),o(ye,xl),h(ko,ye,null),o(q,Ll),o(q,Ao),h(sa,Ao,null),o(Ao,bl),o(Ao,an),o(q,wl),o(q,Me),h(ia,Me,null),o(Me,vl),o(Me,rn),o(Me,$l),h(Ro,Me,null),L(e,rs,x),h(da,e,x),L(e,ns,x),L(e,ke,x),h(la,ke,null),o(ke,yl),o(ke,Po),h(ca,Po,null),o(Po,Ml),o(Po,nn),L(e,ss,x),h(fa,e,x),L(e,is,x),L(e,S,x),h(pa,S,null),o(S,Tl),o(S,sn),o(S,Dl),o(S,dn),h(ma,dn,null),o(S,Sl),o(S,Te),h(ha,Te,null),o(Te,Cl),o(Te,ln),o(Te,kl),o(Te,cn),o(S,Al),o(S,se),h(_a,se,null),o(se,Rl),o(se,fn),o(se,Pl),h(Ho,se,null),o(se,Hl),h(Vo,se,null),o(S,Vl),o(S,De),h(ua,De,null),o(De,Il),o(De,pn),o(De,Fl),h(Io,De,null),o(S,Ul),o(S,Fo),h(ga,Fo,null),o(Fo,Xl),o(Fo,mn),o(S,El),o(S,Uo),h(xa,Uo,null),o(Uo,Wl),o(Uo,hn),o(S,ql),o(S,Se),h(La,Se,null),o(Se,Nl),o(Se,_n),o(Se,zl),h(Xo,Se,null),o(S,Bl),o(S,Ce),h(ba,Ce,null),o(Ce,jl),o(Ce,un),o(Ce,Gl),h(Eo,Ce,null),L(e,ds,x),h(wa,e,x),L(e,ls,x),L(e,Vn,x),cs=!0},p(e,[x]){const N={};x&2&&(N.$$scope={dirty:x,ctx:e}),Ie.$set(N);const va={};x&2&&(va.$$scope={dirty:x,ctx:e}),Xe.$set(va);const $a={};x&2&&($a.$$scope={dirty:x,ctx:e}),Ne.$set($a);const G={};x&2&&(G.$$scope={dirty:x,ctx:e}),Ge.$set(G);const Ae={};x&2&&(Ae.$$scope={dirty:x,ctx:e}),Ze.$set(Ae);const ya={};x&2&&(ya.$$scope={dirty:x,ctx:e}),Qe.$set(ya);const z={};x&2&&(z.$$scope={dirty:x,ctx:e}),eo.$set(z);const Ma={};x&2&&(Ma.$$scope={dirty:x,ctx:e}),oo.$set(Ma);const Ta={};x&2&&(Ta.$$scope={dirty:x,ctx:e}),ro.$set(Ta);const J={};x&2&&(J.$$scope={dirty:x,ctx:e}),so.$set(J);const Re={};x&2&&(Re.$$scope={dirty:x,ctx:e}),co.$set(Re);const Da={};x&2&&(Da.$$scope={dirty:x,ctx:e}),po.$set(Da);const R={};x&2&&(R.$$scope={dirty:x,ctx:e}),_o.$set(R);const Sa={};x&2&&(Sa.$$scope={dirty:x,ctx:e}),go.$set(Sa);const Ca={};x&2&&(Ca.$$scope={dirty:x,ctx:e}),bo.$set(Ca);const ae={};x&2&&(ae.$$scope={dirty:x,ctx:e}),vo.$set(ae);const Pe={};x&2&&(Pe.$$scope={dirty:x,ctx:e}),Mo.$set(Pe);const ka={};x&2&&(ka.$$scope={dirty:x,ctx:e}),Do.$set(ka);const He={};x&2&&(He.$$scope={dirty:x,ctx:e}),ko.$set(He);const A={};x&2&&(A.$$scope={dirty:x,ctx:e}),Ro.$set(A);const Aa={};x&2&&(Aa.$$scope={dirty:x,ctx:e}),Ho.$set(Aa);const Ra={};x&2&&(Ra.$$scope={dirty:x,ctx:e}),Vo.$set(Ra);const re={};x&2&&(re.$$scope={dirty:x,ctx:e}),Io.$set(re);const Ve={};x&2&&(Ve.$$scope={dirty:x,ctx:e}),Xo.$set(Ve);const Pa={};x&2&&(Pa.$$scope={dirty:x,ctx:e}),Eo.$set(Pa)},i(e){cs||(_(M.$$.fragment,e),_(Ie.$$.fragment,e),_(No.$$.fragment,e),_(zo.$$.fragment,e),_(Bo.$$.fragment,e),_(jo.$$.fragment,e),_(Go.$$.fragment,e),_(Jo.$$.fragment,e),_(Xe.$$.fragment,e),_(Zo.$$.fragment,e),_(Yo.$$.fragment,e),_(Oo.$$.fragment,e),_(Qo.$$.fragment,e),_(Ko.$$.fragment,e),_(et.$$.fragment,e),_(ot.$$.fragment,e),_(Ne.$$.fragment,e),_(tt.$$.fragment,e),_(at.$$.fragment,e),_(rt.$$.fragment,e),_(nt.$$.fragment,e),_(st.$$.fragment,e),_(it.$$.fragment,e),_(dt.$$.fragment,e),_(Ge.$$.fragment,e),_(lt.$$.fragment,e),_(ct.$$.fragment,e),_(Ze.$$.fragment,e),_(ft.$$.fragment,e),_(pt.$$.fragment,e),_(mt.$$.fragment,e),_(ht.$$.fragment,e),_(_t.$$.fragment,e),_(ut.$$.fragment,e),_(Qe.$$.fragment,e),_(gt.$$.fragment,e),_(xt.$$.fragment,e),_(eo.$$.fragment,e),_(Lt.$$.fragment,e),_(oo.$$.fragment,e),_(bt.$$.fragment,e),_(wt.$$.fragment,e),_(vt.$$.fragment,e),_($t.$$.fragment,e),_(yt.$$.fragment,e),_(ro.$$.fragment,e),_(Mt.$$.fragment,e),_(Tt.$$.fragment,e),_(so.$$.fragment,e),_(Dt.$$.fragment,e),_(St.$$.fragment,e),_(Ct.$$.fragment,e),_(kt.$$.fragment,e),_(At.$$.fragment,e),_(co.$$.fragment,e),_(Rt.$$.fragment,e),_(Pt.$$.fragment,e),_(po.$$.fragment,e),_(Ht.$$.fragment,e),_(Vt.$$.fragment,e),_(It.$$.fragment,e),_(Ft.$$.fragment,e),_(Ut.$$.fragment,e),_(_o.$$.fragment,e),_(Xt.$$.fragment,e),_(Et.$$.fragment,e),_(go.$$.fragment,e),_(Wt.$$.fragment,e),_(qt.$$.fragment,e),_(Nt.$$.fragment,e),_(zt.$$.fragment,e),_(Bt.$$.fragment,e),_(bo.$$.fragment,e),_(jt.$$.fragment,e),_(Gt.$$.fragment,e),_(vo.$$.fragment,e),_(Jt.$$.fragment,e),_(Zt.$$.fragment,e),_(Yt.$$.fragment,e),_(Ot.$$.fragment,e),_(Qt.$$.fragment,e),_(Mo.$$.fragment,e),_(Kt.$$.fragment,e),_(ea.$$.fragment,e),_(Do.$$.fragment,e),_(oa.$$.fragment,e),_(ta.$$.fragment,e),_(aa.$$.fragment,e),_(ra.$$.fragment,e),_(na.$$.fragment,e),_(ko.$$.fragment,e),_(sa.$$.fragment,e),_(ia.$$.fragment,e),_(Ro.$$.fragment,e),_(da.$$.fragment,e),_(la.$$.fragment,e),_(ca.$$.fragment,e),_(fa.$$.fragment,e),_(pa.$$.fragment,e),_(ma.$$.fragment,e),_(ha.$$.fragment,e),_(_a.$$.fragment,e),_(Ho.$$.fragment,e),_(Vo.$$.fragment,e),_(ua.$$.fragment,e),_(Io.$$.fragment,e),_(ga.$$.fragment,e),_(xa.$$.fragment,e),_(La.$$.fragment,e),_(Xo.$$.fragment,e),_(ba.$$.fragment,e),_(Eo.$$.fragment,e),_(wa.$$.fragment,e),cs=!0)},o(e){u(M.$$.fragment,e),u(Ie.$$.fragment,e),u(No.$$.fragment,e),u(zo.$$.fragment,e),u(Bo.$$.fragment,e),u(jo.$$.fragment,e),u(Go.$$.fragment,e),u(Jo.$$.fragment,e),u(Xe.$$.fragment,e),u(Zo.$$.fragment,e),u(Yo.$$.fragment,e),u(Oo.$$.fragment,e),u(Qo.$$.fragment,e),u(Ko.$$.fragment,e),u(et.$$.fragment,e),u(ot.$$.fragment,e),u(Ne.$$.fragment,e),u(tt.$$.fragment,e),u(at.$$.fragment,e),u(rt.$$.fragment,e),u(nt.$$.fragment,e),u(st.$$.fragment,e),u(it.$$.fragment,e),u(dt.$$.fragment,e),u(Ge.$$.fragment,e),u(lt.$$.fragment,e),u(ct.$$.fragment,e),u(Ze.$$.fragment,e),u(ft.$$.fragment,e),u(pt.$$.fragment,e),u(mt.$$.fragment,e),u(ht.$$.fragment,e),u(_t.$$.fragment,e),u(ut.$$.fragment,e),u(Qe.$$.fragment,e),u(gt.$$.fragment,e),u(xt.$$.fragment,e),u(eo.$$.fragment,e),u(Lt.$$.fragment,e),u(oo.$$.fragment,e),u(bt.$$.fragment,e),u(wt.$$.fragment,e),u(vt.$$.fragment,e),u($t.$$.fragment,e),u(yt.$$.fragment,e),u(ro.$$.fragment,e),u(Mt.$$.fragment,e),u(Tt.$$.fragment,e),u(so.$$.fragment,e),u(Dt.$$.fragment,e),u(St.$$.fragment,e),u(Ct.$$.fragment,e),u(kt.$$.fragment,e),u(At.$$.fragment,e),u(co.$$.fragment,e),u(Rt.$$.fragment,e),u(Pt.$$.fragment,e),u(po.$$.fragment,e),u(Ht.$$.fragment,e),u(Vt.$$.fragment,e),u(It.$$.fragment,e),u(Ft.$$.fragment,e),u(Ut.$$.fragment,e),u(_o.$$.fragment,e),u(Xt.$$.fragment,e),u(Et.$$.fragment,e),u(go.$$.fragment,e),u(Wt.$$.fragment,e),u(qt.$$.fragment,e),u(Nt.$$.fragment,e),u(zt.$$.fragment,e),u(Bt.$$.fragment,e),u(bo.$$.fragment,e),u(jt.$$.fragment,e),u(Gt.$$.fragment,e),u(vo.$$.fragment,e),u(Jt.$$.fragment,e),u(Zt.$$.fragment,e),u(Yt.$$.fragment,e),u(Ot.$$.fragment,e),u(Qt.$$.fragment,e),u(Mo.$$.fragment,e),u(Kt.$$.fragment,e),u(ea.$$.fragment,e),u(Do.$$.fragment,e),u(oa.$$.fragment,e),u(ta.$$.fragment,e),u(aa.$$.fragment,e),u(ra.$$.fragment,e),u(na.$$.fragment,e),u(ko.$$.fragment,e),u(sa.$$.fragment,e),u(ia.$$.fragment,e),u(Ro.$$.fragment,e),u(da.$$.fragment,e),u(la.$$.fragment,e),u(ca.$$.fragment,e),u(fa.$$.fragment,e),u(pa.$$.fragment,e),u(ma.$$.fragment,e),u(ha.$$.fragment,e),u(_a.$$.fragment,e),u(Ho.$$.fragment,e),u(Vo.$$.fragment,e),u(ua.$$.fragment,e),u(Io.$$.fragment,e),u(ga.$$.fragment,e),u(xa.$$.fragment,e),u(La.$$.fragment,e),u(Xo.$$.fragment,e),u(ba.$$.fragment,e),u(Eo.$$.fragment,e),u(wa.$$.fragment,e),cs=!1},d(e){e&&(i(b),i(l),i(c),i(d),i(y),i(In),i(qo),i(Fn),i(Un),i(Xn),i(V),i(En),i(Wn),i(I),i(qn),i(Nn),i(k),i(zn),i(Bn),i(D),i(jn),i(Gn),i(F),i(Jn),i(Zn),i(U),i(Yn),i(On),i(X),i(Qn),i(Kn),i(E),i(es),i(os),i(W),i(ts),i(as),i(q),i(rs),i(ns),i(ke),i(ss),i(is),i(S),i(ds),i(ls),i(Vn)),i(r),g(M,e),g(Ie,e),g(No,e),g(zo),g(Bo),g(jo),g(Go),g(Jo),g(Xe),g(Zo),g(Yo,e),g(Oo),g(Qo),g(Ko),g(et),g(ot),g(Ne),g(tt),g(at,e),g(rt),g(nt),g(st),g(it),g(dt),g(Ge),g(lt),g(ct),g(Ze),g(ft,e),g(pt),g(mt),g(ht),g(_t),g(ut),g(Qe),g(gt),g(xt),g(eo),g(Lt),g(oo),g(bt,e),g(wt),g(vt),g($t),g(yt),g(ro),g(Mt),g(Tt),g(so),g(Dt,e),g(St),g(Ct),g(kt),g(At),g(co),g(Rt),g(Pt),g(po),g(Ht,e),g(Vt),g(It),g(Ft),g(Ut),g(_o),g(Xt),g(Et),g(go),g(Wt,e),g(qt),g(Nt),g(zt),g(Bt),g(bo),g(jt),g(Gt),g(vo),g(Jt,e),g(Zt),g(Yt),g(Ot),g(Qt),g(Mo),g(Kt),g(ea),g(Do),g(oa,e),g(ta),g(aa),g(ra),g(na),g(ko),g(sa),g(ia),g(Ro),g(da,e),g(la),g(ca),g(fa,e),g(pa),g(ma),g(ha),g(_a),g(Ho),g(Vo),g(ua),g(Io),g(ga),g(xa),g(La),g(Xo),g(ba),g(Eo),g(wa,e)}}}const fp='{"title":"LoRA","local":"lora","sections":[{"title":"StableDiffusionLoraLoaderMixin","local":"diffusers.loaders.StableDiffusionLoraLoaderMixin","sections":[],"depth":2},{"title":"StableDiffusionXLLoraLoaderMixin","local":"diffusers.loaders.StableDiffusionXLLoraLoaderMixin","sections":[],"depth":2},{"title":"SD3LoraLoaderMixin","local":"diffusers.loaders.SD3LoraLoaderMixin","sections":[],"depth":2},{"title":"FluxLoraLoaderMixin","local":"diffusers.loaders.FluxLoraLoaderMixin","sections":[],"depth":2},{"title":"CogVideoXLoraLoaderMixin","local":"diffusers.loaders.CogVideoXLoraLoaderMixin","sections":[],"depth":2},{"title":"Mochi1LoraLoaderMixin","local":"diffusers.loaders.Mochi1LoraLoaderMixin","sections":[],"depth":2},{"title":"LTXVideoLoraLoaderMixin","local":"diffusers.loaders.LTXVideoLoraLoaderMixin","sections":[],"depth":2},{"title":"SanaLoraLoaderMixin","local":"diffusers.loaders.SanaLoraLoaderMixin","sections":[],"depth":2},{"title":"HunyuanVideoLoraLoaderMixin","local":"diffusers.loaders.HunyuanVideoLoraLoaderMixin","sections":[],"depth":2},{"title":"Lumina2LoraLoaderMixin","local":"diffusers.loaders.Lumina2LoraLoaderMixin","sections":[],"depth":2},{"title":"AmusedLoraLoaderMixin","local":"diffusers.loaders.AmusedLoraLoaderMixin","sections":[],"depth":2},{"title":"LoraBaseMixin","local":"diffusers.loaders.lora_base.LoraBaseMixin","sections":[],"depth":2}],"depth":1}';function pp(T){return Hf(()=>{new URLSearchParams(window.location.search).get("fw")}),[]}class bp extends Vf{constructor(r){super(),If(this,r,pp,cp,Pf,{})}}export{bp as component}; | |
Xet Storage Details
- Size:
- 232 kB
- Xet hash:
- 8955ea2281c71441ee4b1e781b4108d5d8d5b1ee8fa5376ad9ae775cbf23e727
·
Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.