Buckets:
| import{s as ct,o as pt,n as ut}from"../chunks/scheduler.8c3d61f6.js";import{S as mt,i as ft,g as i,s as o,r as u,A as ht,h as a,f as s,c as r,j as A,u as m,x as p,k as M,y as l,a as n,v as f,d as h,t as g,w as _}from"../chunks/index.da70eac4.js";import{T as gt}from"../chunks/Tip.1d9b8c37.js";import{D as ae}from"../chunks/Docstring.ee4b6913.js";import{C as le}from"../chunks/CodeBlock.00a903b3.js";import{H as Ge,E as _t}from"../chunks/EditOnGithub.1e64e623.js";function bt(de){let d,C="🧪 This is an experimental feature!";return{c(){d=i("p"),d.textContent=C},l(D){d=a(D,"P",{"data-svelte-h":!0}),p(d)!=="svelte-15q3ih4"&&(d.textContent=C)},m(D,K){n(D,d,K)},p:ut,d(D){D&&s(d)}}}function Dt(de){let d,C,D,K,Z,ce,U,Oe='<a href="https://huggingface.co/papers/2010.02502" rel="nofollow">Denoising Diffusion Implicit Models</a> (DDIM) by Jiaming Song, Chenlin Meng and Stefano Ermon.',pe,J,Qe="The abstract from the paper is:",ue,L,Xe=`<em>Denoising diffusion probabilistic models (DDPMs) have achieved high quality image generation without adversarial training, yet they require simulating a Markov chain for many steps to produce a sample. | |
| To accelerate sampling, we present denoising diffusion implicit models (DDIMs), a more efficient class of iterative implicit probabilistic models | |
| with the same training procedure as DDPMs. In DDPMs, the generative process is defined as the reverse of a Markovian diffusion process. | |
| We construct a class of non-Markovian diffusion processes that lead to the same training objective, but whose reverse process can be much faster to sample from. | |
| We empirically demonstrate that DDIMs can produce high quality samples 10× to 50× faster in terms of wall-clock time compared to DDPMs, allow us to trade off computation for sample quality, and can perform semantically meaningful image interpolation directly in the latent space.</em>`,me,j,Fe='The original codebase of this paper can be found at <a href="https://github.com/ermongroup/ddim" rel="nofollow">ermongroup/ddim</a>, and you can contact the author on <a href="https://tsong.me/" rel="nofollow">tsong.me</a>.',fe,H,he,V,Ae='The paper <a href="https://huggingface.co/papers/2305.08891" rel="nofollow">Common Diffusion Noise Schedules and Sample Steps are Flawed</a> claims that a mismatch between the training and inference settings leads to suboptimal inference generation results for Stable Diffusion. To fix this, the authors propose:',ge,$,_e,N,Ke="<li>rescale the noise schedule to enforce zero terminal signal-to-noise ratio (SNR)</li>",be,G,De,y,et='<li>train a model with <code>v_prediction</code> (add the following argument to the <a href="https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/train_text_to_image.py" rel="nofollow">train_text_to_image.py</a> or <a href="https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/train_text_to_image_lora.py" rel="nofollow">train_text_to_image_lora.py</a> scripts)</li>',Me,W,ve,T,tt="<li>change the sampler to always start from the last timestep</li>",$e,k,ye,x,st="<li>rescale classifier-free guidance to prevent over-exposure</li>",Te,R,xe,B,nt="For example:",we,P,Ie,q,Se,c,z,We,ee,ot=`<code>DDIMScheduler</code> extends the denoising procedure introduced in denoising diffusion probabilistic models (DDPMs) with | |
| non-Markovian guidance.`,ke,te,rt=`This model inherits from <a href="/docs/diffusers/main/en/api/schedulers/overview#diffusers.SchedulerMixin">SchedulerMixin</a> and <a href="/docs/diffusers/main/en/api/configuration#diffusers.ConfigMixin">ConfigMixin</a>. Check the superclass documentation for the generic | |
| methods the library implements for all schedulers such as loading and saving.`,Re,w,E,Be,se,it=`Ensures interchangeability with schedulers that need to scale the denoising model input depending on the | |
| current timestep.`,Pe,I,Y,qe,ne,at="Sets the discrete timesteps used for the diffusion chain (to be run before inference).",ze,S,O,Ee,oe,lt=`Predict the sample from the previous timestep by reversing the SDE. This function propagates the diffusion | |
| process from the learned model outputs (most often the predicted noise).`,Ce,Q,Ze,v,X,Ye,re,dt="Output class for the scheduler’s <code>step</code> function output.",Ue,F,Je,ie,Le;return Z=new Ge({props:{title:"DDIMScheduler",local:"ddimscheduler",headingTag:"h1"}}),H=new Ge({props:{title:"Tips",local:"tips",headingTag:"h2"}}),$=new gt({props:{warning:!0,$$slots:{default:[bt]},$$scope:{ctx:de}}}),G=new le({props:{code:"cGlwZS5zY2hlZHVsZXIlMjAlM0QlMjBERElNU2NoZWR1bGVyLmZyb21fY29uZmlnKHBpcGUuc2NoZWR1bGVyLmNvbmZpZyUyQyUyMHJlc2NhbGVfYmV0YXNfemVyb19zbnIlM0RUcnVlKQ==",highlighted:'pipe.scheduler = DDIMScheduler.from_config(pipe.scheduler.config, rescale_betas_zero_snr=<span class="hljs-literal">True</span>)',wrap:!1}}),W=new le({props:{code:"LS1wcmVkaWN0aW9uX3R5cGUlM0QlMjJ2X3ByZWRpY3Rpb24lMjI=",highlighted:'--prediction_type=<span class="hljs-string">"v_prediction"</span>',wrap:!1}}),k=new le({props:{code:"cGlwZS5zY2hlZHVsZXIlMjAlM0QlMjBERElNU2NoZWR1bGVyLmZyb21fY29uZmlnKHBpcGUuc2NoZWR1bGVyLmNvbmZpZyUyQyUyMHRpbWVzdGVwX3NwYWNpbmclM0QlMjJ0cmFpbGluZyUyMik=",highlighted:'pipe.scheduler = DDIMScheduler.from_config(pipe.scheduler.config, timestep_spacing=<span class="hljs-string">"trailing"</span>)',wrap:!1}}),R=new le({props:{code:"aW1hZ2UlMjAlM0QlMjBwaXBlKHByb21wdCUyQyUyMGd1aWRhbmNlX3Jlc2NhbGUlM0QwLjcpLmltYWdlcyU1QjAlNUQ=",highlighted:'image = pipe(prompt, guidance_rescale=<span class="hljs-number">0.7</span>).images[<span class="hljs-number">0</span>]',wrap:!1}}),P=new le({props:{code:"ZnJvbSUyMGRpZmZ1c2VycyUyMGltcG9ydCUyMERpZmZ1c2lvblBpcGVsaW5lJTJDJTIwRERJTVNjaGVkdWxlciUwQWltcG9ydCUyMHRvcmNoJTBBJTBBcGlwZSUyMCUzRCUyMERpZmZ1c2lvblBpcGVsaW5lLmZyb21fcHJldHJhaW5lZCglMjJwdHgwJTJGcHNldWRvLWpvdXJuZXktdjIlMjIlMkMlMjB0b3JjaF9kdHlwZSUzRHRvcmNoLmZsb2F0MTYpJTBBcGlwZS5zY2hlZHVsZXIlMjAlM0QlMjBERElNU2NoZWR1bGVyLmZyb21fY29uZmlnKCUwQSUyMCUyMCUyMCUyMHBpcGUuc2NoZWR1bGVyLmNvbmZpZyUyQyUyMHJlc2NhbGVfYmV0YXNfemVyb19zbnIlM0RUcnVlJTJDJTIwdGltZXN0ZXBfc3BhY2luZyUzRCUyMnRyYWlsaW5nJTIyJTBBKSUwQXBpcGUudG8oJTIyY3VkYSUyMiklMEElMEFwcm9tcHQlMjAlM0QlMjAlMjJBJTIwbGlvbiUyMGluJTIwZ2FsYXhpZXMlMkMlMjBzcGlyYWxzJTJDJTIwbmVidWxhZSUyQyUyMHN0YXJzJTJDJTIwc21va2UlMkMlMjBpcmlkZXNjZW50JTJDJTIwaW50cmljYXRlJTIwZGV0YWlsJTJDJTIwb2N0YW5lJTIwcmVuZGVyJTJDJTIwOGslMjIlMEFpbWFnZSUyMCUzRCUyMHBpcGUocHJvbXB0JTJDJTIwZ3VpZGFuY2VfcmVzY2FsZSUzRDAuNykuaW1hZ2VzJTVCMCU1RCUwQWltYWdl",highlighted:`<span class="hljs-keyword">from</span> diffusers <span class="hljs-keyword">import</span> DiffusionPipeline, DDIMScheduler | |
| <span class="hljs-keyword">import</span> torch | |
| pipe = DiffusionPipeline.from_pretrained(<span class="hljs-string">"ptx0/pseudo-journey-v2"</span>, torch_dtype=torch.float16) | |
| pipe.scheduler = DDIMScheduler.from_config( | |
| pipe.scheduler.config, rescale_betas_zero_snr=<span class="hljs-literal">True</span>, timestep_spacing=<span class="hljs-string">"trailing"</span> | |
| ) | |
| pipe.to(<span class="hljs-string">"cuda"</span>) | |
| prompt = <span class="hljs-string">"A lion in galaxies, spirals, nebulae, stars, smoke, iridescent, intricate detail, octane render, 8k"</span> | |
| image = pipe(prompt, guidance_rescale=<span class="hljs-number">0.7</span>).images[<span class="hljs-number">0</span>] | |
| image`,wrap:!1}}),q=new Ge({props:{title:"DDIMScheduler",local:"diffusers.DDIMScheduler",headingTag:"h2"}}),z=new ae({props:{name:"class diffusers.DDIMScheduler",anchor:"diffusers.DDIMScheduler",parameters:[{name:"num_train_timesteps",val:": int = 1000"},{name:"beta_start",val:": float = 0.0001"},{name:"beta_end",val:": float = 0.02"},{name:"beta_schedule",val:": str = 'linear'"},{name:"trained_betas",val:": Union = None"},{name:"clip_sample",val:": bool = True"},{name:"set_alpha_to_one",val:": bool = True"},{name:"steps_offset",val:": int = 0"},{name:"prediction_type",val:": str = 'epsilon'"},{name:"thresholding",val:": bool = False"},{name:"dynamic_thresholding_ratio",val:": float = 0.995"},{name:"clip_sample_range",val:": float = 1.0"},{name:"sample_max_value",val:": float = 1.0"},{name:"timestep_spacing",val:": str = 'leading'"},{name:"rescale_betas_zero_snr",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.DDIMScheduler.num_train_timesteps",description:`<strong>num_train_timesteps</strong> (<code>int</code>, defaults to 1000) — | |
| The number of diffusion steps to train the model.`,name:"num_train_timesteps"},{anchor:"diffusers.DDIMScheduler.beta_start",description:`<strong>beta_start</strong> (<code>float</code>, defaults to 0.0001) — | |
| The starting <code>beta</code> value of inference.`,name:"beta_start"},{anchor:"diffusers.DDIMScheduler.beta_end",description:`<strong>beta_end</strong> (<code>float</code>, defaults to 0.02) — | |
| The final <code>beta</code> value.`,name:"beta_end"},{anchor:"diffusers.DDIMScheduler.beta_schedule",description:`<strong>beta_schedule</strong> (<code>str</code>, defaults to <code>"linear"</code>) — | |
| The beta schedule, a mapping from a beta range to a sequence of betas for stepping the model. Choose from | |
| <code>linear</code>, <code>scaled_linear</code>, or <code>squaredcos_cap_v2</code>.`,name:"beta_schedule"},{anchor:"diffusers.DDIMScheduler.trained_betas",description:`<strong>trained_betas</strong> (<code>np.ndarray</code>, <em>optional</em>) — | |
| Pass an array of betas directly to the constructor to bypass <code>beta_start</code> and <code>beta_end</code>.`,name:"trained_betas"},{anchor:"diffusers.DDIMScheduler.clip_sample",description:`<strong>clip_sample</strong> (<code>bool</code>, defaults to <code>True</code>) — | |
| Clip the predicted sample for numerical stability.`,name:"clip_sample"},{anchor:"diffusers.DDIMScheduler.clip_sample_range",description:`<strong>clip_sample_range</strong> (<code>float</code>, defaults to 1.0) — | |
| The maximum magnitude for sample clipping. Valid only when <code>clip_sample=True</code>.`,name:"clip_sample_range"},{anchor:"diffusers.DDIMScheduler.set_alpha_to_one",description:`<strong>set_alpha_to_one</strong> (<code>bool</code>, defaults to <code>True</code>) — | |
| Each diffusion step uses the alphas product value at that step and at the previous one. For the final step | |
| there is no previous alpha. When this option is <code>True</code> the previous alpha product is fixed to <code>1</code>, | |
| otherwise it uses the alpha value at step 0.`,name:"set_alpha_to_one"},{anchor:"diffusers.DDIMScheduler.steps_offset",description:`<strong>steps_offset</strong> (<code>int</code>, defaults to 0) — | |
| An offset added to the inference steps, as required by some model families.`,name:"steps_offset"},{anchor:"diffusers.DDIMScheduler.prediction_type",description:`<strong>prediction_type</strong> (<code>str</code>, defaults to <code>epsilon</code>, <em>optional</em>) — | |
| Prediction type of the scheduler function; can be <code>epsilon</code> (predicts the noise of the diffusion process), | |
| <code>sample</code> (directly predicts the noisy sample<code>) or </code>v_prediction\` (see section 2.4 of <a href="https://imagen.research.google/video/paper.pdf" rel="nofollow">Imagen | |
| Video</a> paper).`,name:"prediction_type"},{anchor:"diffusers.DDIMScheduler.thresholding",description:`<strong>thresholding</strong> (<code>bool</code>, defaults to <code>False</code>) — | |
| Whether to use the “dynamic thresholding” method. This is unsuitable for latent-space diffusion models such | |
| as Stable Diffusion.`,name:"thresholding"},{anchor:"diffusers.DDIMScheduler.dynamic_thresholding_ratio",description:`<strong>dynamic_thresholding_ratio</strong> (<code>float</code>, defaults to 0.995) — | |
| The ratio for the dynamic thresholding method. Valid only when <code>thresholding=True</code>.`,name:"dynamic_thresholding_ratio"},{anchor:"diffusers.DDIMScheduler.sample_max_value",description:`<strong>sample_max_value</strong> (<code>float</code>, defaults to 1.0) — | |
| The threshold value for dynamic thresholding. Valid only when <code>thresholding=True</code>.`,name:"sample_max_value"},{anchor:"diffusers.DDIMScheduler.timestep_spacing",description:`<strong>timestep_spacing</strong> (<code>str</code>, defaults to <code>"leading"</code>) — | |
| The way the timesteps should be scaled. Refer to Table 2 of the <a href="https://huggingface.co/papers/2305.08891" rel="nofollow">Common Diffusion Noise Schedules and | |
| Sample Steps are Flawed</a> for more information.`,name:"timestep_spacing"},{anchor:"diffusers.DDIMScheduler.rescale_betas_zero_snr",description:`<strong>rescale_betas_zero_snr</strong> (<code>bool</code>, defaults to <code>False</code>) — | |
| Whether to rescale the betas to have zero terminal SNR. This enables the model to generate very bright and | |
| dark samples instead of limiting it to samples with medium brightness. Loosely related to | |
| <a href="https://github.com/huggingface/diffusers/blob/74fd735eb073eb1d774b1ab4154a0876eb82f055/examples/dreambooth/train_dreambooth.py#L506" rel="nofollow"><code>--offset_noise</code></a>.`,name:"rescale_betas_zero_snr"}],source:"https://github.com/huggingface/diffusers/blob/main/src/diffusers/schedulers/scheduling_ddim.py#L131"}}),E=new ae({props:{name:"scale_model_input",anchor:"diffusers.DDIMScheduler.scale_model_input",parameters:[{name:"sample",val:": Tensor"},{name:"timestep",val:": Optional = None"}],parametersDescription:[{anchor:"diffusers.DDIMScheduler.scale_model_input.sample",description:`<strong>sample</strong> (<code>torch.Tensor</code>) — | |
| The input sample.`,name:"sample"},{anchor:"diffusers.DDIMScheduler.scale_model_input.timestep",description:`<strong>timestep</strong> (<code>int</code>, <em>optional</em>) — | |
| The current timestep in the diffusion chain.`,name:"timestep"}],source:"https://github.com/huggingface/diffusers/blob/main/src/diffusers/schedulers/scheduling_ddim.py#L236",returnDescription:`<script context="module">export const metadata = 'undefined';<\/script> | |
| <p>A scaled input sample.</p> | |
| `,returnType:`<script context="module">export const metadata = 'undefined';<\/script> | |
| <p><code>torch.Tensor</code></p> | |
| `}}),Y=new ae({props:{name:"set_timesteps",anchor:"diffusers.DDIMScheduler.set_timesteps",parameters:[{name:"num_inference_steps",val:": int"},{name:"device",val:": Union = None"}],parametersDescription:[{anchor:"diffusers.DDIMScheduler.set_timesteps.num_inference_steps",description:`<strong>num_inference_steps</strong> (<code>int</code>) — | |
| The number of diffusion steps used when generating samples with a pre-trained model.`,name:"num_inference_steps"}],source:"https://github.com/huggingface/diffusers/blob/main/src/diffusers/schedulers/scheduling_ddim.py#L297"}}),O=new ae({props:{name:"step",anchor:"diffusers.DDIMScheduler.step",parameters:[{name:"model_output",val:": Tensor"},{name:"timestep",val:": int"},{name:"sample",val:": Tensor"},{name:"eta",val:": float = 0.0"},{name:"use_clipped_model_output",val:": bool = False"},{name:"generator",val:" = None"},{name:"variance_noise",val:": Optional = None"},{name:"return_dict",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.DDIMScheduler.step.model_output",description:`<strong>model_output</strong> (<code>torch.Tensor</code>) — | |
| The direct output from learned diffusion model.`,name:"model_output"},{anchor:"diffusers.DDIMScheduler.step.timestep",description:`<strong>timestep</strong> (<code>float</code>) — | |
| The current discrete timestep in the diffusion chain.`,name:"timestep"},{anchor:"diffusers.DDIMScheduler.step.sample",description:`<strong>sample</strong> (<code>torch.Tensor</code>) — | |
| A current instance of a sample created by the diffusion process.`,name:"sample"},{anchor:"diffusers.DDIMScheduler.step.eta",description:`<strong>eta</strong> (<code>float</code>) — | |
| The weight of noise for added noise in diffusion step.`,name:"eta"},{anchor:"diffusers.DDIMScheduler.step.use_clipped_model_output",description:`<strong>use_clipped_model_output</strong> (<code>bool</code>, defaults to <code>False</code>) — | |
| If <code>True</code>, computes “corrected” <code>model_output</code> from the clipped predicted original sample. Necessary | |
| because predicted original sample is clipped to [-1, 1] when <code>self.config.clip_sample</code> is <code>True</code>. If no | |
| clipping has happened, “corrected” <code>model_output</code> would coincide with the one provided as input and | |
| <code>use_clipped_model_output</code> has no effect.`,name:"use_clipped_model_output"},{anchor:"diffusers.DDIMScheduler.step.generator",description:`<strong>generator</strong> (<code>torch.Generator</code>, <em>optional</em>) — | |
| A random number generator.`,name:"generator"},{anchor:"diffusers.DDIMScheduler.step.variance_noise",description:`<strong>variance_noise</strong> (<code>torch.Tensor</code>) — | |
| Alternative to generating noise with <code>generator</code> by directly providing the noise for the variance | |
| itself. Useful for methods such as <code>CycleDiffusion</code>.`,name:"variance_noise"},{anchor:"diffusers.DDIMScheduler.step.return_dict",description:`<strong>return_dict</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether or not to return a <a href="/docs/diffusers/main/en/api/schedulers/ddim#diffusers.schedulers.scheduling_ddim.DDIMSchedulerOutput">DDIMSchedulerOutput</a> or <code>tuple</code>.`,name:"return_dict"}],source:"https://github.com/huggingface/diffusers/blob/main/src/diffusers/schedulers/scheduling_ddim.py#L342",returnDescription:`<script context="module">export const metadata = 'undefined';<\/script> | |
| <p>If return_dict is <code>True</code>, <a | |
| href="/docs/diffusers/main/en/api/schedulers/ddim#diffusers.schedulers.scheduling_ddim.DDIMSchedulerOutput" | |
| >DDIMSchedulerOutput</a> is returned, otherwise a | |
| tuple is returned where the first element is the sample tensor.</p> | |
| `,returnType:`<script context="module">export const metadata = 'undefined';<\/script> | |
| <p><a | |
| href="/docs/diffusers/main/en/api/schedulers/ddim#diffusers.schedulers.scheduling_ddim.DDIMSchedulerOutput" | |
| >DDIMSchedulerOutput</a> or <code>tuple</code></p> | |
| `}}),Q=new Ge({props:{title:"DDIMSchedulerOutput",local:"diffusers.schedulers.scheduling_ddim.DDIMSchedulerOutput",headingTag:"h2"}}),X=new ae({props:{name:"class diffusers.schedulers.scheduling_ddim.DDIMSchedulerOutput",anchor:"diffusers.schedulers.scheduling_ddim.DDIMSchedulerOutput",parameters:[{name:"prev_sample",val:": Tensor"},{name:"pred_original_sample",val:": Optional = None"}],parametersDescription:[{anchor:"diffusers.schedulers.scheduling_ddim.DDIMSchedulerOutput.prev_sample",description:`<strong>prev_sample</strong> (<code>torch.Tensor</code> of shape <code>(batch_size, num_channels, height, width)</code> for images) — | |
| Computed sample <code>(x_{t-1})</code> of previous timestep. <code>prev_sample</code> should be used as next model input in the | |
| denoising loop.`,name:"prev_sample"},{anchor:"diffusers.schedulers.scheduling_ddim.DDIMSchedulerOutput.pred_original_sample",description:`<strong>pred_original_sample</strong> (<code>torch.Tensor</code> of shape <code>(batch_size, num_channels, height, width)</code> for images) — | |
| The predicted denoised sample <code>(x_{0})</code> based on the model output from the current timestep. | |
| <code>pred_original_sample</code> can be used to preview progress or for guidance.`,name:"pred_original_sample"}],source:"https://github.com/huggingface/diffusers/blob/main/src/diffusers/schedulers/scheduling_ddim.py#L31"}}),F=new _t({props:{source:"https://github.com/huggingface/diffusers/blob/main/docs/source/en/api/schedulers/ddim.md"}}),{c(){d=i("meta"),C=o(),D=i("p"),K=o(),u(Z.$$.fragment),ce=o(),U=i("p"),U.innerHTML=Oe,pe=o(),J=i("p"),J.textContent=Qe,ue=o(),L=i("p"),L.innerHTML=Xe,me=o(),j=i("p"),j.innerHTML=Fe,fe=o(),u(H.$$.fragment),he=o(),V=i("p"),V.innerHTML=Ae,ge=o(),u($.$$.fragment),_e=o(),N=i("ol"),N.innerHTML=Ke,be=o(),u(G.$$.fragment),De=o(),y=i("ol"),y.innerHTML=et,Me=o(),u(W.$$.fragment),ve=o(),T=i("ol"),T.innerHTML=tt,$e=o(),u(k.$$.fragment),ye=o(),x=i("ol"),x.innerHTML=st,Te=o(),u(R.$$.fragment),xe=o(),B=i("p"),B.textContent=nt,we=o(),u(P.$$.fragment),Ie=o(),u(q.$$.fragment),Se=o(),c=i("div"),u(z.$$.fragment),We=o(),ee=i("p"),ee.innerHTML=ot,ke=o(),te=i("p"),te.innerHTML=rt,Re=o(),w=i("div"),u(E.$$.fragment),Be=o(),se=i("p"),se.textContent=it,Pe=o(),I=i("div"),u(Y.$$.fragment),qe=o(),ne=i("p"),ne.textContent=at,ze=o(),S=i("div"),u(O.$$.fragment),Ee=o(),oe=i("p"),oe.textContent=lt,Ce=o(),u(Q.$$.fragment),Ze=o(),v=i("div"),u(X.$$.fragment),Ye=o(),re=i("p"),re.innerHTML=dt,Ue=o(),u(F.$$.fragment),Je=o(),ie=i("p"),this.h()},l(e){const t=ht("svelte-u9bgzb",document.head);d=a(t,"META",{name:!0,content:!0}),t.forEach(s),C=r(e),D=a(e,"P",{}),A(D).forEach(s),K=r(e),m(Z.$$.fragment,e),ce=r(e),U=a(e,"P",{"data-svelte-h":!0}),p(U)!=="svelte-1qeb1a3"&&(U.innerHTML=Oe),pe=r(e),J=a(e,"P",{"data-svelte-h":!0}),p(J)!=="svelte-1cwsb16"&&(J.textContent=Qe),ue=r(e),L=a(e,"P",{"data-svelte-h":!0}),p(L)!=="svelte-1h59s10"&&(L.innerHTML=Xe),me=r(e),j=a(e,"P",{"data-svelte-h":!0}),p(j)!=="svelte-1stib0u"&&(j.innerHTML=Fe),fe=r(e),m(H.$$.fragment,e),he=r(e),V=a(e,"P",{"data-svelte-h":!0}),p(V)!=="svelte-nsfv69"&&(V.innerHTML=Ae),ge=r(e),m($.$$.fragment,e),_e=r(e),N=a(e,"OL",{"data-svelte-h":!0}),p(N)!=="svelte-hhk4t"&&(N.innerHTML=Ke),be=r(e),m(G.$$.fragment,e),De=r(e),y=a(e,"OL",{start:!0,"data-svelte-h":!0}),p(y)!=="svelte-18uxl2t"&&(y.innerHTML=et),Me=r(e),m(W.$$.fragment,e),ve=r(e),T=a(e,"OL",{start:!0,"data-svelte-h":!0}),p(T)!=="svelte-17ct3yw"&&(T.innerHTML=tt),$e=r(e),m(k.$$.fragment,e),ye=r(e),x=a(e,"OL",{start:!0,"data-svelte-h":!0}),p(x)!=="svelte-1a6ampg"&&(x.innerHTML=st),Te=r(e),m(R.$$.fragment,e),xe=r(e),B=a(e,"P",{"data-svelte-h":!0}),p(B)!=="svelte-1gkqha7"&&(B.textContent=nt),we=r(e),m(P.$$.fragment,e),Ie=r(e),m(q.$$.fragment,e),Se=r(e),c=a(e,"DIV",{class:!0});var b=A(c);m(z.$$.fragment,b),We=r(b),ee=a(b,"P",{"data-svelte-h":!0}),p(ee)!=="svelte-1wae91k"&&(ee.innerHTML=ot),ke=r(b),te=a(b,"P",{"data-svelte-h":!0}),p(te)!=="svelte-linuuh"&&(te.innerHTML=rt),Re=r(b),w=a(b,"DIV",{class:!0});var je=A(w);m(E.$$.fragment,je),Be=r(je),se=a(je,"P",{"data-svelte-h":!0}),p(se)!=="svelte-1rkfgpx"&&(se.textContent=it),je.forEach(s),Pe=r(b),I=a(b,"DIV",{class:!0});var He=A(I);m(Y.$$.fragment,He),qe=r(He),ne=a(He,"P",{"data-svelte-h":!0}),p(ne)!=="svelte-1vzm9q"&&(ne.textContent=at),He.forEach(s),ze=r(b),S=a(b,"DIV",{class:!0});var Ve=A(S);m(O.$$.fragment,Ve),Ee=r(Ve),oe=a(Ve,"P",{"data-svelte-h":!0}),p(oe)!=="svelte-hi84tp"&&(oe.textContent=lt),Ve.forEach(s),b.forEach(s),Ce=r(e),m(Q.$$.fragment,e),Ze=r(e),v=a(e,"DIV",{class:!0});var Ne=A(v);m(X.$$.fragment,Ne),Ye=r(Ne),re=a(Ne,"P",{"data-svelte-h":!0}),p(re)!=="svelte-id9kic"&&(re.innerHTML=dt),Ne.forEach(s),Ue=r(e),m(F.$$.fragment,e),Je=r(e),ie=a(e,"P",{}),A(ie).forEach(s),this.h()},h(){M(d,"name","hf:doc:metadata"),M(d,"content",Mt),M(y,"start","2"),M(T,"start","3"),M(x,"start","4"),M(w,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),M(I,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),M(S,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),M(c,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),M(v,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8")},m(e,t){l(document.head,d),n(e,C,t),n(e,D,t),n(e,K,t),f(Z,e,t),n(e,ce,t),n(e,U,t),n(e,pe,t),n(e,J,t),n(e,ue,t),n(e,L,t),n(e,me,t),n(e,j,t),n(e,fe,t),f(H,e,t),n(e,he,t),n(e,V,t),n(e,ge,t),f($,e,t),n(e,_e,t),n(e,N,t),n(e,be,t),f(G,e,t),n(e,De,t),n(e,y,t),n(e,Me,t),f(W,e,t),n(e,ve,t),n(e,T,t),n(e,$e,t),f(k,e,t),n(e,ye,t),n(e,x,t),n(e,Te,t),f(R,e,t),n(e,xe,t),n(e,B,t),n(e,we,t),f(P,e,t),n(e,Ie,t),f(q,e,t),n(e,Se,t),n(e,c,t),f(z,c,null),l(c,We),l(c,ee),l(c,ke),l(c,te),l(c,Re),l(c,w),f(E,w,null),l(w,Be),l(w,se),l(c,Pe),l(c,I),f(Y,I,null),l(I,qe),l(I,ne),l(c,ze),l(c,S),f(O,S,null),l(S,Ee),l(S,oe),n(e,Ce,t),f(Q,e,t),n(e,Ze,t),n(e,v,t),f(X,v,null),l(v,Ye),l(v,re),n(e,Ue,t),f(F,e,t),n(e,Je,t),n(e,ie,t),Le=!0},p(e,[t]){const b={};t&2&&(b.$$scope={dirty:t,ctx:e}),$.$set(b)},i(e){Le||(h(Z.$$.fragment,e),h(H.$$.fragment,e),h($.$$.fragment,e),h(G.$$.fragment,e),h(W.$$.fragment,e),h(k.$$.fragment,e),h(R.$$.fragment,e),h(P.$$.fragment,e),h(q.$$.fragment,e),h(z.$$.fragment,e),h(E.$$.fragment,e),h(Y.$$.fragment,e),h(O.$$.fragment,e),h(Q.$$.fragment,e),h(X.$$.fragment,e),h(F.$$.fragment,e),Le=!0)},o(e){g(Z.$$.fragment,e),g(H.$$.fragment,e),g($.$$.fragment,e),g(G.$$.fragment,e),g(W.$$.fragment,e),g(k.$$.fragment,e),g(R.$$.fragment,e),g(P.$$.fragment,e),g(q.$$.fragment,e),g(z.$$.fragment,e),g(E.$$.fragment,e),g(Y.$$.fragment,e),g(O.$$.fragment,e),g(Q.$$.fragment,e),g(X.$$.fragment,e),g(F.$$.fragment,e),Le=!1},d(e){e&&(s(C),s(D),s(K),s(ce),s(U),s(pe),s(J),s(ue),s(L),s(me),s(j),s(fe),s(he),s(V),s(ge),s(_e),s(N),s(be),s(De),s(y),s(Me),s(ve),s(T),s($e),s(ye),s(x),s(Te),s(xe),s(B),s(we),s(Ie),s(Se),s(c),s(Ce),s(Ze),s(v),s(Ue),s(Je),s(ie)),s(d),_(Z,e),_(H,e),_($,e),_(G,e),_(W,e),_(k,e),_(R,e),_(P,e),_(q,e),_(z),_(E),_(Y),_(O),_(Q,e),_(X),_(F,e)}}}const Mt='{"title":"DDIMScheduler","local":"ddimscheduler","sections":[{"title":"Tips","local":"tips","sections":[],"depth":2},{"title":"DDIMScheduler","local":"diffusers.DDIMScheduler","sections":[],"depth":2},{"title":"DDIMSchedulerOutput","local":"diffusers.schedulers.scheduling_ddim.DDIMSchedulerOutput","sections":[],"depth":2}],"depth":1}';function vt(de){return pt(()=>{new URLSearchParams(window.location.search).get("fw")}),[]}class St extends mt{constructor(d){super(),ft(this,d,vt,Dt,ct,{})}}export{St as component}; | |
Xet Storage Details
- Size:
- 25.8 kB
- Xet hash:
- 6cd9f0308f3a0fe5ea734abe0cf9ba85e29c9664ca39939a60a1974f9e711387
·
Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.