Buckets:

rtrm's picture
download
raw
23.6 kB
import{s as qe,o as Ie,n as $e}from"../chunks/scheduler.8c3d61f6.js";import{S as Ne,i as Se,g as l,s,r as _,A as Ae,h as p,f as n,c as i,j as Y,u as M,x as u,k as F,y as x,a,v as b,d as v,t as y,w}from"../chunks/index.da70eac4.js";import{T as He}from"../chunks/Tip.1d9b8c37.js";import{D as ye}from"../chunks/Docstring.eabe339b.js";import{C as Ze}from"../chunks/CodeBlock.a9c4becf.js";import{E as Ee}from"../chunks/ExampleCodeBlock.4ca28371.js";import{H as we,E as We}from"../chunks/getInferenceSnippets.366c2c95.js";function Oe(C){let o,m="This pipeline is deprecated but it can still be used. However, we won’t test the pipeline anymore and won’t accept any changes to it. If you run into any issues, reinstall the last Diffusers version that supported this model.";return{c(){o=l("p"),o.textContent=m},l(r){o=p(r,"P",{"data-svelte-h":!0}),u(o)!=="svelte-ic4u42"&&(o.textContent=m)},m(r,d){a(r,o,d)},p:$e,d(r){r&&n(o)}}}function Xe(C){let o,m='Make sure to check out the Schedulers <a href="../../using-diffusers/schedulers">guide</a> to learn how to explore the tradeoff between scheduler speed and quality, and see the <a href="../../using-diffusers/loading#reuse-a-pipeline">reuse components across pipelines</a> section to learn how to efficiently load the same components into multiple pipelines.';return{c(){o=l("p"),o.innerHTML=m},l(r){o=p(r,"P",{"data-svelte-h":!0}),u(o)!=="svelte-1qn15hi"&&(o.innerHTML=m)},m(r,d){a(r,o,d)},p:$e,d(r){r&&n(o)}}}function ze(C){let o,m="Examples:",r,d,f;return d=new Ze({props:{code:"ZnJvbSUyMGRpZmZ1c2VycyUyMGltcG9ydCUyME11c2ljTERNUGlwZWxpbmUlMEFpbXBvcnQlMjB0b3JjaCUwQWltcG9ydCUyMHNjaXB5JTBBJTBBcmVwb19pZCUyMCUzRCUyMCUyMnVjc2QtcmVhY2glMkZtdXNpY2xkbSUyMiUwQXBpcGUlMjAlM0QlMjBNdXNpY0xETVBpcGVsaW5lLmZyb21fcHJldHJhaW5lZChyZXBvX2lkJTJDJTIwdG9yY2hfZHR5cGUlM0R0b3JjaC5mbG9hdDE2KSUwQXBpcGUlMjAlM0QlMjBwaXBlLnRvKCUyMmN1ZGElMjIpJTBBJTBBcHJvbXB0JTIwJTNEJTIwJTIyVGVjaG5vJTIwbXVzaWMlMjB3aXRoJTIwYSUyMHN0cm9uZyUyQyUyMHVwYmVhdCUyMHRlbXBvJTIwYW5kJTIwaGlnaCUyMG1lbG9kaWMlMjByaWZmcyUyMiUwQWF1ZGlvJTIwJTNEJTIwcGlwZShwcm9tcHQlMkMlMjBudW1faW5mZXJlbmNlX3N0ZXBzJTNEMTAlMkMlMjBhdWRpb19sZW5ndGhfaW5fcyUzRDUuMCkuYXVkaW9zJTVCMCU1RCUwQSUwQSUyMyUyMHNhdmUlMjB0aGUlMjBhdWRpbyUyMHNhbXBsZSUyMGFzJTIwYSUyMC53YXYlMjBmaWxlJTBBc2NpcHkuaW8ud2F2ZmlsZS53cml0ZSglMjJ0ZWNobm8ud2F2JTIyJTJDJTIwcmF0ZSUzRDE2MDAwJTJDJTIwZGF0YSUzRGF1ZGlvKQ==",highlighted:`<span class="hljs-meta">&gt;&gt;&gt; </span><span class="hljs-keyword">from</span> diffusers <span class="hljs-keyword">import</span> MusicLDMPipeline
<span class="hljs-meta">&gt;&gt;&gt; </span><span class="hljs-keyword">import</span> torch
<span class="hljs-meta">&gt;&gt;&gt; </span><span class="hljs-keyword">import</span> scipy
<span class="hljs-meta">&gt;&gt;&gt; </span>repo_id = <span class="hljs-string">&quot;ucsd-reach/musicldm&quot;</span>
<span class="hljs-meta">&gt;&gt;&gt; </span>pipe = MusicLDMPipeline.from_pretrained(repo_id, torch_dtype=torch.float16)
<span class="hljs-meta">&gt;&gt;&gt; </span>pipe = pipe.to(<span class="hljs-string">&quot;cuda&quot;</span>)
<span class="hljs-meta">&gt;&gt;&gt; </span>prompt = <span class="hljs-string">&quot;Techno music with a strong, upbeat tempo and high melodic riffs&quot;</span>
<span class="hljs-meta">&gt;&gt;&gt; </span>audio = pipe(prompt, num_inference_steps=<span class="hljs-number">10</span>, audio_length_in_s=<span class="hljs-number">5.0</span>).audios[<span class="hljs-number">0</span>]
<span class="hljs-meta">&gt;&gt;&gt; </span><span class="hljs-comment"># save the audio sample as a .wav file</span>
<span class="hljs-meta">&gt;&gt;&gt; </span>scipy.io.wavfile.write(<span class="hljs-string">&quot;techno.wav&quot;</span>, rate=<span class="hljs-number">16000</span>, data=audio)`,wrap:!1}}),{c(){o=l("p"),o.textContent=m,r=s(),_(d.$$.fragment)},l(c){o=p(c,"P",{"data-svelte-h":!0}),u(o)!=="svelte-kvfsh7"&&(o.textContent=m),r=i(c),M(d.$$.fragment,c)},m(c,g){a(c,o,g),a(c,r,g),b(d,c,g),f=!0},p:$e,i(c){f||(v(d.$$.fragment,c),f=!0)},o(c){y(d.$$.fragment,c),f=!1},d(c){c&&(n(o),n(r)),w(d,c)}}}function Re(C){let o,m,r,d,f,c,g,K,U,xe=`MusicLDM was proposed in <a href="https://huggingface.co/papers/2308.01546" rel="nofollow">MusicLDM: Enhancing Novelty in Text-to-Music Generation Using Beat-Synchronous Mixup Strategies</a> by Ke Chen, Yusong Wu, Haohe Liu, Marianna Nezhurina, Taylor Berg-Kirkpatrick, Shlomo Dubnov.
MusicLDM takes a text prompt as input and predicts the corresponding music sample.`,ee,j,Te=`Inspired by <a href="https://huggingface.co/docs/diffusers/api/pipelines/stable_diffusion/overview" rel="nofollow">Stable Diffusion</a> and <a href="https://huggingface.co/docs/diffusers/api/pipelines/audioldm" rel="nofollow">AudioLDM</a>,
MusicLDM is a text-to-music <em>latent diffusion model (LDM)</em> that learns continuous audio representations from <a href="https://huggingface.co/docs/transformers/main/model_doc/clap" rel="nofollow">CLAP</a>
latents.`,te,B,Le="MusicLDM is trained on a corpus of 466 hours of music data. Beat-synchronous data augmentation strategies are applied to the music samples, both in the time domain and in the latent space. Using beat-synchronous data augmentation strategies encourages the model to interpolate between the training samples, but stay within the domain of the training data. The result is generated music that is more diverse while staying faithful to the corresponding style.",ne,G,Ce="The abstract of the paper is the following:",ae,J,De="<em>Diffusion models have shown promising results in cross-modal generation tasks, including text-to-image and text-to-audio generation. However, generating music, as a special type of audio, presents unique challenges due to limited availability of music data and sensitive issues related to copyright and plagiarism. In this paper, to tackle these challenges, we first construct a state-of-the-art text-to-music model, MusicLDM, that adapts Stable Diffusion and AudioLDM architectures to the music domain. We achieve this by retraining the contrastive language-audio pretraining model (CLAP) and the Hifi-GAN vocoder, as components of MusicLDM, on a collection of music data samples. Then, to address the limitations of training data and to avoid plagiarism, we leverage a beat tracking model and propose two different mixup strategies for data augmentation: beat-synchronous audio mixup and beat-synchronous latent mixup, which recombine training audio directly or via a latent embeddings space, respectively. Such mixup strategies encourage the model to interpolate between musical training samples and generate new music within the convex hull of the training data, making the generated music more diverse while still staying faithful to the corresponding style. In addition to popular evaluation metrics, we design several new evaluation metrics based on CLAP score to demonstrate that our proposed MusicLDM and beat-synchronous mixup strategies improve both the quality and novelty of generated music, as well as the correspondence between input text and generated music.</em>",oe,H,ke='This pipeline was contributed by <a href="https://huggingface.co/sanchit-gandhi" rel="nofollow">sanchit-gandhi</a>.',se,q,ie,I,Pe="When constructing a prompt, keep in mind:",re,N,Ue="<li>Descriptive prompt inputs work best; use adjectives to describe the sound (for example, “high quality” or “clear”) and make the prompt context specific where possible (e.g. “melodic techno with a fast beat and synths” works better than “techno”).</li> <li>Using a <em>negative prompt</em> can significantly improve the quality of the generated audio. Try using a negative prompt of “low quality, average quality”.</li>",le,S,je="During inference:",pe,A,Be="<li>The <em>quality</em> of the generated audio sample can be controlled by the <code>num_inference_steps</code> argument; higher steps give higher quality audio at the expense of slower inference.</li> <li>Multiple waveforms can be generated in one go: set <code>num_waveforms_per_prompt</code> to a value greater than 1 to enable. Automatic scoring will be performed between the generated waveforms and prompt text, and the audios ranked from best to worst accordingly.</li> <li>The <em>length</em> of the generated audio sample can be controlled by varying the <code>audio_length_in_s</code> argument.</li>",ce,D,de,Z,ue,h,E,he,$,W,_e,R,Ge="The call function to the pipeline for generation.",Me,k,be,P,O,ve,V,Je=`Offloads all models to CPU using accelerate, reducing memory usage with a low impact on performance. Compared
to <code>enable_sequential_cpu_offload</code>, this method moves one whole model at a time to the accelerator when its
<code>forward</code> method is called, and the model remains in accelerator until the next model runs. Memory savings are
lower than with <code>enable_sequential_cpu_offload</code>, but performance is much better due to the iterative execution
of the <code>unet</code>.`,me,X,fe,Q,ge;return f=new He({props:{warning:!0,$$slots:{default:[Oe]},$$scope:{ctx:C}}}),g=new we({props:{title:"MusicLDM",local:"musicldm",headingTag:"h1"}}),q=new we({props:{title:"Tips",local:"tips",headingTag:"h2"}}),D=new He({props:{$$slots:{default:[Xe]},$$scope:{ctx:C}}}),Z=new we({props:{title:"MusicLDMPipeline",local:"diffusers.MusicLDMPipeline",headingTag:"h2"}}),E=new ye({props:{name:"class diffusers.MusicLDMPipeline",anchor:"diffusers.MusicLDMPipeline",parameters:[{name:"vae",val:": AutoencoderKL"},{name:"text_encoder",val:": typing.Union[transformers.models.clap.modeling_clap.ClapTextModelWithProjection, transformers.models.clap.modeling_clap.ClapModel]"},{name:"tokenizer",val:": typing.Union[transformers.models.roberta.tokenization_roberta.RobertaTokenizer, transformers.models.roberta.tokenization_roberta_fast.RobertaTokenizerFast]"},{name:"feature_extractor",val:": typing.Optional[transformers.models.clap.feature_extraction_clap.ClapFeatureExtractor]"},{name:"unet",val:": UNet2DConditionModel"},{name:"scheduler",val:": KarrasDiffusionSchedulers"},{name:"vocoder",val:": SpeechT5HifiGan"}],source:"https://github.com/huggingface/diffusers/blob/vr_11986/src/diffusers/pipelines/musicldm/pipeline_musicldm.py#L79"}}),W=new ye({props:{name:"__call__",anchor:"diffusers.MusicLDMPipeline.__call__",parameters:[{name:"prompt",val:": typing.Union[str, typing.List[str]] = None"},{name:"audio_length_in_s",val:": typing.Optional[float] = None"},{name:"num_inference_steps",val:": int = 200"},{name:"guidance_scale",val:": float = 2.0"},{name:"negative_prompt",val:": typing.Union[str, typing.List[str], NoneType] = None"},{name:"num_waveforms_per_prompt",val:": typing.Optional[int] = 1"},{name:"eta",val:": float = 0.0"},{name:"generator",val:": typing.Union[torch._C.Generator, typing.List[torch._C.Generator], NoneType] = None"},{name:"latents",val:": typing.Optional[torch.Tensor] = None"},{name:"prompt_embeds",val:": typing.Optional[torch.Tensor] = None"},{name:"negative_prompt_embeds",val:": typing.Optional[torch.Tensor] = None"},{name:"return_dict",val:": bool = True"},{name:"callback",val:": typing.Optional[typing.Callable[[int, int, torch.Tensor], NoneType]] = None"},{name:"callback_steps",val:": typing.Optional[int] = 1"},{name:"cross_attention_kwargs",val:": typing.Optional[typing.Dict[str, typing.Any]] = None"},{name:"output_type",val:": typing.Optional[str] = 'np'"}],parametersDescription:[{anchor:"diffusers.MusicLDMPipeline.__call__.prompt",description:`<strong>prompt</strong> (<code>str</code> or <code>List[str]</code>, <em>optional</em>) &#x2014;
The prompt or prompts to guide audio generation. If not defined, you need to pass <code>prompt_embeds</code>.`,name:"prompt"},{anchor:"diffusers.MusicLDMPipeline.__call__.audio_length_in_s",description:`<strong>audio_length_in_s</strong> (<code>int</code>, <em>optional</em>, defaults to 10.24) &#x2014;
The length of the generated audio sample in seconds.`,name:"audio_length_in_s"},{anchor:"diffusers.MusicLDMPipeline.__call__.num_inference_steps",description:`<strong>num_inference_steps</strong> (<code>int</code>, <em>optional</em>, defaults to 200) &#x2014;
The number of denoising steps. More denoising steps usually lead to a higher quality audio at the
expense of slower inference.`,name:"num_inference_steps"},{anchor:"diffusers.MusicLDMPipeline.__call__.guidance_scale",description:`<strong>guidance_scale</strong> (<code>float</code>, <em>optional</em>, defaults to 2.0) &#x2014;
A higher guidance scale value encourages the model to generate audio that is closely linked to the text
<code>prompt</code> at the expense of lower sound quality. Guidance scale is enabled when <code>guidance_scale &gt; 1</code>.`,name:"guidance_scale"},{anchor:"diffusers.MusicLDMPipeline.__call__.negative_prompt",description:`<strong>negative_prompt</strong> (<code>str</code> or <code>List[str]</code>, <em>optional</em>) &#x2014;
The prompt or prompts to guide what to not include in audio generation. If not defined, you need to
pass <code>negative_prompt_embeds</code> instead. Ignored when not using guidance (<code>guidance_scale &lt; 1</code>).`,name:"negative_prompt"},{anchor:"diffusers.MusicLDMPipeline.__call__.num_waveforms_per_prompt",description:`<strong>num_waveforms_per_prompt</strong> (<code>int</code>, <em>optional</em>, defaults to 1) &#x2014;
The number of waveforms to generate per prompt. If <code>num_waveforms_per_prompt &gt; 1</code>, the text encoding
model is a joint text-audio model (<a href="https://huggingface.co/docs/transformers/main/en/model_doc/clap#transformers.ClapModel" rel="nofollow">ClapModel</a>), and the tokenizer is a
<code>[~transformers.ClapProcessor]</code>, then automatic scoring will be performed between the generated outputs
and the input text. This scoring ranks the generated waveforms based on their cosine similarity to text
input in the joint text-audio embedding space.`,name:"num_waveforms_per_prompt"},{anchor:"diffusers.MusicLDMPipeline.__call__.eta",description:`<strong>eta</strong> (<code>float</code>, <em>optional</em>, defaults to 0.0) &#x2014;
Corresponds to parameter eta (&#x3B7;) from the <a href="https://huggingface.co/papers/2010.02502" rel="nofollow">DDIM</a> paper. Only
applies to the <a href="/docs/diffusers/pr_11986/en/api/schedulers/ddim#diffusers.DDIMScheduler">DDIMScheduler</a>, and is ignored in other schedulers.`,name:"eta"},{anchor:"diffusers.MusicLDMPipeline.__call__.generator",description:`<strong>generator</strong> (<code>torch.Generator</code> or <code>List[torch.Generator]</code>, <em>optional</em>) &#x2014;
A <a href="https://pytorch.org/docs/stable/generated/torch.Generator.html" rel="nofollow"><code>torch.Generator</code></a> to make
generation deterministic.`,name:"generator"},{anchor:"diffusers.MusicLDMPipeline.__call__.latents",description:`<strong>latents</strong> (<code>torch.Tensor</code>, <em>optional</em>) &#x2014;
Pre-generated noisy latents sampled from a Gaussian distribution, to be used as inputs for image
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
tensor is generated by sampling using the supplied random <code>generator</code>.`,name:"latents"},{anchor:"diffusers.MusicLDMPipeline.__call__.prompt_embeds",description:`<strong>prompt_embeds</strong> (<code>torch.Tensor</code>, <em>optional</em>) &#x2014;
Pre-generated text embeddings. Can be used to easily tweak text inputs (prompt weighting). If not
provided, text embeddings are generated from the <code>prompt</code> input argument.`,name:"prompt_embeds"},{anchor:"diffusers.MusicLDMPipeline.__call__.negative_prompt_embeds",description:`<strong>negative_prompt_embeds</strong> (<code>torch.Tensor</code>, <em>optional</em>) &#x2014;
Pre-generated negative text embeddings. Can be used to easily tweak text inputs (prompt weighting). If
not provided, <code>negative_prompt_embeds</code> are generated from the <code>negative_prompt</code> input argument.`,name:"negative_prompt_embeds"},{anchor:"diffusers.MusicLDMPipeline.__call__.return_dict",description:`<strong>return_dict</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) &#x2014;
Whether or not to return a <a href="/docs/diffusers/pr_11986/en/api/pipelines/dance_diffusion#diffusers.AudioPipelineOutput">AudioPipelineOutput</a> instead of a plain tuple.`,name:"return_dict"},{anchor:"diffusers.MusicLDMPipeline.__call__.callback",description:`<strong>callback</strong> (<code>Callable</code>, <em>optional</em>) &#x2014;
A function that calls every <code>callback_steps</code> steps during inference. The function is called with the
following arguments: <code>callback(step: int, timestep: int, latents: torch.Tensor)</code>.`,name:"callback"},{anchor:"diffusers.MusicLDMPipeline.__call__.callback_steps",description:`<strong>callback_steps</strong> (<code>int</code>, <em>optional</em>, defaults to 1) &#x2014;
The frequency at which the <code>callback</code> function is called. If not specified, the callback is called at
every step.`,name:"callback_steps"},{anchor:"diffusers.MusicLDMPipeline.__call__.cross_attention_kwargs",description:`<strong>cross_attention_kwargs</strong> (<code>dict</code>, <em>optional</em>) &#x2014;
A kwargs dictionary that if specified is passed along to the <code>AttentionProcessor</code> as defined in
<a href="https://github.com/huggingface/diffusers/blob/main/src/diffusers/models/attention_processor.py" rel="nofollow"><code>self.processor</code></a>.`,name:"cross_attention_kwargs"},{anchor:"diffusers.MusicLDMPipeline.__call__.output_type",description:`<strong>output_type</strong> (<code>str</code>, <em>optional</em>, defaults to <code>&quot;np&quot;</code>) &#x2014;
The output format of the generated audio. Choose between <code>&quot;np&quot;</code> to return a NumPy <code>np.ndarray</code> or
<code>&quot;pt&quot;</code> to return a PyTorch <code>torch.Tensor</code> object. Set to <code>&quot;latent&quot;</code> to return the latent diffusion
model (LDM) output.`,name:"output_type"}],source:"https://github.com/huggingface/diffusers/blob/vr_11986/src/diffusers/pipelines/musicldm/pipeline_musicldm.py#L433",returnDescription:`<script context="module">export const metadata = 'undefined';<\/script>
<p>If <code>return_dict</code> is <code>True</code>, <a
href="/docs/diffusers/pr_11986/en/api/pipelines/dance_diffusion#diffusers.AudioPipelineOutput"
>AudioPipelineOutput</a> is returned, otherwise a <code>tuple</code> is
returned where the first element is a list with the generated audio.</p>
`,returnType:`<script context="module">export const metadata = 'undefined';<\/script>
<p><a
href="/docs/diffusers/pr_11986/en/api/pipelines/dance_diffusion#diffusers.AudioPipelineOutput"
>AudioPipelineOutput</a> or <code>tuple</code></p>
`}}),k=new Ee({props:{anchor:"diffusers.MusicLDMPipeline.__call__.example",$$slots:{default:[ze]},$$scope:{ctx:C}}}),O=new ye({props:{name:"enable_model_cpu_offload",anchor:"diffusers.MusicLDMPipeline.enable_model_cpu_offload",parameters:[{name:"gpu_id",val:" = 0"}],source:"https://github.com/huggingface/diffusers/blob/vr_11986/src/diffusers/pipelines/musicldm/pipeline_musicldm.py#L397"}}),X=new We({props:{source:"https://github.com/huggingface/diffusers/blob/main/docs/source/en/api/pipelines/musicldm.md"}}),{c(){o=l("meta"),m=s(),r=l("p"),d=s(),_(f.$$.fragment),c=s(),_(g.$$.fragment),K=s(),U=l("p"),U.innerHTML=xe,ee=s(),j=l("p"),j.innerHTML=Te,te=s(),B=l("p"),B.textContent=Le,ne=s(),G=l("p"),G.textContent=Ce,ae=s(),J=l("p"),J.innerHTML=De,oe=s(),H=l("p"),H.innerHTML=ke,se=s(),_(q.$$.fragment),ie=s(),I=l("p"),I.textContent=Pe,re=s(),N=l("ul"),N.innerHTML=Ue,le=s(),S=l("p"),S.textContent=je,pe=s(),A=l("ul"),A.innerHTML=Be,ce=s(),_(D.$$.fragment),de=s(),_(Z.$$.fragment),ue=s(),h=l("div"),_(E.$$.fragment),he=s(),$=l("div"),_(W.$$.fragment),_e=s(),R=l("p"),R.textContent=Ge,Me=s(),_(k.$$.fragment),be=s(),P=l("div"),_(O.$$.fragment),ve=s(),V=l("p"),V.innerHTML=Je,me=s(),_(X.$$.fragment),fe=s(),Q=l("p"),this.h()},l(e){const t=Ae("svelte-u9bgzb",document.head);o=p(t,"META",{name:!0,content:!0}),t.forEach(n),m=i(e),r=p(e,"P",{}),Y(r).forEach(n),d=i(e),M(f.$$.fragment,e),c=i(e),M(g.$$.fragment,e),K=i(e),U=p(e,"P",{"data-svelte-h":!0}),u(U)!=="svelte-dgtsvr"&&(U.innerHTML=xe),ee=i(e),j=p(e,"P",{"data-svelte-h":!0}),u(j)!=="svelte-xz1yc5"&&(j.innerHTML=Te),te=i(e),B=p(e,"P",{"data-svelte-h":!0}),u(B)!=="svelte-59771y"&&(B.textContent=Le),ne=i(e),G=p(e,"P",{"data-svelte-h":!0}),u(G)!=="svelte-wu27l3"&&(G.textContent=Ce),ae=i(e),J=p(e,"P",{"data-svelte-h":!0}),u(J)!=="svelte-qsxrno"&&(J.innerHTML=De),oe=i(e),H=p(e,"P",{"data-svelte-h":!0}),u(H)!=="svelte-ig43pw"&&(H.innerHTML=ke),se=i(e),M(q.$$.fragment,e),ie=i(e),I=p(e,"P",{"data-svelte-h":!0}),u(I)!=="svelte-1dvtu0c"&&(I.textContent=Pe),re=i(e),N=p(e,"UL",{"data-svelte-h":!0}),u(N)!=="svelte-1gnv7ni"&&(N.innerHTML=Ue),le=i(e),S=p(e,"P",{"data-svelte-h":!0}),u(S)!=="svelte-1g0t9wk"&&(S.textContent=je),pe=i(e),A=p(e,"UL",{"data-svelte-h":!0}),u(A)!=="svelte-1ynscju"&&(A.innerHTML=Be),ce=i(e),M(D.$$.fragment,e),de=i(e),M(Z.$$.fragment,e),ue=i(e),h=p(e,"DIV",{class:!0});var T=Y(h);M(E.$$.fragment,T),he=i(T),$=p(T,"DIV",{class:!0});var L=Y($);M(W.$$.fragment,L),_e=i(L),R=p(L,"P",{"data-svelte-h":!0}),u(R)!=="svelte-50j04k"&&(R.textContent=Ge),Me=i(L),M(k.$$.fragment,L),L.forEach(n),be=i(T),P=p(T,"DIV",{class:!0});var z=Y(P);M(O.$$.fragment,z),ve=i(z),V=p(z,"P",{"data-svelte-h":!0}),u(V)!=="svelte-9hgqai"&&(V.innerHTML=Je),z.forEach(n),T.forEach(n),me=i(e),M(X.$$.fragment,e),fe=i(e),Q=p(e,"P",{}),Y(Q).forEach(n),this.h()},h(){F(o,"name","hf:doc:metadata"),F(o,"content",Ve),F($,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),F(P,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),F(h,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8")},m(e,t){x(document.head,o),a(e,m,t),a(e,r,t),a(e,d,t),b(f,e,t),a(e,c,t),b(g,e,t),a(e,K,t),a(e,U,t),a(e,ee,t),a(e,j,t),a(e,te,t),a(e,B,t),a(e,ne,t),a(e,G,t),a(e,ae,t),a(e,J,t),a(e,oe,t),a(e,H,t),a(e,se,t),b(q,e,t),a(e,ie,t),a(e,I,t),a(e,re,t),a(e,N,t),a(e,le,t),a(e,S,t),a(e,pe,t),a(e,A,t),a(e,ce,t),b(D,e,t),a(e,de,t),b(Z,e,t),a(e,ue,t),a(e,h,t),b(E,h,null),x(h,he),x(h,$),b(W,$,null),x($,_e),x($,R),x($,Me),b(k,$,null),x(h,be),x(h,P),b(O,P,null),x(P,ve),x(P,V),a(e,me,t),b(X,e,t),a(e,fe,t),a(e,Q,t),ge=!0},p(e,[t]){const T={};t&2&&(T.$$scope={dirty:t,ctx:e}),f.$set(T);const L={};t&2&&(L.$$scope={dirty:t,ctx:e}),D.$set(L);const z={};t&2&&(z.$$scope={dirty:t,ctx:e}),k.$set(z)},i(e){ge||(v(f.$$.fragment,e),v(g.$$.fragment,e),v(q.$$.fragment,e),v(D.$$.fragment,e),v(Z.$$.fragment,e),v(E.$$.fragment,e),v(W.$$.fragment,e),v(k.$$.fragment,e),v(O.$$.fragment,e),v(X.$$.fragment,e),ge=!0)},o(e){y(f.$$.fragment,e),y(g.$$.fragment,e),y(q.$$.fragment,e),y(D.$$.fragment,e),y(Z.$$.fragment,e),y(E.$$.fragment,e),y(W.$$.fragment,e),y(k.$$.fragment,e),y(O.$$.fragment,e),y(X.$$.fragment,e),ge=!1},d(e){e&&(n(m),n(r),n(d),n(c),n(K),n(U),n(ee),n(j),n(te),n(B),n(ne),n(G),n(ae),n(J),n(oe),n(H),n(se),n(ie),n(I),n(re),n(N),n(le),n(S),n(pe),n(A),n(ce),n(de),n(ue),n(h),n(me),n(fe),n(Q)),n(o),w(f,e),w(g,e),w(q,e),w(D,e),w(Z,e),w(E),w(W),w(k),w(O),w(X,e)}}}const Ve='{"title":"MusicLDM","local":"musicldm","sections":[{"title":"Tips","local":"tips","sections":[],"depth":2},{"title":"MusicLDMPipeline","local":"diffusers.MusicLDMPipeline","sections":[],"depth":2}],"depth":1}';function Qe(C){return Ie(()=>{new URLSearchParams(window.location.search).get("fw")}),[]}class ot extends Ne{constructor(o){super(),Se(this,o,Qe,Re,qe,{})}}export{ot as component};

Xet Storage Details

Size:
23.6 kB
·
Xet hash:
003ca049d0363b8bcf39c9e240d97860d6595eea1dab859520b919965f33a7a0

Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.