diff --git "a/assets/worker-DJqCEpj4.js" "b/assets/worker-DJqCEpj4.js" --- "a/assets/worker-DJqCEpj4.js" +++ "b/assets/worker-DJqCEpj4.js" @@ -2842,4 +2842,4 @@ ${w}${b}${k} ${_}${I}`+y.repeat(d)+`${_}`,w}function l(d,u,f,_){return`${u}${_}`+f.repeat(d)+`${u}`}function c(d,u,f,_,y,I){return d===0&&u===0?l(f,_,y,I):a(f,d,u,_,y,I)}class p extends s.Processor{static image_processor_class=n.AutoImageProcessor;static tokenizer_class=o.AutoTokenizer;static uses_processor_config=!0;fake_image_token="";image_token="";global_img_token="";async _call(u,f=null,_={}){_.return_row_col_info??=!0;let y;f&&(y=await this.image_processor(f,_)),Array.isArray(u)||(u=[u]);const I=y.rows??[new Array(u.length).fill(0)],w=y.cols??[new Array(u.length).fill(0)],v=this.config.image_seq_len,k=[],T=[];for(let P=0;Pc(B,O[Y],v,this.fake_image_token,this.image_token,this.global_img_token)),H=x.split(this.image_token);if(H.length===0)throw new Error("The image token should be present in the text.");let W=H[0];for(let B=0;B{t.r(r),t.d(r,{BeitFeatureExtractor:()=>s.BeitFeatureExtractor,BitImageProcessor:()=>n.BitImageProcessor,CLIPFeatureExtractor:()=>i.CLIPFeatureExtractor,CLIPImageProcessor:()=>i.CLIPImageProcessor,ChineseCLIPFeatureExtractor:()=>o.ChineseCLIPFeatureExtractor,ConvNextFeatureExtractor:()=>a.ConvNextFeatureExtractor,ConvNextImageProcessor:()=>a.ConvNextImageProcessor,DINOv3ViTImageProcessor:()=>p.DINOv3ViTImageProcessor,DPTFeatureExtractor:()=>u.DPTFeatureExtractor,DPTImageProcessor:()=>u.DPTImageProcessor,DeiTFeatureExtractor:()=>l.DeiTFeatureExtractor,DeiTImageProcessor:()=>l.DeiTImageProcessor,DetrFeatureExtractor:()=>c.DetrFeatureExtractor,DetrImageProcessor:()=>c.DetrImageProcessor,DonutFeatureExtractor:()=>d.DonutFeatureExtractor,DonutImageProcessor:()=>d.DonutImageProcessor,EfficientNetImageProcessor:()=>f.EfficientNetImageProcessor,GLPNFeatureExtractor:()=>_.GLPNFeatureExtractor,GroundingDinoImageProcessor:()=>y.GroundingDinoImageProcessor,Idefics3ImageProcessor:()=>I.Idefics3ImageProcessor,JinaCLIPImageProcessor:()=>v.JinaCLIPImageProcessor,LlavaOnevisionImageProcessor:()=>k.LlavaOnevisionImageProcessor,Mask2FormerImageProcessor:()=>T.Mask2FormerImageProcessor,MaskFormerFeatureExtractor:()=>b.MaskFormerFeatureExtractor,MaskFormerImageProcessor:()=>b.MaskFormerImageProcessor,MobileNetV1FeatureExtractor:()=>P.MobileNetV1FeatureExtractor,MobileNetV1ImageProcessor:()=>P.MobileNetV1ImageProcessor,MobileNetV2FeatureExtractor:()=>x.MobileNetV2FeatureExtractor,MobileNetV2ImageProcessor:()=>x.MobileNetV2ImageProcessor,MobileNetV3FeatureExtractor:()=>S.MobileNetV3FeatureExtractor,MobileNetV3ImageProcessor:()=>S.MobileNetV3ImageProcessor,MobileNetV4FeatureExtractor:()=>O.MobileNetV4FeatureExtractor,MobileNetV4ImageProcessor:()=>O.MobileNetV4ImageProcessor,MobileViTFeatureExtractor:()=>F.MobileViTFeatureExtractor,MobileViTImageProcessor:()=>F.MobileViTImageProcessor,NougatImageProcessor:()=>H.NougatImageProcessor,OwlViTFeatureExtractor:()=>B.OwlViTFeatureExtractor,OwlViTImageProcessor:()=>B.OwlViTImageProcessor,Owlv2ImageProcessor:()=>W.Owlv2ImageProcessor,Phi3VImageProcessor:()=>Y.Phi3VImageProcessor,PixtralImageProcessor:()=>X.PixtralImageProcessor,PvtImageProcessor:()=>J.PvtImageProcessor,Qwen2VLImageProcessor:()=>re.Qwen2VLImageProcessor,RTDetrImageProcessor:()=>ne.RTDetrImageProcessor,Sam2ImageProcessor:()=>pe.Sam2ImageProcessor,Sam3ImageProcessor:()=>oe.Sam3ImageProcessor,SamImageProcessor:()=>le.SamImageProcessor,SegformerFeatureExtractor:()=>K.SegformerFeatureExtractor,SegformerImageProcessor:()=>K.SegformerImageProcessor,SiglipImageProcessor:()=>j.SiglipImageProcessor,SmolVLMImageProcessor:()=>D.SmolVLMImageProcessor,Swin2SRImageProcessor:()=>te.Swin2SRImageProcessor,VLMImageProcessor:()=>w.VLMImageProcessor,ViTFeatureExtractor:()=>he.ViTFeatureExtractor,ViTImageProcessor:()=>he.ViTImageProcessor,VitMatteImageProcessor:()=>Ae.VitMatteImageProcessor,VitPoseImageProcessor:()=>ke.VitPoseImageProcessor,YolosFeatureExtractor:()=>Ve.YolosFeatureExtractor,YolosImageProcessor:()=>Ve.YolosImageProcessor});var s=t("./src/models/beit/image_processing_beit.js"),n=t("./src/models/bit/image_processing_bit.js"),o=t("./src/models/chinese_clip/image_processing_chinese_clip.js"),i=t("./src/models/clip/image_processing_clip.js"),a=t("./src/models/convnext/image_processing_convnext.js"),l=t("./src/models/deit/image_processing_deit.js"),c=t("./src/models/detr/image_processing_detr.js"),p=t("./src/models/dinov3_vit/image_processing_dinov3_vit.js"),d=t("./src/models/donut/image_processing_donut.js"),u=t("./src/models/dpt/image_processing_dpt.js"),f=t("./src/models/efficientnet/image_processing_efficientnet.js"),_=t("./src/models/glpn/image_processing_glpn.js"),y=t("./src/models/grounding_dino/image_processing_grounding_dino.js"),I=t("./src/models/idefics3/image_processing_idefics3.js"),w=t("./src/models/janus/image_processing_janus.js"),v=t("./src/models/jina_clip/image_processing_jina_clip.js"),k=t("./src/models/llava_onevision/image_processing_llava_onevision.js"),T=t("./src/models/mask2former/image_processing_mask2former.js"),b=t("./src/models/maskformer/image_processing_maskformer.js"),P=t("./src/models/mobilenet_v1/image_processing_mobilenet_v1.js"),x=t("./src/models/mobilenet_v2/image_processing_mobilenet_v2.js"),S=t("./src/models/mobilenet_v3/image_processing_mobilenet_v3.js"),O=t("./src/models/mobilenet_v4/image_processing_mobilenet_v4.js"),F=t("./src/models/mobilevit/image_processing_mobilevit.js"),H=t("./src/models/nougat/image_processing_nougat.js"),W=t("./src/models/owlv2/image_processing_owlv2.js"),B=t("./src/models/owlvit/image_processing_owlvit.js"),Y=t("./src/models/phi3_v/image_processing_phi3_v.js"),X=t("./src/models/pixtral/image_processing_pixtral.js"),J=t("./src/models/pvt/image_processing_pvt.js"),re=t("./src/models/qwen2_vl/image_processing_qwen2_vl.js"),ne=t("./src/models/rt_detr/image_processing_rt_detr.js"),le=t("./src/models/sam/image_processing_sam.js"),pe=t("./src/models/sam2/image_processing_sam2.js"),oe=t("./src/models/sam3/image_processing_sam3.js"),K=t("./src/models/segformer/image_processing_segformer.js"),j=t("./src/models/siglip/image_processing_siglip.js"),D=t("./src/models/smolvlm/image_processing_smolvlm.js"),te=t("./src/models/swin2sr/image_processing_swin2sr.js"),he=t("./src/models/vit/image_processing_vit.js"),Ae=t("./src/models/vitmatte/image_processing_vitmatte.js"),ke=t("./src/models/vitpose/image_processing_vitpose.js"),Ve=t("./src/models/yolos/image_processing_yolos.js")}),"./src/models/janus/image_processing_janus.js":((e,r,t)=>{t.r(r),t.d(r,{VLMImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{constructor(i){super({do_pad:!0,pad_size:{width:i.image_size,height:i.image_size},...i}),this.constant_values=this.config.background_color.map(a=>a*this.rescale_factor)}pad_image(i,a,l,c){return super.pad_image(i,a,l,{constant_values:this.constant_values,center:!0,...c})}}}),"./src/models/janus/processing_janus.js":((e,r,t)=>{t.r(r),t.d(r,{VLChatProcessor:()=>c});var s=t("./src/base/processing_utils.js"),n=t("./src/models/auto/image_processing_auto.js"),o=t("./src/tokenizers.js"),i=t("./src/utils/core.js"),a=t("./src/utils/tensor.js"),l=t("./src/utils/image.js");class c extends s.Processor{static image_processor_class=n.AutoImageProcessor;static tokenizer_class=o.AutoTokenizer;static uses_processor_config=!0;constructor(d,u,f){super(d,u,f),this.image_tag=this.config.image_tag,this.image_start_tag=this.config.image_start_tag,this.image_end_tag=this.config.image_end_tag,this.num_image_tokens=this.config.num_image_tokens}async _call(d,{images:u=null,chat_template:f="default"}={}){u?Array.isArray(u)||(u=[u]):u=await Promise.all(d.filter(F=>F.images).flatMap(F=>F.images).map(F=>l.RawImage.read(F)));const _=this.tokenizer,y=_.apply_chat_template(d,{tokenize:!1,add_generation_prompt:!0,chat_template:f}),I=F=>_.encode(F,{add_special_tokens:!1}),w=y.split(this.image_tag),v=w.length-1;if(u.length!==v)throw new Error(`Number of images provided (${u.length}) does not match number of "${this.image_tag}" image tags (${v})`);const[k,T,b]=_.model.convert_tokens_to_ids([this.image_tag,this.image_start_tag,this.image_end_tag]);let P=I(w[0]),x=new Array(P.length).fill(!1);for(let F=1;F0){const F=await this.image_processor(u);return F.pixel_values.unsqueeze_(0),{...O,...F}}return O}}}),"./src/models/jina_clip/image_processing_jina_clip.js":((e,r,t)=>{t.r(r),t.d(r,{JinaCLIPImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{constructor(i){const{resize_mode:a,fill_color:l,interpolation:c,size:p,...d}=i,u=a==="squash"?{width:p,height:p}:a==="shortest"?{shortest_edge:p}:{longest_edge:p},f=c==="bicubic"?3:2;super({...d,size:u,resample:f,do_center_crop:!0,crop_size:p,do_normalize:!0})}}}),"./src/models/jina_clip/processing_jina_clip.js":((e,r,t)=>{t.r(r),t.d(r,{JinaCLIPProcessor:()=>i});var s=t("./src/base/processing_utils.js"),n=t("./src/models/auto/image_processing_auto.js"),o=t("./src/tokenizers.js");class i extends s.Processor{static tokenizer_class=o.AutoTokenizer;static image_processor_class=n.AutoImageProcessor;async _call(l=null,c=null,p={}){if(!l&&!c)throw new Error("Either text or images must be provided");const d=l?this.tokenizer(l,p):{},u=c?await this.image_processor(c,p):{};return{...d,...u}}}}),"./src/models/llava/processing_llava.js":((e,r,t)=>{t.r(r),t.d(r,{LlavaProcessor:()=>i});var s=t("./src/base/processing_utils.js"),n=t("./src/models/auto/image_processing_auto.js"),o=t("./src/tokenizers.js");class i extends s.Processor{static tokenizer_class=o.AutoTokenizer;static image_processor_class=n.AutoImageProcessor;static uses_processor_config=!0;async _call(l,c=null,p={}){const d=await this.image_processor(l,p);if(c){const[f,_]=d.pixel_values.dims.slice(-2),{image_token:y,patch_size:I,num_additional_image_tokens:w}=this.config,v=Math.floor(f/I)*Math.floor(_/I)+w;c=structuredClone(c),Array.isArray(c)||(c=[c]);for(let k=0;k{t.r(r),t.d(r,{LlavaOnevisionImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{}}),"./src/models/mask2former/image_processing_mask2former.js":((e,r,t)=>{t.r(r),t.d(r,{Mask2FormerImageProcessor:()=>n});var s=t("./src/models/maskformer/image_processing_maskformer.js");class n extends s.MaskFormerImageProcessor{}}),"./src/models/maskformer/image_processing_maskformer.js":((e,r,t)=>{t.r(r),t.d(r,{MaskFormerFeatureExtractor:()=>o,MaskFormerImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{post_process_panoptic_segmentation(...a){return(0,s.post_process_panoptic_segmentation)(...a)}post_process_instance_segmentation(...a){return(0,s.post_process_instance_segmentation)(...a)}}class o extends n{}}),"./src/models/mgp_str/processing_mgp_str.js":((e,r,t)=>{t.r(r),t.d(r,{MgpstrProcessor:()=>l});var s=t("./src/base/processing_utils.js"),n=t("./src/models/auto/image_processing_auto.js"),o=t("./src/tokenizers.js"),i=t("./src/utils/maths.js");const a={char:["char_decode",1],bpe:["bpe_decode",2],wp:["wp_decode",102]};class l extends s.Processor{static tokenizer_class=o.AutoTokenizer;static image_processor_class=n.AutoImageProcessor;get char_tokenizer(){return this.components.char_tokenizer}get bpe_tokenizer(){return this.components.bpe_tokenizer}get wp_tokenizer(){return this.components.wp_tokenizer}_decode_helper(p,d){if(!a.hasOwnProperty(d))throw new Error(`Format ${d} is not supported.`);const[u,f]=a[d],_=this[u].bind(this),[y,I]=p.dims,w=[],v=[],k=p.tolist();for(let b=0;b0?S.reduce((F,H)=>F*H,1):0;v.push(x),w.push(O)}return[_(v),w]}char_decode(p){return this.char_tokenizer.batch_decode(p).map(d=>d.replaceAll(" ",""))}bpe_decode(p){return this.bpe_tokenizer.batch_decode(p)}wp_decode(p){return this.wp_tokenizer.batch_decode(p).map(d=>d.replaceAll(" ",""))}batch_decode([p,d,u]){const[f,_]=this._decode_helper(p,"char"),[y,I]=this._decode_helper(d,"bpe"),[w,v]=this._decode_helper(u,"wp"),k=[],T=[];for(let b=0;b{t.r(r),t.d(r,{MobileNetV1FeatureExtractor:()=>o,MobileNetV1ImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{}class o extends n{}}),"./src/models/mobilenet_v2/image_processing_mobilenet_v2.js":((e,r,t)=>{t.r(r),t.d(r,{MobileNetV2FeatureExtractor:()=>o,MobileNetV2ImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{}class o extends n{}}),"./src/models/mobilenet_v3/image_processing_mobilenet_v3.js":((e,r,t)=>{t.r(r),t.d(r,{MobileNetV3FeatureExtractor:()=>o,MobileNetV3ImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{}class o extends n{}}),"./src/models/mobilenet_v4/image_processing_mobilenet_v4.js":((e,r,t)=>{t.r(r),t.d(r,{MobileNetV4FeatureExtractor:()=>o,MobileNetV4ImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{}class o extends n{}}),"./src/models/mobilevit/image_processing_mobilevit.js":((e,r,t)=>{t.r(r),t.d(r,{MobileViTFeatureExtractor:()=>o,MobileViTImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{}class o extends n{}}),"./src/models/moonshine/feature_extraction_moonshine.js":((e,r,t)=>{t.r(r),t.d(r,{MoonshineFeatureExtractor:()=>o});var s=t("./src/base/feature_extraction_utils.js"),n=t("./src/utils/tensor.js");class o extends s.FeatureExtractor{async _call(a){(0,s.validate_audio_inputs)(a,"MoonshineFeatureExtractor"),a instanceof Float64Array&&(a=new Float32Array(a));const l=[1,a.length];return{input_values:new n.Tensor("float32",a,l)}}}}),"./src/models/moonshine/processing_moonshine.js":((e,r,t)=>{t.r(r),t.d(r,{MoonshineProcessor:()=>i});var s=t("./src/models/auto/feature_extraction_auto.js"),n=t("./src/tokenizers.js"),o=t("./src/base/processing_utils.js");class i extends o.Processor{static tokenizer_class=n.AutoTokenizer;static feature_extractor_class=s.AutoFeatureExtractor;async _call(l){return await this.feature_extractor(l)}}}),"./src/models/nougat/image_processing_nougat.js":((e,r,t)=>{t.r(r),t.d(r,{NougatImageProcessor:()=>n});var s=t("./src/models/donut/image_processing_donut.js");class n extends s.DonutImageProcessor{}}),"./src/models/owlv2/image_processing_owlv2.js":((e,r,t)=>{t.r(r),t.d(r,{Owlv2ImageProcessor:()=>n});var s=t("./src/models/owlvit/image_processing_owlvit.js");class n extends s.OwlViTImageProcessor{}}),"./src/models/owlvit/image_processing_owlvit.js":((e,r,t)=>{t.r(r),t.d(r,{OwlViTFeatureExtractor:()=>o,OwlViTImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{post_process_object_detection(...a){return(0,s.post_process_object_detection)(...a)}}class o extends n{}}),"./src/models/owlvit/processing_owlvit.js":((e,r,t)=>{t.r(r),t.d(r,{OwlViTProcessor:()=>i});var s=t("./src/base/processing_utils.js"),n=t("./src/models/auto/image_processing_auto.js"),o=t("./src/tokenizers.js");class i extends s.Processor{static tokenizer_class=o.AutoTokenizer;static image_processor_class=n.AutoImageProcessor}}),"./src/models/paligemma/processing_paligemma.js":((e,r,t)=>{t.r(r),t.d(r,{PaliGemmaProcessor:()=>l});var s=t("./src/base/processing_utils.js"),n=t("./src/models/auto/image_processing_auto.js"),o=t("./src/tokenizers.js");const i="";function a(c,p,d,u,f){return`${u.repeat(d*f)}${p}${c} `}class l extends s.Processor{static tokenizer_class=o.AutoTokenizer;static image_processor_class=n.AutoImageProcessor;static uses_processor_config=!1;async _call(p,d=null,u={}){d||(console.warn("You are using PaliGemma without a text prefix. It will perform as a picture-captioning model."),d=""),Array.isArray(p)||(p=[p]),Array.isArray(d)||(d=[d]);const f=this.tokenizer.bos_token,_=this.image_processor.config.image_seq_length;let y;d.some(v=>v.includes(i))?y=d.map(v=>{const k=v.replaceAll(i,i.repeat(_)),T=k.lastIndexOf(i),b=T===-1?0:T+i.length;return k.slice(0,b)+f+k.slice(b)+` `}):(console.warn("You are passing both `text` and `images` to `PaliGemmaProcessor`. The processor expects special image tokens in the text, as many tokens as there are images per each text. It is recommended to add `` tokens in the very beginning of your text. For this call, we will infer how many images each text has and add special tokens."),y=d.map(v=>a(v,f,_,i,p.length)));const I=this.tokenizer(y,u);return{...await this.image_processor(p,u),...I}}}}),"./src/models/parakeet/feature_extraction_parakeet.js":((e,r,t)=>{t.r(r),t.d(r,{ParakeetFeatureExtractor:()=>a});var s=t("./src/base/feature_extraction_utils.js"),n=t("./src/utils/tensor.js"),o=t("./src/utils/audio.js");const i=1e-5;class a extends s.FeatureExtractor{constructor(c){super(c),this.config.mel_filters??=(0,o.mel_filter_bank)(Math.floor(1+this.config.n_fft/2),this.config.feature_size,0,this.config.sampling_rate/2,this.config.sampling_rate,"slaney","slaney");const p=(0,o.window_function)(this.config.win_length,"hann",{periodic:!1});this.window=new Float64Array(this.config.n_fft);const d=Math.floor((this.config.n_fft-this.config.win_length)/2);this.window.set(p,d)}async _extract_fbank_features(c){const p=this.config.preemphasis;c=new Float64Array(c);for(let u=c.length-1;u>=1;--u)c[u]-=p*c[u-1];return await(0,o.spectrogram)(c,this.window,this.window.length,this.config.hop_length,{fft_length:this.config.n_fft,power:2,mel_filters:this.config.mel_filters,log_mel:"log",mel_floor:-1/0,pad_mode:"constant",center:!0,transpose:!0,mel_offset:2**-24})}async _call(c){(0,s.validate_audio_inputs)(c,"ParakeetFeatureExtractor");const p=await this._extract_fbank_features(c),d=Math.floor((c.length+Math.floor(this.config.n_fft/2)*2-this.config.n_fft)/this.config.hop_length),u=p.data;u.fill(0,d*p.dims[1]);const[f,_]=p.dims,y=new Float64Array(_),I=new Float64Array(_);for(let k=0;k1?d-1:1;for(let k=0;k<_;++k){const T=y[k]/d,b=(I[k]-d*T*T)/w,x=1/(Math.sqrt(b)+i);for(let S=0;S{t.r(r),t.d(r,{Phi3VImageProcessor:()=>p});var s=t("./src/base/image_processors_utils.js"),n=t("./src/utils/tensor.js");const o=336,i=[2,3],{ceil:a,floor:l,sqrt:c}=Math;class p extends s.ImageProcessor{constructor(u){super({...u,do_normalize:!0,do_pad:!0,pad_size:"custom",do_convert_rgb:!0,do_resize:!0}),this._num_crops=u.num_crops}calc_num_image_tokens_from_image_size(u,f){const{num_img_tokens:_}=this.config;return l((l(f/o)*l(u/o)+1)*_+1+(l(f/o)+1)*c(_))}get_resize_output_image_size(u,f){const _=this._num_crops,[y,I]=u.size;let w=y/I,v=1;for(;v*Math.ceil(v/w)<=_;)v+=1;v-=1;const k=Math.floor(v*336),T=Math.floor(k/w);return[k,T]}pad_image(u,f,_,y={}){const[I,w]=f,v=o*a(I/o),k=o*a(w/o),T=[1,1,1].map((b,P)=>(b-this.image_mean[P])/this.image_std[P]);return super.pad_image(u,f,{width:k,height:v},{center:!0,constant_values:T,...y})}async _call(u,{num_crops:f=null}={}){if(this._num_crops=f??=this.config.num_crops,f<4||c(f)%1!==0)throw new Error("num_crops must be a square number >= 4");Array.isArray(u)||(u=[u]);const _=u.length,y=await Promise.all(u.map(x=>this.preprocess(x))),I=y.map(x=>x.original_size),w=y.map(x=>x.reshaped_input_size),v=[];for(const{pixel_values:x}of y){x.unsqueeze_(0);const[S,O]=x.dims.slice(-2),F=await(0,n.interpolate_4d)(x,{size:[o,o],mode:"bicubic"});if(f>0){const H=[],W=c(f),B=l(O/W),Y=l(S/W);for(let J=0;Jx.map(S=>o*a(S/o))),b=new n.Tensor("int64",T.flat(),[_,2]),P=T.map(([x,S])=>this.calc_num_image_tokens_from_image_size(S,x));return{pixel_values:k,original_sizes:I,reshaped_input_sizes:w,image_sizes:b,num_img_tokens:P}}}}),"./src/models/phi3_v/processing_phi3_v.js":((e,r,t)=>{t.r(r),t.d(r,{Phi3VProcessor:()=>l});var s=t("./src/base/processing_utils.js"),n=t("./src/models/auto/image_processing_auto.js"),o=t("./src/tokenizers.js");t("./src/utils/image.js");const i="<|image|>",a=/<\|image_\d+\|>/g;class l extends s.Processor{static image_processor_class=n.AutoImageProcessor;static tokenizer_class=o.AutoTokenizer;async _call(p,d=null,{padding:u=!0,truncation:f=!0,num_crops:_=null}={}){Array.isArray(p)||(p=[p]);let y,I;if(d){I=await this.image_processor(d,{num_crops:_});const{num_img_tokens:w}=I,v=p.map((T,b)=>T.split(a).join(i.repeat(w[b])));y=this.tokenizer(v,{padding:u,truncation:f});const k=this.tokenizer.model.convert_tokens_to_ids([i])[0];y.input_ids.map_(T=>T==k?-T:T)}else y=this.tokenizer(p);return{...y,...I}}}}),"./src/models/pixtral/image_processing_pixtral.js":((e,r,t)=>{t.r(r),t.d(r,{PixtralImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{get_resize_output_image_size(i,a){const{longest_edge:l}=a;if(l===void 0)throw new Error("size must contain 'longest_edge'");const[c,p]=i.size,d=Math.max(c,p)/l;let u=c,f=p;d>1&&(u=Math.floor(c/d),f=Math.floor(p/d));const{patch_size:_,spatial_merge_size:y}=this.config;if(!y)throw new Error("config must contain 'spatial_merge_size'");const I=_*y,w=Math.floor((u-1)/I)+1,v=Math.floor((f-1)/I)+1;return[w*I,v*I]}}}),"./src/models/pixtral/processing_pixtral.js":((e,r,t)=>{t.r(r),t.d(r,{PixtralProcessor:()=>i});var s=t("./src/base/processing_utils.js"),n=t("./src/models/auto/image_processing_auto.js"),o=t("./src/tokenizers.js");class i extends s.Processor{static tokenizer_class=o.AutoTokenizer;static image_processor_class=n.AutoImageProcessor;static uses_processor_config=!0;async _call(l,c=null,p={}){const d=await this.image_processor(l,p);if(c){const[f,_]=d.pixel_values.dims.slice(-2),{image_token:y,image_break_token:I,image_end_token:w,patch_size:v,spatial_merge_size:k}=this.config,T=v*k,b=Math.floor(f/T),P=Math.floor(_/T);c=structuredClone(c),Array.isArray(c)||(c=[c]);for(let x=0;x{t.r(r),t.d(r,{Florence2Processor:()=>s.Florence2Processor,Gemma3nProcessor:()=>n.Gemma3nProcessor,GroundingDinoProcessor:()=>o.GroundingDinoProcessor,Idefics3Processor:()=>i.Idefics3Processor,JinaCLIPProcessor:()=>l.JinaCLIPProcessor,LlavaProcessor:()=>c.LlavaProcessor,MgpstrProcessor:()=>p.MgpstrProcessor,MoonshineProcessor:()=>d.MoonshineProcessor,OwlViTProcessor:()=>u.OwlViTProcessor,PaliGemmaProcessor:()=>f.PaliGemmaProcessor,Phi3VProcessor:()=>_.Phi3VProcessor,PixtralProcessor:()=>y.PixtralProcessor,PyAnnoteProcessor:()=>I.PyAnnoteProcessor,Qwen2VLProcessor:()=>w.Qwen2VLProcessor,Sam2Processor:()=>k.Sam2Processor,Sam2VideoProcessor:()=>k.Sam2VideoProcessor,SamProcessor:()=>v.SamProcessor,SmolVLMProcessor:()=>T.SmolVLMProcessor,SpeechT5Processor:()=>b.SpeechT5Processor,UltravoxProcessor:()=>P.UltravoxProcessor,VLChatProcessor:()=>a.VLChatProcessor,VoxtralProcessor:()=>x.VoxtralProcessor,Wav2Vec2Processor:()=>S.Wav2Vec2Processor,Wav2Vec2ProcessorWithLM:()=>O.Wav2Vec2ProcessorWithLM,WhisperProcessor:()=>F.WhisperProcessor});var s=t("./src/models/florence2/processing_florence2.js"),n=t("./src/models/gemma3n/processing_gemma3n.js"),o=t("./src/models/grounding_dino/processing_grounding_dino.js"),i=t("./src/models/idefics3/processing_idefics3.js"),a=t("./src/models/janus/processing_janus.js"),l=t("./src/models/jina_clip/processing_jina_clip.js"),c=t("./src/models/llava/processing_llava.js"),p=t("./src/models/mgp_str/processing_mgp_str.js"),d=t("./src/models/moonshine/processing_moonshine.js"),u=t("./src/models/owlvit/processing_owlvit.js"),f=t("./src/models/paligemma/processing_paligemma.js"),_=t("./src/models/phi3_v/processing_phi3_v.js"),y=t("./src/models/pixtral/processing_pixtral.js"),I=t("./src/models/pyannote/processing_pyannote.js"),w=t("./src/models/qwen2_vl/processing_qwen2_vl.js"),v=t("./src/models/sam/processing_sam.js"),k=t("./src/models/sam2/processing_sam2.js"),T=t("./src/models/smolvlm/processing_smolvlm.js"),b=t("./src/models/speecht5/processing_speecht5.js"),P=t("./src/models/ultravox/processing_ultravox.js"),x=t("./src/models/voxtral/processing_voxtral.js"),S=t("./src/models/wav2vec2/processing_wav2vec2.js"),O=t("./src/models/wav2vec2_with_lm/processing_wav2vec2_with_lm.js"),F=t("./src/models/whisper/processing_whisper.js")}),"./src/models/pvt/image_processing_pvt.js":((e,r,t)=>{t.r(r),t.d(r,{PvtImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{}}),"./src/models/pyannote/feature_extraction_pyannote.js":((e,r,t)=>{t.r(r),t.d(r,{PyAnnoteFeatureExtractor:()=>i});var s=t("./src/base/feature_extraction_utils.js"),n=t("./src/utils/tensor.js"),o=t("./src/utils/maths.js");class i extends s.FeatureExtractor{async _call(l){(0,s.validate_audio_inputs)(l,"PyAnnoteFeatureExtractor"),l instanceof Float64Array&&(l=new Float32Array(l));const c=[1,1,l.length];return{input_values:new n.Tensor("float32",l,c)}}samples_to_frames(l){return(l-this.config.offset)/this.config.step}post_process_speaker_diarization(l,c){const p=c/this.samples_to_frames(c)/this.config.sampling_rate,d=[];for(const u of l.tolist()){const f=[];let _=-1;for(let y=0;y({id:y,start:I*p,end:w*p,confidence:v/(w-I)})))}return d}}}),"./src/models/pyannote/processing_pyannote.js":((e,r,t)=>{t.r(r),t.d(r,{PyAnnoteProcessor:()=>o});var s=t("./src/base/processing_utils.js"),n=t("./src/models/pyannote/feature_extraction_pyannote.js");class o extends s.Processor{static feature_extractor_class=n.PyAnnoteFeatureExtractor;async _call(a){return await this.feature_extractor(a)}post_process_speaker_diarization(...a){return this.feature_extractor.post_process_speaker_diarization(...a)}get sampling_rate(){return this.feature_extractor.config.sampling_rate}}}),"./src/models/qwen2_vl/image_processing_qwen2_vl.js":((e,r,t)=>{t.r(r),t.d(r,{Qwen2VLImageProcessor:()=>o});var s=t("./src/base/image_processors_utils.js"),n=t("./src/utils/tensor.js");class o extends s.ImageProcessor{async _call(a,...l){const{pixel_values:c,original_sizes:p,reshaped_input_sizes:d}=await super._call(a,...l);let u=c;const{temporal_patch_size:f,merge_size:_,patch_size:y}=this.config;u.dims[0]===1&&(u=(0,n.cat)(Array.from({length:f},()=>u),0));const I=u.dims[0]/f,w=u.dims[1],v=Math.floor(u.dims[2]/y),k=Math.floor(u.dims[3]/y),T=u.view(I,f,w,Math.floor(v/_),_,y,Math.floor(k/_),_,y).permute(0,3,6,4,7,2,1,5,8).view(I*v*k,w*f*y*y),b=new n.Tensor("int64",[I,v,k],[1,3]);return{pixel_values:T,image_grid_thw:b,original_sizes:p,reshaped_input_sizes:d}}}}),"./src/models/qwen2_vl/processing_qwen2_vl.js":((e,r,t)=>{t.r(r),t.d(r,{Qwen2VLProcessor:()=>i});var s=t("./src/base/processing_utils.js"),n=t("./src/models/auto/image_processing_auto.js"),o=t("./src/tokenizers.js");t("./src/utils/image.js");class i extends s.Processor{static image_processor_class=n.AutoImageProcessor;static tokenizer_class=o.AutoTokenizer;async _call(l,c=null,...p){Array.isArray(l)||(l=[l]);let d,u;if(c&&(d=await this.image_processor(c),u=d.image_grid_thw),u){let _=this.image_processor.config.merge_size**2,y=0;const I=u.tolist();l=l.map(w=>{for(;w.includes("<|image_pad|>");){const v=Number(I[y++].reduce((k,T)=>k*T,1n));w=w.replace("<|image_pad|>","<|placeholder|>".repeat(Math.floor(v/_)))}return w.replaceAll("<|placeholder|>","<|image_pad|>")})}return{...this.tokenizer(l),...d}}}}),"./src/models/rt_detr/image_processing_rt_detr.js":((e,r,t)=>{t.r(r),t.d(r,{RTDetrImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{post_process_object_detection(...i){return(0,s.post_process_object_detection)(...i)}}}),"./src/models/sam/image_processing_sam.js":((e,r,t)=>{t.r(r),t.d(r,{SamImageProcessor:()=>i});var s=t("./src/base/image_processors_utils.js"),n=t("./src/utils/core.js"),o=t("./src/utils/tensor.js");class i extends s.ImageProcessor{reshape_input_points(l,c,p,d=!1){l=structuredClone(l);let u=(0,n.calculateDimensions)(l);if(u.length===3)d||(u=[1,...u]),l=[l];else if(u.length!==4)throw Error("The input_points must be a 4D tensor of shape `batch_size`, `point_batch_size`, `nb_points_per_image`, `2`.");for(let f=0;fd!==c.dims[u]))throw Error(`The first ${p.length} dimensions of 'input_points' and 'input_labels' must be the same.`);return new o.Tensor("int64",l.flat(1/0).map(BigInt),p)}async _call(l,{input_points:c=null,input_labels:p=null,input_boxes:d=null}={}){const u=await super._call(l);if(c&&(u.input_points=this.reshape_input_points(c,u.original_sizes,u.reshaped_input_sizes)),p){if(!u.input_points)throw Error("`input_points` must be provided if `input_labels` are provided.");u.input_labels=this.add_input_labels(p,u.input_points)}return d&&(u.input_boxes=this.reshape_input_points(d,u.original_sizes,u.reshaped_input_sizes,!0)),u}async post_process_masks(l,c,p,{mask_threshold:d=0,binarize:u=!0,pad_size:f=null}={}){const _=[];f=f??this.pad_size??this.size;const y=[f.height,f.width];for(let I=0;Id&&(b[P]=1);k=new o.Tensor("bool",b,k.dims)}_.push(k)}return _}generate_crop_boxes(l,c,{crop_n_layers:p=0,overlap_ratio:d=512/1500,points_per_crop:u=32,crop_n_points_downscale_factor:f=1}={}){}}}),"./src/models/sam/processing_sam.js":((e,r,t)=>{t.r(r),t.d(r,{SamProcessor:()=>o});var s=t("./src/base/processing_utils.js"),n=t("./src/models/auto/image_processing_auto.js");class o extends s.Processor{static image_processor_class=n.AutoImageProcessor;async _call(...a){return await this.image_processor(...a)}post_process_masks(...a){return this.image_processor.post_process_masks(...a)}reshape_input_points(...a){return this.image_processor.reshape_input_points(...a)}}}),"./src/models/sam2/image_processing_sam2.js":((e,r,t)=>{t.r(r),t.d(r,{Sam2ImageProcessor:()=>s.SamImageProcessor});var s=t("./src/models/sam/image_processing_sam.js")}),"./src/models/sam2/processing_sam2.js":((e,r,t)=>{t.r(r),t.d(r,{Sam2Processor:()=>n,Sam2VideoProcessor:()=>o});var s=t("./src/models/sam/processing_sam.js");class n extends s.SamProcessor{}class o extends n{}}),"./src/models/sam3/image_processing_sam3.js":((e,r,t)=>{t.r(r),t.d(r,{Sam3ImageProcessor:()=>s.Sam2ImageProcessor});var s=t("./src/models/sam2/image_processing_sam2.js")}),"./src/models/seamless_m4t/feature_extraction_seamless_m4t.js":((e,r,t)=>{t.r(r),t.d(r,{SeamlessM4TFeatureExtractor:()=>i});var s=t("./src/base/feature_extraction_utils.js"),n=t("./src/utils/tensor.js"),o=t("./src/utils/audio.js");class i extends s.FeatureExtractor{constructor(l){super(l);const c=this.config.sampling_rate,p=(0,o.mel_filter_bank)(257,this.config.num_mel_bins,20,Math.floor(c/2),c,null,"kaldi",!0);this.mel_filters=p,this.window=(0,o.window_function)(400,"povey",{periodic:!1})}async _extract_fbank_features(l,c){return l=l.map(p=>p*32768),(0,o.spectrogram)(l,this.window,400,160,{fft_length:512,power:2,center:!1,preemphasis:.97,mel_filters:this.mel_filters,log_mel:"log",mel_floor:1192092955078125e-22,remove_dc_offset:!0,max_num_frames:c,transpose:!0})}async _call(l,{padding:c=!0,pad_to_multiple_of:p=2,do_normalize_per_mel_bins:d=!0,return_attention_mask:u=!0}={}){(0,s.validate_audio_inputs)(l,"SeamlessM4TFeatureExtractor");let f=await this._extract_fbank_features(l,this.config.max_length);if(d){const[b,P]=f.dims,x=f.data;for(let S=0;S0){const O=new Float32Array(P*(b+S));O.set(x),O.fill(this.config.padding_value,x.length);const F=b+S;f=new n.Tensor(f.type,O,[F,P]),u&&(_=new n.Tensor("int64",new BigInt64Array(F),[1,F]),_.data.fill(1n,0,b))}}const[y,I]=f.dims,w=this.config.stride;if(y%w!==0)throw new Error(`The number of frames (${y}) must be a multiple of the stride (${w}).`);const k=f.view(1,Math.floor(y/w),I*w),T={input_features:k};if(u){const b=k.dims[1],P=new BigInt64Array(b);if(_){const x=_.data;for(let S=1,O=0;S{t.r(r),t.d(r,{SegformerFeatureExtractor:()=>o,SegformerImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{post_process_semantic_segmentation(...a){return(0,s.post_process_semantic_segmentation)(...a)}}class o extends n{}}),"./src/models/siglip/image_processing_siglip.js":((e,r,t)=>{t.r(r),t.d(r,{SiglipImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{}}),"./src/models/smolvlm/image_processing_smolvlm.js":((e,r,t)=>{t.r(r),t.d(r,{SmolVLMImageProcessor:()=>s.Idefics3ImageProcessor});var s=t("./src/models/idefics3/image_processing_idefics3.js")}),"./src/models/smolvlm/processing_smolvlm.js":((e,r,t)=>{t.r(r),t.d(r,{SmolVLMProcessor:()=>s.Idefics3Processor});var s=t("./src/models/idefics3/processing_idefics3.js")}),"./src/models/snac/feature_extraction_snac.js":((e,r,t)=>{t.r(r),t.d(r,{SnacFeatureExtractor:()=>n});var s=t("./src/models/dac/feature_extraction_dac.js");class n extends s.DacFeatureExtractor{}}),"./src/models/speecht5/feature_extraction_speecht5.js":((e,r,t)=>{t.r(r),t.d(r,{SpeechT5FeatureExtractor:()=>n});var s=t("./src/base/feature_extraction_utils.js");class n extends s.FeatureExtractor{}}),"./src/models/speecht5/processing_speecht5.js":((e,r,t)=>{t.r(r),t.d(r,{SpeechT5Processor:()=>i});var s=t("./src/base/processing_utils.js"),n=t("./src/tokenizers.js"),o=t("./src/models/auto/feature_extraction_auto.js");class i extends s.Processor{static tokenizer_class=n.AutoTokenizer;static feature_extractor_class=o.AutoFeatureExtractor;async _call(l){return await this.feature_extractor(l)}}}),"./src/models/swin2sr/image_processing_swin2sr.js":((e,r,t)=>{t.r(r),t.d(r,{Swin2SRImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{pad_image(i,a,l,c={}){const[p,d,u]=a;return super.pad_image(i,a,{width:d+(l-d%l)%l,height:p+(l-p%l)%l},{mode:"symmetric",center:!1,constant_values:-1,...c})}}}),"./src/models/ultravox/processing_ultravox.js":((e,r,t)=>{t.r(r),t.d(r,{UltravoxProcessor:()=>i});var s=t("./src/models/auto/feature_extraction_auto.js"),n=t("./src/tokenizers.js"),o=t("./src/base/processing_utils.js");class i extends o.Processor{static tokenizer_class=n.AutoTokenizer;static feature_extractor_class=s.AutoFeatureExtractor;static uses_processor_config=!0;async _call(l,c=null,p={}){if(Array.isArray(l))throw new Error("Batched inputs are not supported yet.");let d={};if(c){const f=c.length,{input_features:_}=await this.feature_extractor(c,{...p,max_length:f}),y=Math.round(f/this.config.encoder_ds_factor+1e-4),I=1+Math.ceil(y/this.config.stack_factor);d.audio_token_len=[I],d.audio_values=_;const w=this.config.audio_placeholder;if(!l.includes(w))throw new Error(`The input text does not contain the image token ${w}.`);l=l.replaceAll(w,w.repeat(I))}return{...this.tokenizer(l,{add_special_tokens:!1,...p}),...d}}}}),"./src/models/vit/image_processing_vit.js":((e,r,t)=>{t.r(r),t.d(r,{ViTFeatureExtractor:()=>o,ViTImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{}class o extends n{}}),"./src/models/vitmatte/image_processing_vitmatte.js":((e,r,t)=>{t.r(r),t.d(r,{VitMatteImageProcessor:()=>o});var s=t("./src/base/image_processors_utils.js"),n=t("./src/utils/tensor.js");class o extends s.ImageProcessor{async _call(a,l){Array.isArray(a)||(a=[a]),Array.isArray(l)||(l=[l]);const c=await Promise.all(a.map(u=>this.preprocess(u))),p=await Promise.all(l.map(u=>this.preprocess(u,{do_normalize:!1,do_convert_rgb:!1,do_convert_grayscale:!0})));return{pixel_values:(0,n.stack)(c.map((u,f)=>(0,n.cat)([u.pixel_values,p[f].pixel_values],0)),0),original_sizes:c.map(u=>u.original_size),reshaped_input_sizes:c.map(u=>u.reshaped_input_size)}}}}),"./src/models/vitpose/image_processing_vitpose.js":((e,r,t)=>{t.r(r),t.d(r,{VitPoseImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{post_process_pose_estimation(i,a,{threshold:l=null}={}){const c=i.tolist(),[p,d,u,f]=i.dims,_=[];for(let y=0;y{t.r(r),t.d(r,{VoxtralProcessor:()=>d});var s=t("./src/models/auto/feature_extraction_auto.js"),n=t("./src/tokenizers.js"),o=t("./src/base/processing_utils.js"),i=t("./src/utils/tensor.js");const a="[AUDIO]",l="[BEGIN_AUDIO]",c=375;function p(u,f){const _=[];for(let y=0;yp(F,T)),P=b.map(F=>F.length),x=b.flat(),S=(await Promise.all(x.map(F=>this.feature_extractor(F,y)))).map(F=>F.input_features);I.audio_values=S.length>1?(0,i.cat)(S,0):S[0];let O=v[0];for(let F=0;F{t.r(r),t.d(r,{Wav2Vec2FeatureExtractor:()=>o});var s=t("./src/base/feature_extraction_utils.js"),n=t("./src/utils/tensor.js");class o extends s.FeatureExtractor{_zero_mean_unit_var_norm(a){const c=a.reduce((d,u)=>d+u,0)/a.length,p=a.reduce((d,u)=>d+(u-c)**2,0)/a.length;return a.map(d=>(d-c)/Math.sqrt(p+1e-7))}async _call(a){(0,s.validate_audio_inputs)(a,"Wav2Vec2FeatureExtractor"),a instanceof Float64Array&&(a=new Float32Array(a));let l=a;this.config.do_normalize&&(l=this._zero_mean_unit_var_norm(l));const c=[1,l.length];return{input_values:new n.Tensor("float32",l,c),attention_mask:new n.Tensor("int64",new BigInt64Array(l.length).fill(1n),c)}}}}),"./src/models/wav2vec2/processing_wav2vec2.js":((e,r,t)=>{t.r(r),t.d(r,{Wav2Vec2Processor:()=>i});var s=t("./src/tokenizers.js"),n=t("./src/models/auto/feature_extraction_auto.js"),o=t("./src/base/processing_utils.js");class i extends o.Processor{static tokenizer_class=s.AutoTokenizer;static feature_extractor_class=n.AutoFeatureExtractor;async _call(l){return await this.feature_extractor(l)}}}),"./src/models/wav2vec2_with_lm/processing_wav2vec2_with_lm.js":((e,r,t)=>{t.r(r),t.d(r,{Wav2Vec2ProcessorWithLM:()=>i});var s=t("./src/tokenizers.js"),n=t("./src/models/auto/feature_extraction_auto.js"),o=t("./src/base/processing_utils.js");class i extends o.Processor{static tokenizer_class=s.AutoTokenizer;static feature_extractor_class=n.AutoFeatureExtractor;async _call(l){return await this.feature_extractor(l)}}}),"./src/models/wespeaker/feature_extraction_wespeaker.js":((e,r,t)=>{t.r(r),t.d(r,{WeSpeakerFeatureExtractor:()=>o});var s=t("./src/base/feature_extraction_utils.js");t("./src/utils/tensor.js");var n=t("./src/utils/audio.js");class o extends s.FeatureExtractor{constructor(a){super(a);const l=this.config.sampling_rate,c=(0,n.mel_filter_bank)(257,this.config.num_mel_bins,20,Math.floor(l/2),l,null,"kaldi",!0);this.mel_filters=c,this.window=(0,n.window_function)(400,"hamming",{periodic:!1}),this.min_num_frames=this.config.min_num_frames}async _extract_fbank_features(a){return a=a.map(l=>l*32768),(0,n.spectrogram)(a,this.window,400,160,{fft_length:512,power:2,center:!1,preemphasis:.97,mel_filters:this.mel_filters,log_mel:"log",mel_floor:1192092955078125e-22,remove_dc_offset:!0,transpose:!0,min_num_frames:this.min_num_frames})}async _call(a){(0,s.validate_audio_inputs)(a,"WeSpeakerFeatureExtractor");const l=(await this._extract_fbank_features(a)).unsqueeze_(0);if(this.config.fbank_centering_span===null){const c=l.mean(1).data,p=l.data,[d,u,f]=l.dims;for(let _=0;_{t.r(r),t.d(r,{WHISPER_LANGUAGE_MAPPING:()=>n,WHISPER_TO_LANGUAGE_CODE_MAPPING:()=>o,whisper_language_to_code:()=>i});const s=[["en","english"],["zh","chinese"],["de","german"],["es","spanish"],["ru","russian"],["ko","korean"],["fr","french"],["ja","japanese"],["pt","portuguese"],["tr","turkish"],["pl","polish"],["ca","catalan"],["nl","dutch"],["ar","arabic"],["sv","swedish"],["it","italian"],["id","indonesian"],["hi","hindi"],["fi","finnish"],["vi","vietnamese"],["he","hebrew"],["uk","ukrainian"],["el","greek"],["ms","malay"],["cs","czech"],["ro","romanian"],["da","danish"],["hu","hungarian"],["ta","tamil"],["no","norwegian"],["th","thai"],["ur","urdu"],["hr","croatian"],["bg","bulgarian"],["lt","lithuanian"],["la","latin"],["mi","maori"],["ml","malayalam"],["cy","welsh"],["sk","slovak"],["te","telugu"],["fa","persian"],["lv","latvian"],["bn","bengali"],["sr","serbian"],["az","azerbaijani"],["sl","slovenian"],["kn","kannada"],["et","estonian"],["mk","macedonian"],["br","breton"],["eu","basque"],["is","icelandic"],["hy","armenian"],["ne","nepali"],["mn","mongolian"],["bs","bosnian"],["kk","kazakh"],["sq","albanian"],["sw","swahili"],["gl","galician"],["mr","marathi"],["pa","punjabi"],["si","sinhala"],["km","khmer"],["sn","shona"],["yo","yoruba"],["so","somali"],["af","afrikaans"],["oc","occitan"],["ka","georgian"],["be","belarusian"],["tg","tajik"],["sd","sindhi"],["gu","gujarati"],["am","amharic"],["yi","yiddish"],["lo","lao"],["uz","uzbek"],["fo","faroese"],["ht","haitian creole"],["ps","pashto"],["tk","turkmen"],["nn","nynorsk"],["mt","maltese"],["sa","sanskrit"],["lb","luxembourgish"],["my","myanmar"],["bo","tibetan"],["tl","tagalog"],["mg","malagasy"],["as","assamese"],["tt","tatar"],["haw","hawaiian"],["ln","lingala"],["ha","hausa"],["ba","bashkir"],["jw","javanese"],["su","sundanese"]],n=new Map(s),o=new Map([...s.map(([a,l])=>[l,a]),["burmese","my"],["valencian","ca"],["flemish","nl"],["haitian","ht"],["letzeburgesch","lb"],["pushto","ps"],["panjabi","pa"],["moldavian","ro"],["moldovan","ro"],["sinhalese","si"],["castilian","es"]]);function i(a){a=a.toLowerCase();let l=o.get(a);if(l===void 0){const c=a.match(/^<\|([a-z]{2})\|>$/);if(c&&(a=c[1]),n.has(a))l=a;else{const d=a.length===2?n.keys():n.values();throw new Error(`Language "${a}" is not supported. Must be one of: ${JSON.stringify(Array.from(d))}`)}}return l}}),"./src/models/whisper/feature_extraction_whisper.js":((e,r,t)=>{t.r(r),t.d(r,{WhisperFeatureExtractor:()=>i});var s=t("./src/base/feature_extraction_utils.js");t("./src/utils/tensor.js");var n=t("./src/utils/audio.js"),o=t("./src/utils/maths.js");class i extends s.FeatureExtractor{constructor(l){super(l),this.config.mel_filters??=(0,n.mel_filter_bank)(Math.floor(1+this.config.n_fft/2),this.config.feature_size,0,8e3,this.config.sampling_rate,"slaney","slaney"),this.window=(0,n.window_function)(this.config.n_fft,"hann")}async _extract_fbank_features(l){const c=await(0,n.spectrogram)(l,this.window,this.config.n_fft,this.config.hop_length,{power:2,mel_filters:this.config.mel_filters,log_mel:"log10",max_num_frames:Math.min(Math.floor(l.length/this.config.hop_length),this.config.nb_max_frames)}),p=c.data,d=(0,o.max)(p)[0];for(let u=0;ud?(l.length>this.config.n_samples&&console.warn("Attempting to extract features for audio longer than 30 seconds. If using a pipeline to extract transcript from a long audio clip, remember to specify `chunk_length_s` and/or `stride_length_s`."),p=l.slice(0,d)):(p=new Float32Array(d),p.set(l)),{input_features:(await this._extract_fbank_features(p)).unsqueeze_(0)}}}}),"./src/models/whisper/generation_whisper.js":((e,r,t)=>{t.r(r),t.d(r,{WhisperGenerationConfig:()=>n});var s=t("./src/generation/configuration_utils.js");class n extends s.GenerationConfig{return_timestamps=null;return_token_timestamps=null;num_frames=null;alignment_heads=null;task=null;language=null;no_timestamps_token_id=null;prompt_ids=null;is_multilingual=null;lang_to_id=null;task_to_id=null;max_initial_timestamp_index=1}}),"./src/models/whisper/processing_whisper.js":((e,r,t)=>{t.r(r),t.d(r,{WhisperProcessor:()=>i});var s=t("./src/models/auto/feature_extraction_auto.js"),n=t("./src/tokenizers.js"),o=t("./src/base/processing_utils.js");class i extends o.Processor{static tokenizer_class=n.AutoTokenizer;static feature_extractor_class=s.AutoFeatureExtractor;async _call(l){return await this.feature_extractor(l)}}}),"./src/models/yolos/image_processing_yolos.js":((e,r,t)=>{t.r(r),t.d(r,{YolosFeatureExtractor:()=>o,YolosImageProcessor:()=>n});var s=t("./src/base/image_processors_utils.js");class n extends s.ImageProcessor{post_process_object_detection(...a){return(0,s.post_process_object_detection)(...a)}}class o extends n{}}),"./src/ops/registry.js":((e,r,t)=>{t.r(r),t.d(r,{TensorOpRegistry:()=>i});var s=t("./src/backends/onnx.js"),n=t("./src/utils/tensor.js");const o=async(a,l,c)=>{const p=await(0,s.createInferenceSession)(new Uint8Array(a),l);return(async d=>{const u=(0,s.isONNXProxy)(),f=Object.fromEntries(Object.entries(d).map(([y,I])=>[y,(u?I.clone():I).ort_tensor])),_=await(0,s.runInferenceSession)(p,f);return Array.isArray(c)?c.map(y=>new n.Tensor(_[y])):new n.Tensor(_[c])})};class i{static session_options={};static get nearest_interpolate_4d(){return this._nearest_interpolate_4d||(this._nearest_interpolate_4d=o([8,10,18,0,58,129,1,10,41,10,1,120,10,0,10,0,10,1,115,18,1,121,34,6,82,101,115,105,122,101,42,18,10,4,109,111,100,101,34,7,110,101,97,114,101,115,116,160,1,3,18,1,114,90,31,10,1,120,18,26,10,24,8,1,18,20,10,3,18,1,98,10,3,18,1,99,10,3,18,1,104,10,3,18,1,119,90,15,10,1,115,18,10,10,8,8,7,18,4,10,2,8,4,98,31,10,1,121,18,26,10,24,8,1,18,20,10,3,18,1,98,10,3,18,1,99,10,3,18,1,104,10,3,18,1,119,66,2,16,21],this.session_options,"y")),this._nearest_interpolate_4d}static get bilinear_interpolate_4d(){return this._bilinear_interpolate_4d||(this._bilinear_interpolate_4d=o([8,9,18,0,58,128,1,10,40,10,1,120,10,0,10,0,10,1,115,18,1,121,34,6,82,101,115,105,122,101,42,17,10,4,109,111,100,101,34,6,108,105,110,101,97,114,160,1,3,18,1,114,90,31,10,1,120,18,26,10,24,8,1,18,20,10,3,18,1,98,10,3,18,1,99,10,3,18,1,104,10,3,18,1,119,90,15,10,1,115,18,10,10,8,8,7,18,4,10,2,8,4,98,31,10,1,121,18,26,10,24,8,1,18,20,10,3,18,1,98,10,3,18,1,99,10,3,18,1,104,10,3,18,1,119,66,2,16,20],this.session_options,"y")),this._bilinear_interpolate_4d}static get bicubic_interpolate_4d(){return this._bicubic_interpolate_4d||(this._bicubic_interpolate_4d=o([8,9,18,0,58,127,10,39,10,1,120,10,0,10,0,10,1,115,18,1,121,34,6,82,101,115,105,122,101,42,16,10,4,109,111,100,101,34,5,99,117,98,105,99,160,1,3,18,1,114,90,31,10,1,120,18,26,10,24,8,1,18,20,10,3,18,1,98,10,3,18,1,99,10,3,18,1,104,10,3,18,1,119,90,15,10,1,115,18,10,10,8,8,7,18,4,10,2,8,4,98,31,10,1,121,18,26,10,24,8,1,18,20,10,3,18,1,98,10,3,18,1,99,10,3,18,1,104,10,3,18,1,119,66,2,16,20],this.session_options,"y")),this._bicubic_interpolate_4d}static get matmul(){return this._matmul||(this._matmul=o([8,9,18,0,58,55,10,17,10,1,97,10,1,98,18,1,99,34,6,77,97,116,77,117,108,18,1,114,90,9,10,1,97,18,4,10,2,8,1,90,9,10,1,98,18,4,10,2,8,1,98,9,10,1,99,18,4,10,2,8,1,66,2,16,20],this.session_options,"c")),this._matmul}static get stft(){return this._stft||(this._stft=o([8,7,18,0,58,148,1,10,38,10,1,115,10,1,106,10,1,119,10,1,108,18,1,111,34,4,83,84,70,84,42,15,10,8,111,110,101,115,105,100,101,100,24,1,160,1,2,18,1,115,90,26,10,1,115,18,21,10,19,8,1,18,15,10,3,18,1,98,10,3,18,1,115,10,3,18,1,99,90,11,10,1,106,18,6,10,4,8,7,18,0,90,16,10,1,119,18,11,10,9,8,1,18,5,10,3,18,1,119,90,11,10,1,108,18,6,10,4,8,7,18,0,98,31,10,1,111,18,26,10,24,8,1,18,20,10,3,18,1,98,10,3,18,1,102,10,3,18,1,100,10,3,18,1,99,66,2,16,17],this.session_options,"o")),this._stft}static get rfft(){return this._rfft||(this._rfft=o([8,9,18,0,58,97,10,33,10,1,120,10,0,10,1,97,18,1,121,34,3,68,70,84,42,15,10,8,111,110,101,115,105,100,101,100,24,1,160,1,2,18,1,100,90,21,10,1,120,18,16,10,14,8,1,18,10,10,3,18,1,115,10,3,18,1,99,90,11,10,1,97,18,6,10,4,8,7,18,0,98,21,10,1,121,18,16,10,14,8,1,18,10,10,3,18,1,115,10,3,18,1,99,66,2,16,20],this.session_options,"y")),this._rfft}static get top_k(){return this._top_k||(this._top_k=o([8,10,18,0,58,73,10,18,10,1,120,10,1,107,18,1,118,18,1,105,34,4,84,111,112,75,18,1,116,90,9,10,1,120,18,4,10,2,8,1,90,15,10,1,107,18,10,10,8,8,7,18,4,10,2,8,1,98,9,10,1,118,18,4,10,2,8,1,98,9,10,1,105,18,4,10,2,8,7,66,2,16,21],this.session_options,["v","i"])),this._top_k}static get slice(){return this._slice||(this._slice=o([8,7,18,0,58,96,10,25,10,1,120,10,1,115,10,1,101,10,1,97,10,1,116,18,1,121,34,5,83,108,105,99,101,18,1,114,90,9,10,1,120,18,4,10,2,8,1,90,9,10,1,115,18,4,10,2,8,7,90,9,10,1,101,18,4,10,2,8,7,90,9,10,1,97,18,4,10,2,8,7,90,9,10,1,116,18,4,10,2,8,7,98,9,10,1,121,18,4,10,2,8,1,66,2,16,13],this.session_options,"y")),this._slice}}}),"./src/pipelines.js":((e,r,t)=>{t.r(r),t.d(r,{AudioClassificationPipeline:()=>W,AutomaticSpeechRecognitionPipeline:()=>Y,BackgroundRemovalPipeline:()=>ne,DepthEstimationPipeline:()=>te,DocumentQuestionAnsweringPipeline:()=>K,FeatureExtractionPipeline:()=>F,FillMaskPipeline:()=>k,ImageClassificationPipeline:()=>J,ImageFeatureExtractionPipeline:()=>H,ImageSegmentationPipeline:()=>re,ImageToImagePipeline:()=>D,ImageToTextPipeline:()=>X,ObjectDetectionPipeline:()=>pe,Pipeline:()=>y,QuestionAnsweringPipeline:()=>v,SummarizationPipeline:()=>b,Text2TextGenerationPipeline:()=>T,TextClassificationPipeline:()=>I,TextGenerationPipeline:()=>S,TextToAudioPipeline:()=>j,TokenClassificationPipeline:()=>w,TranslationPipeline:()=>P,ZeroShotAudioClassificationPipeline:()=>B,ZeroShotClassificationPipeline:()=>O,ZeroShotImageClassificationPipeline:()=>le,ZeroShotObjectDetectionPipeline:()=>oe,pipeline:()=>ke});var s=t("./src/tokenizers.js"),n=t("./src/models.js"),o=t("./src/models/auto/processing_auto.js");t("./src/base/processing_utils.js");var i=t("./src/utils/generic.js"),a=t("./src/utils/core.js"),l=t("./src/utils/maths.js"),c=t("./src/utils/audio.js"),p=t("./src/utils/tensor.js"),d=t("./src/utils/image.js");async function u(Te){return Array.isArray(Te)||(Te=[Te]),await Promise.all(Te.map(Q=>d.RawImage.read(Q)))}async function f(Te,Q){return Array.isArray(Te)||(Te=[Te]),await Promise.all(Te.map(z=>typeof z=="string"||z instanceof URL?(0,c.read_audio)(z,Q):z instanceof Float64Array?new Float32Array(z):z))}function _(Te,Q){Q&&(Te=Te.map(xe=>xe|0));const[z,de,be,ve]=Te;return{xmin:z,ymin:de,xmax:be,ymax:ve}}class y extends i.Callable{constructor({task:Q,model:z,tokenizer:de=null,processor:be=null}){super(),this.task=Q,this.model=z,this.tokenizer=de,this.processor=be}async dispose(){await this.model.dispose()}}class I extends y{constructor(Q){super(Q)}async _call(Q,{top_k:z=1}={}){const de=this.tokenizer(Q,{padding:!0,truncation:!0}),be=await this.model(de),ve=this.model.config.problem_type==="multi_label_classification"?ge=>ge.sigmoid():ge=>new p.Tensor("float32",(0,l.softmax)(ge.data),ge.dims),xe=this.model.config.id2label,Ce=[];for(const ge of be.logits){const De=ve(ge),fe=await(0,p.topk)(De,z),Pe=fe[0].tolist(),Fe=fe[1].tolist().map((tt,Re)=>({label:xe?xe[tt]:`LABEL_${tt}`,score:Pe[Re]}));z===1?Ce.push(...Fe):Ce.push(Fe)}return Array.isArray(Q)||z===1?Ce:Ce[0]}}class w extends y{constructor(Q){super(Q)}async _call(Q,{ignore_labels:z=["O"]}={}){const de=Array.isArray(Q),be=this.tokenizer(de?Q:[Q],{padding:!0,truncation:!0}),xe=(await this.model(be)).logits,Ce=this.model.config.id2label,ge=[];for(let De=0;DeOe==this.tokenizer.sep_token_id);ge[Pe].map((Oe,at)=>Oe==1&&(at===0||at>Fe&&De.findIndex(ht=>ht==We[at])===-1));const tt=ve[Pe].tolist(),Re=xe[Pe].tolist();for(let Oe=1;Oeat==We[Oe])!==-1)&&(tt[Oe]=-1/0,Re[Oe]=-1/0);const rt=(0,l.softmax)(tt).map((Oe,at)=>[Oe,at]),Ze=(0,l.softmax)(Re).map((Oe,at)=>[Oe,at]);rt[0][0]=0,Ze[0][0]=0;const je=(0,a.product)(rt,Ze).filter(Oe=>Oe[0][1]<=Oe[1][1]).map(Oe=>[Oe[0][1],Oe[1][1],Oe[0][0]*Oe[1][0]]).sort((Oe,at)=>at[2]-Oe[2]);for(let Oe=0;Oett==this.tokenizer.mask_token_id);if(De===-1)throw Error(`Mask token (${this.tokenizer.mask_token}) not found in text.`);const fe=be[Ce][De],Pe=await(0,p.topk)(new p.Tensor("float32",(0,l.softmax)(fe.data),fe.dims),z),We=Pe[0].tolist(),Fe=Pe[1].tolist();ve.push(Fe.map((tt,Re)=>{const rt=ge.slice();return rt[De]=tt,{score:We[Re],token:Number(tt),token_str:this.tokenizer.decode([tt]),sequence:this.tokenizer.decode(rt,{skip_special_tokens:!0})}}))}return Array.isArray(Q)?ve:ve[0]}}class T extends y{_key="generated_text";constructor(Q){super(Q)}async _call(Q,z={}){Array.isArray(Q)||(Q=[Q]),this.model.config.prefix&&(Q=Q.map(ge=>this.model.config.prefix+ge));const de=this.model.config.task_specific_params;de&&de[this.task]&&de[this.task].prefix&&(Q=Q.map(ge=>de[this.task].prefix+ge));const be=this.tokenizer,ve={padding:!0,truncation:!0};let xe;this instanceof P&&"_build_translation_inputs"in be?xe=be._build_translation_inputs(Q,ve,z):xe=be(Q,ve);const Ce=await this.model.generate({...xe,...z});return be.batch_decode(Ce,{skip_special_tokens:!0}).map(ge=>({[this._key]:ge}))}}class b extends T{_key="summary_text";constructor(Q){super(Q)}}class P extends T{_key="translation_text";constructor(Q){super(Q)}}function x(Te){return Array.isArray(Te)&&Te.every(Q=>"role"in Q&&"content"in Q)}class S extends y{constructor(Q){super(Q)}async _call(Q,z={}){let de=!1,be=!1,ve=z.add_special_tokens??(this.tokenizer.add_bos_token||this.tokenizer.add_eos_token)??!1,xe;if(typeof Q=="string")xe=Q=[Q];else if(Array.isArray(Q)&&Q.every(Fe=>typeof Fe=="string"))de=!0,xe=Q;else{if(x(Q))Q=[Q];else if(Array.isArray(Q)&&Q.every(x))de=!0;else throw new Error("Input must be a string, an array of strings, a Chat, or an array of Chats");be=!0,xe=Q.map(Fe=>this.tokenizer.apply_chat_template(Fe,{tokenize:!1,add_generation_prompt:!0})),ve=!1}const Ce=be?!1:z.return_full_text??!0;this.tokenizer.padding_side="left";const ge=this.tokenizer(xe,{add_special_tokens:ve,padding:!0,truncation:!0}),De=await this.model.generate({...ge,...z}),fe=this.tokenizer.batch_decode(De,{skip_special_tokens:!0});let Pe;!Ce&&ge.input_ids.dims.at(-1)>0&&(Pe=this.tokenizer.batch_decode(ge.input_ids,{skip_special_tokens:!0}).map(Fe=>Fe.length));const We=Array.from({length:Q.length},Fe=>[]);for(let Fe=0;Fe[z.toLowerCase(),de])),this.entailment_id=this.label2id.entailment,this.entailment_id===void 0&&(console.warn("Could not find 'entailment' in label2id mapping. Using 2 as entailment_id."),this.entailment_id=2),this.contradiction_id=this.label2id.contradiction??this.label2id.not_entailment,this.contradiction_id===void 0&&(console.warn("Could not find 'contradiction' in label2id mapping. Using 0 as contradiction_id."),this.contradiction_id=0)}async _call(Q,z,{hypothesis_template:de="This example is {}.",multi_label:be=!1}={}){const ve=Array.isArray(Q);ve||(Q=[Q]),Array.isArray(z)||(z=[z]);const xe=z.map(De=>de.replace("{}",De)),Ce=be||z.length===1,ge=[];for(const De of Q){const fe=[];for(const Fe of xe){const tt=this.tokenizer(De,{text_pair:Fe,padding:!0,truncation:!0}),Re=await this.model(tt);Ce?fe.push([Re.logits.data[this.contradiction_id],Re.logits.data[this.entailment_id]]):fe.push(Re.logits.data[this.entailment_id])}const We=(Ce?fe.map(Fe=>(0,l.softmax)(Fe)[1]):(0,l.softmax)(fe)).map((Fe,tt)=>[Fe,tt]).sort((Fe,tt)=>tt[0]-Fe[0]);ge.push({sequence:De,labels:We.map(Fe=>z[Fe[1]]),scores:We.map(Fe=>Fe[0])})}return ve?ge:ge[0]}}class F extends y{constructor(Q){super(Q)}async _call(Q,{pooling:z="none",normalize:de=!1,quantize:be=!1,precision:ve="binary"}={}){const xe=this.tokenizer(Q,{padding:!0,truncation:!0}),Ce=await this.model(xe);let ge=Ce.last_hidden_state??Ce.logits??Ce.token_embeddings;switch(z){case"none":break;case"mean":ge=(0,p.mean_pooling)(ge,xe.attention_mask);break;case"first_token":case"cls":ge=ge.slice(null,0);break;case"last_token":case"eos":ge=ge.slice(null,-1);break;default:throw Error(`Pooling method '${z}' not supported.`)}return de&&(ge=ge.normalize(2,-1)),be&&(ge=(0,p.quantize_embeddings)(ge,ve)),ge}}class H extends y{constructor(Q){super(Q)}async _call(Q,{pool:z=null}={}){const de=await u(Q),{pixel_values:be}=await this.processor(de),ve=await this.model({pixel_values:be});let xe;if(z){if(!("pooler_output"in ve))throw Error("No pooled output was returned. Make sure the model has a 'pooler' layer when using the 'pool' option.");xe=ve.pooler_output}else xe=ve.last_hidden_state??ve.logits??ve.image_embeds;return xe}}class W extends y{constructor(Q){super(Q)}async _call(Q,{top_k:z=5}={}){const de=this.processor.feature_extractor.config.sampling_rate,be=await f(Q,de),ve=this.model.config.id2label,xe=[];for(const Ce of be){const ge=await this.processor(Ce),fe=(await this.model(ge)).logits[0],Pe=await(0,p.topk)(new p.Tensor("float32",(0,l.softmax)(fe.data),fe.dims),z),We=Pe[0].tolist(),tt=Pe[1].tolist().map((Re,rt)=>({label:ve?ve[Re]:`LABEL_${Re}`,score:We[rt]}));xe.push(tt)}return Array.isArray(Q)?xe:xe[0]}}class B extends y{constructor(Q){super(Q)}async _call(Q,z,{hypothesis_template:de="This is a sound of {}."}={}){const be=!Array.isArray(Q);be&&(Q=[Q]);const ve=z.map(fe=>de.replace("{}",fe)),xe=this.tokenizer(ve,{padding:!0,truncation:!0}),Ce=this.processor.feature_extractor.config.sampling_rate,ge=await f(Q,Ce),De=[];for(const fe of ge){const Pe=await this.processor(fe),We=await this.model({...xe,...Pe}),Fe=(0,l.softmax)(We.logits_per_audio.data);De.push([...Fe].map((tt,Re)=>({score:tt,label:z[Re]})))}return be?De[0]:De}}class Y extends y{constructor(Q){super(Q)}async _call(Q,z={}){switch(this.model.config.model_type){case"whisper":case"lite-whisper":return this._call_whisper(Q,z);case"wav2vec2":case"wav2vec2-bert":case"unispeech":case"unispeech-sat":case"hubert":case"parakeet_ctc":return this._call_wav2vec2(Q,z);case"moonshine":return this._call_moonshine(Q,z);default:throw new Error(`AutomaticSpeechRecognitionPipeline does not support model type '${this.model.config.model_type}'.`)}}async _call_wav2vec2(Q,z){z.language&&console.warn('`language` parameter is not yet supported for `wav2vec2` models, defaulting to "English".'),z.task&&console.warn('`task` parameter is not yet supported for `wav2vec2` models, defaulting to "transcribe".');const de=!Array.isArray(Q);de&&(Q=[Q]);const be=this.processor.feature_extractor.config.sampling_rate,ve=await f(Q,be),xe=[];for(const Ce of ve){const ge=await this.processor(Ce),fe=(await this.model(ge)).logits[0],Pe=[];for(const Fe of fe)Pe.push((0,l.max)(Fe.data)[1]);const We=this.tokenizer.decode(Pe,{skip_special_tokens:!0}).trim();xe.push({text:We})}return de?xe[0]:xe}async _call_whisper(Q,z){const de=z.return_timestamps??!1,be=z.chunk_length_s??0,ve=z.force_full_sequences??!1;let xe=z.stride_length_s??null;const Ce={...z};de==="word"&&(Ce.return_token_timestamps=!0,Ce.return_timestamps=!1);const ge=!Array.isArray(Q);ge&&(Q=[Q]);const De=this.processor.feature_extractor.config.chunk_length/this.model.config.max_source_positions,fe=this.processor.feature_extractor.config.hop_length,Pe=this.processor.feature_extractor.config.sampling_rate,We=await f(Q,Pe),Fe=[];for(const tt of We){let Re=[];if(be>0){if(xe===null)xe=be/6;else if(be<=xe)throw Error("`chunk_length_s` must be larger than `stride_length_s`.");const je=Pe*be,Oe=Pe*xe,at=je-2*Oe;let ht=0;for(;;){const Nt=ht+je,kt=tt.subarray(ht,Nt),gr=await this.processor(kt),Or=ht===0,Bt=Nt>=tt.length;if(Re.push({stride:[kt.length,Or?0:Oe,Bt?0:Oe],input_features:gr.input_features,is_last:Bt}),Bt)break;ht+=at}}else Re=[{stride:[tt.length,0,0],input_features:(await this.processor(tt)).input_features,is_last:!0}];for(const je of Re){Ce.num_frames=Math.floor(je.stride[0]/fe);const Oe=await this.model.generate({inputs:je.input_features,...Ce});de==="word"?(je.tokens=Oe.sequences.tolist()[0],je.token_timestamps=Oe.token_timestamps.tolist()[0].map(at=>(0,l.round)(at,2))):je.tokens=Oe[0].tolist(),je.stride=je.stride.map(at=>at/Pe)}const[rt,Ze]=this.tokenizer._decode_asr(Re,{time_precision:De,return_timestamps:de,force_full_sequences:ve});Fe.push({text:rt,...Ze})}return ge?Fe[0]:Fe}async _call_moonshine(Q,z){const de=!Array.isArray(Q);de&&(Q=[Q]);const be=this.processor.feature_extractor.config.sampling_rate,ve=await f(Q,be),xe=[];for(const Ce of ve){const ge=await this.processor(Ce),De=Math.floor(Ce.length/be)*6,fe=await this.model.generate({max_new_tokens:De,...z,...ge}),Pe=this.processor.batch_decode(fe,{skip_special_tokens:!0})[0];xe.push({text:Pe})}return de?xe[0]:xe}}class X extends y{constructor(Q){super(Q)}async _call(Q,z={}){const de=Array.isArray(Q),be=await u(Q),{pixel_values:ve}=await this.processor(be),xe=[];for(const Ce of ve){Ce.dims=[1,...Ce.dims];const ge=await this.model.generate({inputs:Ce,...z}),De=this.tokenizer.batch_decode(ge,{skip_special_tokens:!0}).map(fe=>({generated_text:fe.trim()}));xe.push(De)}return de?xe:xe[0]}}class J extends y{constructor(Q){super(Q)}async _call(Q,{top_k:z=5}={}){const de=await u(Q),{pixel_values:be}=await this.processor(de),ve=await this.model({pixel_values:be}),xe=this.model.config.id2label,Ce=[];for(const ge of ve.logits){const De=await(0,p.topk)(new p.Tensor("float32",(0,l.softmax)(ge.data),ge.dims),z),fe=De[0].tolist(),We=De[1].tolist().map((Fe,tt)=>({label:xe?xe[Fe]:`LABEL_${Fe}`,score:fe[tt]}));Ce.push(We)}return Array.isArray(Q)?Ce:Ce[0]}}class re extends y{constructor(Q){super(Q),this.subtasks_mapping={panoptic:"post_process_panoptic_segmentation",instance:"post_process_instance_segmentation",semantic:"post_process_semantic_segmentation"}}async _call(Q,{threshold:z=.5,mask_threshold:de=.5,overlap_mask_area_threshold:be=.8,label_ids_to_fuse:ve=null,target_sizes:xe=null,subtask:Ce=null}={}){if(Array.isArray(Q)&&Q.length!==1)throw Error("Image segmentation pipeline currently only supports a batch size of 1.");const De=await u(Q),fe=De.map(je=>[je.height,je.width]),Pe=await this.processor(De),{inputNames:We,outputNames:Fe}=this.model.sessions.model;if(!We.includes("pixel_values")){if(We.length!==1)throw Error(`Expected a single input name, but got ${We.length} inputs: ${We}.`);const je=We[0];if(je in Pe)throw Error(`Input name ${je} already exists in the inputs.`);Pe[je]=Pe.pixel_values}const tt=await this.model(Pe);let Re=null;if(Ce!==null)Re=this.subtasks_mapping[Ce];else if(this.processor.image_processor){for(const[je,Oe]of Object.entries(this.subtasks_mapping))if(Oe in this.processor.image_processor){Re=this.processor.image_processor[Oe].bind(this.processor.image_processor),Ce=je;break}}const rt=this.model.config.id2label,Ze=[];if(Ce)if(Ce==="panoptic"||Ce==="instance"){const je=Re(tt,z,de,be,ve,xe??fe)[0],Oe=je.segmentation;for(const at of je.segments_info){const ht=new Uint8ClampedArray(Oe.data.length);for(let kt=0;ktgr<-1e-5||gr>1+1e-5)&&Nt.sigmoid_();const kt=await d.RawImage.fromTensor(Nt.mul_(255).to("uint8")).resize(ht[1],ht[0]);Ze.push({label:null,score:null,mask:kt})}}return Ze}}class ne extends re{constructor(Q){super(Q)}async _call(Q,z={}){if(Array.isArray(Q)&&Q.length!==1)throw Error("Background removal pipeline currently only supports a batch size of 1.");const be=await u(Q),ve=await super._call(Q,z);return be.map((Ce,ge)=>{const De=Ce.clone();return De.putAlpha(ve[ge].mask),De})}}class le extends y{constructor(Q){super(Q)}async _call(Q,z,{hypothesis_template:de="This is a photo of {}"}={}){const be=Array.isArray(Q),ve=await u(Q),xe=z.map(We=>de.replace("{}",We)),Ce=this.tokenizer(xe,{padding:this.model.config.model_type==="siglip"?"max_length":!0,truncation:!0}),{pixel_values:ge}=await this.processor(ve),De=await this.model({...Ce,pixel_values:ge}),fe=this.model.config.model_type==="siglip"?We=>We.sigmoid().data:We=>(0,l.softmax)(We.data),Pe=[];for(const We of De.logits_per_image){const tt=[...fe(We)].map((Re,rt)=>({score:Re,label:z[rt]}));tt.sort((Re,rt)=>rt.score-Re.score),Pe.push(tt)}return be?Pe:Pe[0]}}class pe extends y{constructor(Q){super(Q)}async _call(Q,{threshold:z=.9,percentage:de=!1}={}){const be=Array.isArray(Q);if(be&&Q.length!==1)throw Error("Object detection pipeline currently only supports a batch size of 1.");const ve=await u(Q),xe=de?null:ve.map(Fe=>[Fe.height,Fe.width]),{pixel_values:Ce,pixel_mask:ge}=await this.processor(ve),De=await this.model({pixel_values:Ce,pixel_mask:ge}),fe=this.processor.image_processor.post_process_object_detection(De,z,xe),Pe=this.model.config.id2label,We=fe.map(Fe=>Fe.boxes.map((tt,Re)=>({score:Fe.scores[Re],label:Pe[Fe.classes[Re]],box:_(tt,!de)})));return be?We:We[0]}}class oe extends y{constructor(Q){super(Q)}async _call(Q,z,{threshold:de=.1,top_k:be=null,percentage:ve=!1}={}){const xe=Array.isArray(Q),Ce=await u(Q),ge=this.tokenizer(z,{padding:!0,truncation:!0}),De=await this.processor(Ce),fe=[];for(let Pe=0;Pe({score:Ze.scores[Oe],label:Ze.labels[Oe],box:_(je,!ve)}))}else{const Ze=this.processor.image_processor.post_process_object_detection(Re,de,Fe,!0)[0];rt=Ze.boxes.map((je,Oe)=>({score:Ze.scores[Oe],label:z[Ze.classes[Oe]],box:_(je,!ve)}))}rt.sort((Ze,je)=>je.score-Ze.score),be!==null&&(rt=rt.slice(0,be)),fe.push(rt)}return xe?fe:fe[0]}}class K extends y{constructor(Q){super(Q)}async _call(Q,z,de={}){const be=(await u(Q))[0],{pixel_values:ve}=await this.processor(be),xe=`${z}`,Ce=this.tokenizer(xe,{add_special_tokens:!1,padding:!0,truncation:!0}).input_ids,ge=await this.model.generate({inputs:ve,max_length:this.model.config.decoder.max_position_embeddings,decoder_input_ids:Ce,...de}),fe=this.tokenizer.batch_decode(ge)[0].match(/(.*?)<\/s_answer>/);let Pe=null;return fe&&fe.length>=2&&(Pe=fe[1].trim()),[{answer:Pe}]}}class j extends y{DEFAULT_VOCODER_ID="Xenova/speecht5_hifigan";constructor(Q){super(Q),this.vocoder=Q.vocoder??null}async _prepare_speaker_embeddings(Q){if((typeof Q=="string"||Q instanceof URL)&&(Q=new Float32Array(await(await fetch(Q)).arrayBuffer())),Q instanceof Float32Array)Q=new p.Tensor("float32",Q,[Q.length]);else if(!(Q instanceof p.Tensor))throw new Error("Speaker embeddings must be a `Tensor`, `Float32Array`, `string`, or `URL`.");return Q}async _call(Q,{speaker_embeddings:z=null,num_inference_steps:de,speed:be}={}){return this.processor?this._call_text_to_spectrogram(Q,{speaker_embeddings:z}):this.model.config.model_type==="supertonic"?this._call_supertonic(Q,{speaker_embeddings:z,num_inference_steps:de,speed:be}):this._call_text_to_waveform(Q)}async _call_supertonic(Q,{speaker_embeddings:z,num_inference_steps:de,speed:be}){if(!z)throw new Error("Speaker embeddings must be provided for Supertonic models.");z=await this._prepare_speaker_embeddings(z);const{sampling_rate:ve,style_dim:xe}=this.model.config;z=z.view(1,-1,xe);const Ce=this.tokenizer(Q,{padding:!0,truncation:!0}),{waveform:ge}=await this.model.generate_speech({...Ce,style:z,num_inference_steps:de,speed:be});return new c.RawAudio(ge.data,ve)}async _call_text_to_waveform(Q){const z=this.tokenizer(Q,{padding:!0,truncation:!0}),{waveform:de}=await this.model(z),be=this.model.config.sampling_rate;return new c.RawAudio(de.data,be)}async _call_text_to_spectrogram(Q,{speaker_embeddings:z}){this.vocoder||(console.log("No vocoder specified, using default HifiGan vocoder."),this.vocoder=await n.AutoModel.from_pretrained(this.DEFAULT_VOCODER_ID,{dtype:"fp32"}));const{input_ids:de}=this.tokenizer(Q,{padding:!0,truncation:!0});z=await this._prepare_speaker_embeddings(z),z=z.view(1,-1);const{waveform:be}=await this.model.generate_speech(de,z,{vocoder:this.vocoder}),ve=this.processor.feature_extractor.config.sampling_rate;return new c.RawAudio(be.data,ve)}}class D extends y{constructor(Q){super(Q)}async _call(Q){const z=await u(Q),de=await this.processor(z),be=await this.model(de),ve=[];for(const xe of be.reconstruction){const Ce=xe.squeeze().clamp_(0,1).mul_(255).round_().to("uint8");ve.push(d.RawImage.fromTensor(Ce))}return ve.length>1?ve:ve[0]}}class te extends y{constructor(Q){super(Q)}async _call(Q){const z=await u(Q),de=await this.processor(z),{predicted_depth:be}=await this.model(de),ve=[];for(let xe=0;xe1?ve:ve[0]}}const he=Object.freeze({"text-classification":{tokenizer:s.AutoTokenizer,pipeline:I,model:n.AutoModelForSequenceClassification,default:{model:"Xenova/distilbert-base-uncased-finetuned-sst-2-english"},type:"text"},"token-classification":{tokenizer:s.AutoTokenizer,pipeline:w,model:n.AutoModelForTokenClassification,default:{model:"Xenova/bert-base-multilingual-cased-ner-hrl"},type:"text"},"question-answering":{tokenizer:s.AutoTokenizer,pipeline:v,model:n.AutoModelForQuestionAnswering,default:{model:"Xenova/distilbert-base-cased-distilled-squad"},type:"text"},"fill-mask":{tokenizer:s.AutoTokenizer,pipeline:k,model:n.AutoModelForMaskedLM,default:{model:"Xenova/bert-base-uncased"},type:"text"},summarization:{tokenizer:s.AutoTokenizer,pipeline:b,model:n.AutoModelForSeq2SeqLM,default:{model:"Xenova/distilbart-cnn-6-6"},type:"text"},translation:{tokenizer:s.AutoTokenizer,pipeline:P,model:n.AutoModelForSeq2SeqLM,default:{model:"Xenova/t5-small"},type:"text"},"text2text-generation":{tokenizer:s.AutoTokenizer,pipeline:T,model:n.AutoModelForSeq2SeqLM,default:{model:"Xenova/flan-t5-small"},type:"text"},"text-generation":{tokenizer:s.AutoTokenizer,pipeline:S,model:n.AutoModelForCausalLM,default:{model:"Xenova/gpt2"},type:"text"},"zero-shot-classification":{tokenizer:s.AutoTokenizer,pipeline:O,model:n.AutoModelForSequenceClassification,default:{model:"Xenova/distilbert-base-uncased-mnli"},type:"text"},"audio-classification":{pipeline:W,model:n.AutoModelForAudioClassification,processor:o.AutoProcessor,default:{model:"Xenova/wav2vec2-base-superb-ks"},type:"audio"},"zero-shot-audio-classification":{tokenizer:s.AutoTokenizer,pipeline:B,model:n.AutoModel,processor:o.AutoProcessor,default:{model:"Xenova/clap-htsat-unfused"},type:"multimodal"},"automatic-speech-recognition":{tokenizer:s.AutoTokenizer,pipeline:Y,model:[n.AutoModelForSpeechSeq2Seq,n.AutoModelForCTC],processor:o.AutoProcessor,default:{model:"Xenova/whisper-tiny.en"},type:"multimodal"},"text-to-audio":{tokenizer:s.AutoTokenizer,pipeline:j,model:[n.AutoModelForTextToWaveform,n.AutoModelForTextToSpectrogram],processor:[o.AutoProcessor,null],default:{model:"Xenova/speecht5_tts"},type:"text"},"image-to-text":{tokenizer:s.AutoTokenizer,pipeline:X,model:n.AutoModelForVision2Seq,processor:o.AutoProcessor,default:{model:"Xenova/vit-gpt2-image-captioning"},type:"multimodal"},"image-classification":{pipeline:J,model:n.AutoModelForImageClassification,processor:o.AutoProcessor,default:{model:"Xenova/vit-base-patch16-224"},type:"multimodal"},"image-segmentation":{pipeline:re,model:[n.AutoModelForImageSegmentation,n.AutoModelForSemanticSegmentation,n.AutoModelForUniversalSegmentation],processor:o.AutoProcessor,default:{model:"Xenova/detr-resnet-50-panoptic"},type:"multimodal"},"background-removal":{pipeline:ne,model:[n.AutoModelForImageSegmentation,n.AutoModelForSemanticSegmentation,n.AutoModelForUniversalSegmentation],processor:o.AutoProcessor,default:{model:"Xenova/modnet"},type:"image"},"zero-shot-image-classification":{tokenizer:s.AutoTokenizer,pipeline:le,model:n.AutoModel,processor:o.AutoProcessor,default:{model:"Xenova/clip-vit-base-patch32"},type:"multimodal"},"object-detection":{pipeline:pe,model:n.AutoModelForObjectDetection,processor:o.AutoProcessor,default:{model:"Xenova/detr-resnet-50"},type:"multimodal"},"zero-shot-object-detection":{tokenizer:s.AutoTokenizer,pipeline:oe,model:n.AutoModelForZeroShotObjectDetection,processor:o.AutoProcessor,default:{model:"Xenova/owlvit-base-patch32"},type:"multimodal"},"document-question-answering":{tokenizer:s.AutoTokenizer,pipeline:K,model:n.AutoModelForDocumentQuestionAnswering,processor:o.AutoProcessor,default:{model:"Xenova/donut-base-finetuned-docvqa"},type:"multimodal"},"image-to-image":{pipeline:D,model:n.AutoModelForImageToImage,processor:o.AutoProcessor,default:{model:"Xenova/swin2SR-classical-sr-x2-64"},type:"image"},"depth-estimation":{pipeline:te,model:n.AutoModelForDepthEstimation,processor:o.AutoProcessor,default:{model:"Xenova/dpt-large"},type:"image"},"feature-extraction":{tokenizer:s.AutoTokenizer,pipeline:F,model:n.AutoModel,default:{model:"Xenova/all-MiniLM-L6-v2"},type:"text"},"image-feature-extraction":{processor:o.AutoProcessor,pipeline:H,model:[n.AutoModelForImageFeatureExtraction,n.AutoModel],default:{model:"Xenova/vit-base-patch16-224-in21k"},type:"image"}}),Ae=Object.freeze({"sentiment-analysis":"text-classification",ner:"token-classification",asr:"automatic-speech-recognition","text-to-speech":"text-to-audio",embeddings:"feature-extraction"});async function ke(Te,Q=null,{progress_callback:z=null,config:de=null,cache_dir:be=null,local_files_only:ve=!1,revision:xe="main",device:Ce=null,dtype:ge=null,subfolder:De="onnx",use_external_data_format:fe=null,model_file_name:Pe=null,session_options:We={}}={}){Te=Ae[Te]??Te;const Fe=he[Te.split("_",1)[0]];if(!Fe)throw Error(`Unsupported pipeline: ${Te}. Must be one of [${Object.keys(he)}]`);Q||(Q=Fe.default.model,console.log(`No model specified. Using default model: "${Q}".`));const tt={progress_callback:z,config:de,cache_dir:be,local_files_only:ve,revision:xe,device:Ce,dtype:ge,subfolder:De,use_external_data_format:fe,model_file_name:Pe,session_options:We},Re=new Map([["tokenizer",Fe.tokenizer],["model",Fe.model],["processor",Fe.processor]]),rt=await Ve(Re,Q,tt);rt.task=Te,(0,a.dispatchCallback)(z,{status:"ready",task:Te,model:Q});const Ze=Fe.pipeline;return new Ze(rt)}async function Ve(Te,Q,z){const de=Object.create(null),be=[];for(const[ve,xe]of Te.entries()){if(!xe)continue;let Ce;Array.isArray(xe)?Ce=new Promise(async(ge,De)=>{let fe;for(const Pe of xe){if(Pe===null){ge(null);return}try{ge(await Pe.from_pretrained(Q,z));return}catch(We){if(We.message?.includes("Unsupported model type"))fe=We;else if(We.message?.includes("Could not locate file"))fe=We;else{De(We);return}}}De(fe)}):Ce=xe.from_pretrained(Q,z),de[ve]=Ce,be.push(Ce)}await Promise.all(be);for(const[ve,xe]of Object.entries(de))de[ve]=await xe;return de}}),"./src/tokenizers.js":((e,r,t)=>{t.r(r),t.d(r,{AlbertTokenizer:()=>Ts,AutoTokenizer:()=>Nn,BartTokenizer:()=>He,BertTokenizer:()=>qt,BlenderbotSmallTokenizer:()=>Ue,BlenderbotTokenizer:()=>ze,BloomTokenizer:()=>It,CLIPTokenizer:()=>$s,CamembertTokenizer:()=>Z,CodeGenTokenizer:()=>ks,CodeLlamaTokenizer:()=>ir,CohereTokenizer:()=>As,ConvBertTokenizer:()=>q,DebertaTokenizer:()=>yt,DebertaV2Tokenizer:()=>Es,DistilBertTokenizer:()=>G,ElectraTokenizer:()=>ye,EsmTokenizer:()=>Lr,FalconTokenizer:()=>Dr,GPT2Tokenizer:()=>ut,GPTNeoXTokenizer:()=>Ss,GemmaTokenizer:()=>ns,Grok1Tokenizer:()=>wr,HerbertTokenizer:()=>C,LlamaTokenizer:()=>pr,M2M100Tokenizer:()=>os,MBart50Tokenizer:()=>qe,MBartTokenizer:()=>Mt,MPNetTokenizer:()=>Cs,MarianTokenizer:()=>as,MgpstrTokenizer:()=>Ns,MobileBertTokenizer:()=>Ps,NllbTokenizer:()=>Kr,NougatTokenizer:()=>Kt,PreTrainedTokenizer:()=>ft,Qwen2Tokenizer:()=>zr,RoFormerTokenizer:()=>R,RobertaTokenizer:()=>Pt,SiglipTokenizer:()=>Is,SpeechT5Tokenizer:()=>nt,SqueezeBertTokenizer:()=>Gr,T5Tokenizer:()=>et,TokenizerModel:()=>H,VitsTokenizer:()=>js,Wav2Vec2CTCTokenizer:()=>Nr,WhisperTokenizer:()=>Rs,XLMRobertaTokenizer:()=>Tr,XLMTokenizer:()=>ce,is_chinese_char:()=>k});var s=t("./src/utils/generic.js"),n=t("./src/utils/core.js"),o=t("./src/utils/hub.js"),i=t("./src/utils/maths.js"),a=t("./src/utils/tensor.js"),l=t("./src/utils/data-structures.js"),c=t("./node_modules/@huggingface/jinja/dist/index.js"),p=t("./src/models/whisper/common_whisper.js");async function d(ue,$){const U=await Promise.all([(0,o.getModelJSON)(ue,"tokenizer.json",!0,$),(0,o.getModelJSON)(ue,"tokenizer_config.json",!0,$)]);return $.legacy!==null&&(U[1].legacy=$.legacy),U}function u(ue,$){const U=[];let ee=0;for(const se of ue.matchAll($)){const Me=se[0];ee0&&U.push(Me),ee=se.index+Me.length}return ee=19968&&ue<=40959||ue>=13312&&ue<=19903||ue>=131072&&ue<=173791||ue>=173824&&ue<=177983||ue>=177984&&ue<=178207||ue>=178208&&ue<=183983||ue>=63744&&ue<=64255||ue>=194560&&ue<=195103}function T(ue,$,U){const ee=[];let se=0;for(;sethis.tokens_to_ids.get(U)??this.unk_token_id)}convert_ids_to_tokens($){return $.map(U=>this.vocab[U]??this.unk_token)}}class W extends H{constructor($){super($),this.tokens_to_ids=_($.vocab),this.unk_token_id=this.tokens_to_ids.get($.unk_token),this.unk_token=$.unk_token,this.max_input_chars_per_word=$.max_input_chars_per_word??100,this.vocab=new Array(this.tokens_to_ids.size);for(const[U,ee]of this.tokens_to_ids)this.vocab[ee]=U}encode($){const U=[];for(const ee of $){const se=[...ee];if(se.length>this.max_input_chars_per_word){U.push(this.unk_token);continue}let Me=!1,$e=0;const Xe=[];for(;$e0&&(Ke=this.config.continuing_subword_prefix+Ke),this.tokens_to_ids.has(Ke)){Ye=Ke;break}--Je}if(Ye===null){Me=!0;break}Xe.push(Ye),$e=Je}Me?U.push(this.unk_token):U.push(...Xe)}return U}}class B extends H{constructor($,U){super($);const ee=$.vocab.length;this.vocab=new Array(ee),this.scores=new Array(ee);for(let se=0;se[se,Me])),this.bos_token=" ",this.bos_token_id=this.tokens_to_ids.get(this.bos_token),this.eos_token=U.eos_token,this.eos_token_id=this.tokens_to_ids.get(this.eos_token),this.unk_token=this.vocab[this.unk_token_id],this.minScore=(0,i.min)(this.scores)[0],this.unk_score=this.minScore-10,this.scores[this.unk_token_id]=this.unk_score,this.trie=new l.CharTrie,this.trie.extend(this.vocab),this.fuse_unk=!0}populateNodes($){const U=$.chars,ee=1;let se=0;for(;se{const ue=[...Array.from({length:94},(se,Me)=>Me+33),...Array.from({length:12},(se,Me)=>Me+161),...Array.from({length:82},(se,Me)=>Me+174)],$=ue.slice();let U=0;for(let se=0;se<256;++se)ue.includes(se)||(ue.push(se),$.push(256+U),U+=1);const ee=$.map(se=>String.fromCharCode(se));return Object.fromEntries(ue.map((se,Me)=>[se,ee[Me]]))})(),X=(0,n.reverseDictionary)(Y);class J extends H{constructor($){super($),this.tokens_to_ids=_($.vocab),this.unk_token_id=this.tokens_to_ids.get($.unk_token),this.unk_token=$.unk_token,this.vocab=new Array(this.tokens_to_ids.size);for(const[ee,se]of this.tokens_to_ids)this.vocab[se]=ee;const U=Array.isArray($.merges[0]);this.merges=U?$.merges:$.merges.map(ee=>ee.split(" ",2)),this.bpe_ranks=new Map(this.merges.map((ee,se)=>[JSON.stringify(ee),se])),this.end_of_word_suffix=$.end_of_word_suffix,this.continuing_subword_suffix=$.continuing_subword_suffix??null,this.byte_fallback=this.config.byte_fallback??!1,this.byte_fallback&&(this.text_encoder=new TextEncoder),this.ignore_merges=this.config.ignore_merges??!1,this.max_length_to_cache=256,this.cache_capacity=1e4,this.cache=new l.LRUCache(this.cache_capacity)}clear_cache(){this.cache.clear()}bpe($){if($.length===0)return[];const U=this.cache.get($);if(U!==void 0)return U;const ee=Array.from($);this.end_of_word_suffix&&(ee[ee.length-1]+=this.end_of_word_suffix);let se=[];if(ee.length>1){const Me=new l.PriorityQueue((Je,Ye)=>Je.score`<0x${Xe.toString(16).toUpperCase().padStart(2,"0")}>`);$e.every(Xe=>this.tokens_to_ids.has(Xe))?U.push(...$e):U.push(this.unk_token)}else U.push(this.unk_token)}return U}}class re extends H{constructor($,U){super($),this.tokens_to_ids=_(U.target_lang?$.vocab[U.target_lang]:$.vocab),this.bos_token=U.bos_token,this.bos_token_id=this.tokens_to_ids.get(this.bos_token),this.eos_token=U.eos_token,this.eos_token_id=this.tokens_to_ids.get(this.eos_token),this.pad_token=U.pad_token,this.pad_token_id=this.tokens_to_ids.get(this.pad_token),this.unk_token=U.unk_token,this.unk_token_id=this.tokens_to_ids.get(this.unk_token),this.vocab=new Array(this.tokens_to_ids.size);for(const[ee,se]of this.tokens_to_ids)this.vocab[se]=ee}encode($){return $}}class ne extends s.Callable{constructor($){super(),this.config=$}static fromConfig($){if($===null)return null;switch($.type){case"BertNormalizer":return new Te($);case"Precompiled":return new Bt($);case"Sequence":return new Ve($);case"Replace":return new le($);case"NFC":return new oe($);case"NFD":return new K($);case"NFKC":return new j($);case"NFKD":return new D($);case"Strip":return new te($);case"StripAccents":return new he($);case"Lowercase":return new Ae($);case"Prepend":return new ke($);default:throw new Error(`Unknown Normalizer type: ${$.type}`)}}normalize($){throw Error("normalize should be implemented in subclass.")}_call($){return this.normalize($)}}class le extends ne{normalize($){const U=f(this.config.pattern);return U===null?$:$.replaceAll(U,this.config.content)}}class pe extends ne{form=void 0;normalize($){return $=$.normalize(this.form),$}}class oe extends pe{form="NFC"}class K extends pe{form="NFD"}class j extends pe{form="NFKC"}class D extends pe{form="NFKD"}class te extends ne{normalize($){return this.config.strip_left&&this.config.strip_right?$=$.trim():(this.config.strip_left&&($=$.trimStart()),this.config.strip_right&&($=$.trimEnd())),$}}class he extends ne{normalize($){return $=w($),$}}class Ae extends ne{normalize($){return $=$.toLowerCase(),$}}class ke extends ne{normalize($){return $=this.config.prepend+$,$}}class Ve extends ne{constructor($){super($),this.normalizers=$.normalizers.map(U=>ne.fromConfig(U))}normalize($){return this.normalizers.reduce((U,ee)=>ee.normalize(U),$)}}class Te extends ne{_tokenize_chinese_chars($){const U=[];for(let ee=0;ee<$.length;++ee){const se=$[ee],Me=se.charCodeAt(0);k(Me)?(U.push(" "),U.push(se),U.push(" ")):U.push(se)}return U.join("")}stripAccents($){return $.normalize("NFD").replace(new RegExp("\\p{Mn}","gu"),"")}_is_control($){switch($){case" ":case` -`:case"\r":return!1;default:return new RegExp("^\\p{Cc}|\\p{Cf}|\\p{Co}|\\p{Cs}$","u").test($)}}_clean_text($){const U=[];for(const ee of $){const se=ee.charCodeAt(0);se===0||se===65533||this._is_control(ee)||(/^\s$/.test(ee)?U.push(" "):U.push(ee))}return U.join("")}normalize($){return this.config.clean_text&&($=this._clean_text($)),this.config.handle_chinese_chars&&($=this._tokenize_chinese_chars($)),this.config.lowercase?($=$.toLowerCase(),this.config.strip_accents!==!1&&($=this.stripAccents($))):this.config.strip_accents&&($=this.stripAccents($)),$}}class Q extends s.Callable{static fromConfig($){if($===null)return null;switch($.type){case"BertPreTokenizer":return new z($);case"Sequence":return new jr($);case"Whitespace":return new Qs($);case"WhitespaceSplit":return new Xs($);case"Metaspace":return new gr($);case"ByteLevel":return new de($);case"Split":return new be($);case"Punctuation":return new ve($);case"Digits":return new xe($);case"Replace":return new Js($);case"FixedLength":return new ar($);default:throw new Error(`Unknown PreTokenizer type: ${$.type}`)}}pre_tokenize_text($,U){throw Error("pre_tokenize_text should be implemented in subclass.")}pre_tokenize($,U){return(Array.isArray($)?$.map(ee=>this.pre_tokenize_text(ee,U)):this.pre_tokenize_text($,U)).flat()}_call($,U){return this.pre_tokenize($,U)}}class z extends Q{constructor($){super(),this.pattern=new RegExp(`[^\\s${P}]+|[${P}]`,"gu")}pre_tokenize_text($,U){return $.trim().match(this.pattern)||[]}}class de extends Q{constructor($){super(),this.config=$,this.add_prefix_space=this.config.add_prefix_space,this.trim_offsets=this.config.trim_offsets,this.use_regex=this.config.use_regex??!0,this.pattern=new RegExp("'s|'t|'re|'ve|'m|'ll|'d| ?\\p{L}+| ?\\p{N}+| ?[^\\s\\p{L}\\p{N}]+|\\s+(?!\\S)|\\s+","gu"),this.byte_encoder=Y,this.text_encoder=new TextEncoder}pre_tokenize_text($,U){return this.add_prefix_space&&!$.startsWith(" ")&&($=" "+$),(this.use_regex?$.match(this.pattern)||[]:[$]).map(se=>Array.from(this.text_encoder.encode(se),Me=>this.byte_encoder[Me]).join(""))}}class be extends Q{constructor($){super(),this.config=$,this.pattern=f(this.config.pattern,this.config.invert)}pre_tokenize_text($,U){return this.pattern===null?[]:this.config.invert?$.match(this.pattern)||[]:this.config.behavior?.toLowerCase()==="removed"?$.split(this.pattern).filter(ee=>ee):u($,this.pattern)}}class ve extends Q{constructor($){super(),this.config=$,this.pattern=new RegExp(`[^${P}]+|[${P}]+`,"gu")}pre_tokenize_text($,U){return $.match(this.pattern)||[]}}class xe extends Q{constructor($){super(),this.config=$;const U=`[^\\d]+|\\d${this.config.individual_digits?"":"+"}`;this.pattern=new RegExp(U,"gu")}pre_tokenize_text($,U){return $.match(this.pattern)||[]}}class Ce extends s.Callable{constructor($){super(),this.config=$}static fromConfig($){if($===null)return null;switch($.type){case"TemplateProcessing":return new fe($);case"ByteLevel":return new Pe($);case"RobertaProcessing":return new De($);case"BertProcessing":return new ge($);case"Sequence":return new We($);default:throw new Error(`Unknown PostProcessor type: ${$.type}`)}}post_process($,...U){throw Error("post_process should be implemented in subclass.")}_call($,...U){return this.post_process($,...U)}}class ge extends Ce{constructor($){super($),this.cls=$.cls[0],this.sep=$.sep[0]}post_process($,U=null,{add_special_tokens:ee=!0}={}){ee&&($=(0,n.mergeArrays)([this.cls],$,[this.sep]));let se=new Array($.length).fill(0);if(U!==null){const Me=ee&&this instanceof De?[this.sep]:[],$e=ee?[this.sep]:[];$=(0,n.mergeArrays)($,Me,U,$e),se=(0,n.mergeArrays)(se,new Array(U.length+Me.length+$e.length).fill(1))}return{tokens:$,token_type_ids:se}}}class De extends ge{}class fe extends Ce{constructor($){super($),this.single=$.single,this.pair=$.pair}post_process($,U=null,{add_special_tokens:ee=!0}={}){const se=U===null?this.single:this.pair;let Me=[],$e=[];for(const Xe of se)"SpecialToken"in Xe?ee&&(Me.push(Xe.SpecialToken.id),$e.push(Xe.SpecialToken.type_id)):"Sequence"in Xe&&(Xe.Sequence.id==="A"?(Me=(0,n.mergeArrays)(Me,$),$e=(0,n.mergeArrays)($e,new Array($.length).fill(Xe.Sequence.type_id))):Xe.Sequence.id==="B"&&(Me=(0,n.mergeArrays)(Me,U),$e=(0,n.mergeArrays)($e,new Array(U.length).fill(Xe.Sequence.type_id))));return{tokens:Me,token_type_ids:$e}}}class Pe extends Ce{post_process($,U=null){return U&&($=(0,n.mergeArrays)($,U)),{tokens:$}}}class We extends Ce{constructor($){super($),this.processors=$.processors.map(U=>Ce.fromConfig(U))}post_process($,U=null,ee={}){let se;for(const Me of this.processors)if(Me instanceof Pe)$=Me.post_process($).tokens,U&&(U=Me.post_process(U).tokens);else{const $e=Me.post_process($,U,ee);$=$e.tokens,se=$e.token_type_ids}return{tokens:$,token_type_ids:se}}}class Fe extends s.Callable{constructor($){super(),this.config=$,this.added_tokens=[],this.end_of_word_suffix=null,this.trim_offsets=$.trim_offsets}static fromConfig($){if($===null)return null;switch($.type){case"WordPiece":return new je($);case"Metaspace":return new Or($);case"ByteLevel":return new Oe($);case"Replace":return new tt($);case"ByteFallback":return new Re($);case"Fuse":return new rt($);case"Strip":return new Ze($);case"Sequence":return new ht($);case"CTC":return new at($);case"BPEDecoder":return new Nt($);default:throw new Error(`Unknown Decoder type: ${$.type}`)}}_call($){return this.decode($)}decode($){return this.decode_chain($).join("")}decode_chain($){throw Error("`decode_chain` should be implemented in subclass.")}}class tt extends Fe{decode_chain($){const U=f(this.config.pattern);return U===null?$:$.map(ee=>ee.replaceAll(U,this.config.content))}}class Re extends Fe{constructor($){super($),this.text_decoder=new TextDecoder}decode_chain($){const U=[];let ee=[];for(const se of $){let Me=null;if(se.length===6&&se.startsWith("<0x")&&se.endsWith(">")){const $e=parseInt(se.slice(3,5),16);isNaN($e)||(Me=$e)}if(Me!==null)ee.push(Me);else{if(ee.length>0){const $e=this.text_decoder.decode(Uint8Array.from(ee));U.push($e),ee=[]}U.push(se)}}if(ee.length>0){const se=this.text_decoder.decode(Uint8Array.from(ee));U.push(se),ee=[]}return U}}class rt extends Fe{decode_chain($){return[$.join("")]}}class Ze extends Fe{constructor($){super($),this.content=this.config.content,this.start=this.config.start,this.stop=this.config.stop}decode_chain($){return $.map(U=>{let ee=0;for(let Me=0;Me(ee!==0&&(U.startsWith(this.config.prefix)?U=U.replace(this.config.prefix,""):U=" "+U),this.cleanup&&(U=I(U)),U))}}class Oe extends Fe{constructor($){super($),this.byte_decoder=X,this.text_decoder=new TextDecoder("utf-8",{fatal:!1,ignoreBOM:!0}),this.end_of_word_suffix=null}convert_tokens_to_string($){const U=$.join(""),ee=new Uint8Array([...U].map(Me=>this.byte_decoder[Me]));return this.text_decoder.decode(ee)}decode_chain($){const U=[];let ee=[];for(const se of $)this.added_tokens.find(Me=>Me.content===se)!==void 0?(ee.length>0&&(U.push(this.convert_tokens_to_string(ee)),ee=[]),U.push(se)):ee.push(se);return ee.length>0&&U.push(this.convert_tokens_to_string(ee)),U}}class at extends Fe{constructor($){super($),this.pad_token=this.config.pad_token,this.word_delimiter_token=this.config.word_delimiter_token,this.cleanup=this.config.cleanup}convert_tokens_to_string($){if($.length===0)return"";const U=[$[0]];for(let Me=1;Me<$.length;++Me)$[Me]!==U.at(-1)&&U.push($[Me]);let se=U.filter(Me=>Me!==this.pad_token).join("");return this.cleanup&&(se=I(se).replaceAll(this.word_delimiter_token," ").trim()),se}decode_chain($){return[this.convert_tokens_to_string($)]}}class ht extends Fe{constructor($){super($),this.decoders=$.decoders.map(U=>Fe.fromConfig(U))}decode_chain($){return this.decoders.reduce((U,ee)=>ee.decode_chain(U),$)}}class Nt extends Fe{constructor($){super($),this.suffix=this.config.suffix}decode_chain($){return $.map((U,ee)=>U.replaceAll(this.suffix,ee===$.length-1?"":" "))}}class kt extends Fe{decode_chain($){let U="";for(let ee=1;ee<$.length;ee+=2)U+=$[ee];return[U]}}class gr extends Q{constructor($){super(),this.replacement=$.replacement,this.strRep=$.str_rep||this.replacement,this.prepend_scheme=$.prepend_scheme??"always"}pre_tokenize_text($,{section_index:U=void 0}={}){let ee=$.replaceAll(" ",this.strRep);return!ee.startsWith(this.replacement)&&(this.prepend_scheme==="always"||this.prepend_scheme==="first"&&U===0)&&(ee=this.strRep+ee),[ee]}}class Or extends Fe{constructor($){super($),this.replacement=$.replacement}decode_chain($){const U=[];for(let ee=0;ee<$.length;++ee){let se=$[ee].replaceAll(this.replacement," ");ee==0&&se.startsWith(" ")&&(se=se.substring(1)),U.push(se)}return U}}class Bt extends ne{constructor($){super($),this.charsmap=$.precompiled_charsmap}normalize($){return $=$.replace(/[\u0001-\u0008\u000B\u000E-\u001F\u007F\u008F\u009F]/gm,""),$=$.replace(/[\u0009\u000A\u000C\u000D\u00A0\u1680\u2000-\u200F\u2028\u2029\u202F\u205F\u2581\u3000\uFEFF\uFFFD]/gm," "),$.includes("~")?$=$.split("~").map(ee=>ee.normalize("NFKC")).join("~"):$=$.normalize("NFKC"),$}}class jr extends Q{constructor($){super(),this.tokenizers=$.pretokenizers.map(U=>Q.fromConfig(U))}pre_tokenize_text($,U){return this.tokenizers.reduce((ee,se)=>se.pre_tokenize(ee,U),[$])}}class Qs extends Q{constructor($){super()}pre_tokenize_text($,U){return $.match(/\w+|[^\w\s]+/g)||[]}}class Xs extends Q{constructor($){super()}pre_tokenize_text($,U){return b($)}}class Js extends Q{constructor($){super(),this.config=$,this.pattern=f(this.config.pattern),this.content=this.config.content}pre_tokenize_text($,U){return this.pattern===null?[$]:[$.replaceAll(this.pattern,this.config.content)]}}class ar extends Q{constructor($){super(),this._length=$.length}pre_tokenize_text($,U){const ee=[];for(let se=0;se<$.length;se+=this._length)ee.push($.slice(se,se+this._length));return ee}}const kr=["bos_token","eos_token","unk_token","sep_token","pad_token","cls_token","mask_token"];function Jr(ue,$,U,ee){for(const se of Object.keys(ue)){const Me=$-ue[se].length,$e=U(se),Xe=new Array(Me).fill($e);ue[se]=ee==="right"?(0,n.mergeArrays)(ue[se],Xe):(0,n.mergeArrays)(Xe,ue[se])}}function Bs(ue,$){for(const U of Object.keys(ue))ue[U].length=$}class ft extends s.Callable{return_token_type_ids=!1;padding_side="right";constructor($,U){super(),this.config=U,this.normalizer=ne.fromConfig($.normalizer),this.pre_tokenizer=Q.fromConfig($.pre_tokenizer),this.model=H.fromConfig($.model,U),this.post_processor=Ce.fromConfig($.post_processor),this.decoder=Fe.fromConfig($.decoder),this.special_tokens=[],this.all_special_ids=[],this.added_tokens=[];for(const ee of $.added_tokens){const se=new F(ee);this.added_tokens.push(se),this.model.tokens_to_ids.set(se.content,se.id),this.model.vocab[se.id]=se.content,se.special&&(this.special_tokens.push(se.content),this.all_special_ids.push(se.id))}if(this.additional_special_tokens=U.additional_special_tokens??[],this.special_tokens.push(...this.additional_special_tokens),this.special_tokens=[...new Set(this.special_tokens)],this.decoder&&(this.decoder.added_tokens=this.added_tokens,this.decoder.end_of_word_suffix=this.model.end_of_word_suffix),this.added_tokens_splitter=new l.DictionarySplitter(this.added_tokens.map(ee=>ee.content)),this.added_tokens_map=new Map(this.added_tokens.map(ee=>[ee.content,ee])),this.mask_token=this.getToken("mask_token"),this.mask_token_id=this.model.tokens_to_ids.get(this.mask_token),this.pad_token=this.getToken("pad_token","eos_token"),this.pad_token_id=this.model.tokens_to_ids.get(this.pad_token),this.sep_token=this.getToken("sep_token"),this.sep_token_id=this.model.tokens_to_ids.get(this.sep_token),this.unk_token=this.getToken("unk_token"),this.unk_token_id=this.model.tokens_to_ids.get(this.unk_token),this.bos_token=this.getToken("bos_token"),this.bos_token_id=this.model.tokens_to_ids.get(this.bos_token),this.eos_token=this.getToken("eos_token"),this.eos_token_id=this.model.tokens_to_ids.get(this.eos_token),this.model_max_length=U.model_max_length,this.remove_space=U.remove_space,this.clean_up_tokenization_spaces=U.clean_up_tokenization_spaces??!0,this.do_lowercase_and_remove_accent=U.do_lowercase_and_remove_accent??!1,U.padding_side&&(this.padding_side=U.padding_side),this.add_bos_token=U.add_bos_token,this.add_eos_token=U.add_eos_token,this.legacy=!1,this.chat_template=U.chat_template??null,Array.isArray(this.chat_template)){const ee=Object.create(null);for(const{name:se,template:Me}of this.chat_template){if(typeof se!="string"||typeof Me!="string")throw new Error('Chat template must be a list of objects with "name" and "template" properties');ee[se]=Me}this.chat_template=ee}this._compiled_template_cache=new Map}getToken(...$){for(const U of $){const ee=this.config[U];if(ee)if(typeof ee=="object"){if(ee.__type==="AddedToken")return ee.content;throw Error(`Unknown token: ${ee}`)}else return ee}return null}static async from_pretrained($,{progress_callback:U=null,config:ee=null,cache_dir:se=null,local_files_only:Me=!1,revision:$e="main",legacy:Xe=null}={}){const Je=await d($,{progress_callback:U,config:ee,cache_dir:se,local_files_only:Me,revision:$e,legacy:Xe});return new this(...Je)}_call($,{text_pair:U=null,add_special_tokens:ee=!0,padding:se=!1,truncation:Me=null,max_length:$e=null,return_tensor:Xe=!0,return_token_type_ids:Je=null}={}){const Ye=Array.isArray($);let Ke;if(Ye){if($.length===0)throw Error("text array must be non-empty");if(U!==null){if(Array.isArray(U)){if($.length!==U.length)throw Error("text and text_pair must have the same length")}else throw Error("text_pair must also be an array");Ke=$.map((Et,rr)=>this._encode_plus(Et,{text_pair:U[rr],add_special_tokens:ee,return_token_type_ids:Je}))}else Ke=$.map(Et=>this._encode_plus(Et,{add_special_tokens:ee,return_token_type_ids:Je}))}else{if($==null)throw Error("text may not be null or undefined");if(Array.isArray(U))throw Error("When specifying `text_pair`, since `text` is a string, `text_pair` must also be a string (i.e., not an array).");Ke=[this._encode_plus($,{text_pair:U,add_special_tokens:ee,return_token_type_ids:Je})]}if($e===null?$e=this.model_max_length:Me===null&&(se===!0?(console.warn("`max_length` is ignored when `padding: true` and there is no truncation strategy. To pad to max length, use `padding: 'max_length'`."),$e=this.model_max_length):se===!1&&(console.warn("Truncation was not explicitly activated but `max_length` is provided a specific value, please use `truncation: true` to explicitly truncate examples to max length."),Me=!0)),se===!0&&($e=Math.min((0,i.max)(Ke.map(Et=>Et.input_ids.length))[0],$e??1/0)),$e=Math.min($e,this.model_max_length??1/0),se||Me)for(let Et=0;Et$e?Me&&Bs(Ke[Et],$e):se&&Jr(Ke[Et],$e,rr=>rr==="input_ids"?this.pad_token_id:0,this.padding_side));const $t={};if(Xe){if(!(se&&Me)&&Ke.some(rr=>{for(const br of Object.keys(rr))if(rr[br].length!==Ke[0][br]?.length)return!0;return!1}))throw Error("Unable to create tensor, you should probably activate truncation and/or padding with 'padding=true' and 'truncation=true' to have batched tensors with the same length.");const Et=[Ke.length,Ke[0].input_ids.length];for(const rr of Object.keys(Ke[0]))$t[rr]=new a.Tensor("int64",BigInt64Array.from(Ke.flatMap(br=>br[rr]).map(BigInt)),Et)}else{for(const Et of Object.keys(Ke[0]))$t[Et]=Ke.map(rr=>rr[Et]);if(!Ye)for(const Et of Object.keys($t))$t[Et]=$t[Et][0]}return $t}_encode_text($){if($===null)return null;const U=this.added_tokens_splitter.split($);for(let se=0;se0&&(U[se-1]=U[se-1].trimEnd()),Me.rstrip&&se{if(se.length===0)return[];if(this.added_tokens_map.has(se))return[se];if(this.remove_space===!0&&(se=se.trim().split(/\s+/).join(" ")),this.do_lowercase_and_remove_accent&&(se=v(se)),this.normalizer!==null&&(se=this.normalizer(se)),se.length===0)return[];const $e=this.pre_tokenizer!==null?this.pre_tokenizer(se,{section_index:Me}):[se];return this.model($e)})}_encode_plus($,{text_pair:U=null,add_special_tokens:ee=!0,return_token_type_ids:se=null}={}){const{tokens:Me,token_type_ids:$e}=this._tokenize_helper($,{pair:U,add_special_tokens:ee}),Xe=this.model.convert_tokens_to_ids(Me),Je={input_ids:Xe,attention_mask:new Array(Xe.length).fill(1)};return(se??this.return_token_type_ids)&&$e&&(Je.token_type_ids=$e),Je}_tokenize_helper($,{pair:U=null,add_special_tokens:ee=!1}={}){const se=this._encode_text($),Me=this._encode_text(U);return this.post_processor?this.post_processor(se,Me,{add_special_tokens:ee}):{tokens:(0,n.mergeArrays)(se??[],Me??[])}}tokenize($,{pair:U=null,add_special_tokens:ee=!1}={}){return this._tokenize_helper($,{pair:U,add_special_tokens:ee}).tokens}encode($,{text_pair:U=null,add_special_tokens:ee=!0,return_token_type_ids:se=null}={}){return this._encode_plus($,{text_pair:U,add_special_tokens:ee,return_token_type_ids:se}).input_ids}batch_decode($,U={}){return $ instanceof a.Tensor&&($=$.tolist()),$.map(ee=>this.decode(ee,U))}decode($,U={}){if($ instanceof a.Tensor&&($=y($)),!Array.isArray($)||$.length===0||!(0,n.isIntegralNumber)($[0]))throw Error("token_ids must be a non-empty array of integers.");return this.decode_single($,U)}decode_single($,{skip_special_tokens:U=!1,clean_up_tokenization_spaces:ee=null}){let se=this.model.convert_ids_to_tokens($);U&&(se=se.filter($e=>!this.special_tokens.includes($e)));let Me=this.decoder?this.decoder(se):se.join(" ");return this.decoder&&this.decoder.end_of_word_suffix&&(Me=Me.replaceAll(this.decoder.end_of_word_suffix," "),U&&(Me=Me.trim())),(ee??this.clean_up_tokenization_spaces)&&(Me=I(Me)),Me}get_chat_template({chat_template:$=null,tools:U=null}={}){if(this.chat_template&&typeof this.chat_template=="object"){const ee=this.chat_template;if($!==null&&Object.hasOwn(ee,$))$=ee[$];else if($===null)if(U!==null&&"tool_use"in ee)$=ee.tool_use;else if("default"in ee)$=ee.default;else throw Error(`This model has multiple chat templates with no default specified! Please either pass a chat template or the name of the template you wish to use to the 'chat_template' argument. Available template names are ${Object.keys(ee).sort()}.`)}else if($===null)if(this.chat_template)$=this.chat_template;else throw Error("Cannot use apply_chat_template() because tokenizer.chat_template is not set and no template argument was passed! For information about writing templates and setting the tokenizer.chat_template attribute, please see the documentation at https://huggingface.co/docs/transformers/main/en/chat_templating");return $}apply_chat_template($,{tools:U=null,documents:ee=null,chat_template:se=null,add_generation_prompt:Me=!1,tokenize:$e=!0,padding:Xe=!1,truncation:Je=!1,max_length:Ye=null,return_tensor:Ke=!0,return_dict:$t=!1,tokenizer_kwargs:Et={},...rr}={}){if(se=this.get_chat_template({chat_template:se,tools:U}),typeof se!="string")throw Error(`chat_template must be a string, but got ${typeof se}`);let br=this._compiled_template_cache.get(se);br===void 0&&(br=new c.Template(se),this._compiled_template_cache.set(se,br));const Jt=Object.create(null);for(const Ht of kr){const Hr=this.getToken(Ht);Hr&&(Jt[Ht]=Hr)}const sr=br.render({messages:$,add_generation_prompt:Me,tools:U,documents:ee,...Jt,...rr});if($e){const Ht=this._call(sr,{add_special_tokens:!1,padding:Xe,truncation:Je,max_length:Ye,return_tensor:Ke,...Et});return $t?Ht:Ht.input_ids}return sr}}class qt extends ft{return_token_type_ids=!0}class Ts extends ft{return_token_type_ids=!0}class Ps extends ft{return_token_type_ids=!0}class Gr extends ft{return_token_type_ids=!0}class yt extends ft{return_token_type_ids=!0}class Es extends ft{return_token_type_ids=!0}class C extends ft{return_token_type_ids=!0}class q extends ft{return_token_type_ids=!0}class R extends ft{return_token_type_ids=!0}class G extends ft{}class Z extends ft{}class ce extends ft{return_token_type_ids=!0;constructor($,U){super($,U),console.warn('WARNING: `XLMTokenizer` is not yet supported by Hugging Face\'s "fast" tokenizers library. Therefore, you may experience slightly inaccurate results.')}}class ye extends ft{return_token_type_ids=!0}class et extends ft{}class ut extends ft{}class He extends ft{}class Mt extends ft{constructor($,U){super($,U),this.languageRegex=/^[a-z]{2}_[A-Z]{2}$/,this.language_codes=this.special_tokens.filter(ee=>this.languageRegex.test(ee)),this.lang_to_token=ee=>ee}_build_translation_inputs($,U,ee){return lr(this,$,U,ee)}}class qe extends Mt{}class Pt extends ft{}class It extends ft{}const Mr="▁";class pr extends ft{padding_side="left";constructor($,U){super($,U),this.legacy=U.legacy??!0,this.legacy||(this.normalizer=null,this.pre_tokenizer=new gr({replacement:Mr,prepend_scheme:"first"}))}_encode_text($){if($===null)return null;if(this.legacy||$.length===0)return super._encode_text($);let U=super._encode_text(Mr+$.replaceAll(Mr," "));return U.length>1&&U[0]===Mr&&this.special_tokens.includes(U[1])&&(U=U.slice(1)),U}}class ir extends ft{}class Tr extends ft{}class Cs extends ft{}class Dr extends ft{}class Ss extends ft{}class Lr extends ft{}class zr extends ft{}class ns extends ft{}class wr extends ft{}function lr(ue,$,U,ee){if(!("language_codes"in ue)||!Array.isArray(ue.language_codes))throw new Error("Tokenizer must have `language_codes` attribute set and it should be an array of language ids.");if(!("languageRegex"in ue)||!(ue.languageRegex instanceof RegExp))throw new Error("Tokenizer must have `languageRegex` attribute set and it should be a regular expression.");if(!("lang_to_token"in ue)||typeof ue.lang_to_token!="function")throw new Error("Tokenizer must have `lang_to_token` attribute set and it should be a function.");const se=ee.src_lang,Me=ee.tgt_lang;if(!ue.language_codes.includes(Me))throw new Error(`Target language code "${Me}" is not valid. Must be one of: {${ue.language_codes.join(", ")}}`);if(se!==void 0){if(!ue.language_codes.includes(se))throw new Error(`Source language code "${se}" is not valid. Must be one of: {${ue.language_codes.join(", ")}}`);for(const $e of ue.post_processor.config.single)if("SpecialToken"in $e&&ue.languageRegex.test($e.SpecialToken.id)){$e.SpecialToken.id=ue.lang_to_token(se);break}}return ee.forced_bos_token_id=ue.model.convert_tokens_to_ids([ue.lang_to_token(Me)])[0],ue._call($,U)}class Kr extends ft{constructor($,U){super($,U),this.languageRegex=/^[a-z]{3}_[A-Z][a-z]{3}$/,this.language_codes=this.special_tokens.filter(ee=>this.languageRegex.test(ee)),this.lang_to_token=ee=>ee}_build_translation_inputs($,U,ee){return lr(this,$,U,ee)}}class os extends ft{constructor($,U){super($,U),this.languageRegex=/^__[a-z]{2,3}__$/,this.language_codes=this.special_tokens.filter(ee=>this.languageRegex.test(ee)).map(ee=>ee.slice(2,-2)),this.lang_to_token=ee=>`__${ee}__`}_build_translation_inputs($,U,ee){return lr(this,$,U,ee)}}class Rs extends ft{get timestamp_begin(){return this.model.convert_tokens_to_ids(["<|notimestamps|>"])[0]+1}_decode_asr($,{return_timestamps:U=!1,return_language:ee=!1,time_precision:se=null,force_full_sequences:Me=!0}={}){if(se===null)throw Error("Must specify time_precision");let $e=null;const Xe=U==="word";function Je(){return{language:$e,timestamp:[null,null],text:""}}const Ye=[];let Ke=Je(),$t=0;const Et=this.timestamp_begin,br=Et+1500;let Jt=[],sr=[],Ht=!1,Hr=null;const is=new Set(this.all_special_ids);for(const Yt of $){const hr=Yt.tokens,$r=Xe?Yt.token_timestamps:null;let Yr=null,bs=Et;if("stride"in Yt){const[vr,Zt,_r]=Yt.stride;if($t-=Zt,Hr=vr-_r,Zt&&(bs=Zt/se+Et),_r)for(let cr=hr.length-1;cr>=0;--cr){const Ur=Number(hr[cr]);if(Ur>=Et){if(Yr!==null&&(Ur-Et)*se=Et&&Zt<=br){const _r=(Zt-Et)*se+$t,cr=(0,i.round)(_r,2);if(Yr!==null&&Zt>=Yr)Ht=!0;else if(Ht||Jt.length>0&&Zt0?(Jt.push(yr),Xe&&sr.push(ls)):Jt.every(vr=>vr.length===0)&&(Ke=Je(),Jt=[],yr=[],sr=[],ls=[])}if(Jt.length>0){if(Me&&U)throw new Error("Whisper did not predict an ending timestamp, which can happen if audio is cut off in the middle of a word. Also make sure WhisperTimeStampLogitsProcessor was used during generation.");const[Yt,hr]=this.findLongestCommonSequence(Jt,sr),$r=this.decode(Yt);Ke.text=$r,Xe&&(Ke.words=this.collateWordTimestamps(Yt,hr,$e)),Ye.push(Ke)}let Pr=Object.create(null);const ws=Ye.map(Yt=>Yt.text).join("");if(U||ee){for(let Yt=0;Yt0;let Xe=$e?[]:null,Je=$e?U[0]:null;for(let Ye=1;Ye<$.length;++Ye){const Ke=$[Ye];let $t=0,Et=[se,se,0,0];const rr=Ke.length;for(let Pr=1;PrZt===bs[_r]&&Je[ws+_r]<=U[Ye][$r+_r]).length:yr=hr.filter((Zt,_r)=>Zt===bs[_r]).length;const ls=Pr/1e4,vr=yr/Pr+ls;yr>1&&vr>$t&&($t=vr,Et=[ws,Yt,$r,Yr])}const[br,Jt,sr,Ht]=Et,Hr=Math.floor((Jt+br)/2),is=Math.floor((Ht+sr)/2);Me.push(...ee.slice(0,Hr)),ee=Ke.slice(is),se=ee.length,$e&&(Xe.push(...Je.slice(0,Hr)),Je=U[Ye].slice(is))}return Me.push(...ee),$e?(Xe.push(...Je),[Me,Xe]):[Me,[]]}collateWordTimestamps($,U,ee){const[se,Me,$e]=this.combineTokensIntoWords($,ee),Xe=[];for(let Je=0;Je=se){const Xe=(($e-se)*ee).toFixed(2);Me.push(`<|${Xe}|>`),Me.push([])}else Me[Me.length-1].push($e);return Me=Me.map($e=>typeof $e=="string"?$e:super.decode($e,U)),Me.join("")}splitTokensOnUnicode($){const U=this.decode($,{decode_with_timestamps:!0}),ee="�",se=[],Me=[],$e=[];let Xe=[],Je=[],Ye=0;for(let Ke=0;Ke<$.length;++Ke){const $t=$[Ke];Xe.push($t),Je.push(Ke);const Et=this.decode(Xe,{decode_with_timestamps:!0});(!Et.includes(ee)||U[Ye+Et.indexOf(ee)]===ee)&&(se.push(Et),Me.push(Xe),$e.push(Je),Xe=[],Je=[],Ye+=Et.length)}return[se,Me,$e]}splitTokensOnSpaces($){const[U,ee,se]=this.splitTokensOnUnicode($),Me=[],$e=[],Xe=[],Je=new RegExp(`^[${P}]$`,"gu");for(let Ye=0;Ye=this.model.tokens_to_ids.get("<|endoftext|>"),br=Ke.startsWith(" "),Jt=Ke.trim(),sr=Je.test(Jt);if(rr||br||sr||Me.length===0)Me.push(Ke),$e.push($t),Xe.push(Et);else{const Ht=Me.length-1;Me[Ht]+=Ke,$e[Ht].push(...$t),Xe[Ht].push(...Et)}}return[Me,$e,Xe]}mergePunctuations($,U,ee,se,Me){const $e=structuredClone($),Xe=structuredClone(U),Je=structuredClone(ee);let Ye=$e.length-2,Ke=$e.length-1;for(;Ye>=0;)$e[Ye].startsWith(" ")&&se.includes($e[Ye].trim())?($e[Ke]=$e[Ye]+$e[Ke],Xe[Ke]=(0,n.mergeArrays)(Xe[Ye],Xe[Ke]),Je[Ke]=(0,n.mergeArrays)(Je[Ye],Je[Ke]),$e[Ye]="",Xe[Ye]=[],Je[Ye]=[]):Ke=Ye,--Ye;for(Ye=0,Ke=1;Ke<$e.length;)!$e[Ye].endsWith(" ")&&Me.includes($e[Ke])?($e[Ye]+=$e[Ke],Xe[Ye]=(0,n.mergeArrays)(Xe[Ye],Xe[Ke]),Je[Ye]=(0,n.mergeArrays)(Je[Ye],Je[Ke]),$e[Ke]="",Xe[Ke]=[],Je[Ke]=[]):Ye=Ke,++Ke;return[$e.filter($t=>$t),Xe.filter($t=>$t.length>0),Je.filter($t=>$t.length>0)]}}class ks extends ft{}class $s extends ft{}class Is extends ft{}class as extends ft{constructor($,U){super($,U),this.languageRegex=/^(>>\w+<<)\s*/g,this.supported_language_codes=this.model.vocab.filter(ee=>this.languageRegex.test(ee)),console.warn('WARNING: `MarianTokenizer` is not yet supported by Hugging Face\'s "fast" tokenizers library. Therefore, you may experience slightly inaccurate results.')}_encode_text($){if($===null)return null;const[U,...ee]=$.trim().split(this.languageRegex);if(ee.length===0)return super._encode_text(U);if(ee.length===2){const[se,Me]=ee;return this.supported_language_codes.includes(se)||console.warn(`Unsupported language code "${se}" detected, which may lead to unexpected behavior. Should be one of: ${JSON.stringify(this.supported_language_codes)}`),(0,n.mergeArrays)([se],super._encode_text(Me))}}}class Nr extends ft{}class ze extends ft{}class Ue extends ft{}class nt extends ft{}class Kt extends ft{}class js extends ft{constructor($,U){super($,U),this.decoder=new kt({})}}class As extends ft{}class Ns extends ft{}class Nn{static TOKENIZER_CLASS_MAPPING={T5Tokenizer:et,DistilBertTokenizer:G,CamembertTokenizer:Z,DebertaTokenizer:yt,DebertaV2Tokenizer:Es,BertTokenizer:qt,HerbertTokenizer:C,ConvBertTokenizer:q,RoFormerTokenizer:R,XLMTokenizer:ce,ElectraTokenizer:ye,MobileBertTokenizer:Ps,SqueezeBertTokenizer:Gr,AlbertTokenizer:Ts,GPT2Tokenizer:ut,BartTokenizer:He,MBartTokenizer:Mt,MBart50Tokenizer:qe,RobertaTokenizer:Pt,WhisperTokenizer:Rs,CodeGenTokenizer:ks,CLIPTokenizer:$s,SiglipTokenizer:Is,MarianTokenizer:as,BloomTokenizer:It,NllbTokenizer:Kr,M2M100Tokenizer:os,LlamaTokenizer:pr,CodeLlamaTokenizer:ir,XLMRobertaTokenizer:Tr,MPNetTokenizer:Cs,FalconTokenizer:Dr,GPTNeoXTokenizer:Ss,EsmTokenizer:Lr,Wav2Vec2CTCTokenizer:Nr,BlenderbotTokenizer:ze,BlenderbotSmallTokenizer:Ue,SpeechT5Tokenizer:nt,NougatTokenizer:Kt,VitsTokenizer:js,Qwen2Tokenizer:zr,GemmaTokenizer:ns,Grok1Tokenizer:wr,CohereTokenizer:As,MgpstrTokenizer:Ns,PreTrainedTokenizer:ft};static async from_pretrained($,{progress_callback:U=null,config:ee=null,cache_dir:se=null,local_files_only:Me=!1,revision:$e="main",legacy:Xe=null}={}){const[Je,Ye]=await d($,{progress_callback:U,config:ee,cache_dir:se,local_files_only:Me,revision:$e,legacy:Xe}),Ke=Ye.tokenizer_class?.replace(/Fast$/,"")??"PreTrainedTokenizer";let $t=this.TOKENIZER_CLASS_MAPPING[Ke];return $t||(console.warn(`Unknown tokenizer class "${Ke}", attempting to construct from base class.`),$t=ft),new $t(Je,Ye)}}}),"./src/utils/audio.js":((e,r,t)=>{t.r(r),t.d(r,{RawAudio:()=>W,hamming:()=>u,hanning:()=>d,mel_filter_bank:()=>k,read_audio:()=>c,spectrogram:()=>S,window_function:()=>O});var s=t("./src/utils/hub.js"),n=t("./src/utils/maths.js"),o=t("./src/utils/core.js"),i=t("./src/env.js"),a=t("./src/utils/tensor.js"),l=t("?7992");async function c(B,Y){if(typeof AudioContext>"u")throw Error("Unable to load audio from path/URL since `AudioContext` is not available in your environment. Instead, audio data should be passed directly to the pipeline/processor. For more information and some example code, see https://huggingface.co/docs/transformers.js/guides/node-audio-processing.");const X=await(await(0,s.getFile)(B)).arrayBuffer(),J=new AudioContext({sampleRate:Y});typeof Y>"u"&&console.warn(`No sampling rate provided, using default of ${J.sampleRate}Hz.`);const re=await J.decodeAudioData(X);let ne;if(re.numberOfChannels===2){const le=Math.sqrt(2),pe=re.getChannelData(0),oe=re.getChannelData(1);ne=new Float32Array(pe.length);for(let K=0;K2595*Math.log10(1+B/700),kaldi:B=>1127*Math.log(1+B/700),slaney:(B,Y=1e3,X=15,J=27/Math.log(6.4))=>B>=Y?X+Math.log(B/Y)*J:3*B/200};function _(B,Y="htk"){const X=f[Y];if(!X)throw new Error('mel_scale should be one of "htk", "slaney" or "kaldi".');return typeof B=="number"?X(B):B.map(J=>X(J))}const y={htk:B=>700*(10**(B/2595)-1),kaldi:B=>700*(Math.exp(B/1127)-1),slaney:(B,Y=1e3,X=15,J=Math.log(6.4)/27)=>B>=X?Y*Math.exp(J*(B-X)):200*B/3};function I(B,Y="htk"){const X=y[Y];if(!X)throw new Error('mel_scale should be one of "htk", "slaney" or "kaldi".');return typeof B=="number"?X(B):B.map(J=>X(J))}function w(B,Y){const X=Float64Array.from({length:Y.length-1},(le,pe)=>Y[pe+1]-Y[pe]),J=Array.from({length:B.length},()=>new Array(Y.length));for(let le=0;lenew Array(B.length));for(let le=0;leB+J*ne)}function k(B,Y,X,J,re,ne=null,le="htk",pe=!1){if(ne!==null&&ne!=="slaney")throw new Error('norm must be one of null or "slaney"');if(B<2)throw new Error(`Require num_frequency_bins: ${B} >= 2`);if(X>J)throw new Error(`Require min_frequency: ${X} <= max_frequency: ${J}`);const oe=_(X,le),K=_(J,le),j=v(oe,K,Y+2);let D=I(j,le),te;if(pe){const Ae=re/((B-1)*2);te=_(Float64Array.from({length:B},(ke,Ve)=>Ve*Ae),le),D=j}else te=v(0,Math.floor(re/2),B);const he=w(te,D);if(ne!==null&&ne==="slaney")for(let Ae=0;Aere)throw Error(`frame_length (${X}) may not be larger than fft_length (${re})`);if(xe!==X)throw new Error(`Length of the window (${xe}) must equal frame_length (${X})`);if(J<=0)throw new Error("hop_length must be greater than zero");if(ne===null&&D!==null)throw new Error("You have provided `mel_filters` but `power` is `None`. Mel spectrogram computation is not yet supported for complex-valued spectrogram. Specify `power` to fix this issue.");if(!j)throw new Error("`preemphasis_htk_flavor=false` is not currently supported.");if(le)switch(pe){case"reflect":{const je=Math.floor((re-1)/2)+1;B=T(B,je,je);break}case"constant":{const je=Math.floor(re/2),Oe=new B.constructor(B.length+2*je);Oe.set(B,je),B=Oe;break}default:throw new Error(`pad_mode="${pe}" not implemented yet.`)}let Ce=Math.floor(1+Math.floor((B.length-X)/J));Q!==null&&CeCe?de&&(fe=z):fe=De=z);const Pe=new n.FFT(re),We=new Float64Array(re),Fe=new Float64Array(Pe.outputBufferSize),tt=new Float32Array(ge*fe);for(let je=0;je=1;--ht)We[ht]-=K*We[ht-1];We[0]*=1-K}for(let ht=0;htMath.pow(pe,.85));break;default:throw new Error(`Unknown window type ${Y}.`)}if(X&&(le=le.subarray(0,B)),J===null)return le;if(B>J)throw new Error(`Length of the window (${B}) may not be larger than frame_length (${J})`);return le}function F(B,Y){let X=44;const J=new ArrayBuffer(X+B.length*4),re=new DataView(J);H(re,0,"RIFF"),re.setUint32(4,36+B.length*4,!0),H(re,8,"WAVE"),H(re,12,"fmt "),re.setUint32(16,16,!0),re.setUint16(20,3,!0),re.setUint16(22,1,!0),re.setUint32(24,Y,!0),re.setUint32(28,Y*4,!0),re.setUint16(32,4,!0),re.setUint16(34,32,!0),H(re,36,"data"),re.setUint32(40,B.length*4,!0);for(let ne=0;ne{let ne=await re.arrayBuffer();l.writeFileSync(J,Buffer.from(ne))};else throw new Error("Unable to save because filesystem is disabled in this environment.");await X(Y,this.toBlob())}}}),"./src/utils/constants.js":((e,r,t)=>{t.r(r),t.d(r,{CHAT_TEMPLATE_NAME:()=>l,CONFIG_NAME:()=>n,FEATURE_EXTRACTOR_NAME:()=>o,GENERATION_CONFIG_NAME:()=>c,GITHUB_ISSUE_URL:()=>s,IMAGE_PROCESSOR_NAME:()=>i,PROCESSOR_NAME:()=>a});const s="https://github.com/huggingface/transformers.js/issues/new/choose",n="config.json",o="preprocessor_config.json",i=o,a="processor_config.json",l="chat_template.jinja",c="generation_config.json"}),"./src/utils/core.js":((e,r,t)=>{t.r(r),t.d(r,{calculateDimensions:()=>c,calculateReflectOffset:()=>f,count:()=>w,dispatchCallback:()=>s,escapeRegExp:()=>o,isIntegralNumber:()=>a,isNullishDimension:()=>l,isTypedArray:()=>i,len:()=>I,mergeArrays:()=>d,pick:()=>y,pop:()=>p,product:()=>u,reverseDictionary:()=>n,saveBlob:()=>_});function s(v,k){v&&v(k)}function n(v){return Object.fromEntries(Object.entries(v).map(([k,T])=>[T,k]))}function o(v){return v.replace(/[.*+?^${}()|[\]\\]/g,"\\$&")}function i(v){return v?.prototype?.__proto__?.constructor?.name==="TypedArray"}function a(v){return Number.isInteger(v)||typeof v=="bigint"}function l(v){return v==null||v===-1}function c(v){const k=[];let T=v;for(;Array.isArray(T);)k.push(T.length),T=T[0];return k}function p(v,k,T=void 0){const b=v[k];if(b!==void 0)return delete v[k],b;if(T===void 0)throw Error(`Key ${k} does not exist in object.`);return T}function d(...v){return Array.prototype.concat.apply([],v)}function u(...v){return v.reduce((k,T)=>k.flatMap(b=>T.map(P=>[b,P])))}function f(v,k){return Math.abs((v+k)%(2*k)-k)}function _(v,k){const T=URL.createObjectURL(k),b=document.createElement("a");b.href=T,b.download=v,b.click(),b.remove(),URL.revokeObjectURL(T)}function y(v,k){return Object.assign({},...k.map(T=>{if(v[T]!==void 0)return{[T]:v[T]}}))}function I(v){let k=0;for(const T of v)++k;return k}function w(v,k){let T=0;for(const b of v)b===k&&++T;return T}}),"./src/utils/data-structures.js":((e,r,t)=>{t.r(r),t.d(r,{CharTrie:()=>n,DictionarySplitter:()=>l,LRUCache:()=>c,PriorityQueue:()=>s,TokenLattice:()=>i});class s{constructor(d=(f,_)=>f>_,u=1/0){this._heap=[],this._comparator=d,this._maxSize=u}get size(){return this._heap.length}isEmpty(){return this.size===0}peek(){return this._heap[0]}push(...d){return this.extend(d)}extend(d){for(const u of d)if(this.size0&&this._swap(0,u),this._heap.pop(),this._siftDown(),d}replace(d){const u=this.peek();return this._heap[0]=d,this._siftDown(),u}_parent(d){return(d+1>>>1)-1}_left(d){return(d<<1)+1}_right(d){return d+1<<1}_greater(d,u){return this._comparator(this._heap[d],this._heap[u])}_swap(d,u){const f=this._heap[d];this._heap[d]=this._heap[u],this._heap[u]=f}_siftUp(){this._siftUpFrom(this.size-1)}_siftUpFrom(d){for(;d>0&&this._greater(d,this._parent(d));)this._swap(d,this._parent(d)),d=this._parent(d)}_siftDown(){let d=0;for(;this._left(d)[]),this.endNodes=Array.from({length:this.len+1},()=>[]);const _=new a(this.bosTokenId,0,0,0,0),y=new a(this.eosTokenId,1,this.len,0,0);this.nodes.push(_.clone()),this.nodes.push(y.clone()),this.beginNodes[this.len].push(y),this.endNodes[0].push(_)}insert(d,u,f,_){const y=this.nodes.length,I=new a(_,y,d,u,f);this.beginNodes[d].push(I),this.endNodes[d+u].push(I),this.nodes.push(I)}viterbi(){const d=this.len;let u=0;for(;u<=d;){if(this.beginNodes[u].length==0)return[];for(let w of this.beginNodes[u]){w.prev=null;let v=0,k=null;for(let T of this.endNodes[u]){const b=T.backtraceScore+w.score;(k===null||b>v)&&(k=T.clone(),v=b)}if(k!==null)w.prev=k,w.backtraceScore=v;else return[]}++u}const f=[],y=this.beginNodes[d][0].prev;if(y===null)return[];let I=y.clone();for(;I.prev!==null;)f.push(I.clone()),I=I.clone().prev.clone();return f.reverse(),f}piece(d){return this.chars.slice(d.pos,d.pos+d.length).join("")}tokens(){return this.viterbi().map(u=>this.piece(u))}tokenIds(){return this.viterbi().map(u=>u.tokenId)}}class a{constructor(d,u,f,_,y){this.tokenId=d,this.nodeId=u,this.pos=f,this.length=_,this.score=y,this.prev=null,this.backtraceScore=0}clone(){const d=new a(this.tokenId,this.nodeId,this.pos,this.length,this.score);return d.prev=this.prev,d.backtraceScore=this.backtraceScore,d}}class l{constructor(d){this.trie=this._buildTrie(d)}_buildTrie(d){const u=Object.create(null);for(const f of d){let _=u;for(let y=0;y_&&u.push(d.slice(_,y)),u.push(w),y+=w.length,_=y):++y}return _this.capacity&&this.cache.delete(this.cache.keys().next().value)}clear(){this.cache.clear()}}}),"./src/utils/devices.js":((e,r,t)=>{t.r(r),t.d(r,{DEVICE_TYPES:()=>s});const s=Object.freeze({auto:"auto",gpu:"gpu",cpu:"cpu",wasm:"wasm",webgpu:"webgpu",cuda:"cuda",dml:"dml",webnn:"webnn","webnn-npu":"webnn-npu","webnn-gpu":"webnn-gpu","webnn-cpu":"webnn-cpu"})}),"./src/utils/dtypes.js":((e,r,t)=>{t.r(r),t.d(r,{DATA_TYPES:()=>i,DEFAULT_DEVICE_DTYPE_MAPPING:()=>a,DEFAULT_DTYPE_SUFFIX_MAPPING:()=>l,isWebGpuFp16Supported:()=>o});var s=t("./src/env.js"),n=t("./src/utils/devices.js");const o=(function(){let c;return async function(){if(c===void 0)if(!s.apis.IS_WEBGPU_AVAILABLE)c=!1;else try{c=(await navigator.gpu.requestAdapter()).features.has("shader-f16")}catch{c=!1}return c}})(),i=Object.freeze({auto:"auto",fp32:"fp32",fp16:"fp16",q8:"q8",int8:"int8",uint8:"uint8",q4:"q4",bnb4:"bnb4",q4f16:"q4f16"}),a=Object.freeze({[n.DEVICE_TYPES.wasm]:i.q8}),l=Object.freeze({[i.fp32]:"",[i.fp16]:"_fp16",[i.int8]:"_int8",[i.uint8]:"_uint8",[i.q8]:"_quantized",[i.q4]:"_q4",[i.q4f16]:"_q4f16",[i.bnb4]:"_bnb4"})}),"./src/utils/generic.js":((e,r,t)=>{t.r(r),t.d(r,{Callable:()=>s});const s=class{constructor(){let n=function(...o){return n._call(...o)};return Object.setPrototypeOf(n,new.target.prototype)}_call(...n){throw Error("Must implement _call method in subclass")}}}),"./src/utils/hub.js":((e,r,t)=>{t.r(r),t.d(r,{MAX_EXTERNAL_DATA_CHUNKS:()=>a,getFile:()=>f,getModelFile:()=>v,getModelJSON:()=>T,getModelText:()=>k});var s=t("?7992"),n=t("?5af5"),o=t("./src/env.js"),i=t("./src/utils/core.js");const a=100,l={txt:"text/plain",html:"text/html",css:"text/css",js:"text/javascript",json:"application/json",png:"image/png",jpg:"image/jpeg",jpeg:"image/jpeg",gif:"image/gif"};class c{constructor(S){if(this.filePath=S,this.headers=new Headers,this.exists=s.existsSync(S),this.exists){this.status=200,this.statusText="OK";let O=s.statSync(S);this.headers.set("content-length",O.size.toString()),this.updateContentType();const F=s.createReadStream(S);this.body=new ReadableStream({start(H){F.on("data",W=>H.enqueue(W)),F.on("end",()=>H.close()),F.on("error",W=>H.error(W))},cancel(){F.destroy()}})}else this.status=404,this.statusText="Not Found",this.body=null}updateContentType(){const S=this.filePath.toString().split(".").pop().toLowerCase();this.headers.set("content-type",l[S]??"application/octet-stream")}clone(){let S=new c(this.filePath);return S.exists=this.exists,S.status=this.status,S.statusText=this.statusText,S.headers=new Headers(this.headers),S}async arrayBuffer(){return(await s.promises.readFile(this.filePath)).buffer}async blob(){const S=await s.promises.readFile(this.filePath);return new Blob([S],{type:this.headers.get("content-type")})}async text(){return await s.promises.readFile(this.filePath,"utf8")}async json(){return JSON.parse(await this.text())}}function p(x,S=null,O=null){let F;try{F=new URL(x)}catch{return!1}return!(S&&!S.includes(F.protocol)||O&&!O.includes(F.hostname))}const d=/^(\b[\w\-.]+\b\/)?\b[\w\-.]{1,96}\b$/;function u(x){return!(!d.test(x)||x.includes("..")||x.includes("--")||x.endsWith(".git")||x.endsWith(".ipynb"))}async function f(x){if(o.env.useFS&&!p(x,["http:","https:","blob:"]))return new c(x instanceof URL?x.protocol==="file:"?x.pathname:x.toString():x);if(typeof process<"u"&&process?.release?.name==="node"){const S=!!pc?.TESTING_REMOTELY,O=o.env.version,F=new Headers;if(F.set("User-Agent",`transformers.js/${O}; is_ci/${S};`),p(x,["http:","https:"],["huggingface.co","hf.co"])){const W=pc?.HF_TOKEN??pc?.HF_ACCESS_TOKEN;W&&F.set("Authorization",`Bearer ${W}`)}return fetch(x,{headers:F})}else return fetch(x)}const _={400:"Bad request error occurred while trying to load file",401:"Unauthorized access to file",403:"Forbidden access to file",404:"Could not locate file",408:"Request timeout error occurred while trying to load file",500:"Internal server error error occurred while trying to load file",502:"Bad gateway error occurred while trying to load file",503:"Service unavailable error occurred while trying to load file",504:"Gateway timeout error occurred while trying to load file"};function y(x,S,O){if(!O)return null;const F=_[x]??`Error (${x}) occurred while trying to load file`;throw Error(`${F}: "${S}".`)}class I{constructor(S){this.path=S}async match(S){let O=n.join(this.path,S),F=new c(O);if(F.exists)return F}async put(S,O,F=void 0){let H=n.join(this.path,S);try{const W=O.headers.get("Content-Length"),B=parseInt(W??"0");let Y=0;await s.promises.mkdir(n.dirname(H),{recursive:!0});const X=s.createWriteStream(H),J=O.body.getReader();for(;;){const{done:re,value:ne}=await J.read();if(re)break;await new Promise((pe,oe)=>{X.write(ne,K=>{if(K){oe(K);return}pe()})}),Y+=ne.length;const le=B?Y/B*100:0;F?.({progress:le,loaded:Y,total:B})}X.close()}catch(W){try{await s.promises.unlink(H)}catch{}throw W}}}async function w(x,...S){for(let O of S)try{let F=await x.match(O);if(F)return F}catch{continue}}async function v(x,S,O=!0,F={},H=!1){if(!o.env.allowLocalModels){if(F.local_files_only)throw Error("Invalid configuration detected: local models are disabled (`env.allowLocalModels=false`) but you have requested to only use local models (`local_files_only=true`).");if(!o.env.allowRemoteModels)throw Error("Invalid configuration detected: both local and remote models are disabled. Fix by setting `env.allowLocalModels` or `env.allowRemoteModels` to `true`.")}(0,i.dispatchCallback)(F.progress_callback,{status:"initiate",name:x,file:S});let W;if(!W&&o.env.useCustomCache){if(!o.env.customCache)throw Error("`env.useCustomCache=true`, but `env.customCache` is not defined.");if(!o.env.customCache.match||!o.env.customCache.put)throw new Error("`env.customCache` must be an object which implements the `match` and `put` functions of the Web Cache API. For more information, see https://developer.mozilla.org/en-US/docs/Web/API/Cache");W=o.env.customCache}if(!W&&o.env.useBrowserCache){if(typeof caches>"u")throw Error("Browser cache is not available in this environment.");try{W=await caches.open("transformers-cache")}catch(te){console.warn("An error occurred while opening the browser cache:",te)}}if(!W&&o.env.useFSCache){if(!o.apis.IS_FS_AVAILABLE)throw Error("File System Cache is not available in this environment.");W=new I(F.cache_dir??o.env.cacheDir)}const B=F.revision??"main",Y=P(x,S),X=u(x),J=X?P(o.env.localModelPath,Y):Y,re=P(o.env.remoteHost,o.env.remotePathTemplate.replaceAll("{model}",x).replaceAll("{revision}",encodeURIComponent(B)),S);let ne;const le=W instanceof I?B==="main"?Y:P(x,B,S):re;let pe=!1,oe;W&&(oe=await w(W,J,le));const K=oe!==void 0;if(oe===void 0){if(o.env.allowLocalModels)if(p(Y,["http:","https:"])){if(F.local_files_only)throw new Error(`\`local_files_only=true\`, but attempted to load a remote file from: ${Y}.`);if(!o.env.allowRemoteModels)throw new Error(`\`env.allowRemoteModels=false\`, but attempted to load a remote file from: ${Y}.`)}else try{oe=await f(J),ne=J}catch(he){console.warn(`Unable to load from local path "${J}": "${he}"`)}if(oe===void 0||oe.status===404){if(F.local_files_only||!o.env.allowRemoteModels){if(O)throw Error(`\`local_files_only=true\` or \`env.allowRemoteModels=false\` and file was not found locally at "${J}".`);return null}if(!X)throw Error(`Local file missing at "${J}" and download aborted due to invalid model ID "${x}".`);if(oe=await f(re),oe.status!==200)return y(oe.status,re,O);ne=le}pe=W&&typeof Response<"u"&&oe instanceof Response&&oe.status===200}(0,i.dispatchCallback)(F.progress_callback,{status:"download",name:x,file:S});let j;if(!(o.apis.IS_NODE_ENV&&H)){let te;F.progress_callback?K&&typeof navigator<"u"&&/firefox/i.test(navigator.userAgent)?(te=new Uint8Array(await oe.arrayBuffer()),(0,i.dispatchCallback)(F.progress_callback,{status:"progress",name:x,file:S,progress:100,loaded:te.length,total:te.length})):te=await b(oe,he=>{(0,i.dispatchCallback)(F.progress_callback,{status:"progress",name:x,file:S,...he})}):te=new Uint8Array(await oe.arrayBuffer()),j=te}if(pe&&ne&&await W.match(ne)===void 0)if(j)await W.put(ne,new Response(j,{headers:oe.headers})).catch(te=>{console.warn(`Unable to add response to browser cache: ${te}.`)});else{const te=F.progress_callback?he=>(0,i.dispatchCallback)(F.progress_callback,{status:"progress",name:x,file:S,...he}):void 0;await W.put(ne,oe,te)}if((0,i.dispatchCallback)(F.progress_callback,{status:"done",name:x,file:S}),j){if(!o.apis.IS_NODE_ENV&&H)throw new Error("Cannot return path in a browser environment.");return j}if(oe instanceof c)return oe.filePath;const D=await W?.match(ne);if(D instanceof c)return D.filePath;if(D instanceof Response)return new Uint8Array(await D.arrayBuffer());if(typeof D=="string")return D;throw new Error("Unable to get model file path or buffer.")}async function k(x,S,O=!0,F={}){const H=await v(x,S,O,F,!1);return H===null?null:new TextDecoder("utf-8").decode(H)}async function T(x,S,O=!0,F={}){const H=await k(x,S,O,F);return H===null?{}:JSON.parse(H)}async function b(x,S){const O=x.headers.get("Content-Length");O===null&&console.warn("Unable to determine content-length from response headers. Will expand buffer when needed.");let F=parseInt(O??"0"),H=new Uint8Array(F),W=0;const B=x.body.getReader();async function Y(){const{done:X,value:J}=await B.read();if(X)return;const re=W+J.length;if(re>F){F=re;const le=new Uint8Array(F);le.set(H),H=le}H.set(J,W),W=re;const ne=W/F*100;return S({progress:ne,loaded:W,total:F}),Y()}return await Y(),H}function P(...x){return x=x.map((S,O)=>(O&&(S=S.replace(new RegExp("^/"),"")),O!==x.length-1&&(S=S.replace(new RegExp("/$"),"")),S)),x.join("/")}}),"./src/utils/image.js":((e,r,t)=>{t.r(r),t.d(r,{RawImage:()=>_,load_image:()=>y});var s=t("./src/utils/core.js"),n=t("./src/utils/hub.js"),o=t("./src/env.js"),i=t("./src/utils/tensor.js"),a=t("?2b25");let l,c,p;const d=o.apis.IS_BROWSER_ENV||o.apis.IS_WEBWORKER_ENV;if(d)l=(I,w)=>{if(!self.OffscreenCanvas)throw new Error("OffscreenCanvas not supported by this browser.");return new self.OffscreenCanvas(I,w)},p=self.createImageBitmap,c=self.ImageData;else if(a)p=async I=>{const v=(await I.metadata()).channels,{data:k,info:T}=await I.rotate().raw().toBuffer({resolveWithObject:!0}),b=new _(new Uint8ClampedArray(k),T.width,T.height,T.channels);return v!==void 0&&v!==T.channels&&b.convert(v),b};else throw new Error("Unable to load image processing library.");const u={0:"nearest",1:"lanczos",2:"bilinear",3:"bicubic",4:"box",5:"hamming"},f=new Map([["png","image/png"],["jpg","image/jpeg"],["jpeg","image/jpeg"],["gif","image/gif"]]);class _{constructor(w,v,k,T){this.data=w,this.width=v,this.height=k,this.channels=T}get size(){return[this.width,this.height]}static async read(w){if(w instanceof _)return w;if(typeof w=="string"||w instanceof URL)return await this.fromURL(w);if(w instanceof Blob)return await this.fromBlob(w);if(typeof HTMLCanvasElement<"u"&&w instanceof HTMLCanvasElement||typeof OffscreenCanvas<"u"&&w instanceof OffscreenCanvas)return this.fromCanvas(w);throw new Error(`Unsupported input type: ${typeof w}`)}static fromCanvas(w){if(!d)throw new Error("fromCanvas() is only supported in browser environments.");const k=w.getContext("2d").getImageData(0,0,w.width,w.height).data;return new _(k,w.width,w.height,4)}static async fromURL(w){const v=await(0,n.getFile)(w);if(v.status!==200)throw new Error(`Unable to read image from "${w}" (${v.status} ${v.statusText})`);const k=await v.blob();return this.fromBlob(k)}static async fromBlob(w){if(d){const v=await p(w),k=l(v.width,v.height).getContext("2d");return k.drawImage(v,0,0),new this(k.getImageData(0,0,v.width,v.height).data,v.width,v.height,4)}else{const v=a(await w.arrayBuffer());return await p(v)}}static fromTensor(w,v="CHW"){if(w.dims.length!==3)throw new Error(`Tensor should have 3 dimensions, but has ${w.dims.length} dimensions.`);if(v==="CHW")w=w.transpose(1,2,0);else if(v!=="HWC")throw new Error(`Unsupported channel format: ${v}`);if(!(w.data instanceof Uint8ClampedArray||w.data instanceof Uint8Array))throw new Error(`Unsupported tensor type: ${w.type}`);switch(w.dims[2]){case 1:case 2:case 3:case 4:return new _(w.data,w.dims[1],w.dims[0],w.dims[2]);default:throw new Error(`Unsupported number of channels: ${w.dims[2]}`)}}grayscale(){if(this.channels===1)return this;const w=new Uint8ClampedArray(this.width*this.height*1);switch(this.channels){case 3:case 4:for(let v=0,k=0;v=0?S=k:F=-k,T>=0?O=T:H=-T,x.drawImage(P,S,O,w,v,F,H,w,v),new _(x.getImageData(0,0,w,v).data,w,v,4).convert(b)}else{let b=this.toSharp();if(k>=0&&T>=0)b=b.extract({left:Math.floor(k),top:Math.floor(T),width:w,height:v});else if(k<=0&&T<=0){const P=Math.floor(-T),x=Math.floor(-k);b=b.extend({top:P,left:x,right:w-this.width-x,bottom:v-this.height-P})}else{let P=[0,0],x=0;T<0?(P[0]=Math.floor(-T),P[1]=v-this.height-P[0]):x=Math.floor(T);let S=[0,0],O=0;k<0?(S[0]=Math.floor(-k),S[1]=w-this.width-S[0]):O=Math.floor(k),b=b.extend({top:P[0],bottom:P[1],left:S[0],right:S[1]}).extract({left:O,top:x,width:w,height:v})}return await p(b)}}async toBlob(w="image/png",v=1){if(!d)throw new Error("toBlob() is only supported in browser environments.");return await this.toCanvas().convertToBlob({type:w,quality:v})}toTensor(w="CHW"){let v=new i.Tensor("uint8",new Uint8Array(this.data),[this.height,this.width,this.channels]);if(w!=="HWC")if(w==="CHW")v=v.permute(2,0,1);else throw new Error(`Unsupported channel format: ${w}`);return v}toCanvas(){if(!d)throw new Error("toCanvas() is only supported in browser environments.");const w=this.clone().rgba(),v=l(w.width,w.height),k=new c(w.data,w.width,w.height);return v.getContext("2d").putImageData(k,0,0),v}split(){const{data:w,width:v,height:k,channels:T}=this,b=w.constructor,P=w.length/T,x=Array.from({length:T},()=>new b(P));for(let S=0;Snew _(S,v,k,1))}_update(w,v,k,T=null){return this.data=w,this.width=v,this.height=k,T!==null&&(this.channels=T),this}clone(){return new _(this.data.slice(),this.width,this.height,this.channels)}convert(w){if(this.channels===w)return this;switch(w){case 1:this.grayscale();break;case 3:this.rgb();break;case 4:this.rgba();break;default:throw new Error(`Conversion failed due to unsupported number of channels: ${this.channels}`)}return this}async save(w){if(d){if(o.apis.IS_WEBWORKER_ENV)throw new Error("Unable to save an image from a Web Worker.");const v=w.split(".").pop().toLowerCase(),k=f.get(v)??"image/png",T=await this.toBlob(k);(0,s.saveBlob)(w,T)}else{if(o.apis.IS_FS_AVAILABLE)return await this.toSharp().toFile(w);throw new Error("Unable to save the image because filesystem is disabled in this environment.")}}toSharp(){if(d)throw new Error("toSharp() is only supported in server-side environments.");return a(this.data,{raw:{width:this.width,height:this.height,channels:this.channels}})}}const y=_.read.bind(_)}),"./src/utils/maths.js":((e,r,t)=>{t.r(r),t.d(r,{FFT:()=>y,bankers_round:()=>v,cos_sim:()=>l,dot:()=>a,dynamic_time_warping:()=>k,interpolate_data:()=>s,log_softmax:()=>i,magnitude:()=>c,max:()=>d,medianFilter:()=>I,min:()=>p,permute_data:()=>n,round:()=>w,softmax:()=>o});function s(T,[b,P,x],[S,O],F="bilinear",H=!1){const W=O/x,B=S/P,Y=new T.constructor(S*O*b),X=P*x,J=S*O;for(let re=0;re=0;--H)S[H]=W,x[H]=b[P[H]],W*=x[H];const O=P.map((H,W)=>S[P.indexOf(W)]),F=new T.constructor(T.length);for(let H=0;H=0;--B)W+=Y%b[B]*O[B],Y=Math.floor(Y/b[B]);F[W]=T[H]}return[F,x]}function o(T){const b=d(T)[0],P=T.map(O=>Math.exp(O-b)),x=P.reduce((O,F)=>O+F,0);return P.map(O=>O/x)}function i(T){const b=d(T)[0];let P=0;for(let O=0;OO-b-x)}function a(T,b){let P=0;for(let x=0;xb+P*P,0))}function p(T){if(T.length===0)throw Error("Array must not be empty");let b=T[0],P=0;for(let x=1;xb&&(b=T[x],P=x);return[b,P]}function u(T){return T>0&&(T&T-1)===0}class f{constructor(b){if(this.size=b|0,this.size<=1||!u(this.size))throw new Error("FFT size must be a power of two larger than 1");this._csize=b<<1,this.table=new Float64Array(this.size*2);for(let x=0;xx;x<<=1)++P;this._width=P%2===0?P-1:P,this._bitrev=new Int32Array(1<>>S&3)<>>1);for(let S=0;S>>1]=b[S];return x}toComplexArray(b,P){const x=P||this.createComplexArray();for(let S=0;S>>1],x[S+1]=0;return x}transform(b,P){if(b===P)throw new Error("Input and output buffers must be different");this._transform4(b,P,1)}realTransform(b,P){if(b===P)throw new Error("Input and output buffers must be different");this._realTransform4(b,P,1)}inverseTransform(b,P){if(b===P)throw new Error("Input and output buffers must be different");this._transform4(b,P,-1);for(let x=0;x>=2;F>=2;F>>=2){H=S/F<<1;const J=H>>>2;for(W=0;W>>1,F>>>1)}else for(W=0,B=0;W>>1,F>>>1,x)}const X=this.table;for(F>>=2;F>=2;F>>=2){H=S/F<<1;const re=H>>>1,ne=re>>>1,le=ne>>>1;for(W=0;W>>1;for(let re=2;re>1;++Y){const X=(Y+1-b)**2/2,J=Math.sqrt(W**2+B**2)**X,re=X*Math.atan2(B,W),ne=2*Y;O[ne]=J*Math.cos(re),O[ne+1]=J*Math.sin(re),F[ne]=O[ne],F[ne+1]=-O[ne+1]}this._slicedChirpBuffer=O.subarray(P,x),this._f=new f(S>>1),this._f.transform(this._chirpBuffer,F)}_transform(b,P,x){const S=this._buffer1,O=this._buffer2,F=this._outBuffer1,H=this._outBuffer2,W=this._chirpBuffer,B=this._slicedChirpBuffer,Y=this._a;if(x)for(let X=0;X>1,ne=P[re];S[X]=ne*B[X],S[J]=ne*B[J]}else for(let X=0;X=T.length&&(W=2*(T.length-1)-W),x[F++]=T[W]}x.sort(),P[O]=x[S]}return P}function w(T,b){const P=Math.pow(10,b);return Math.round(T*P)/P}function v(T){const b=Math.round(T);return Math.abs(T)%1===.5?b%2===0?b:b-1:b}function k(T){const b=T.length,P=T[0].length,x=[b+1,P+1],S=Array.from({length:x[0]},()=>Array(x[1]).fill(1/0));S[0][0]=0;const O=Array.from({length:x[0]},()=>Array(x[1]).fill(-1));for(let Y=1;Y0||H>0;)switch(W.push(F-1),B.push(H-1),O[F][H]){case 0:--F,--H;break;case 1:--F;break;case 2:--H;break;default:throw new Error(`Internal error in dynamic time warping. Unexpected trace[${F}, ${H}]. Please file a bug report.`)}return W.reverse(),B.reverse(),[W,B]}}),"./src/utils/tensor.js":((e,r,t)=>{t.r(r),t.d(r,{DataTypeMap:()=>i,Tensor:()=>a,cat:()=>P,full:()=>B,full_like:()=>Y,interpolate:()=>p,interpolate_4d:()=>d,layer_norm:()=>v,matmul:()=>u,mean:()=>F,mean_pooling:()=>w,ones:()=>X,ones_like:()=>J,permute:()=>c,quantize_embeddings:()=>oe,rand:()=>le,randn:()=>pe,rfft:()=>f,slice:()=>I,stack:()=>x,std_mean:()=>O,topk:()=>_,zeros:()=>re,zeros_like:()=>ne});var s=t("./src/utils/maths.js"),n=t("./src/backends/onnx.js"),o=t("./src/ops/registry.js");const i=Object.freeze({float32:Float32Array,float16:typeof Float16Array<"u"?Float16Array:Uint16Array,float64:Float64Array,string:Array,int8:Int8Array,uint8:Uint8Array,int16:Int16Array,uint16:Uint16Array,int32:Int32Array,uint32:Uint32Array,int64:BigInt64Array,uint64:BigUint64Array,bool:Uint8Array,uint4:Uint8Array,int4:Int8Array});class a{get dims(){return this.ort_tensor.dims}set dims(j){this.ort_tensor.dims=j}get type(){return this.ort_tensor.type}get data(){return this.ort_tensor.data}get size(){return this.ort_tensor.size}get location(){return this.ort_tensor.location}ort_tensor;constructor(...j){return(0,n.isONNXTensor)(j[0])?this.ort_tensor=j[0]:this.ort_tensor=new n.Tensor(j[0],j[1],j[2]),new Proxy(this,{get:(D,te)=>{if(typeof te=="string"){let he=Number(te);if(Number.isInteger(he))return D._getitem(he)}return D[te]},set:(D,te,he)=>D[te]=he})}dispose(){this.ort_tensor.dispose()}*[Symbol.iterator](){const[j,...D]=this.dims;if(D.length>0){const te=D.reduce((he,Ae)=>he*Ae);for(let he=0;he0){const he=te.reduce((Ae,ke)=>Ae*ke);return this._subarray(j,he,te)}else return new a(this.type,[this.data[j]],te)}indexOf(j){const D=this.data;for(let te=0;teve)throw new Error(`Invalid slice: ${de}`);const xe=[Math.max(be,0),Math.min(ve,this.dims[z])];te.push(xe),D.push(xe[1]-xe[0])}else throw new Error(`Invalid slice: ${de}`)}const he=te.map(([z,de])=>de-z),Ae=he.reduce((z,de)=>z*de),ke=this.data,Ve=new ke.constructor(Ae),Te=this.stride();let Q=!0;for(let z=1;z=0;--be){const xe=he[be];de+=(ve%xe+te[be][0])*Te[be],ve=Math.floor(ve/xe)}Ve[z]=ke[de]}return new a(this.type,Ve,D)}permute(...j){return c(this,j)}transpose(...j){return this.permute(...j)}sum(j=null,D=!1){return this.norm(1,j,D)}norm(j="fro",D=null,te=!1){if(j==="fro")j=2;else if(typeof j=="string")throw Error(`Unsupported norm: ${j}`);const he=this.data,Ae=(Q,z)=>Q+z**j;if(D===null){const Q=he.reduce(Ae,0)**(1/j);return new a(this.type,[Q],[])}const[ke,Ve,Te]=S(Ae,this,D,te);if(j!==1)for(let Q=0;Q=0;--Te){const de=this.dims[Te];if(Te!==D){const be=Q%de;Ve+=be*z,z*=this.dims[Te]}Q=Math.floor(Q/de)}he[ke]/=Ae[Ve]}return this}normalize(j=2,D=1){return this.clone().normalize_(j,D)}stride(){return H(this.dims)}squeeze(j=null){return new a(this.type,this.data,k(this.dims,j))}squeeze_(j=null){return this.dims=k(this.dims,j),this}unsqueeze(j=null){return new a(this.type,this.data,T(this.dims,j))}unsqueeze_(j=null){return this.dims=T(this.dims,j),this}flatten_(j=0,D=-1){D=(D+this.dims.length)%this.dims.length;let te=this.dims.slice(0,j),he=this.dims.slice(j,D+1),Ae=this.dims.slice(D+1);return this.dims=[...te,he.reduce((ke,Ve)=>ke*Ve,1),...Ae],this}flatten(j=0,D=-1){return this.clone().flatten_(j,D)}view(...j){let D=-1;for(let he=0;heVe!==D?Ae*ke:Ae,1);j[D]=te.length/he}return new a(this.type,te,j)}neg_(){const j=this.data;for(let D=0;Dj?1:0;return new a("bool",D,this.dims)}lt(j){const D=new Uint8Array(this.data.length),te=this.data;for(let he=0;heMath.min(ke,Ve),this,j,D,1/0);return new a(te,he,Ae)}max(j=null,D=!1){if(j===null){const ke=(0,s.max)(this.data)[0];return new a(this.type,[ke],[])}const[te,he,Ae]=S((ke,Ve)=>Math.max(ke,Ve),this,j,D,-1/0);return new a(te,he,Ae)}argmin(j=null,D=!1){if(j!==null)throw new Error("`dim !== null` not yet implemented.");const te=(0,s.min)(this.data)[1];return new a("int64",[BigInt(te)],[])}argmax(j=null,D=!1){if(j!==null)throw new Error("`dim !== null` not yet implemented.");const te=(0,s.max)(this.data)[1];return new a("int64",[BigInt(te)],[])}to(j){if(this.type===j)return this;if(!i.hasOwnProperty(j))throw new Error(`Unsupported type: ${j}`);let D;const te=["int64","uint64"].includes(this.type),he=["int64","uint64"].includes(j);return te&&!he?D=Number:!te&&he&&(["float16","float32","float64"].includes(this.type)?D=Ae=>BigInt(Math.floor(Ae)):D=BigInt),new a(j,i[j].from(this.data,D),this.dims)}}function l(K,j){const D=K.length,te=j.reduce((Ae,ke)=>Ae*ke);if(D!==te)throw Error(`cannot reshape array of size ${D} into shape (${j})`);let he=K;for(let Ae=j.length-1;Ae>=0;Ae--)he=he.reduce((ke,Ve)=>{let Te=ke[ke.length-1];return Te.lengthnew a("int64",K,[K.length]);async function I(K,j,D,te,he){return await(await o.TensorOpRegistry.slice)({x:K,s:y(j),e:y(D),a:y(te),t:y(he??new Array(te.length).fill(1))})}function w(K,j){const D=K.data,te=j.data,he=[K.dims[0],K.dims[2]],Ae=new D.constructor(he[0]*he[1]),[ke,Ve,Te]=K.dims;let Q=0;for(let z=0;zD!==1):typeof j=="number"?K[j]===1&&K.splice(j,1):Array.isArray(j)&&(K=K.filter((D,te)=>D!==1||!j.includes(te))),K}function T(K,j){return j=b(j,K.length+1),K=K.slice(),K.splice(j,0,1),K}function b(K,j,D=null,te=!0){if(K<-j||K>=j){if(te)throw new Error(`IndexError: index ${K} is out of bounds for dimension${D===null?"":" "+D} with size ${j}`);return K<-j?0:j}return K<0&&(K=(K%j+j)%j),K}function P(K,j=0){j=b(j,K[0].dims.length);const D=K[0].dims.slice();D[j]=K.reduce((ke,Ve)=>ke+Ve.dims[j],0);const te=D.reduce((ke,Ve)=>ke*Ve,1),he=new K[0].data.constructor(te),Ae=K[0].type;if(j===0){let ke=0;for(const Ve of K){const Te=Ve.data;he.set(Te,ke),ke+=Te.length}}else{let ke=0;for(let Ve=0;Ve=0;--be){const Ce=Q[be];let ge=ve%Ce;be===j&&(ge+=ke),de+=ge*xe,xe*=D[be],ve=Math.floor(ve/Ce)}he[de]=Te[z]}ke+=Q[j]}}return new a(Ae,he,D)}function x(K,j=0){return P(K.map(D=>D.unsqueeze(j)),j)}function S(K,j,D=null,te=!1,he=null){const Ae=j.data,ke=j.dims;D=b(D,ke.length);const Ve=ke.slice();Ve[D]=1;const Te=new Ae.constructor(Ae.length/ke[D]);he!==null&&Te.fill(he);for(let Q=0;Q=0;--de){const xe=ke[de];if(de!==D){const Ce=be%xe;z+=Ce*ve,ve*=Ve[de]}be=Math.floor(be/xe)}Te[z]=K(Te[z],Ae[Q],Q,z)}return te||Ve.splice(D,1),[j.type,Te,Ve]}function O(K,j=null,D=1,te=!1){const he=K.data,Ae=K.dims;if(j===null){const ve=he.reduce((De,fe)=>De+fe,0)/he.length,xe=Math.sqrt(he.reduce((De,fe)=>De+(fe-ve)**2,0)/(he.length-D)),Ce=new a(K.type,[ve],[]);return[new a(K.type,[xe],[]),Ce]}j=b(j,Ae.length);const ke=F(K,j,te),Ve=ke.data,[Te,Q,z]=S((be,ve,xe,Ce)=>be+(ve-Ve[Ce])**2,K,j,te);for(let be=0;beQ+z,0);return new a(K.type,[Te/he.length],[])}j=b(j,te.length);const[Ae,ke,Ve]=S((Te,Q)=>Te+Q,K,j,D);if(te[j]!==1)for(let Te=0;Te=0;--D)j[D]=te,te*=K[D];return j}function W(K,j,D,te){const he=K.reduce((Ae,ke)=>Ae*ke,1);return new a(D,new te(he).fill(j),K)}function B(K,j){let D,te;if(typeof j=="number")D="float32",te=Float32Array;else if(typeof j=="bigint")D="int64",te=BigInt64Array;else if(typeof j=="boolean")D="bool",te=Uint8Array;else throw new Error(`Unsupported data type: ${typeof j}`);return W(K,j,D,te)}function Y(K,j){return B(K.dims,j)}function X(K){return W(K,1n,"int64",BigInt64Array)}function J(K){return X(K.dims)}function re(K){return W(K,0n,"int64",BigInt64Array)}function ne(K){return re(K.dims)}function le(K){const j=K.reduce((D,te)=>D*te,1);return new a("float32",Float32Array.from({length:j},()=>Math.random()),K)}function pe(K){const j=K.reduce((te,he)=>te*he,1);function D(){const te=1-Math.random(),he=1-Math.random();return Math.sqrt(-2*Math.log(te))*Math.cos(2*Math.PI*he)}return new a("float32",Float32Array.from({length:j},()=>D()),K)}function oe(K,j){if(K.dims.length!==2)throw new Error("The tensor must have 2 dimensions");if(K.dims.at(-1)%8!==0)throw new Error("The last dimension of the tensor must be a multiple of 8");if(!["binary","ubinary"].includes(j))throw new Error("The precision must be either 'binary' or 'ubinary'");const D=j==="binary",te=D?"int8":"uint8",he=D?Int8Array:Uint8Array,Ae=K.data,ke=new he(Ae.length/8);for(let Ve=0;Ve0?1:0,Q=Math.floor(Ve/8),z=Ve%8;ke[Q]|=Te<<7-z,D&&z===0&&(ke[Q]-=128)}return new a(te,ke,[K.dims[0],K.dims[1]/8])}}),"./src/utils/video.js":((e,r,t)=>{t.r(r),t.d(r,{RawVideo:()=>i,RawVideoFrame:()=>o,load_video:()=>a});var s=t("./src/utils/image.js"),n=t("./src/env.js");class o{constructor(c,p){this.image=c,this.timestamp=p}}class i{constructor(c,p){c.length>0&&c[0]instanceof s.RawImage&&(c=c.map((d,u)=>new o(d,(u+1)/(c.length+1)*p))),this.frames=c,this.duration=p}get width(){return this.frames[0].image.width}get height(){return this.frames[0].image.height}get fps(){return this.frames.length/this.duration}}async function a(l,{num_frames:c=null,fps:p=null}={}){if(!n.apis.IS_BROWSER_ENV)throw new Error("`load_video` is currently only supported in browser environments.");if(c==null&&p==null)throw new Error("Either num_frames or fps must be provided.");const d=[],u=document.createElement("video");if(u.crossOrigin="anonymous",u.muted=!0,typeof l=="string")u.src=l;else if(l instanceof Blob)u.src=URL.createObjectURL(l);else if(l instanceof HTMLVideoElement)u.src=l.src;else throw new Error("Invalid URL or video element provided.");if(await new Promise(k=>u.onloadedmetadata=k),u.seekable.start(0)===u.seekable.end(0)){const T=await(await fetch(u.src)).blob();u.src=URL.createObjectURL(T),await new Promise(b=>u.onloadedmetadata=b)}const f=u.duration;let _,y;c!=null?(_=c,y=c===1?0:f/(c-1)):(y=1/p,_=Math.floor(f/y));let I=[];for(let k=0;k<_;++k)I.push(c===1?f/2:k*y);const w=document.createElement("canvas");w.width=u.videoWidth,w.height=u.videoHeight;const v=w.getContext("2d",{willReadFrequently:!0});for(const k of I){u.currentTime=k,await new Promise(x=>{u.onseeked=x}),v.drawImage(u,0,0,w.width,w.height);const T=v.getImageData(0,0,w.width,w.height),b=new s.RawImage(T.data,w.width,w.height,4),P=new o(b,k);d.push(P)}return u.remove(),new i(d,f)}})},lM={};function jt(e){var r=lM[e];if(r!==void 0)return r.exports;var t=lM[e]={exports:{}};return mT[e](t,t.exports,jt),t.exports}(()=>{var e=Object.getPrototypeOf?t=>Object.getPrototypeOf(t):t=>t.__proto__,r;jt.t=function(t,s){if(s&1&&(t=this(t)),s&8||typeof t=="object"&&t&&(s&4&&t.__esModule||s&16&&typeof t.then=="function"))return t;var n=Object.create(null);jt.r(n);var o={};r=r||[null,e({}),e([]),e(e)];for(var i=s&2&&t;typeof i=="object"&&!~r.indexOf(i);i=e(i))Object.getOwnPropertyNames(i).forEach(a=>o[a]=()=>t[a]);return o.default=()=>t,jt.d(n,o),n}})(),jt.d=(e,r)=>{for(var t in r)jt.o(r,t)&&!jt.o(e,t)&&Object.defineProperty(e,t,{enumerable:!0,get:r[t]})},jt.o=(e,r)=>Object.prototype.hasOwnProperty.call(e,r),jt.r=e=>{typeof Symbol<"u"&&Symbol.toStringTag&&Object.defineProperty(e,Symbol.toStringTag,{value:"Module"}),Object.defineProperty(e,"__esModule",{value:!0})};var m={};(()=>{jt.r(m),jt.d(m,{ASTFeatureExtractor:()=>d.ASTFeatureExtractor,ASTForAudioClassification:()=>t.ASTForAudioClassification,ASTModel:()=>t.ASTModel,ASTPreTrainedModel:()=>t.ASTPreTrainedModel,AlbertForMaskedLM:()=>t.AlbertForMaskedLM,AlbertForQuestionAnswering:()=>t.AlbertForQuestionAnswering,AlbertForSequenceClassification:()=>t.AlbertForSequenceClassification,AlbertModel:()=>t.AlbertModel,AlbertPreTrainedModel:()=>t.AlbertPreTrainedModel,AlbertTokenizer:()=>s.AlbertTokenizer,ArceeForCausalLM:()=>t.ArceeForCausalLM,ArceeModel:()=>t.ArceeModel,ArceePreTrainedModel:()=>t.ArceePreTrainedModel,AudioClassificationPipeline:()=>r.AudioClassificationPipeline,AutoConfig:()=>n.AutoConfig,AutoFeatureExtractor:()=>u.AutoFeatureExtractor,AutoImageProcessor:()=>y.AutoImageProcessor,AutoModel:()=>t.AutoModel,AutoModelForAudioClassification:()=>t.AutoModelForAudioClassification,AutoModelForAudioFrameClassification:()=>t.AutoModelForAudioFrameClassification,AutoModelForAudioTextToText:()=>t.AutoModelForAudioTextToText,AutoModelForCTC:()=>t.AutoModelForCTC,AutoModelForCausalLM:()=>t.AutoModelForCausalLM,AutoModelForDepthEstimation:()=>t.AutoModelForDepthEstimation,AutoModelForDocumentQuestionAnswering:()=>t.AutoModelForDocumentQuestionAnswering,AutoModelForImageClassification:()=>t.AutoModelForImageClassification,AutoModelForImageFeatureExtraction:()=>t.AutoModelForImageFeatureExtraction,AutoModelForImageMatting:()=>t.AutoModelForImageMatting,AutoModelForImageSegmentation:()=>t.AutoModelForImageSegmentation,AutoModelForImageTextToText:()=>t.AutoModelForImageTextToText,AutoModelForImageToImage:()=>t.AutoModelForImageToImage,AutoModelForMaskGeneration:()=>t.AutoModelForMaskGeneration,AutoModelForMaskedLM:()=>t.AutoModelForMaskedLM,AutoModelForNormalEstimation:()=>t.AutoModelForNormalEstimation,AutoModelForObjectDetection:()=>t.AutoModelForObjectDetection,AutoModelForPoseEstimation:()=>t.AutoModelForPoseEstimation,AutoModelForQuestionAnswering:()=>t.AutoModelForQuestionAnswering,AutoModelForSemanticSegmentation:()=>t.AutoModelForSemanticSegmentation,AutoModelForSeq2SeqLM:()=>t.AutoModelForSeq2SeqLM,AutoModelForSequenceClassification:()=>t.AutoModelForSequenceClassification,AutoModelForSpeechSeq2Seq:()=>t.AutoModelForSpeechSeq2Seq,AutoModelForTextToSpectrogram:()=>t.AutoModelForTextToSpectrogram,AutoModelForTextToWaveform:()=>t.AutoModelForTextToWaveform,AutoModelForTokenClassification:()=>t.AutoModelForTokenClassification,AutoModelForUniversalSegmentation:()=>t.AutoModelForUniversalSegmentation,AutoModelForVision2Seq:()=>t.AutoModelForVision2Seq,AutoModelForXVector:()=>t.AutoModelForXVector,AutoModelForZeroShotObjectDetection:()=>t.AutoModelForZeroShotObjectDetection,AutoProcessor:()=>v.AutoProcessor,AutoTokenizer:()=>s.AutoTokenizer,AutomaticSpeechRecognitionPipeline:()=>r.AutomaticSpeechRecognitionPipeline,BackgroundRemovalPipeline:()=>r.BackgroundRemovalPipeline,BartForConditionalGeneration:()=>t.BartForConditionalGeneration,BartForSequenceClassification:()=>t.BartForSequenceClassification,BartModel:()=>t.BartModel,BartPretrainedModel:()=>t.BartPretrainedModel,BartTokenizer:()=>s.BartTokenizer,BaseModelOutput:()=>t.BaseModelOutput,BaseStreamer:()=>k.BaseStreamer,BeitFeatureExtractor:()=>_.BeitFeatureExtractor,BeitForImageClassification:()=>t.BeitForImageClassification,BeitModel:()=>t.BeitModel,BeitPreTrainedModel:()=>t.BeitPreTrainedModel,BertForMaskedLM:()=>t.BertForMaskedLM,BertForQuestionAnswering:()=>t.BertForQuestionAnswering,BertForSequenceClassification:()=>t.BertForSequenceClassification,BertForTokenClassification:()=>t.BertForTokenClassification,BertModel:()=>t.BertModel,BertPreTrainedModel:()=>t.BertPreTrainedModel,BertTokenizer:()=>s.BertTokenizer,BitImageProcessor:()=>_.BitImageProcessor,BlenderbotForConditionalGeneration:()=>t.BlenderbotForConditionalGeneration,BlenderbotModel:()=>t.BlenderbotModel,BlenderbotPreTrainedModel:()=>t.BlenderbotPreTrainedModel,BlenderbotSmallForConditionalGeneration:()=>t.BlenderbotSmallForConditionalGeneration,BlenderbotSmallModel:()=>t.BlenderbotSmallModel,BlenderbotSmallPreTrainedModel:()=>t.BlenderbotSmallPreTrainedModel,BlenderbotSmallTokenizer:()=>s.BlenderbotSmallTokenizer,BlenderbotTokenizer:()=>s.BlenderbotTokenizer,BloomForCausalLM:()=>t.BloomForCausalLM,BloomModel:()=>t.BloomModel,BloomPreTrainedModel:()=>t.BloomPreTrainedModel,BloomTokenizer:()=>s.BloomTokenizer,CLIPFeatureExtractor:()=>_.CLIPFeatureExtractor,CLIPImageProcessor:()=>_.CLIPImageProcessor,CLIPModel:()=>t.CLIPModel,CLIPPreTrainedModel:()=>t.CLIPPreTrainedModel,CLIPSegForImageSegmentation:()=>t.CLIPSegForImageSegmentation,CLIPSegModel:()=>t.CLIPSegModel,CLIPSegPreTrainedModel:()=>t.CLIPSegPreTrainedModel,CLIPTextModel:()=>t.CLIPTextModel,CLIPTextModelWithProjection:()=>t.CLIPTextModelWithProjection,CLIPTokenizer:()=>s.CLIPTokenizer,CLIPVisionModel:()=>t.CLIPVisionModel,CLIPVisionModelWithProjection:()=>t.CLIPVisionModelWithProjection,CamembertForMaskedLM:()=>t.CamembertForMaskedLM,CamembertForQuestionAnswering:()=>t.CamembertForQuestionAnswering,CamembertForSequenceClassification:()=>t.CamembertForSequenceClassification,CamembertForTokenClassification:()=>t.CamembertForTokenClassification,CamembertModel:()=>t.CamembertModel,CamembertPreTrainedModel:()=>t.CamembertPreTrainedModel,CamembertTokenizer:()=>s.CamembertTokenizer,CausalLMOutput:()=>t.CausalLMOutput,CausalLMOutputWithPast:()=>t.CausalLMOutputWithPast,ChineseCLIPFeatureExtractor:()=>_.ChineseCLIPFeatureExtractor,ChineseCLIPModel:()=>t.ChineseCLIPModel,ChineseCLIPPreTrainedModel:()=>t.ChineseCLIPPreTrainedModel,ClapAudioModelWithProjection:()=>t.ClapAudioModelWithProjection,ClapFeatureExtractor:()=>d.ClapFeatureExtractor,ClapModel:()=>t.ClapModel,ClapPreTrainedModel:()=>t.ClapPreTrainedModel,ClapTextModelWithProjection:()=>t.ClapTextModelWithProjection,ClassifierFreeGuidanceLogitsProcessor:()=>b.ClassifierFreeGuidanceLogitsProcessor,CodeGenForCausalLM:()=>t.CodeGenForCausalLM,CodeGenModel:()=>t.CodeGenModel,CodeGenPreTrainedModel:()=>t.CodeGenPreTrainedModel,CodeGenTokenizer:()=>s.CodeGenTokenizer,CodeLlamaTokenizer:()=>s.CodeLlamaTokenizer,CohereForCausalLM:()=>t.CohereForCausalLM,CohereModel:()=>t.CohereModel,CoherePreTrainedModel:()=>t.CoherePreTrainedModel,CohereTokenizer:()=>s.CohereTokenizer,ConvBertForMaskedLM:()=>t.ConvBertForMaskedLM,ConvBertForQuestionAnswering:()=>t.ConvBertForQuestionAnswering,ConvBertForSequenceClassification:()=>t.ConvBertForSequenceClassification,ConvBertForTokenClassification:()=>t.ConvBertForTokenClassification,ConvBertModel:()=>t.ConvBertModel,ConvBertPreTrainedModel:()=>t.ConvBertPreTrainedModel,ConvBertTokenizer:()=>s.ConvBertTokenizer,ConvNextFeatureExtractor:()=>_.ConvNextFeatureExtractor,ConvNextForImageClassification:()=>t.ConvNextForImageClassification,ConvNextImageProcessor:()=>_.ConvNextImageProcessor,ConvNextModel:()=>t.ConvNextModel,ConvNextPreTrainedModel:()=>t.ConvNextPreTrainedModel,ConvNextV2ForImageClassification:()=>t.ConvNextV2ForImageClassification,ConvNextV2Model:()=>t.ConvNextV2Model,ConvNextV2PreTrainedModel:()=>t.ConvNextV2PreTrainedModel,DFineForObjectDetection:()=>t.DFineForObjectDetection,DFineModel:()=>t.DFineModel,DFinePreTrainedModel:()=>t.DFinePreTrainedModel,DINOv3ConvNextModel:()=>t.DINOv3ConvNextModel,DINOv3ConvNextPreTrainedModel:()=>t.DINOv3ConvNextPreTrainedModel,DINOv3ViTImageProcessor:()=>_.DINOv3ViTImageProcessor,DINOv3ViTModel:()=>t.DINOv3ViTModel,DINOv3ViTPreTrainedModel:()=>t.DINOv3ViTPreTrainedModel,DPTFeatureExtractor:()=>_.DPTFeatureExtractor,DPTForDepthEstimation:()=>t.DPTForDepthEstimation,DPTImageProcessor:()=>_.DPTImageProcessor,DPTModel:()=>t.DPTModel,DPTPreTrainedModel:()=>t.DPTPreTrainedModel,DacDecoderModel:()=>t.DacDecoderModel,DacDecoderOutput:()=>t.DacDecoderOutput,DacEncoderModel:()=>t.DacEncoderModel,DacEncoderOutput:()=>t.DacEncoderOutput,DacFeatureExtractor:()=>d.DacFeatureExtractor,DacModel:()=>t.DacModel,DacPreTrainedModel:()=>t.DacPreTrainedModel,DataTypeMap:()=>l.DataTypeMap,DebertaForMaskedLM:()=>t.DebertaForMaskedLM,DebertaForQuestionAnswering:()=>t.DebertaForQuestionAnswering,DebertaForSequenceClassification:()=>t.DebertaForSequenceClassification,DebertaForTokenClassification:()=>t.DebertaForTokenClassification,DebertaModel:()=>t.DebertaModel,DebertaPreTrainedModel:()=>t.DebertaPreTrainedModel,DebertaTokenizer:()=>s.DebertaTokenizer,DebertaV2ForMaskedLM:()=>t.DebertaV2ForMaskedLM,DebertaV2ForQuestionAnswering:()=>t.DebertaV2ForQuestionAnswering,DebertaV2ForSequenceClassification:()=>t.DebertaV2ForSequenceClassification,DebertaV2ForTokenClassification:()=>t.DebertaV2ForTokenClassification,DebertaV2Model:()=>t.DebertaV2Model,DebertaV2PreTrainedModel:()=>t.DebertaV2PreTrainedModel,DebertaV2Tokenizer:()=>s.DebertaV2Tokenizer,DecisionTransformerModel:()=>t.DecisionTransformerModel,DecisionTransformerPreTrainedModel:()=>t.DecisionTransformerPreTrainedModel,DeiTFeatureExtractor:()=>_.DeiTFeatureExtractor,DeiTForImageClassification:()=>t.DeiTForImageClassification,DeiTImageProcessor:()=>_.DeiTImageProcessor,DeiTModel:()=>t.DeiTModel,DeiTPreTrainedModel:()=>t.DeiTPreTrainedModel,DepthAnythingForDepthEstimation:()=>t.DepthAnythingForDepthEstimation,DepthAnythingPreTrainedModel:()=>t.DepthAnythingPreTrainedModel,DepthEstimationPipeline:()=>r.DepthEstimationPipeline,DepthProForDepthEstimation:()=>t.DepthProForDepthEstimation,DepthProPreTrainedModel:()=>t.DepthProPreTrainedModel,DetrFeatureExtractor:()=>_.DetrFeatureExtractor,DetrForObjectDetection:()=>t.DetrForObjectDetection,DetrForSegmentation:()=>t.DetrForSegmentation,DetrImageProcessor:()=>_.DetrImageProcessor,DetrModel:()=>t.DetrModel,DetrObjectDetectionOutput:()=>t.DetrObjectDetectionOutput,DetrPreTrainedModel:()=>t.DetrPreTrainedModel,DetrSegmentationOutput:()=>t.DetrSegmentationOutput,Dinov2ForImageClassification:()=>t.Dinov2ForImageClassification,Dinov2Model:()=>t.Dinov2Model,Dinov2PreTrainedModel:()=>t.Dinov2PreTrainedModel,Dinov2WithRegistersForImageClassification:()=>t.Dinov2WithRegistersForImageClassification,Dinov2WithRegistersModel:()=>t.Dinov2WithRegistersModel,Dinov2WithRegistersPreTrainedModel:()=>t.Dinov2WithRegistersPreTrainedModel,DistilBertForMaskedLM:()=>t.DistilBertForMaskedLM,DistilBertForQuestionAnswering:()=>t.DistilBertForQuestionAnswering,DistilBertForSequenceClassification:()=>t.DistilBertForSequenceClassification,DistilBertForTokenClassification:()=>t.DistilBertForTokenClassification,DistilBertModel:()=>t.DistilBertModel,DistilBertPreTrainedModel:()=>t.DistilBertPreTrainedModel,DistilBertTokenizer:()=>s.DistilBertTokenizer,DocumentQuestionAnsweringPipeline:()=>r.DocumentQuestionAnsweringPipeline,DonutFeatureExtractor:()=>_.DonutFeatureExtractor,DonutImageProcessor:()=>_.DonutImageProcessor,DonutSwinModel:()=>t.DonutSwinModel,DonutSwinPreTrainedModel:()=>t.DonutSwinPreTrainedModel,EdgeTamModel:()=>t.EdgeTamModel,EfficientNetForImageClassification:()=>t.EfficientNetForImageClassification,EfficientNetImageProcessor:()=>_.EfficientNetImageProcessor,EfficientNetModel:()=>t.EfficientNetModel,EfficientNetPreTrainedModel:()=>t.EfficientNetPreTrainedModel,ElectraForMaskedLM:()=>t.ElectraForMaskedLM,ElectraForQuestionAnswering:()=>t.ElectraForQuestionAnswering,ElectraForSequenceClassification:()=>t.ElectraForSequenceClassification,ElectraForTokenClassification:()=>t.ElectraForTokenClassification,ElectraModel:()=>t.ElectraModel,ElectraPreTrainedModel:()=>t.ElectraPreTrainedModel,ElectraTokenizer:()=>s.ElectraTokenizer,EncodecFeatureExtractor:()=>d.EncodecFeatureExtractor,EosTokenCriteria:()=>T.EosTokenCriteria,Ernie4_5ForCausalLM:()=>t.Ernie4_5ForCausalLM,Ernie4_5Model:()=>t.Ernie4_5Model,Ernie4_5PreTrainedModel:()=>t.Ernie4_5PreTrainedModel,EsmForMaskedLM:()=>t.EsmForMaskedLM,EsmForSequenceClassification:()=>t.EsmForSequenceClassification,EsmForTokenClassification:()=>t.EsmForTokenClassification,EsmModel:()=>t.EsmModel,EsmPreTrainedModel:()=>t.EsmPreTrainedModel,EsmTokenizer:()=>s.EsmTokenizer,ExaoneForCausalLM:()=>t.ExaoneForCausalLM,ExaoneModel:()=>t.ExaoneModel,ExaonePreTrainedModel:()=>t.ExaonePreTrainedModel,FFT:()=>c.FFT,FalconForCausalLM:()=>t.FalconForCausalLM,FalconModel:()=>t.FalconModel,FalconPreTrainedModel:()=>t.FalconPreTrainedModel,FalconTokenizer:()=>s.FalconTokenizer,FastViTForImageClassification:()=>t.FastViTForImageClassification,FastViTModel:()=>t.FastViTModel,FastViTPreTrainedModel:()=>t.FastViTPreTrainedModel,FeatureExtractionPipeline:()=>r.FeatureExtractionPipeline,FeatureExtractor:()=>p.FeatureExtractor,FillMaskPipeline:()=>r.FillMaskPipeline,Florence2ForConditionalGeneration:()=>t.Florence2ForConditionalGeneration,Florence2PreTrainedModel:()=>t.Florence2PreTrainedModel,Florence2Processor:()=>w.Florence2Processor,ForcedBOSTokenLogitsProcessor:()=>b.ForcedBOSTokenLogitsProcessor,ForcedEOSTokenLogitsProcessor:()=>b.ForcedEOSTokenLogitsProcessor,GLPNFeatureExtractor:()=>_.GLPNFeatureExtractor,GLPNForDepthEstimation:()=>t.GLPNForDepthEstimation,GLPNModel:()=>t.GLPNModel,GLPNPreTrainedModel:()=>t.GLPNPreTrainedModel,GPT2LMHeadModel:()=>t.GPT2LMHeadModel,GPT2Model:()=>t.GPT2Model,GPT2PreTrainedModel:()=>t.GPT2PreTrainedModel,GPT2Tokenizer:()=>s.GPT2Tokenizer,GPTBigCodeForCausalLM:()=>t.GPTBigCodeForCausalLM,GPTBigCodeModel:()=>t.GPTBigCodeModel,GPTBigCodePreTrainedModel:()=>t.GPTBigCodePreTrainedModel,GPTJForCausalLM:()=>t.GPTJForCausalLM,GPTJModel:()=>t.GPTJModel,GPTJPreTrainedModel:()=>t.GPTJPreTrainedModel,GPTNeoForCausalLM:()=>t.GPTNeoForCausalLM,GPTNeoModel:()=>t.GPTNeoModel,GPTNeoPreTrainedModel:()=>t.GPTNeoPreTrainedModel,GPTNeoXForCausalLM:()=>t.GPTNeoXForCausalLM,GPTNeoXModel:()=>t.GPTNeoXModel,GPTNeoXPreTrainedModel:()=>t.GPTNeoXPreTrainedModel,GPTNeoXTokenizer:()=>s.GPTNeoXTokenizer,Gemma2ForCausalLM:()=>t.Gemma2ForCausalLM,Gemma2Model:()=>t.Gemma2Model,Gemma2PreTrainedModel:()=>t.Gemma2PreTrainedModel,Gemma3ForCausalLM:()=>t.Gemma3ForCausalLM,Gemma3Model:()=>t.Gemma3Model,Gemma3PreTrainedModel:()=>t.Gemma3PreTrainedModel,Gemma3nAudioFeatureExtractor:()=>d.Gemma3nAudioFeatureExtractor,Gemma3nForConditionalGeneration:()=>t.Gemma3nForConditionalGeneration,Gemma3nPreTrainedModel:()=>t.Gemma3nPreTrainedModel,Gemma3nProcessor:()=>w.Gemma3nProcessor,GemmaForCausalLM:()=>t.GemmaForCausalLM,GemmaModel:()=>t.GemmaModel,GemmaPreTrainedModel:()=>t.GemmaPreTrainedModel,GemmaTokenizer:()=>s.GemmaTokenizer,GlmForCausalLM:()=>t.GlmForCausalLM,GlmModel:()=>t.GlmModel,GlmPreTrainedModel:()=>t.GlmPreTrainedModel,GraniteForCausalLM:()=>t.GraniteForCausalLM,GraniteModel:()=>t.GraniteModel,GraniteMoeHybridForCausalLM:()=>t.GraniteMoeHybridForCausalLM,GraniteMoeHybridModel:()=>t.GraniteMoeHybridModel,GraniteMoeHybridPreTrainedModel:()=>t.GraniteMoeHybridPreTrainedModel,GranitePreTrainedModel:()=>t.GranitePreTrainedModel,Grok1Tokenizer:()=>s.Grok1Tokenizer,GroundingDinoForObjectDetection:()=>t.GroundingDinoForObjectDetection,GroundingDinoImageProcessor:()=>_.GroundingDinoImageProcessor,GroundingDinoPreTrainedModel:()=>t.GroundingDinoPreTrainedModel,GroundingDinoProcessor:()=>w.GroundingDinoProcessor,GroupViTModel:()=>t.GroupViTModel,GroupViTPreTrainedModel:()=>t.GroupViTPreTrainedModel,HeliumForCausalLM:()=>t.HeliumForCausalLM,HeliumModel:()=>t.HeliumModel,HeliumPreTrainedModel:()=>t.HeliumPreTrainedModel,HerbertTokenizer:()=>s.HerbertTokenizer,HieraForImageClassification:()=>t.HieraForImageClassification,HieraModel:()=>t.HieraModel,HieraPreTrainedModel:()=>t.HieraPreTrainedModel,HubertForCTC:()=>t.HubertForCTC,HubertForSequenceClassification:()=>t.HubertForSequenceClassification,HubertModel:()=>t.HubertModel,HubertPreTrainedModel:()=>t.HubertPreTrainedModel,IJepaForImageClassification:()=>t.IJepaForImageClassification,IJepaModel:()=>t.IJepaModel,IJepaPreTrainedModel:()=>t.IJepaPreTrainedModel,Idefics3ForConditionalGeneration:()=>t.Idefics3ForConditionalGeneration,Idefics3ImageProcessor:()=>_.Idefics3ImageProcessor,Idefics3PreTrainedModel:()=>t.Idefics3PreTrainedModel,Idefics3Processor:()=>w.Idefics3Processor,ImageClassificationPipeline:()=>r.ImageClassificationPipeline,ImageFeatureExtractionPipeline:()=>r.ImageFeatureExtractionPipeline,ImageFeatureExtractor:()=>d.ImageFeatureExtractor,ImageMattingOutput:()=>t.ImageMattingOutput,ImageProcessor:()=>f.ImageProcessor,ImageSegmentationPipeline:()=>r.ImageSegmentationPipeline,ImageToImagePipeline:()=>r.ImageToImagePipeline,ImageToTextPipeline:()=>r.ImageToTextPipeline,InterruptableStoppingCriteria:()=>T.InterruptableStoppingCriteria,JAISLMHeadModel:()=>t.JAISLMHeadModel,JAISModel:()=>t.JAISModel,JAISPreTrainedModel:()=>t.JAISPreTrainedModel,JinaCLIPImageProcessor:()=>_.JinaCLIPImageProcessor,JinaCLIPModel:()=>t.JinaCLIPModel,JinaCLIPPreTrainedModel:()=>t.JinaCLIPPreTrainedModel,JinaCLIPProcessor:()=>w.JinaCLIPProcessor,JinaCLIPTextModel:()=>t.JinaCLIPTextModel,JinaCLIPVisionModel:()=>t.JinaCLIPVisionModel,Lfm2ForCausalLM:()=>t.Lfm2ForCausalLM,Lfm2Model:()=>t.Lfm2Model,Lfm2PreTrainedModel:()=>t.Lfm2PreTrainedModel,LiteWhisperForConditionalGeneration:()=>t.LiteWhisperForConditionalGeneration,Llama4ForCausalLM:()=>t.Llama4ForCausalLM,Llama4PreTrainedModel:()=>t.Llama4PreTrainedModel,LlamaForCausalLM:()=>t.LlamaForCausalLM,LlamaModel:()=>t.LlamaModel,LlamaPreTrainedModel:()=>t.LlamaPreTrainedModel,LlamaTokenizer:()=>s.LlamaTokenizer,LlavaForConditionalGeneration:()=>t.LlavaForConditionalGeneration,LlavaOnevisionForConditionalGeneration:()=>t.LlavaOnevisionForConditionalGeneration,LlavaOnevisionImageProcessor:()=>_.LlavaOnevisionImageProcessor,LlavaPreTrainedModel:()=>t.LlavaPreTrainedModel,LlavaProcessor:()=>w.LlavaProcessor,LlavaQwen2ForCausalLM:()=>t.LlavaQwen2ForCausalLM,LogitsProcessor:()=>b.LogitsProcessor,LogitsProcessorList:()=>b.LogitsProcessorList,LogitsWarper:()=>b.LogitsWarper,LongT5ForConditionalGeneration:()=>t.LongT5ForConditionalGeneration,LongT5Model:()=>t.LongT5Model,LongT5PreTrainedModel:()=>t.LongT5PreTrainedModel,M2M100ForConditionalGeneration:()=>t.M2M100ForConditionalGeneration,M2M100Model:()=>t.M2M100Model,M2M100PreTrainedModel:()=>t.M2M100PreTrainedModel,M2M100Tokenizer:()=>s.M2M100Tokenizer,MBart50Tokenizer:()=>s.MBart50Tokenizer,MBartForCausalLM:()=>t.MBartForCausalLM,MBartForConditionalGeneration:()=>t.MBartForConditionalGeneration,MBartForSequenceClassification:()=>t.MBartForSequenceClassification,MBartModel:()=>t.MBartModel,MBartPreTrainedModel:()=>t.MBartPreTrainedModel,MBartTokenizer:()=>s.MBartTokenizer,MPNetForMaskedLM:()=>t.MPNetForMaskedLM,MPNetForQuestionAnswering:()=>t.MPNetForQuestionAnswering,MPNetForSequenceClassification:()=>t.MPNetForSequenceClassification,MPNetForTokenClassification:()=>t.MPNetForTokenClassification,MPNetModel:()=>t.MPNetModel,MPNetPreTrainedModel:()=>t.MPNetPreTrainedModel,MPNetTokenizer:()=>s.MPNetTokenizer,MT5ForConditionalGeneration:()=>t.MT5ForConditionalGeneration,MT5Model:()=>t.MT5Model,MT5PreTrainedModel:()=>t.MT5PreTrainedModel,MarianMTModel:()=>t.MarianMTModel,MarianModel:()=>t.MarianModel,MarianPreTrainedModel:()=>t.MarianPreTrainedModel,MarianTokenizer:()=>s.MarianTokenizer,Mask2FormerImageProcessor:()=>_.Mask2FormerImageProcessor,MaskFormerFeatureExtractor:()=>_.MaskFormerFeatureExtractor,MaskFormerForInstanceSegmentation:()=>t.MaskFormerForInstanceSegmentation,MaskFormerImageProcessor:()=>_.MaskFormerImageProcessor,MaskFormerModel:()=>t.MaskFormerModel,MaskFormerPreTrainedModel:()=>t.MaskFormerPreTrainedModel,MaskedLMOutput:()=>t.MaskedLMOutput,MaxLengthCriteria:()=>T.MaxLengthCriteria,Metric3DForDepthEstimation:()=>t.Metric3DForDepthEstimation,Metric3DPreTrainedModel:()=>t.Metric3DPreTrainedModel,Metric3Dv2ForDepthEstimation:()=>t.Metric3Dv2ForDepthEstimation,Metric3Dv2PreTrainedModel:()=>t.Metric3Dv2PreTrainedModel,MgpstrForSceneTextRecognition:()=>t.MgpstrForSceneTextRecognition,MgpstrModelOutput:()=>t.MgpstrModelOutput,MgpstrPreTrainedModel:()=>t.MgpstrPreTrainedModel,MgpstrProcessor:()=>w.MgpstrProcessor,MgpstrTokenizer:()=>s.MgpstrTokenizer,MimiDecoderModel:()=>t.MimiDecoderModel,MimiDecoderOutput:()=>t.MimiDecoderOutput,MimiEncoderModel:()=>t.MimiEncoderModel,MimiEncoderOutput:()=>t.MimiEncoderOutput,MimiModel:()=>t.MimiModel,MimiPreTrainedModel:()=>t.MimiPreTrainedModel,MinLengthLogitsProcessor:()=>b.MinLengthLogitsProcessor,MinNewTokensLengthLogitsProcessor:()=>b.MinNewTokensLengthLogitsProcessor,Ministral3ForCausalLM:()=>t.Ministral3ForCausalLM,Ministral3Model:()=>t.Ministral3Model,Ministral3PreTrainedModel:()=>t.Ministral3PreTrainedModel,MinistralForCausalLM:()=>t.MinistralForCausalLM,MinistralModel:()=>t.MinistralModel,MinistralPreTrainedModel:()=>t.MinistralPreTrainedModel,Mistral3ForConditionalGeneration:()=>t.Mistral3ForConditionalGeneration,MistralForCausalLM:()=>t.MistralForCausalLM,MistralModel:()=>t.MistralModel,MistralPreTrainedModel:()=>t.MistralPreTrainedModel,MobileBertForMaskedLM:()=>t.MobileBertForMaskedLM,MobileBertForQuestionAnswering:()=>t.MobileBertForQuestionAnswering,MobileBertForSequenceClassification:()=>t.MobileBertForSequenceClassification,MobileBertModel:()=>t.MobileBertModel,MobileBertPreTrainedModel:()=>t.MobileBertPreTrainedModel,MobileBertTokenizer:()=>s.MobileBertTokenizer,MobileLLMForCausalLM:()=>t.MobileLLMForCausalLM,MobileLLMModel:()=>t.MobileLLMModel,MobileLLMPreTrainedModel:()=>t.MobileLLMPreTrainedModel,MobileNetV1FeatureExtractor:()=>_.MobileNetV1FeatureExtractor,MobileNetV1ForImageClassification:()=>t.MobileNetV1ForImageClassification,MobileNetV1ForSemanticSegmentation:()=>t.MobileNetV1ForSemanticSegmentation,MobileNetV1ImageProcessor:()=>_.MobileNetV1ImageProcessor,MobileNetV1Model:()=>t.MobileNetV1Model,MobileNetV1PreTrainedModel:()=>t.MobileNetV1PreTrainedModel,MobileNetV2FeatureExtractor:()=>_.MobileNetV2FeatureExtractor,MobileNetV2ForImageClassification:()=>t.MobileNetV2ForImageClassification,MobileNetV2ForSemanticSegmentation:()=>t.MobileNetV2ForSemanticSegmentation,MobileNetV2ImageProcessor:()=>_.MobileNetV2ImageProcessor,MobileNetV2Model:()=>t.MobileNetV2Model,MobileNetV2PreTrainedModel:()=>t.MobileNetV2PreTrainedModel,MobileNetV3FeatureExtractor:()=>_.MobileNetV3FeatureExtractor,MobileNetV3ForImageClassification:()=>t.MobileNetV3ForImageClassification,MobileNetV3ForSemanticSegmentation:()=>t.MobileNetV3ForSemanticSegmentation,MobileNetV3ImageProcessor:()=>_.MobileNetV3ImageProcessor,MobileNetV3Model:()=>t.MobileNetV3Model,MobileNetV3PreTrainedModel:()=>t.MobileNetV3PreTrainedModel,MobileNetV4FeatureExtractor:()=>_.MobileNetV4FeatureExtractor,MobileNetV4ForImageClassification:()=>t.MobileNetV4ForImageClassification,MobileNetV4ForSemanticSegmentation:()=>t.MobileNetV4ForSemanticSegmentation,MobileNetV4ImageProcessor:()=>_.MobileNetV4ImageProcessor,MobileNetV4Model:()=>t.MobileNetV4Model,MobileNetV4PreTrainedModel:()=>t.MobileNetV4PreTrainedModel,MobileViTFeatureExtractor:()=>_.MobileViTFeatureExtractor,MobileViTForImageClassification:()=>t.MobileViTForImageClassification,MobileViTImageProcessor:()=>_.MobileViTImageProcessor,MobileViTModel:()=>t.MobileViTModel,MobileViTPreTrainedModel:()=>t.MobileViTPreTrainedModel,MobileViTV2ForImageClassification:()=>t.MobileViTV2ForImageClassification,MobileViTV2Model:()=>t.MobileViTV2Model,MobileViTV2PreTrainedModel:()=>t.MobileViTV2PreTrainedModel,ModelOutput:()=>t.ModelOutput,ModernBertDecoderForCausalLM:()=>t.ModernBertDecoderForCausalLM,ModernBertDecoderModel:()=>t.ModernBertDecoderModel,ModernBertDecoderPreTrainedModel:()=>t.ModernBertDecoderPreTrainedModel,ModernBertForMaskedLM:()=>t.ModernBertForMaskedLM,ModernBertForSequenceClassification:()=>t.ModernBertForSequenceClassification,ModernBertForTokenClassification:()=>t.ModernBertForTokenClassification,ModernBertModel:()=>t.ModernBertModel,ModernBertPreTrainedModel:()=>t.ModernBertPreTrainedModel,Moondream1ForConditionalGeneration:()=>t.Moondream1ForConditionalGeneration,MoonshineFeatureExtractor:()=>d.MoonshineFeatureExtractor,MoonshineForConditionalGeneration:()=>t.MoonshineForConditionalGeneration,MoonshineModel:()=>t.MoonshineModel,MoonshinePreTrainedModel:()=>t.MoonshinePreTrainedModel,MoonshineProcessor:()=>w.MoonshineProcessor,MptForCausalLM:()=>t.MptForCausalLM,MptModel:()=>t.MptModel,MptPreTrainedModel:()=>t.MptPreTrainedModel,MultiModalityCausalLM:()=>t.MultiModalityCausalLM,MultiModalityPreTrainedModel:()=>t.MultiModalityPreTrainedModel,MusicgenForCausalLM:()=>t.MusicgenForCausalLM,MusicgenForConditionalGeneration:()=>t.MusicgenForConditionalGeneration,MusicgenModel:()=>t.MusicgenModel,MusicgenPreTrainedModel:()=>t.MusicgenPreTrainedModel,NanoChatForCausalLM:()=>t.NanoChatForCausalLM,NanoChatModel:()=>t.NanoChatModel,NanoChatPreTrainedModel:()=>t.NanoChatPreTrainedModel,NeoBertForMaskedLM:()=>t.NeoBertForMaskedLM,NeoBertForQuestionAnswering:()=>t.NeoBertForQuestionAnswering,NeoBertForSequenceClassification:()=>t.NeoBertForSequenceClassification,NeoBertForTokenClassification:()=>t.NeoBertForTokenClassification,NeoBertModel:()=>t.NeoBertModel,NeoBertPreTrainedModel:()=>t.NeoBertPreTrainedModel,NllbTokenizer:()=>s.NllbTokenizer,NoBadWordsLogitsProcessor:()=>b.NoBadWordsLogitsProcessor,NoRepeatNGramLogitsProcessor:()=>b.NoRepeatNGramLogitsProcessor,NomicBertModel:()=>t.NomicBertModel,NomicBertPreTrainedModel:()=>t.NomicBertPreTrainedModel,NougatImageProcessor:()=>_.NougatImageProcessor,NougatTokenizer:()=>s.NougatTokenizer,OPTForCausalLM:()=>t.OPTForCausalLM,OPTModel:()=>t.OPTModel,OPTPreTrainedModel:()=>t.OPTPreTrainedModel,ObjectDetectionPipeline:()=>r.ObjectDetectionPipeline,Olmo2ForCausalLM:()=>t.Olmo2ForCausalLM,Olmo2Model:()=>t.Olmo2Model,Olmo2PreTrainedModel:()=>t.Olmo2PreTrainedModel,OlmoForCausalLM:()=>t.OlmoForCausalLM,OlmoModel:()=>t.OlmoModel,OlmoPreTrainedModel:()=>t.OlmoPreTrainedModel,OpenELMForCausalLM:()=>t.OpenELMForCausalLM,OpenELMModel:()=>t.OpenELMModel,OpenELMPreTrainedModel:()=>t.OpenELMPreTrainedModel,OwlViTFeatureExtractor:()=>_.OwlViTFeatureExtractor,OwlViTForObjectDetection:()=>t.OwlViTForObjectDetection,OwlViTImageProcessor:()=>_.OwlViTImageProcessor,OwlViTModel:()=>t.OwlViTModel,OwlViTPreTrainedModel:()=>t.OwlViTPreTrainedModel,OwlViTProcessor:()=>w.OwlViTProcessor,Owlv2ForObjectDetection:()=>t.Owlv2ForObjectDetection,Owlv2ImageProcessor:()=>_.Owlv2ImageProcessor,Owlv2Model:()=>t.Owlv2Model,Owlv2PreTrainedModel:()=>t.Owlv2PreTrainedModel,PaliGemmaForConditionalGeneration:()=>t.PaliGemmaForConditionalGeneration,PaliGemmaPreTrainedModel:()=>t.PaliGemmaPreTrainedModel,PaliGemmaProcessor:()=>w.PaliGemmaProcessor,ParakeetFeatureExtractor:()=>d.ParakeetFeatureExtractor,ParakeetForCTC:()=>t.ParakeetForCTC,ParakeetPreTrainedModel:()=>t.ParakeetPreTrainedModel,PatchTSMixerForPrediction:()=>t.PatchTSMixerForPrediction,PatchTSMixerModel:()=>t.PatchTSMixerModel,PatchTSMixerPreTrainedModel:()=>t.PatchTSMixerPreTrainedModel,PatchTSTForPrediction:()=>t.PatchTSTForPrediction,PatchTSTModel:()=>t.PatchTSTModel,PatchTSTPreTrainedModel:()=>t.PatchTSTPreTrainedModel,Phi3ForCausalLM:()=>t.Phi3ForCausalLM,Phi3Model:()=>t.Phi3Model,Phi3PreTrainedModel:()=>t.Phi3PreTrainedModel,Phi3VForCausalLM:()=>t.Phi3VForCausalLM,Phi3VImageProcessor:()=>_.Phi3VImageProcessor,Phi3VPreTrainedModel:()=>t.Phi3VPreTrainedModel,Phi3VProcessor:()=>w.Phi3VProcessor,PhiForCausalLM:()=>t.PhiForCausalLM,PhiModel:()=>t.PhiModel,PhiPreTrainedModel:()=>t.PhiPreTrainedModel,Pipeline:()=>r.Pipeline,PixtralImageProcessor:()=>_.PixtralImageProcessor,PixtralProcessor:()=>w.PixtralProcessor,PreTrainedModel:()=>t.PreTrainedModel,PreTrainedTokenizer:()=>s.PreTrainedTokenizer,PretrainedConfig:()=>n.PretrainedConfig,PretrainedMixin:()=>t.PretrainedMixin,Processor:()=>I.Processor,PvtForImageClassification:()=>t.PvtForImageClassification,PvtImageProcessor:()=>_.PvtImageProcessor,PvtModel:()=>t.PvtModel,PvtPreTrainedModel:()=>t.PvtPreTrainedModel,PyAnnoteFeatureExtractor:()=>d.PyAnnoteFeatureExtractor,PyAnnoteForAudioFrameClassification:()=>t.PyAnnoteForAudioFrameClassification,PyAnnoteModel:()=>t.PyAnnoteModel,PyAnnotePreTrainedModel:()=>t.PyAnnotePreTrainedModel,PyAnnoteProcessor:()=>w.PyAnnoteProcessor,QuestionAnsweringModelOutput:()=>t.QuestionAnsweringModelOutput,QuestionAnsweringPipeline:()=>r.QuestionAnsweringPipeline,Qwen2ForCausalLM:()=>t.Qwen2ForCausalLM,Qwen2Model:()=>t.Qwen2Model,Qwen2PreTrainedModel:()=>t.Qwen2PreTrainedModel,Qwen2Tokenizer:()=>s.Qwen2Tokenizer,Qwen2VLForConditionalGeneration:()=>t.Qwen2VLForConditionalGeneration,Qwen2VLImageProcessor:()=>_.Qwen2VLImageProcessor,Qwen2VLPreTrainedModel:()=>t.Qwen2VLPreTrainedModel,Qwen2VLProcessor:()=>w.Qwen2VLProcessor,Qwen3ForCausalLM:()=>t.Qwen3ForCausalLM,Qwen3Model:()=>t.Qwen3Model,Qwen3PreTrainedModel:()=>t.Qwen3PreTrainedModel,RFDetrForObjectDetection:()=>t.RFDetrForObjectDetection,RFDetrModel:()=>t.RFDetrModel,RFDetrObjectDetectionOutput:()=>t.RFDetrObjectDetectionOutput,RFDetrPreTrainedModel:()=>t.RFDetrPreTrainedModel,RTDetrForObjectDetection:()=>t.RTDetrForObjectDetection,RTDetrImageProcessor:()=>_.RTDetrImageProcessor,RTDetrModel:()=>t.RTDetrModel,RTDetrObjectDetectionOutput:()=>t.RTDetrObjectDetectionOutput,RTDetrPreTrainedModel:()=>t.RTDetrPreTrainedModel,RTDetrV2ForObjectDetection:()=>t.RTDetrV2ForObjectDetection,RTDetrV2Model:()=>t.RTDetrV2Model,RTDetrV2ObjectDetectionOutput:()=>t.RTDetrV2ObjectDetectionOutput,RTDetrV2PreTrainedModel:()=>t.RTDetrV2PreTrainedModel,RawAudio:()=>o.RawAudio,RawImage:()=>i.RawImage,RawVideo:()=>a.RawVideo,RawVideoFrame:()=>a.RawVideoFrame,RepetitionPenaltyLogitsProcessor:()=>b.RepetitionPenaltyLogitsProcessor,ResNetForImageClassification:()=>t.ResNetForImageClassification,ResNetModel:()=>t.ResNetModel,ResNetPreTrainedModel:()=>t.ResNetPreTrainedModel,RoFormerForMaskedLM:()=>t.RoFormerForMaskedLM,RoFormerForQuestionAnswering:()=>t.RoFormerForQuestionAnswering,RoFormerForSequenceClassification:()=>t.RoFormerForSequenceClassification,RoFormerForTokenClassification:()=>t.RoFormerForTokenClassification,RoFormerModel:()=>t.RoFormerModel,RoFormerPreTrainedModel:()=>t.RoFormerPreTrainedModel,RoFormerTokenizer:()=>s.RoFormerTokenizer,RobertaForMaskedLM:()=>t.RobertaForMaskedLM,RobertaForQuestionAnswering:()=>t.RobertaForQuestionAnswering,RobertaForSequenceClassification:()=>t.RobertaForSequenceClassification,RobertaForTokenClassification:()=>t.RobertaForTokenClassification,RobertaModel:()=>t.RobertaModel,RobertaPreTrainedModel:()=>t.RobertaPreTrainedModel,RobertaTokenizer:()=>s.RobertaTokenizer,Sam2ImageProcessor:()=>_.Sam2ImageProcessor,Sam2ImageSegmentationOutput:()=>t.Sam2ImageSegmentationOutput,Sam2Model:()=>t.Sam2Model,Sam2PreTrainedModel:()=>t.Sam2PreTrainedModel,Sam2Processor:()=>w.Sam2Processor,Sam2VideoProcessor:()=>w.Sam2VideoProcessor,Sam3ImageProcessor:()=>_.Sam3ImageProcessor,Sam3TrackerModel:()=>t.Sam3TrackerModel,SamImageProcessor:()=>_.SamImageProcessor,SamImageSegmentationOutput:()=>t.SamImageSegmentationOutput,SamModel:()=>t.SamModel,SamPreTrainedModel:()=>t.SamPreTrainedModel,SamProcessor:()=>w.SamProcessor,SapiensForDepthEstimation:()=>t.SapiensForDepthEstimation,SapiensForNormalEstimation:()=>t.SapiensForNormalEstimation,SapiensForSemanticSegmentation:()=>t.SapiensForSemanticSegmentation,SapiensPreTrainedModel:()=>t.SapiensPreTrainedModel,SeamlessM4TFeatureExtractor:()=>d.SeamlessM4TFeatureExtractor,SegformerFeatureExtractor:()=>_.SegformerFeatureExtractor,SegformerForImageClassification:()=>t.SegformerForImageClassification,SegformerForSemanticSegmentation:()=>t.SegformerForSemanticSegmentation,SegformerImageProcessor:()=>_.SegformerImageProcessor,SegformerModel:()=>t.SegformerModel,SegformerPreTrainedModel:()=>t.SegformerPreTrainedModel,Seq2SeqLMOutput:()=>t.Seq2SeqLMOutput,SequenceClassifierOutput:()=>t.SequenceClassifierOutput,SiglipImageProcessor:()=>_.SiglipImageProcessor,SiglipModel:()=>t.SiglipModel,SiglipPreTrainedModel:()=>t.SiglipPreTrainedModel,SiglipTextModel:()=>t.SiglipTextModel,SiglipTokenizer:()=>s.SiglipTokenizer,SiglipVisionModel:()=>t.SiglipVisionModel,SmolLM3ForCausalLM:()=>t.SmolLM3ForCausalLM,SmolLM3Model:()=>t.SmolLM3Model,SmolLM3PreTrainedModel:()=>t.SmolLM3PreTrainedModel,SmolVLMForConditionalGeneration:()=>t.SmolVLMForConditionalGeneration,SmolVLMImageProcessor:()=>_.SmolVLMImageProcessor,SmolVLMProcessor:()=>w.SmolVLMProcessor,SnacDecoderModel:()=>t.SnacDecoderModel,SnacEncoderModel:()=>t.SnacEncoderModel,SnacFeatureExtractor:()=>d.SnacFeatureExtractor,SnacModel:()=>t.SnacModel,SnacPreTrainedModel:()=>t.SnacPreTrainedModel,SpeechT5FeatureExtractor:()=>d.SpeechT5FeatureExtractor,SpeechT5ForSpeechToText:()=>t.SpeechT5ForSpeechToText,SpeechT5ForTextToSpeech:()=>t.SpeechT5ForTextToSpeech,SpeechT5HifiGan:()=>t.SpeechT5HifiGan,SpeechT5Model:()=>t.SpeechT5Model,SpeechT5PreTrainedModel:()=>t.SpeechT5PreTrainedModel,SpeechT5Processor:()=>w.SpeechT5Processor,SpeechT5Tokenizer:()=>s.SpeechT5Tokenizer,SqueezeBertForMaskedLM:()=>t.SqueezeBertForMaskedLM,SqueezeBertForQuestionAnswering:()=>t.SqueezeBertForQuestionAnswering,SqueezeBertForSequenceClassification:()=>t.SqueezeBertForSequenceClassification,SqueezeBertModel:()=>t.SqueezeBertModel,SqueezeBertPreTrainedModel:()=>t.SqueezeBertPreTrainedModel,SqueezeBertTokenizer:()=>s.SqueezeBertTokenizer,StableLmForCausalLM:()=>t.StableLmForCausalLM,StableLmModel:()=>t.StableLmModel,StableLmPreTrainedModel:()=>t.StableLmPreTrainedModel,Starcoder2ForCausalLM:()=>t.Starcoder2ForCausalLM,Starcoder2Model:()=>t.Starcoder2Model,Starcoder2PreTrainedModel:()=>t.Starcoder2PreTrainedModel,StoppingCriteria:()=>T.StoppingCriteria,StoppingCriteriaList:()=>T.StoppingCriteriaList,StyleTextToSpeech2Model:()=>t.StyleTextToSpeech2Model,StyleTextToSpeech2PreTrainedModel:()=>t.StyleTextToSpeech2PreTrainedModel,SummarizationPipeline:()=>r.SummarizationPipeline,SupertonicForConditionalGeneration:()=>t.SupertonicForConditionalGeneration,SupertonicPreTrainedModel:()=>t.SupertonicPreTrainedModel,SuppressTokensAtBeginLogitsProcessor:()=>b.SuppressTokensAtBeginLogitsProcessor,Swin2SRForImageSuperResolution:()=>t.Swin2SRForImageSuperResolution,Swin2SRImageProcessor:()=>_.Swin2SRImageProcessor,Swin2SRModel:()=>t.Swin2SRModel,Swin2SRPreTrainedModel:()=>t.Swin2SRPreTrainedModel,SwinForImageClassification:()=>t.SwinForImageClassification,SwinForSemanticSegmentation:()=>t.SwinForSemanticSegmentation,SwinModel:()=>t.SwinModel,SwinPreTrainedModel:()=>t.SwinPreTrainedModel,T5ForConditionalGeneration:()=>t.T5ForConditionalGeneration,T5Model:()=>t.T5Model,T5PreTrainedModel:()=>t.T5PreTrainedModel,T5Tokenizer:()=>s.T5Tokenizer,TableTransformerForObjectDetection:()=>t.TableTransformerForObjectDetection,TableTransformerModel:()=>t.TableTransformerModel,TableTransformerObjectDetectionOutput:()=>t.TableTransformerObjectDetectionOutput,TableTransformerPreTrainedModel:()=>t.TableTransformerPreTrainedModel,TemperatureLogitsWarper:()=>b.TemperatureLogitsWarper,Tensor:()=>l.Tensor,Text2TextGenerationPipeline:()=>r.Text2TextGenerationPipeline,TextClassificationPipeline:()=>r.TextClassificationPipeline,TextGenerationPipeline:()=>r.TextGenerationPipeline,TextStreamer:()=>k.TextStreamer,TextToAudioPipeline:()=>r.TextToAudioPipeline,TokenClassificationPipeline:()=>r.TokenClassificationPipeline,TokenClassifierOutput:()=>t.TokenClassifierOutput,TokenizerModel:()=>s.TokenizerModel,TopKLogitsWarper:()=>b.TopKLogitsWarper,TopPLogitsWarper:()=>b.TopPLogitsWarper,TrOCRForCausalLM:()=>t.TrOCRForCausalLM,TrOCRPreTrainedModel:()=>t.TrOCRPreTrainedModel,TranslationPipeline:()=>r.TranslationPipeline,UltravoxModel:()=>t.UltravoxModel,UltravoxPreTrainedModel:()=>t.UltravoxPreTrainedModel,UltravoxProcessor:()=>w.UltravoxProcessor,UniSpeechForCTC:()=>t.UniSpeechForCTC,UniSpeechForSequenceClassification:()=>t.UniSpeechForSequenceClassification,UniSpeechModel:()=>t.UniSpeechModel,UniSpeechPreTrainedModel:()=>t.UniSpeechPreTrainedModel,UniSpeechSatForAudioFrameClassification:()=>t.UniSpeechSatForAudioFrameClassification,UniSpeechSatForCTC:()=>t.UniSpeechSatForCTC,UniSpeechSatForSequenceClassification:()=>t.UniSpeechSatForSequenceClassification,UniSpeechSatModel:()=>t.UniSpeechSatModel,UniSpeechSatPreTrainedModel:()=>t.UniSpeechSatPreTrainedModel,VLChatProcessor:()=>w.VLChatProcessor,VLMImageProcessor:()=>_.VLMImageProcessor,VaultGemmaForCausalLM:()=>t.VaultGemmaForCausalLM,VaultGemmaModel:()=>t.VaultGemmaModel,VaultGemmaPreTrainedModel:()=>t.VaultGemmaPreTrainedModel,ViTFeatureExtractor:()=>_.ViTFeatureExtractor,ViTForImageClassification:()=>t.ViTForImageClassification,ViTImageProcessor:()=>_.ViTImageProcessor,ViTMAEModel:()=>t.ViTMAEModel,ViTMAEPreTrainedModel:()=>t.ViTMAEPreTrainedModel,ViTMSNForImageClassification:()=>t.ViTMSNForImageClassification,ViTMSNModel:()=>t.ViTMSNModel,ViTMSNPreTrainedModel:()=>t.ViTMSNPreTrainedModel,ViTModel:()=>t.ViTModel,ViTPreTrainedModel:()=>t.ViTPreTrainedModel,VisionEncoderDecoderModel:()=>t.VisionEncoderDecoderModel,VitMatteForImageMatting:()=>t.VitMatteForImageMatting,VitMatteImageProcessor:()=>_.VitMatteImageProcessor,VitMattePreTrainedModel:()=>t.VitMattePreTrainedModel,VitPoseForPoseEstimation:()=>t.VitPoseForPoseEstimation,VitPoseImageProcessor:()=>_.VitPoseImageProcessor,VitPosePreTrainedModel:()=>t.VitPosePreTrainedModel,VitsModel:()=>t.VitsModel,VitsModelOutput:()=>t.VitsModelOutput,VitsPreTrainedModel:()=>t.VitsPreTrainedModel,VitsTokenizer:()=>s.VitsTokenizer,VoxtralForConditionalGeneration:()=>t.VoxtralForConditionalGeneration,VoxtralProcessor:()=>w.VoxtralProcessor,Wav2Vec2BertForCTC:()=>t.Wav2Vec2BertForCTC,Wav2Vec2BertForSequenceClassification:()=>t.Wav2Vec2BertForSequenceClassification,Wav2Vec2BertModel:()=>t.Wav2Vec2BertModel,Wav2Vec2BertPreTrainedModel:()=>t.Wav2Vec2BertPreTrainedModel,Wav2Vec2CTCTokenizer:()=>s.Wav2Vec2CTCTokenizer,Wav2Vec2FeatureExtractor:()=>d.Wav2Vec2FeatureExtractor,Wav2Vec2ForAudioFrameClassification:()=>t.Wav2Vec2ForAudioFrameClassification,Wav2Vec2ForCTC:()=>t.Wav2Vec2ForCTC,Wav2Vec2ForSequenceClassification:()=>t.Wav2Vec2ForSequenceClassification,Wav2Vec2Model:()=>t.Wav2Vec2Model,Wav2Vec2PreTrainedModel:()=>t.Wav2Vec2PreTrainedModel,Wav2Vec2Processor:()=>w.Wav2Vec2Processor,Wav2Vec2ProcessorWithLM:()=>w.Wav2Vec2ProcessorWithLM,WavLMForAudioFrameClassification:()=>t.WavLMForAudioFrameClassification,WavLMForCTC:()=>t.WavLMForCTC,WavLMForSequenceClassification:()=>t.WavLMForSequenceClassification,WavLMForXVector:()=>t.WavLMForXVector,WavLMModel:()=>t.WavLMModel,WavLMPreTrainedModel:()=>t.WavLMPreTrainedModel,WeSpeakerFeatureExtractor:()=>d.WeSpeakerFeatureExtractor,WeSpeakerResNetModel:()=>t.WeSpeakerResNetModel,WeSpeakerResNetPreTrainedModel:()=>t.WeSpeakerResNetPreTrainedModel,WhisperFeatureExtractor:()=>d.WhisperFeatureExtractor,WhisperForConditionalGeneration:()=>t.WhisperForConditionalGeneration,WhisperModel:()=>t.WhisperModel,WhisperPreTrainedModel:()=>t.WhisperPreTrainedModel,WhisperProcessor:()=>w.WhisperProcessor,WhisperTextStreamer:()=>k.WhisperTextStreamer,WhisperTimeStampLogitsProcessor:()=>b.WhisperTimeStampLogitsProcessor,WhisperTokenizer:()=>s.WhisperTokenizer,XLMForQuestionAnswering:()=>t.XLMForQuestionAnswering,XLMForSequenceClassification:()=>t.XLMForSequenceClassification,XLMForTokenClassification:()=>t.XLMForTokenClassification,XLMModel:()=>t.XLMModel,XLMPreTrainedModel:()=>t.XLMPreTrainedModel,XLMRobertaForMaskedLM:()=>t.XLMRobertaForMaskedLM,XLMRobertaForQuestionAnswering:()=>t.XLMRobertaForQuestionAnswering,XLMRobertaForSequenceClassification:()=>t.XLMRobertaForSequenceClassification,XLMRobertaForTokenClassification:()=>t.XLMRobertaForTokenClassification,XLMRobertaModel:()=>t.XLMRobertaModel,XLMRobertaPreTrainedModel:()=>t.XLMRobertaPreTrainedModel,XLMRobertaTokenizer:()=>s.XLMRobertaTokenizer,XLMTokenizer:()=>s.XLMTokenizer,XLMWithLMHeadModel:()=>t.XLMWithLMHeadModel,XVectorOutput:()=>t.XVectorOutput,YolosFeatureExtractor:()=>_.YolosFeatureExtractor,YolosForObjectDetection:()=>t.YolosForObjectDetection,YolosImageProcessor:()=>_.YolosImageProcessor,YolosModel:()=>t.YolosModel,YolosObjectDetectionOutput:()=>t.YolosObjectDetectionOutput,YolosPreTrainedModel:()=>t.YolosPreTrainedModel,ZeroShotAudioClassificationPipeline:()=>r.ZeroShotAudioClassificationPipeline,ZeroShotClassificationPipeline:()=>r.ZeroShotClassificationPipeline,ZeroShotImageClassificationPipeline:()=>r.ZeroShotImageClassificationPipeline,ZeroShotObjectDetectionPipeline:()=>r.ZeroShotObjectDetectionPipeline,bankers_round:()=>c.bankers_round,cat:()=>l.cat,cos_sim:()=>c.cos_sim,dot:()=>c.dot,dynamic_time_warping:()=>c.dynamic_time_warping,env:()=>e.env,full:()=>l.full,full_like:()=>l.full_like,getCacheShapes:()=>n.getCacheShapes,hamming:()=>o.hamming,hanning:()=>o.hanning,interpolate:()=>l.interpolate,interpolate_4d:()=>l.interpolate_4d,interpolate_data:()=>c.interpolate_data,is_chinese_char:()=>s.is_chinese_char,layer_norm:()=>l.layer_norm,load_image:()=>i.load_image,load_video:()=>a.load_video,log_softmax:()=>c.log_softmax,magnitude:()=>c.magnitude,matmul:()=>l.matmul,max:()=>c.max,mean:()=>l.mean,mean_pooling:()=>l.mean_pooling,medianFilter:()=>c.medianFilter,mel_filter_bank:()=>o.mel_filter_bank,min:()=>c.min,ones:()=>l.ones,ones_like:()=>l.ones_like,permute:()=>l.permute,permute_data:()=>c.permute_data,pipeline:()=>r.pipeline,quantize_embeddings:()=>l.quantize_embeddings,rand:()=>l.rand,randn:()=>l.randn,read_audio:()=>o.read_audio,rfft:()=>l.rfft,round:()=>c.round,slice:()=>l.slice,softmax:()=>c.softmax,spectrogram:()=>o.spectrogram,stack:()=>l.stack,std_mean:()=>l.std_mean,topk:()=>l.topk,window_function:()=>o.window_function,zeros:()=>l.zeros,zeros_like:()=>l.zeros_like});var e=jt("./src/env.js"),r=jt("./src/pipelines.js"),t=jt("./src/models.js"),s=jt("./src/tokenizers.js"),n=jt("./src/configs.js"),o=jt("./src/utils/audio.js"),i=jt("./src/utils/image.js"),a=jt("./src/utils/video.js"),l=jt("./src/utils/tensor.js"),c=jt("./src/utils/maths.js"),p=jt("./src/base/feature_extraction_utils.js"),d=jt("./src/models/feature_extractors.js"),u=jt("./src/models/auto/feature_extraction_auto.js"),f=jt("./src/base/image_processors_utils.js"),_=jt("./src/models/image_processors.js"),y=jt("./src/models/auto/image_processing_auto.js"),I=jt("./src/base/processing_utils.js"),w=jt("./src/models/processors.js"),v=jt("./src/models/auto/processing_auto.js"),k=jt("./src/generation/streamers.js"),T=jt("./src/generation/stopping_criteria.js"),b=jt("./src/generation/logits_process.js")})(),m.ASTFeatureExtractor,m.ASTForAudioClassification,m.ASTModel,m.ASTPreTrainedModel,m.AlbertForMaskedLM,m.AlbertForQuestionAnswering,m.AlbertForSequenceClassification,m.AlbertModel,m.AlbertPreTrainedModel,m.AlbertTokenizer,m.ArceeForCausalLM,m.ArceeModel,m.ArceePreTrainedModel,m.AudioClassificationPipeline,m.AutoConfig,m.AutoFeatureExtractor,m.AutoImageProcessor,m.AutoModel,m.AutoModelForAudioClassification,m.AutoModelForAudioFrameClassification,m.AutoModelForAudioTextToText,m.AutoModelForCTC;var hT=m.AutoModelForCausalLM;m.AutoModelForDepthEstimation,m.AutoModelForDocumentQuestionAnswering,m.AutoModelForImageClassification,m.AutoModelForImageFeatureExtraction,m.AutoModelForImageMatting,m.AutoModelForImageSegmentation,m.AutoModelForImageTextToText,m.AutoModelForImageToImage,m.AutoModelForMaskGeneration,m.AutoModelForMaskedLM,m.AutoModelForNormalEstimation,m.AutoModelForObjectDetection,m.AutoModelForPoseEstimation,m.AutoModelForQuestionAnswering,m.AutoModelForSemanticSegmentation,m.AutoModelForSeq2SeqLM,m.AutoModelForSequenceClassification,m.AutoModelForSpeechSeq2Seq,m.AutoModelForTextToSpectrogram,m.AutoModelForTextToWaveform,m.AutoModelForTokenClassification,m.AutoModelForUniversalSegmentation,m.AutoModelForVision2Seq,m.AutoModelForXVector,m.AutoModelForZeroShotObjectDetection,m.AutoProcessor;var _T=m.AutoTokenizer;m.AutomaticSpeechRecognitionPipeline,m.BackgroundRemovalPipeline,m.BartForConditionalGeneration,m.BartForSequenceClassification,m.BartModel,m.BartPretrainedModel,m.BartTokenizer,m.BaseModelOutput,m.BaseStreamer,m.BeitFeatureExtractor,m.BeitForImageClassification,m.BeitModel,m.BeitPreTrainedModel,m.BertForMaskedLM,m.BertForQuestionAnswering,m.BertForSequenceClassification,m.BertForTokenClassification,m.BertModel,m.BertPreTrainedModel,m.BertTokenizer,m.BitImageProcessor,m.BlenderbotForConditionalGeneration,m.BlenderbotModel,m.BlenderbotPreTrainedModel,m.BlenderbotSmallForConditionalGeneration,m.BlenderbotSmallModel,m.BlenderbotSmallPreTrainedModel,m.BlenderbotSmallTokenizer,m.BlenderbotTokenizer,m.BloomForCausalLM,m.BloomModel,m.BloomPreTrainedModel,m.BloomTokenizer,m.CLIPFeatureExtractor,m.CLIPImageProcessor,m.CLIPModel,m.CLIPPreTrainedModel,m.CLIPSegForImageSegmentation,m.CLIPSegModel,m.CLIPSegPreTrainedModel,m.CLIPTextModel,m.CLIPTextModelWithProjection,m.CLIPTokenizer,m.CLIPVisionModel,m.CLIPVisionModelWithProjection,m.CamembertForMaskedLM,m.CamembertForQuestionAnswering,m.CamembertForSequenceClassification,m.CamembertForTokenClassification,m.CamembertModel,m.CamembertPreTrainedModel,m.CamembertTokenizer,m.CausalLMOutput,m.CausalLMOutputWithPast,m.ChineseCLIPFeatureExtractor,m.ChineseCLIPModel,m.ChineseCLIPPreTrainedModel,m.ClapAudioModelWithProjection,m.ClapFeatureExtractor,m.ClapModel,m.ClapPreTrainedModel,m.ClapTextModelWithProjection,m.ClassifierFreeGuidanceLogitsProcessor,m.CodeGenForCausalLM,m.CodeGenModel,m.CodeGenPreTrainedModel,m.CodeGenTokenizer,m.CodeLlamaTokenizer,m.CohereForCausalLM,m.CohereModel,m.CoherePreTrainedModel,m.CohereTokenizer,m.ConvBertForMaskedLM,m.ConvBertForQuestionAnswering,m.ConvBertForSequenceClassification,m.ConvBertForTokenClassification,m.ConvBertModel,m.ConvBertPreTrainedModel,m.ConvBertTokenizer,m.ConvNextFeatureExtractor,m.ConvNextForImageClassification,m.ConvNextImageProcessor,m.ConvNextModel,m.ConvNextPreTrainedModel,m.ConvNextV2ForImageClassification,m.ConvNextV2Model,m.ConvNextV2PreTrainedModel,m.DFineForObjectDetection,m.DFineModel,m.DFinePreTrainedModel,m.DINOv3ConvNextModel,m.DINOv3ConvNextPreTrainedModel,m.DINOv3ViTImageProcessor,m.DINOv3ViTModel,m.DINOv3ViTPreTrainedModel,m.DPTFeatureExtractor,m.DPTForDepthEstimation,m.DPTImageProcessor,m.DPTModel,m.DPTPreTrainedModel,m.DacDecoderModel,m.DacDecoderOutput,m.DacEncoderModel,m.DacEncoderOutput,m.DacFeatureExtractor,m.DacModel,m.DacPreTrainedModel,m.DataTypeMap,m.DebertaForMaskedLM,m.DebertaForQuestionAnswering,m.DebertaForSequenceClassification,m.DebertaForTokenClassification,m.DebertaModel,m.DebertaPreTrainedModel,m.DebertaTokenizer,m.DebertaV2ForMaskedLM,m.DebertaV2ForQuestionAnswering,m.DebertaV2ForSequenceClassification,m.DebertaV2ForTokenClassification,m.DebertaV2Model,m.DebertaV2PreTrainedModel,m.DebertaV2Tokenizer,m.DecisionTransformerModel,m.DecisionTransformerPreTrainedModel,m.DeiTFeatureExtractor,m.DeiTForImageClassification,m.DeiTImageProcessor,m.DeiTModel,m.DeiTPreTrainedModel,m.DepthAnythingForDepthEstimation,m.DepthAnythingPreTrainedModel,m.DepthEstimationPipeline,m.DepthProForDepthEstimation,m.DepthProPreTrainedModel,m.DetrFeatureExtractor,m.DetrForObjectDetection,m.DetrForSegmentation,m.DetrImageProcessor,m.DetrModel,m.DetrObjectDetectionOutput,m.DetrPreTrainedModel,m.DetrSegmentationOutput,m.Dinov2ForImageClassification,m.Dinov2Model,m.Dinov2PreTrainedModel,m.Dinov2WithRegistersForImageClassification,m.Dinov2WithRegistersModel,m.Dinov2WithRegistersPreTrainedModel,m.DistilBertForMaskedLM,m.DistilBertForQuestionAnswering,m.DistilBertForSequenceClassification,m.DistilBertForTokenClassification,m.DistilBertModel,m.DistilBertPreTrainedModel,m.DistilBertTokenizer,m.DocumentQuestionAnsweringPipeline,m.DonutFeatureExtractor,m.DonutImageProcessor,m.DonutSwinModel,m.DonutSwinPreTrainedModel,m.EdgeTamModel,m.EfficientNetForImageClassification,m.EfficientNetImageProcessor,m.EfficientNetModel,m.EfficientNetPreTrainedModel,m.ElectraForMaskedLM,m.ElectraForQuestionAnswering,m.ElectraForSequenceClassification,m.ElectraForTokenClassification,m.ElectraModel,m.ElectraPreTrainedModel,m.ElectraTokenizer,m.EncodecFeatureExtractor,m.EosTokenCriteria,m.Ernie4_5ForCausalLM,m.Ernie4_5Model,m.Ernie4_5PreTrainedModel,m.EsmForMaskedLM,m.EsmForSequenceClassification,m.EsmForTokenClassification,m.EsmModel,m.EsmPreTrainedModel,m.EsmTokenizer,m.ExaoneForCausalLM,m.ExaoneModel,m.ExaonePreTrainedModel,m.FFT,m.FalconForCausalLM,m.FalconModel,m.FalconPreTrainedModel,m.FalconTokenizer,m.FastViTForImageClassification,m.FastViTModel,m.FastViTPreTrainedModel,m.FeatureExtractionPipeline,m.FeatureExtractor,m.FillMaskPipeline,m.Florence2ForConditionalGeneration,m.Florence2PreTrainedModel,m.Florence2Processor,m.ForcedBOSTokenLogitsProcessor,m.ForcedEOSTokenLogitsProcessor,m.GLPNFeatureExtractor,m.GLPNForDepthEstimation,m.GLPNModel,m.GLPNPreTrainedModel,m.GPT2LMHeadModel,m.GPT2Model,m.GPT2PreTrainedModel,m.GPT2Tokenizer,m.GPTBigCodeForCausalLM,m.GPTBigCodeModel,m.GPTBigCodePreTrainedModel,m.GPTJForCausalLM,m.GPTJModel,m.GPTJPreTrainedModel,m.GPTNeoForCausalLM,m.GPTNeoModel,m.GPTNeoPreTrainedModel,m.GPTNeoXForCausalLM,m.GPTNeoXModel,m.GPTNeoXPreTrainedModel,m.GPTNeoXTokenizer,m.Gemma2ForCausalLM,m.Gemma2Model,m.Gemma2PreTrainedModel,m.Gemma3ForCausalLM,m.Gemma3Model,m.Gemma3PreTrainedModel,m.Gemma3nAudioFeatureExtractor,m.Gemma3nForConditionalGeneration,m.Gemma3nPreTrainedModel,m.Gemma3nProcessor,m.GemmaForCausalLM,m.GemmaModel,m.GemmaPreTrainedModel,m.GemmaTokenizer,m.GlmForCausalLM,m.GlmModel,m.GlmPreTrainedModel,m.GraniteForCausalLM,m.GraniteModel,m.GraniteMoeHybridForCausalLM,m.GraniteMoeHybridModel,m.GraniteMoeHybridPreTrainedModel,m.GranitePreTrainedModel,m.Grok1Tokenizer,m.GroundingDinoForObjectDetection,m.GroundingDinoImageProcessor,m.GroundingDinoPreTrainedModel,m.GroundingDinoProcessor,m.GroupViTModel,m.GroupViTPreTrainedModel,m.HeliumForCausalLM,m.HeliumModel,m.HeliumPreTrainedModel,m.HerbertTokenizer,m.HieraForImageClassification,m.HieraModel,m.HieraPreTrainedModel,m.HubertForCTC,m.HubertForSequenceClassification,m.HubertModel,m.HubertPreTrainedModel,m.IJepaForImageClassification,m.IJepaModel,m.IJepaPreTrainedModel,m.Idefics3ForConditionalGeneration,m.Idefics3ImageProcessor,m.Idefics3PreTrainedModel,m.Idefics3Processor,m.ImageClassificationPipeline,m.ImageFeatureExtractionPipeline,m.ImageFeatureExtractor,m.ImageMattingOutput,m.ImageProcessor,m.ImageSegmentationPipeline,m.ImageToImagePipeline,m.ImageToTextPipeline,m.InterruptableStoppingCriteria,m.JAISLMHeadModel,m.JAISModel,m.JAISPreTrainedModel,m.JinaCLIPImageProcessor,m.JinaCLIPModel,m.JinaCLIPPreTrainedModel,m.JinaCLIPProcessor,m.JinaCLIPTextModel,m.JinaCLIPVisionModel,m.Lfm2ForCausalLM,m.Lfm2Model,m.Lfm2PreTrainedModel,m.LiteWhisperForConditionalGeneration,m.Llama4ForCausalLM,m.Llama4PreTrainedModel,m.LlamaForCausalLM,m.LlamaModel,m.LlamaPreTrainedModel,m.LlamaTokenizer,m.LlavaForConditionalGeneration,m.LlavaOnevisionForConditionalGeneration,m.LlavaOnevisionImageProcessor,m.LlavaPreTrainedModel,m.LlavaProcessor,m.LlavaQwen2ForCausalLM,m.LogitsProcessor,m.LogitsProcessorList,m.LogitsWarper,m.LongT5ForConditionalGeneration,m.LongT5Model,m.LongT5PreTrainedModel,m.M2M100ForConditionalGeneration,m.M2M100Model,m.M2M100PreTrainedModel,m.M2M100Tokenizer,m.MBart50Tokenizer,m.MBartForCausalLM,m.MBartForConditionalGeneration,m.MBartForSequenceClassification,m.MBartModel,m.MBartPreTrainedModel,m.MBartTokenizer,m.MPNetForMaskedLM,m.MPNetForQuestionAnswering,m.MPNetForSequenceClassification,m.MPNetForTokenClassification,m.MPNetModel,m.MPNetPreTrainedModel,m.MPNetTokenizer,m.MT5ForConditionalGeneration,m.MT5Model,m.MT5PreTrainedModel,m.MarianMTModel,m.MarianModel,m.MarianPreTrainedModel,m.MarianTokenizer,m.Mask2FormerImageProcessor,m.MaskFormerFeatureExtractor,m.MaskFormerForInstanceSegmentation,m.MaskFormerImageProcessor,m.MaskFormerModel,m.MaskFormerPreTrainedModel,m.MaskedLMOutput,m.MaxLengthCriteria,m.Metric3DForDepthEstimation,m.Metric3DPreTrainedModel,m.Metric3Dv2ForDepthEstimation,m.Metric3Dv2PreTrainedModel,m.MgpstrForSceneTextRecognition,m.MgpstrModelOutput,m.MgpstrPreTrainedModel,m.MgpstrProcessor,m.MgpstrTokenizer,m.MimiDecoderModel,m.MimiDecoderOutput,m.MimiEncoderModel,m.MimiEncoderOutput,m.MimiModel,m.MimiPreTrainedModel,m.MinLengthLogitsProcessor,m.MinNewTokensLengthLogitsProcessor,m.Ministral3ForCausalLM,m.Ministral3Model,m.Ministral3PreTrainedModel,m.MinistralForCausalLM,m.MinistralModel,m.MinistralPreTrainedModel,m.Mistral3ForConditionalGeneration,m.MistralForCausalLM,m.MistralModel,m.MistralPreTrainedModel,m.MobileBertForMaskedLM,m.MobileBertForQuestionAnswering,m.MobileBertForSequenceClassification,m.MobileBertModel,m.MobileBertPreTrainedModel,m.MobileBertTokenizer,m.MobileLLMForCausalLM,m.MobileLLMModel,m.MobileLLMPreTrainedModel,m.MobileNetV1FeatureExtractor,m.MobileNetV1ForImageClassification,m.MobileNetV1ForSemanticSegmentation,m.MobileNetV1ImageProcessor,m.MobileNetV1Model,m.MobileNetV1PreTrainedModel,m.MobileNetV2FeatureExtractor,m.MobileNetV2ForImageClassification,m.MobileNetV2ForSemanticSegmentation,m.MobileNetV2ImageProcessor,m.MobileNetV2Model,m.MobileNetV2PreTrainedModel,m.MobileNetV3FeatureExtractor,m.MobileNetV3ForImageClassification,m.MobileNetV3ForSemanticSegmentation,m.MobileNetV3ImageProcessor,m.MobileNetV3Model,m.MobileNetV3PreTrainedModel,m.MobileNetV4FeatureExtractor,m.MobileNetV4ForImageClassification,m.MobileNetV4ForSemanticSegmentation,m.MobileNetV4ImageProcessor,m.MobileNetV4Model,m.MobileNetV4PreTrainedModel,m.MobileViTFeatureExtractor,m.MobileViTForImageClassification,m.MobileViTImageProcessor,m.MobileViTModel,m.MobileViTPreTrainedModel,m.MobileViTV2ForImageClassification,m.MobileViTV2Model,m.MobileViTV2PreTrainedModel,m.ModelOutput,m.ModernBertDecoderForCausalLM,m.ModernBertDecoderModel,m.ModernBertDecoderPreTrainedModel,m.ModernBertForMaskedLM,m.ModernBertForSequenceClassification,m.ModernBertForTokenClassification,m.ModernBertModel,m.ModernBertPreTrainedModel,m.Moondream1ForConditionalGeneration,m.MoonshineFeatureExtractor,m.MoonshineForConditionalGeneration,m.MoonshineModel,m.MoonshinePreTrainedModel,m.MoonshineProcessor,m.MptForCausalLM,m.MptModel,m.MptPreTrainedModel,m.MultiModalityCausalLM,m.MultiModalityPreTrainedModel,m.MusicgenForCausalLM,m.MusicgenForConditionalGeneration,m.MusicgenModel,m.MusicgenPreTrainedModel,m.NanoChatForCausalLM,m.NanoChatModel,m.NanoChatPreTrainedModel,m.NeoBertForMaskedLM,m.NeoBertForQuestionAnswering,m.NeoBertForSequenceClassification,m.NeoBertForTokenClassification,m.NeoBertModel,m.NeoBertPreTrainedModel,m.NllbTokenizer,m.NoBadWordsLogitsProcessor,m.NoRepeatNGramLogitsProcessor,m.NomicBertModel,m.NomicBertPreTrainedModel,m.NougatImageProcessor,m.NougatTokenizer,m.OPTForCausalLM,m.OPTModel,m.OPTPreTrainedModel,m.ObjectDetectionPipeline,m.Olmo2ForCausalLM,m.Olmo2Model,m.Olmo2PreTrainedModel,m.OlmoForCausalLM,m.OlmoModel,m.OlmoPreTrainedModel,m.OpenELMForCausalLM,m.OpenELMModel,m.OpenELMPreTrainedModel,m.OwlViTFeatureExtractor,m.OwlViTForObjectDetection,m.OwlViTImageProcessor,m.OwlViTModel,m.OwlViTPreTrainedModel,m.OwlViTProcessor,m.Owlv2ForObjectDetection,m.Owlv2ImageProcessor,m.Owlv2Model,m.Owlv2PreTrainedModel,m.PaliGemmaForConditionalGeneration,m.PaliGemmaPreTrainedModel,m.PaliGemmaProcessor,m.ParakeetFeatureExtractor,m.ParakeetForCTC,m.ParakeetPreTrainedModel,m.PatchTSMixerForPrediction,m.PatchTSMixerModel,m.PatchTSMixerPreTrainedModel,m.PatchTSTForPrediction,m.PatchTSTModel,m.PatchTSTPreTrainedModel,m.Phi3ForCausalLM,m.Phi3Model,m.Phi3PreTrainedModel,m.Phi3VForCausalLM,m.Phi3VImageProcessor,m.Phi3VPreTrainedModel,m.Phi3VProcessor,m.PhiForCausalLM,m.PhiModel,m.PhiPreTrainedModel,m.Pipeline,m.PixtralImageProcessor,m.PixtralProcessor,m.PreTrainedModel,m.PreTrainedTokenizer,m.PretrainedConfig,m.PretrainedMixin,m.Processor,m.PvtForImageClassification,m.PvtImageProcessor,m.PvtModel,m.PvtPreTrainedModel,m.PyAnnoteFeatureExtractor,m.PyAnnoteForAudioFrameClassification,m.PyAnnoteModel,m.PyAnnotePreTrainedModel,m.PyAnnoteProcessor,m.QuestionAnsweringModelOutput,m.QuestionAnsweringPipeline,m.Qwen2ForCausalLM,m.Qwen2Model,m.Qwen2PreTrainedModel,m.Qwen2Tokenizer,m.Qwen2VLForConditionalGeneration,m.Qwen2VLImageProcessor,m.Qwen2VLPreTrainedModel,m.Qwen2VLProcessor,m.Qwen3ForCausalLM,m.Qwen3Model,m.Qwen3PreTrainedModel,m.RFDetrForObjectDetection,m.RFDetrModel,m.RFDetrObjectDetectionOutput,m.RFDetrPreTrainedModel,m.RTDetrForObjectDetection,m.RTDetrImageProcessor,m.RTDetrModel,m.RTDetrObjectDetectionOutput,m.RTDetrPreTrainedModel,m.RTDetrV2ForObjectDetection,m.RTDetrV2Model,m.RTDetrV2ObjectDetectionOutput,m.RTDetrV2PreTrainedModel,m.RawAudio,m.RawImage,m.RawVideo,m.RawVideoFrame,m.RepetitionPenaltyLogitsProcessor,m.ResNetForImageClassification,m.ResNetModel,m.ResNetPreTrainedModel,m.RoFormerForMaskedLM,m.RoFormerForQuestionAnswering,m.RoFormerForSequenceClassification,m.RoFormerForTokenClassification,m.RoFormerModel,m.RoFormerPreTrainedModel,m.RoFormerTokenizer,m.RobertaForMaskedLM,m.RobertaForQuestionAnswering,m.RobertaForSequenceClassification,m.RobertaForTokenClassification,m.RobertaModel,m.RobertaPreTrainedModel,m.RobertaTokenizer,m.Sam2ImageProcessor,m.Sam2ImageSegmentationOutput,m.Sam2Model,m.Sam2PreTrainedModel,m.Sam2Processor,m.Sam2VideoProcessor,m.Sam3ImageProcessor,m.Sam3TrackerModel,m.SamImageProcessor,m.SamImageSegmentationOutput,m.SamModel,m.SamPreTrainedModel,m.SamProcessor,m.SapiensForDepthEstimation,m.SapiensForNormalEstimation,m.SapiensForSemanticSegmentation,m.SapiensPreTrainedModel,m.SeamlessM4TFeatureExtractor,m.SegformerFeatureExtractor,m.SegformerForImageClassification,m.SegformerForSemanticSegmentation,m.SegformerImageProcessor,m.SegformerModel,m.SegformerPreTrainedModel,m.Seq2SeqLMOutput,m.SequenceClassifierOutput,m.SiglipImageProcessor,m.SiglipModel,m.SiglipPreTrainedModel,m.SiglipTextModel,m.SiglipTokenizer,m.SiglipVisionModel,m.SmolLM3ForCausalLM,m.SmolLM3Model,m.SmolLM3PreTrainedModel,m.SmolVLMForConditionalGeneration,m.SmolVLMImageProcessor,m.SmolVLMProcessor,m.SnacDecoderModel,m.SnacEncoderModel,m.SnacFeatureExtractor,m.SnacModel,m.SnacPreTrainedModel,m.SpeechT5FeatureExtractor,m.SpeechT5ForSpeechToText,m.SpeechT5ForTextToSpeech,m.SpeechT5HifiGan,m.SpeechT5Model,m.SpeechT5PreTrainedModel,m.SpeechT5Processor,m.SpeechT5Tokenizer,m.SqueezeBertForMaskedLM,m.SqueezeBertForQuestionAnswering,m.SqueezeBertForSequenceClassification,m.SqueezeBertModel,m.SqueezeBertPreTrainedModel,m.SqueezeBertTokenizer,m.StableLmForCausalLM,m.StableLmModel,m.StableLmPreTrainedModel,m.Starcoder2ForCausalLM,m.Starcoder2Model,m.Starcoder2PreTrainedModel,m.StoppingCriteria,m.StoppingCriteriaList,m.StyleTextToSpeech2Model,m.StyleTextToSpeech2PreTrainedModel,m.SummarizationPipeline,m.SupertonicForConditionalGeneration,m.SupertonicPreTrainedModel,m.SuppressTokensAtBeginLogitsProcessor,m.Swin2SRForImageSuperResolution,m.Swin2SRImageProcessor,m.Swin2SRModel,m.Swin2SRPreTrainedModel,m.SwinForImageClassification,m.SwinForSemanticSegmentation,m.SwinModel,m.SwinPreTrainedModel,m.T5ForConditionalGeneration,m.T5Model,m.T5PreTrainedModel,m.T5Tokenizer,m.TableTransformerForObjectDetection,m.TableTransformerModel,m.TableTransformerObjectDetectionOutput,m.TableTransformerPreTrainedModel,m.TemperatureLogitsWarper,m.Tensor,m.Text2TextGenerationPipeline,m.TextClassificationPipeline,m.TextGenerationPipeline,m.TextStreamer,m.TextToAudioPipeline,m.TokenClassificationPipeline,m.TokenClassifierOutput,m.TokenizerModel,m.TopKLogitsWarper,m.TopPLogitsWarper,m.TrOCRForCausalLM,m.TrOCRPreTrainedModel,m.TranslationPipeline,m.UltravoxModel,m.UltravoxPreTrainedModel,m.UltravoxProcessor,m.UniSpeechForCTC,m.UniSpeechForSequenceClassification,m.UniSpeechModel,m.UniSpeechPreTrainedModel,m.UniSpeechSatForAudioFrameClassification,m.UniSpeechSatForCTC,m.UniSpeechSatForSequenceClassification,m.UniSpeechSatModel,m.UniSpeechSatPreTrainedModel,m.VLChatProcessor,m.VLMImageProcessor,m.VaultGemmaForCausalLM,m.VaultGemmaModel,m.VaultGemmaPreTrainedModel,m.ViTFeatureExtractor,m.ViTForImageClassification,m.ViTImageProcessor,m.ViTMAEModel,m.ViTMAEPreTrainedModel,m.ViTMSNForImageClassification,m.ViTMSNModel,m.ViTMSNPreTrainedModel,m.ViTModel,m.ViTPreTrainedModel,m.VisionEncoderDecoderModel,m.VitMatteForImageMatting,m.VitMatteImageProcessor,m.VitMattePreTrainedModel,m.VitPoseForPoseEstimation,m.VitPoseImageProcessor,m.VitPosePreTrainedModel,m.VitsModel,m.VitsModelOutput,m.VitsPreTrainedModel,m.VitsTokenizer,m.VoxtralForConditionalGeneration,m.VoxtralProcessor,m.Wav2Vec2BertForCTC,m.Wav2Vec2BertForSequenceClassification,m.Wav2Vec2BertModel,m.Wav2Vec2BertPreTrainedModel,m.Wav2Vec2CTCTokenizer,m.Wav2Vec2FeatureExtractor,m.Wav2Vec2ForAudioFrameClassification,m.Wav2Vec2ForCTC,m.Wav2Vec2ForSequenceClassification,m.Wav2Vec2Model,m.Wav2Vec2PreTrainedModel,m.Wav2Vec2Processor,m.Wav2Vec2ProcessorWithLM,m.WavLMForAudioFrameClassification,m.WavLMForCTC,m.WavLMForSequenceClassification,m.WavLMForXVector,m.WavLMModel,m.WavLMPreTrainedModel,m.WeSpeakerFeatureExtractor,m.WeSpeakerResNetModel,m.WeSpeakerResNetPreTrainedModel,m.WhisperFeatureExtractor,m.WhisperForConditionalGeneration,m.WhisperModel,m.WhisperPreTrainedModel,m.WhisperProcessor,m.WhisperTextStreamer,m.WhisperTimeStampLogitsProcessor,m.WhisperTokenizer,m.XLMForQuestionAnswering,m.XLMForSequenceClassification,m.XLMForTokenClassification,m.XLMModel,m.XLMPreTrainedModel,m.XLMRobertaForMaskedLM,m.XLMRobertaForQuestionAnswering,m.XLMRobertaForSequenceClassification,m.XLMRobertaForTokenClassification,m.XLMRobertaModel,m.XLMRobertaPreTrainedModel,m.XLMRobertaTokenizer,m.XLMTokenizer,m.XLMWithLMHeadModel,m.XVectorOutput,m.YolosFeatureExtractor,m.YolosForObjectDetection,m.YolosImageProcessor,m.YolosModel,m.YolosObjectDetectionOutput,m.YolosPreTrainedModel,m.ZeroShotAudioClassificationPipeline,m.ZeroShotClassificationPipeline,m.ZeroShotImageClassificationPipeline,m.ZeroShotObjectDetectionPipeline,m.bankers_round,m.cat,m.cos_sim,m.dot,m.dynamic_time_warping,m.env,m.full,m.full_like,m.getCacheShapes,m.hamming,m.hanning,m.interpolate,m.interpolate_4d,m.interpolate_data,m.is_chinese_char,m.layer_norm,m.load_image,m.load_video,m.log_softmax,m.magnitude,m.matmul,m.max,m.mean,m.mean_pooling,m.medianFilter,m.mel_filter_bank,m.min,m.ones,m.ones_like,m.permute,m.permute_data,m.pipeline,m.quantize_embeddings,m.rand,m.randn,m.read_audio,m.rfft,m.round,m.slice,m.softmax,m.spectrogram,m.stack,m.std_mean,m.topk,m.window_function,m.zeros,m.zeros_like;const cM=Symbol("Comlink.proxy"),fT=Symbol("Comlink.endpoint"),gT=Symbol("Comlink.releaseProxy"),mc=Symbol("Comlink.finalizer"),ya=Symbol("Comlink.thrown"),uM=e=>typeof e=="object"&&e!==null||typeof e=="function",MT={canHandle:e=>uM(e)&&e[cM],serialize(e){const{port1:r,port2:t}=new MessageChannel;return hc(e,r),[t,[t]]},deserialize(e){return e.start(),vT(e)}},wT={canHandle:e=>uM(e)&&ya in e,serialize({value:e}){let r;return e instanceof Error?r={isError:!0,value:{message:e.message,name:e.name,stack:e.stack}}:r={isError:!1,value:e},[r,[]]},deserialize(e){throw e.isError?Object.assign(new Error(e.value.message),e.value):e.value}},dM=new Map([["proxy",MT],["throw",wT]]);function bT(e,r){for(const t of e)if(r===t||t==="*"||t instanceof RegExp&&t.test(r))return!0;return!1}function hc(e,r=globalThis,t=["*"]){r.addEventListener("message",function s(n){if(!n||!n.data)return;if(!bT(t,n.origin)){console.warn(`Invalid origin '${n.origin}' for comlink proxy`);return}const{id:o,type:i,path:a}=Object.assign({path:[]},n.data),l=(n.data.argumentList||[]).map(wn);let c;try{const p=a.slice(0,-1).reduce((u,f)=>u[f],e),d=a.reduce((u,f)=>u[f],e);switch(i){case"GET":c=d;break;case"SET":p[a.slice(-1)[0]]=wn(n.data.value),c=!0;break;case"APPLY":c=d.apply(p,l);break;case"CONSTRUCT":{const u=new d(...l);c=CT(u)}break;case"ENDPOINT":{const{port1:u,port2:f}=new MessageChannel;hc(e,f),c=ET(u,[u])}break;case"RELEASE":c=void 0;break;default:return}}catch(p){c={value:p,[ya]:0}}Promise.resolve(c).catch(p=>({value:p,[ya]:0})).then(p=>{const[d,u]=Pa(p);r.postMessage(Object.assign(Object.assign({},d),{id:o}),u),i==="RELEASE"&&(r.removeEventListener("message",s),pM(r),mc in e&&typeof e[mc]=="function"&&e[mc]())}).catch(p=>{const[d,u]=Pa({value:new TypeError("Unserializable return value"),[ya]:0});r.postMessage(Object.assign(Object.assign({},d),{id:o}),u)})}),r.start&&r.start()}function yT(e){return e.constructor.name==="MessagePort"}function pM(e){yT(e)&&e.close()}function vT(e,r){const t=new Map;return e.addEventListener("message",function(n){const{data:o}=n;if(!o||!o.id)return;const i=t.get(o.id);if(i)try{i(o)}finally{t.delete(o.id)}}),_c(e,t,[],r)}function va(e){if(e)throw new Error("Proxy has been released and is not useable")}function mM(e){return jn(e,new Map,{type:"RELEASE"}).then(()=>{pM(e)})}const xa=new WeakMap,Ta="FinalizationRegistry"in globalThis&&new FinalizationRegistry(e=>{const r=(xa.get(e)||0)-1;xa.set(e,r),r===0&&mM(e)});function xT(e,r){const t=(xa.get(r)||0)+1;xa.set(r,t),Ta&&Ta.register(e,r,e)}function TT(e){Ta&&Ta.unregister(e)}function _c(e,r,t=[],s=function(){}){let n=!1;const o=new Proxy(s,{get(i,a){if(va(n),a===gT)return()=>{TT(o),mM(e),r.clear(),n=!0};if(a==="then"){if(t.length===0)return{then:()=>o};const l=jn(e,r,{type:"GET",path:t.map(c=>c.toString())}).then(wn);return l.then.bind(l)}return _c(e,r,[...t,a])},set(i,a,l){va(n);const[c,p]=Pa(l);return jn(e,r,{type:"SET",path:[...t,a].map(d=>d.toString()),value:c},p).then(wn)},apply(i,a,l){va(n);const c=t[t.length-1];if(c===fT)return jn(e,r,{type:"ENDPOINT"}).then(wn);if(c==="bind")return _c(e,r,t.slice(0,-1));const[p,d]=hM(l);return jn(e,r,{type:"APPLY",path:t.map(u=>u.toString()),argumentList:p},d).then(wn)},construct(i,a){va(n);const[l,c]=hM(a);return jn(e,r,{type:"CONSTRUCT",path:t.map(p=>p.toString()),argumentList:l},c).then(wn)}});return xT(o,e),o}function PT(e){return Array.prototype.concat.apply([],e)}function hM(e){const r=e.map(Pa);return[r.map(t=>t[0]),PT(r.map(t=>t[1]))]}const _M=new WeakMap;function ET(e,r){return _M.set(e,r),e}function CT(e){return Object.assign(e,{[cM]:!0})}function Pa(e){for(const[r,t]of dM)if(t.canHandle(e)){const[s,n]=t.serialize(e);return[{type:"HANDLER",name:r,value:s},n]}return[{type:"RAW",value:e},_M.get(e)||[]]}function wn(e){switch(e.type){case"HANDLER":return dM.get(e.name).deserialize(e.value);case"RAW":return e.value}}function jn(e,r,t,s){return new Promise(n=>{const o=ST();r.set(o,n),e.start&&e.start(),e.postMessage(Object.assign({id:o},t),s)})}function ST(){return new Array(4).fill(0).map(()=>Math.floor(Math.random()*Number.MAX_SAFE_INTEGER).toString(16)).join("-")}const fM="onnx-community/functiongemma-270m-it-ONNX";let gM=null;function MM(e){return gM??=(async()=>{console.log("[Worker] Loading model...");const[r,t]=await Promise.all([_T.from_pretrained(fM,{progress_callback:e}),hT.from_pretrained(fM,{dtype:"q4f16",device:"webgpu",progress_callback:e})]);return console.log("[Worker] Model loaded!"),{model:t,tokenizer:r}})(),gM}hc({async initModel(e){console.log("[Worker] initModel called"),await MM(e),console.log("[Worker] initModel completed")},async generate(e,r){console.log("[Worker] ========== GENERATE =========="),console.log("[Worker] Messages:",JSON.stringify(e,null,2));const{model:t,tokenizer:s}=await MM(),n=s.apply_chat_template(e,{tools:r,tokenize:!0,add_generation_prompt:!0,return_dict:!0});console.log("[Worker] Input token count:",n.input_ids.dims[1]);const o=await t.generate({...n,max_new_tokens:512}),i=n.input_ids.dims[1],a=s.decode(o.slice(0,[i,null]),{skip_special_tokens:!1});return console.log("[Worker] Raw output:",a),a}})})(); +`:case"\r":return!1;default:return new RegExp("^\\p{Cc}|\\p{Cf}|\\p{Co}|\\p{Cs}$","u").test($)}}_clean_text($){const U=[];for(const ee of $){const se=ee.charCodeAt(0);se===0||se===65533||this._is_control(ee)||(/^\s$/.test(ee)?U.push(" "):U.push(ee))}return U.join("")}normalize($){return this.config.clean_text&&($=this._clean_text($)),this.config.handle_chinese_chars&&($=this._tokenize_chinese_chars($)),this.config.lowercase?($=$.toLowerCase(),this.config.strip_accents!==!1&&($=this.stripAccents($))):this.config.strip_accents&&($=this.stripAccents($)),$}}class Q extends s.Callable{static fromConfig($){if($===null)return null;switch($.type){case"BertPreTokenizer":return new z($);case"Sequence":return new jr($);case"Whitespace":return new Qs($);case"WhitespaceSplit":return new Xs($);case"Metaspace":return new gr($);case"ByteLevel":return new de($);case"Split":return new be($);case"Punctuation":return new ve($);case"Digits":return new xe($);case"Replace":return new Js($);case"FixedLength":return new ar($);default:throw new Error(`Unknown PreTokenizer type: ${$.type}`)}}pre_tokenize_text($,U){throw Error("pre_tokenize_text should be implemented in subclass.")}pre_tokenize($,U){return(Array.isArray($)?$.map(ee=>this.pre_tokenize_text(ee,U)):this.pre_tokenize_text($,U)).flat()}_call($,U){return this.pre_tokenize($,U)}}class z extends Q{constructor($){super(),this.pattern=new RegExp(`[^\\s${P}]+|[${P}]`,"gu")}pre_tokenize_text($,U){return $.trim().match(this.pattern)||[]}}class de extends Q{constructor($){super(),this.config=$,this.add_prefix_space=this.config.add_prefix_space,this.trim_offsets=this.config.trim_offsets,this.use_regex=this.config.use_regex??!0,this.pattern=new RegExp("'s|'t|'re|'ve|'m|'ll|'d| ?\\p{L}+| ?\\p{N}+| ?[^\\s\\p{L}\\p{N}]+|\\s+(?!\\S)|\\s+","gu"),this.byte_encoder=Y,this.text_encoder=new TextEncoder}pre_tokenize_text($,U){return this.add_prefix_space&&!$.startsWith(" ")&&($=" "+$),(this.use_regex?$.match(this.pattern)||[]:[$]).map(se=>Array.from(this.text_encoder.encode(se),Me=>this.byte_encoder[Me]).join(""))}}class be extends Q{constructor($){super(),this.config=$,this.pattern=f(this.config.pattern,this.config.invert)}pre_tokenize_text($,U){return this.pattern===null?[]:this.config.invert?$.match(this.pattern)||[]:this.config.behavior?.toLowerCase()==="removed"?$.split(this.pattern).filter(ee=>ee):u($,this.pattern)}}class ve extends Q{constructor($){super(),this.config=$,this.pattern=new RegExp(`[^${P}]+|[${P}]+`,"gu")}pre_tokenize_text($,U){return $.match(this.pattern)||[]}}class xe extends Q{constructor($){super(),this.config=$;const U=`[^\\d]+|\\d${this.config.individual_digits?"":"+"}`;this.pattern=new RegExp(U,"gu")}pre_tokenize_text($,U){return $.match(this.pattern)||[]}}class Ce extends s.Callable{constructor($){super(),this.config=$}static fromConfig($){if($===null)return null;switch($.type){case"TemplateProcessing":return new fe($);case"ByteLevel":return new Pe($);case"RobertaProcessing":return new De($);case"BertProcessing":return new ge($);case"Sequence":return new We($);default:throw new Error(`Unknown PostProcessor type: ${$.type}`)}}post_process($,...U){throw Error("post_process should be implemented in subclass.")}_call($,...U){return this.post_process($,...U)}}class ge extends Ce{constructor($){super($),this.cls=$.cls[0],this.sep=$.sep[0]}post_process($,U=null,{add_special_tokens:ee=!0}={}){ee&&($=(0,n.mergeArrays)([this.cls],$,[this.sep]));let se=new Array($.length).fill(0);if(U!==null){const Me=ee&&this instanceof De?[this.sep]:[],$e=ee?[this.sep]:[];$=(0,n.mergeArrays)($,Me,U,$e),se=(0,n.mergeArrays)(se,new Array(U.length+Me.length+$e.length).fill(1))}return{tokens:$,token_type_ids:se}}}class De extends ge{}class fe extends Ce{constructor($){super($),this.single=$.single,this.pair=$.pair}post_process($,U=null,{add_special_tokens:ee=!0}={}){const se=U===null?this.single:this.pair;let Me=[],$e=[];for(const Xe of se)"SpecialToken"in Xe?ee&&(Me.push(Xe.SpecialToken.id),$e.push(Xe.SpecialToken.type_id)):"Sequence"in Xe&&(Xe.Sequence.id==="A"?(Me=(0,n.mergeArrays)(Me,$),$e=(0,n.mergeArrays)($e,new Array($.length).fill(Xe.Sequence.type_id))):Xe.Sequence.id==="B"&&(Me=(0,n.mergeArrays)(Me,U),$e=(0,n.mergeArrays)($e,new Array(U.length).fill(Xe.Sequence.type_id))));return{tokens:Me,token_type_ids:$e}}}class Pe extends Ce{post_process($,U=null){return U&&($=(0,n.mergeArrays)($,U)),{tokens:$}}}class We extends Ce{constructor($){super($),this.processors=$.processors.map(U=>Ce.fromConfig(U))}post_process($,U=null,ee={}){let se;for(const Me of this.processors)if(Me instanceof Pe)$=Me.post_process($).tokens,U&&(U=Me.post_process(U).tokens);else{const $e=Me.post_process($,U,ee);$=$e.tokens,se=$e.token_type_ids}return{tokens:$,token_type_ids:se}}}class Fe extends s.Callable{constructor($){super(),this.config=$,this.added_tokens=[],this.end_of_word_suffix=null,this.trim_offsets=$.trim_offsets}static fromConfig($){if($===null)return null;switch($.type){case"WordPiece":return new je($);case"Metaspace":return new Or($);case"ByteLevel":return new Oe($);case"Replace":return new tt($);case"ByteFallback":return new Re($);case"Fuse":return new rt($);case"Strip":return new Ze($);case"Sequence":return new ht($);case"CTC":return new at($);case"BPEDecoder":return new Nt($);default:throw new Error(`Unknown Decoder type: ${$.type}`)}}_call($){return this.decode($)}decode($){return this.decode_chain($).join("")}decode_chain($){throw Error("`decode_chain` should be implemented in subclass.")}}class tt extends Fe{decode_chain($){const U=f(this.config.pattern);return U===null?$:$.map(ee=>ee.replaceAll(U,this.config.content))}}class Re extends Fe{constructor($){super($),this.text_decoder=new TextDecoder}decode_chain($){const U=[];let ee=[];for(const se of $){let Me=null;if(se.length===6&&se.startsWith("<0x")&&se.endsWith(">")){const $e=parseInt(se.slice(3,5),16);isNaN($e)||(Me=$e)}if(Me!==null)ee.push(Me);else{if(ee.length>0){const $e=this.text_decoder.decode(Uint8Array.from(ee));U.push($e),ee=[]}U.push(se)}}if(ee.length>0){const se=this.text_decoder.decode(Uint8Array.from(ee));U.push(se),ee=[]}return U}}class rt extends Fe{decode_chain($){return[$.join("")]}}class Ze extends Fe{constructor($){super($),this.content=this.config.content,this.start=this.config.start,this.stop=this.config.stop}decode_chain($){return $.map(U=>{let ee=0;for(let Me=0;Me(ee!==0&&(U.startsWith(this.config.prefix)?U=U.replace(this.config.prefix,""):U=" "+U),this.cleanup&&(U=I(U)),U))}}class Oe extends Fe{constructor($){super($),this.byte_decoder=X,this.text_decoder=new TextDecoder("utf-8",{fatal:!1,ignoreBOM:!0}),this.end_of_word_suffix=null}convert_tokens_to_string($){const U=$.join(""),ee=new Uint8Array([...U].map(Me=>this.byte_decoder[Me]));return this.text_decoder.decode(ee)}decode_chain($){const U=[];let ee=[];for(const se of $)this.added_tokens.find(Me=>Me.content===se)!==void 0?(ee.length>0&&(U.push(this.convert_tokens_to_string(ee)),ee=[]),U.push(se)):ee.push(se);return ee.length>0&&U.push(this.convert_tokens_to_string(ee)),U}}class at extends Fe{constructor($){super($),this.pad_token=this.config.pad_token,this.word_delimiter_token=this.config.word_delimiter_token,this.cleanup=this.config.cleanup}convert_tokens_to_string($){if($.length===0)return"";const U=[$[0]];for(let Me=1;Me<$.length;++Me)$[Me]!==U.at(-1)&&U.push($[Me]);let se=U.filter(Me=>Me!==this.pad_token).join("");return this.cleanup&&(se=I(se).replaceAll(this.word_delimiter_token," ").trim()),se}decode_chain($){return[this.convert_tokens_to_string($)]}}class ht extends Fe{constructor($){super($),this.decoders=$.decoders.map(U=>Fe.fromConfig(U))}decode_chain($){return this.decoders.reduce((U,ee)=>ee.decode_chain(U),$)}}class Nt extends Fe{constructor($){super($),this.suffix=this.config.suffix}decode_chain($){return $.map((U,ee)=>U.replaceAll(this.suffix,ee===$.length-1?"":" "))}}class kt extends Fe{decode_chain($){let U="";for(let ee=1;ee<$.length;ee+=2)U+=$[ee];return[U]}}class gr extends Q{constructor($){super(),this.replacement=$.replacement,this.strRep=$.str_rep||this.replacement,this.prepend_scheme=$.prepend_scheme??"always"}pre_tokenize_text($,{section_index:U=void 0}={}){let ee=$.replaceAll(" ",this.strRep);return!ee.startsWith(this.replacement)&&(this.prepend_scheme==="always"||this.prepend_scheme==="first"&&U===0)&&(ee=this.strRep+ee),[ee]}}class Or extends Fe{constructor($){super($),this.replacement=$.replacement}decode_chain($){const U=[];for(let ee=0;ee<$.length;++ee){let se=$[ee].replaceAll(this.replacement," ");ee==0&&se.startsWith(" ")&&(se=se.substring(1)),U.push(se)}return U}}class Bt extends ne{constructor($){super($),this.charsmap=$.precompiled_charsmap}normalize($){return $=$.replace(/[\u0001-\u0008\u000B\u000E-\u001F\u007F\u008F\u009F]/gm,""),$=$.replace(/[\u0009\u000A\u000C\u000D\u00A0\u1680\u2000-\u200F\u2028\u2029\u202F\u205F\u2581\u3000\uFEFF\uFFFD]/gm," "),$.includes("~")?$=$.split("~").map(ee=>ee.normalize("NFKC")).join("~"):$=$.normalize("NFKC"),$}}class jr extends Q{constructor($){super(),this.tokenizers=$.pretokenizers.map(U=>Q.fromConfig(U))}pre_tokenize_text($,U){return this.tokenizers.reduce((ee,se)=>se.pre_tokenize(ee,U),[$])}}class Qs extends Q{constructor($){super()}pre_tokenize_text($,U){return $.match(/\w+|[^\w\s]+/g)||[]}}class Xs extends Q{constructor($){super()}pre_tokenize_text($,U){return b($)}}class Js extends Q{constructor($){super(),this.config=$,this.pattern=f(this.config.pattern),this.content=this.config.content}pre_tokenize_text($,U){return this.pattern===null?[$]:[$.replaceAll(this.pattern,this.config.content)]}}class ar extends Q{constructor($){super(),this._length=$.length}pre_tokenize_text($,U){const ee=[];for(let se=0;se<$.length;se+=this._length)ee.push($.slice(se,se+this._length));return ee}}const kr=["bos_token","eos_token","unk_token","sep_token","pad_token","cls_token","mask_token"];function Jr(ue,$,U,ee){for(const se of Object.keys(ue)){const Me=$-ue[se].length,$e=U(se),Xe=new Array(Me).fill($e);ue[se]=ee==="right"?(0,n.mergeArrays)(ue[se],Xe):(0,n.mergeArrays)(Xe,ue[se])}}function Bs(ue,$){for(const U of Object.keys(ue))ue[U].length=$}class ft extends s.Callable{return_token_type_ids=!1;padding_side="right";constructor($,U){super(),this.config=U,this.normalizer=ne.fromConfig($.normalizer),this.pre_tokenizer=Q.fromConfig($.pre_tokenizer),this.model=H.fromConfig($.model,U),this.post_processor=Ce.fromConfig($.post_processor),this.decoder=Fe.fromConfig($.decoder),this.special_tokens=[],this.all_special_ids=[],this.added_tokens=[];for(const ee of $.added_tokens){const se=new F(ee);this.added_tokens.push(se),this.model.tokens_to_ids.set(se.content,se.id),this.model.vocab[se.id]=se.content,se.special&&(this.special_tokens.push(se.content),this.all_special_ids.push(se.id))}if(this.additional_special_tokens=U.additional_special_tokens??[],this.special_tokens.push(...this.additional_special_tokens),this.special_tokens=[...new Set(this.special_tokens)],this.decoder&&(this.decoder.added_tokens=this.added_tokens,this.decoder.end_of_word_suffix=this.model.end_of_word_suffix),this.added_tokens_splitter=new l.DictionarySplitter(this.added_tokens.map(ee=>ee.content)),this.added_tokens_map=new Map(this.added_tokens.map(ee=>[ee.content,ee])),this.mask_token=this.getToken("mask_token"),this.mask_token_id=this.model.tokens_to_ids.get(this.mask_token),this.pad_token=this.getToken("pad_token","eos_token"),this.pad_token_id=this.model.tokens_to_ids.get(this.pad_token),this.sep_token=this.getToken("sep_token"),this.sep_token_id=this.model.tokens_to_ids.get(this.sep_token),this.unk_token=this.getToken("unk_token"),this.unk_token_id=this.model.tokens_to_ids.get(this.unk_token),this.bos_token=this.getToken("bos_token"),this.bos_token_id=this.model.tokens_to_ids.get(this.bos_token),this.eos_token=this.getToken("eos_token"),this.eos_token_id=this.model.tokens_to_ids.get(this.eos_token),this.model_max_length=U.model_max_length,this.remove_space=U.remove_space,this.clean_up_tokenization_spaces=U.clean_up_tokenization_spaces??!0,this.do_lowercase_and_remove_accent=U.do_lowercase_and_remove_accent??!1,U.padding_side&&(this.padding_side=U.padding_side),this.add_bos_token=U.add_bos_token,this.add_eos_token=U.add_eos_token,this.legacy=!1,this.chat_template=U.chat_template??null,Array.isArray(this.chat_template)){const ee=Object.create(null);for(const{name:se,template:Me}of this.chat_template){if(typeof se!="string"||typeof Me!="string")throw new Error('Chat template must be a list of objects with "name" and "template" properties');ee[se]=Me}this.chat_template=ee}this._compiled_template_cache=new Map}getToken(...$){for(const U of $){const ee=this.config[U];if(ee)if(typeof ee=="object"){if(ee.__type==="AddedToken")return ee.content;throw Error(`Unknown token: ${ee}`)}else return ee}return null}static async from_pretrained($,{progress_callback:U=null,config:ee=null,cache_dir:se=null,local_files_only:Me=!1,revision:$e="main",legacy:Xe=null}={}){const Je=await d($,{progress_callback:U,config:ee,cache_dir:se,local_files_only:Me,revision:$e,legacy:Xe});return new this(...Je)}_call($,{text_pair:U=null,add_special_tokens:ee=!0,padding:se=!1,truncation:Me=null,max_length:$e=null,return_tensor:Xe=!0,return_token_type_ids:Je=null}={}){const Ye=Array.isArray($);let Ke;if(Ye){if($.length===0)throw Error("text array must be non-empty");if(U!==null){if(Array.isArray(U)){if($.length!==U.length)throw Error("text and text_pair must have the same length")}else throw Error("text_pair must also be an array");Ke=$.map((Et,rr)=>this._encode_plus(Et,{text_pair:U[rr],add_special_tokens:ee,return_token_type_ids:Je}))}else Ke=$.map(Et=>this._encode_plus(Et,{add_special_tokens:ee,return_token_type_ids:Je}))}else{if($==null)throw Error("text may not be null or undefined");if(Array.isArray(U))throw Error("When specifying `text_pair`, since `text` is a string, `text_pair` must also be a string (i.e., not an array).");Ke=[this._encode_plus($,{text_pair:U,add_special_tokens:ee,return_token_type_ids:Je})]}if($e===null?$e=this.model_max_length:Me===null&&(se===!0?(console.warn("`max_length` is ignored when `padding: true` and there is no truncation strategy. To pad to max length, use `padding: 'max_length'`."),$e=this.model_max_length):se===!1&&(console.warn("Truncation was not explicitly activated but `max_length` is provided a specific value, please use `truncation: true` to explicitly truncate examples to max length."),Me=!0)),se===!0&&($e=Math.min((0,i.max)(Ke.map(Et=>Et.input_ids.length))[0],$e??1/0)),$e=Math.min($e,this.model_max_length??1/0),se||Me)for(let Et=0;Et$e?Me&&Bs(Ke[Et],$e):se&&Jr(Ke[Et],$e,rr=>rr==="input_ids"?this.pad_token_id:0,this.padding_side));const $t={};if(Xe){if(!(se&&Me)&&Ke.some(rr=>{for(const br of Object.keys(rr))if(rr[br].length!==Ke[0][br]?.length)return!0;return!1}))throw Error("Unable to create tensor, you should probably activate truncation and/or padding with 'padding=true' and 'truncation=true' to have batched tensors with the same length.");const Et=[Ke.length,Ke[0].input_ids.length];for(const rr of Object.keys(Ke[0]))$t[rr]=new a.Tensor("int64",BigInt64Array.from(Ke.flatMap(br=>br[rr]).map(BigInt)),Et)}else{for(const Et of Object.keys(Ke[0]))$t[Et]=Ke.map(rr=>rr[Et]);if(!Ye)for(const Et of Object.keys($t))$t[Et]=$t[Et][0]}return $t}_encode_text($){if($===null)return null;const U=this.added_tokens_splitter.split($);for(let se=0;se0&&(U[se-1]=U[se-1].trimEnd()),Me.rstrip&&se{if(se.length===0)return[];if(this.added_tokens_map.has(se))return[se];if(this.remove_space===!0&&(se=se.trim().split(/\s+/).join(" ")),this.do_lowercase_and_remove_accent&&(se=v(se)),this.normalizer!==null&&(se=this.normalizer(se)),se.length===0)return[];const $e=this.pre_tokenizer!==null?this.pre_tokenizer(se,{section_index:Me}):[se];return this.model($e)})}_encode_plus($,{text_pair:U=null,add_special_tokens:ee=!0,return_token_type_ids:se=null}={}){const{tokens:Me,token_type_ids:$e}=this._tokenize_helper($,{pair:U,add_special_tokens:ee}),Xe=this.model.convert_tokens_to_ids(Me),Je={input_ids:Xe,attention_mask:new Array(Xe.length).fill(1)};return(se??this.return_token_type_ids)&&$e&&(Je.token_type_ids=$e),Je}_tokenize_helper($,{pair:U=null,add_special_tokens:ee=!1}={}){const se=this._encode_text($),Me=this._encode_text(U);return this.post_processor?this.post_processor(se,Me,{add_special_tokens:ee}):{tokens:(0,n.mergeArrays)(se??[],Me??[])}}tokenize($,{pair:U=null,add_special_tokens:ee=!1}={}){return this._tokenize_helper($,{pair:U,add_special_tokens:ee}).tokens}encode($,{text_pair:U=null,add_special_tokens:ee=!0,return_token_type_ids:se=null}={}){return this._encode_plus($,{text_pair:U,add_special_tokens:ee,return_token_type_ids:se}).input_ids}batch_decode($,U={}){return $ instanceof a.Tensor&&($=$.tolist()),$.map(ee=>this.decode(ee,U))}decode($,U={}){if($ instanceof a.Tensor&&($=y($)),!Array.isArray($)||$.length===0||!(0,n.isIntegralNumber)($[0]))throw Error("token_ids must be a non-empty array of integers.");return this.decode_single($,U)}decode_single($,{skip_special_tokens:U=!1,clean_up_tokenization_spaces:ee=null}){let se=this.model.convert_ids_to_tokens($);U&&(se=se.filter($e=>!this.special_tokens.includes($e)));let Me=this.decoder?this.decoder(se):se.join(" ");return this.decoder&&this.decoder.end_of_word_suffix&&(Me=Me.replaceAll(this.decoder.end_of_word_suffix," "),U&&(Me=Me.trim())),(ee??this.clean_up_tokenization_spaces)&&(Me=I(Me)),Me}get_chat_template({chat_template:$=null,tools:U=null}={}){if(this.chat_template&&typeof this.chat_template=="object"){const ee=this.chat_template;if($!==null&&Object.hasOwn(ee,$))$=ee[$];else if($===null)if(U!==null&&"tool_use"in ee)$=ee.tool_use;else if("default"in ee)$=ee.default;else throw Error(`This model has multiple chat templates with no default specified! Please either pass a chat template or the name of the template you wish to use to the 'chat_template' argument. Available template names are ${Object.keys(ee).sort()}.`)}else if($===null)if(this.chat_template)$=this.chat_template;else throw Error("Cannot use apply_chat_template() because tokenizer.chat_template is not set and no template argument was passed! For information about writing templates and setting the tokenizer.chat_template attribute, please see the documentation at https://huggingface.co/docs/transformers/main/en/chat_templating");return $}apply_chat_template($,{tools:U=null,documents:ee=null,chat_template:se=null,add_generation_prompt:Me=!1,tokenize:$e=!0,padding:Xe=!1,truncation:Je=!1,max_length:Ye=null,return_tensor:Ke=!0,return_dict:$t=!1,tokenizer_kwargs:Et={},...rr}={}){if(se=this.get_chat_template({chat_template:se,tools:U}),typeof se!="string")throw Error(`chat_template must be a string, but got ${typeof se}`);let br=this._compiled_template_cache.get(se);br===void 0&&(br=new c.Template(se),this._compiled_template_cache.set(se,br));const Jt=Object.create(null);for(const Ht of kr){const Hr=this.getToken(Ht);Hr&&(Jt[Ht]=Hr)}const sr=br.render({messages:$,add_generation_prompt:Me,tools:U,documents:ee,...Jt,...rr});if($e){const Ht=this._call(sr,{add_special_tokens:!1,padding:Xe,truncation:Je,max_length:Ye,return_tensor:Ke,...Et});return $t?Ht:Ht.input_ids}return sr}}class qt extends ft{return_token_type_ids=!0}class Ts extends ft{return_token_type_ids=!0}class Ps extends ft{return_token_type_ids=!0}class Gr extends ft{return_token_type_ids=!0}class yt extends ft{return_token_type_ids=!0}class Es extends ft{return_token_type_ids=!0}class C extends ft{return_token_type_ids=!0}class q extends ft{return_token_type_ids=!0}class R extends ft{return_token_type_ids=!0}class G extends ft{}class Z extends ft{}class ce extends ft{return_token_type_ids=!0;constructor($,U){super($,U),console.warn('WARNING: `XLMTokenizer` is not yet supported by Hugging Face\'s "fast" tokenizers library. Therefore, you may experience slightly inaccurate results.')}}class ye extends ft{return_token_type_ids=!0}class et extends ft{}class ut extends ft{}class He extends ft{}class Mt extends ft{constructor($,U){super($,U),this.languageRegex=/^[a-z]{2}_[A-Z]{2}$/,this.language_codes=this.special_tokens.filter(ee=>this.languageRegex.test(ee)),this.lang_to_token=ee=>ee}_build_translation_inputs($,U,ee){return lr(this,$,U,ee)}}class qe extends Mt{}class Pt extends ft{}class It extends ft{}const Mr="▁";class pr extends ft{padding_side="left";constructor($,U){super($,U),this.legacy=U.legacy??!0,this.legacy||(this.normalizer=null,this.pre_tokenizer=new gr({replacement:Mr,prepend_scheme:"first"}))}_encode_text($){if($===null)return null;if(this.legacy||$.length===0)return super._encode_text($);let U=super._encode_text(Mr+$.replaceAll(Mr," "));return U.length>1&&U[0]===Mr&&this.special_tokens.includes(U[1])&&(U=U.slice(1)),U}}class ir extends ft{}class Tr extends ft{}class Cs extends ft{}class Dr extends ft{}class Ss extends ft{}class Lr extends ft{}class zr extends ft{}class ns extends ft{}class wr extends ft{}function lr(ue,$,U,ee){if(!("language_codes"in ue)||!Array.isArray(ue.language_codes))throw new Error("Tokenizer must have `language_codes` attribute set and it should be an array of language ids.");if(!("languageRegex"in ue)||!(ue.languageRegex instanceof RegExp))throw new Error("Tokenizer must have `languageRegex` attribute set and it should be a regular expression.");if(!("lang_to_token"in ue)||typeof ue.lang_to_token!="function")throw new Error("Tokenizer must have `lang_to_token` attribute set and it should be a function.");const se=ee.src_lang,Me=ee.tgt_lang;if(!ue.language_codes.includes(Me))throw new Error(`Target language code "${Me}" is not valid. Must be one of: {${ue.language_codes.join(", ")}}`);if(se!==void 0){if(!ue.language_codes.includes(se))throw new Error(`Source language code "${se}" is not valid. Must be one of: {${ue.language_codes.join(", ")}}`);for(const $e of ue.post_processor.config.single)if("SpecialToken"in $e&&ue.languageRegex.test($e.SpecialToken.id)){$e.SpecialToken.id=ue.lang_to_token(se);break}}return ee.forced_bos_token_id=ue.model.convert_tokens_to_ids([ue.lang_to_token(Me)])[0],ue._call($,U)}class Kr extends ft{constructor($,U){super($,U),this.languageRegex=/^[a-z]{3}_[A-Z][a-z]{3}$/,this.language_codes=this.special_tokens.filter(ee=>this.languageRegex.test(ee)),this.lang_to_token=ee=>ee}_build_translation_inputs($,U,ee){return lr(this,$,U,ee)}}class os extends ft{constructor($,U){super($,U),this.languageRegex=/^__[a-z]{2,3}__$/,this.language_codes=this.special_tokens.filter(ee=>this.languageRegex.test(ee)).map(ee=>ee.slice(2,-2)),this.lang_to_token=ee=>`__${ee}__`}_build_translation_inputs($,U,ee){return lr(this,$,U,ee)}}class Rs extends ft{get timestamp_begin(){return this.model.convert_tokens_to_ids(["<|notimestamps|>"])[0]+1}_decode_asr($,{return_timestamps:U=!1,return_language:ee=!1,time_precision:se=null,force_full_sequences:Me=!0}={}){if(se===null)throw Error("Must specify time_precision");let $e=null;const Xe=U==="word";function Je(){return{language:$e,timestamp:[null,null],text:""}}const Ye=[];let Ke=Je(),$t=0;const Et=this.timestamp_begin,br=Et+1500;let Jt=[],sr=[],Ht=!1,Hr=null;const is=new Set(this.all_special_ids);for(const Yt of $){const hr=Yt.tokens,$r=Xe?Yt.token_timestamps:null;let Yr=null,bs=Et;if("stride"in Yt){const[vr,Zt,_r]=Yt.stride;if($t-=Zt,Hr=vr-_r,Zt&&(bs=Zt/se+Et),_r)for(let cr=hr.length-1;cr>=0;--cr){const Ur=Number(hr[cr]);if(Ur>=Et){if(Yr!==null&&(Ur-Et)*se=Et&&Zt<=br){const _r=(Zt-Et)*se+$t,cr=(0,i.round)(_r,2);if(Yr!==null&&Zt>=Yr)Ht=!0;else if(Ht||Jt.length>0&&Zt0?(Jt.push(yr),Xe&&sr.push(ls)):Jt.every(vr=>vr.length===0)&&(Ke=Je(),Jt=[],yr=[],sr=[],ls=[])}if(Jt.length>0){if(Me&&U)throw new Error("Whisper did not predict an ending timestamp, which can happen if audio is cut off in the middle of a word. Also make sure WhisperTimeStampLogitsProcessor was used during generation.");const[Yt,hr]=this.findLongestCommonSequence(Jt,sr),$r=this.decode(Yt);Ke.text=$r,Xe&&(Ke.words=this.collateWordTimestamps(Yt,hr,$e)),Ye.push(Ke)}let Pr=Object.create(null);const ws=Ye.map(Yt=>Yt.text).join("");if(U||ee){for(let Yt=0;Yt0;let Xe=$e?[]:null,Je=$e?U[0]:null;for(let Ye=1;Ye<$.length;++Ye){const Ke=$[Ye];let $t=0,Et=[se,se,0,0];const rr=Ke.length;for(let Pr=1;PrZt===bs[_r]&&Je[ws+_r]<=U[Ye][$r+_r]).length:yr=hr.filter((Zt,_r)=>Zt===bs[_r]).length;const ls=Pr/1e4,vr=yr/Pr+ls;yr>1&&vr>$t&&($t=vr,Et=[ws,Yt,$r,Yr])}const[br,Jt,sr,Ht]=Et,Hr=Math.floor((Jt+br)/2),is=Math.floor((Ht+sr)/2);Me.push(...ee.slice(0,Hr)),ee=Ke.slice(is),se=ee.length,$e&&(Xe.push(...Je.slice(0,Hr)),Je=U[Ye].slice(is))}return Me.push(...ee),$e?(Xe.push(...Je),[Me,Xe]):[Me,[]]}collateWordTimestamps($,U,ee){const[se,Me,$e]=this.combineTokensIntoWords($,ee),Xe=[];for(let Je=0;Je=se){const Xe=(($e-se)*ee).toFixed(2);Me.push(`<|${Xe}|>`),Me.push([])}else Me[Me.length-1].push($e);return Me=Me.map($e=>typeof $e=="string"?$e:super.decode($e,U)),Me.join("")}splitTokensOnUnicode($){const U=this.decode($,{decode_with_timestamps:!0}),ee="�",se=[],Me=[],$e=[];let Xe=[],Je=[],Ye=0;for(let Ke=0;Ke<$.length;++Ke){const $t=$[Ke];Xe.push($t),Je.push(Ke);const Et=this.decode(Xe,{decode_with_timestamps:!0});(!Et.includes(ee)||U[Ye+Et.indexOf(ee)]===ee)&&(se.push(Et),Me.push(Xe),$e.push(Je),Xe=[],Je=[],Ye+=Et.length)}return[se,Me,$e]}splitTokensOnSpaces($){const[U,ee,se]=this.splitTokensOnUnicode($),Me=[],$e=[],Xe=[],Je=new RegExp(`^[${P}]$`,"gu");for(let Ye=0;Ye=this.model.tokens_to_ids.get("<|endoftext|>"),br=Ke.startsWith(" "),Jt=Ke.trim(),sr=Je.test(Jt);if(rr||br||sr||Me.length===0)Me.push(Ke),$e.push($t),Xe.push(Et);else{const Ht=Me.length-1;Me[Ht]+=Ke,$e[Ht].push(...$t),Xe[Ht].push(...Et)}}return[Me,$e,Xe]}mergePunctuations($,U,ee,se,Me){const $e=structuredClone($),Xe=structuredClone(U),Je=structuredClone(ee);let Ye=$e.length-2,Ke=$e.length-1;for(;Ye>=0;)$e[Ye].startsWith(" ")&&se.includes($e[Ye].trim())?($e[Ke]=$e[Ye]+$e[Ke],Xe[Ke]=(0,n.mergeArrays)(Xe[Ye],Xe[Ke]),Je[Ke]=(0,n.mergeArrays)(Je[Ye],Je[Ke]),$e[Ye]="",Xe[Ye]=[],Je[Ye]=[]):Ke=Ye,--Ye;for(Ye=0,Ke=1;Ke<$e.length;)!$e[Ye].endsWith(" ")&&Me.includes($e[Ke])?($e[Ye]+=$e[Ke],Xe[Ye]=(0,n.mergeArrays)(Xe[Ye],Xe[Ke]),Je[Ye]=(0,n.mergeArrays)(Je[Ye],Je[Ke]),$e[Ke]="",Xe[Ke]=[],Je[Ke]=[]):Ye=Ke,++Ke;return[$e.filter($t=>$t),Xe.filter($t=>$t.length>0),Je.filter($t=>$t.length>0)]}}class ks extends ft{}class $s extends ft{}class Is extends ft{}class as extends ft{constructor($,U){super($,U),this.languageRegex=/^(>>\w+<<)\s*/g,this.supported_language_codes=this.model.vocab.filter(ee=>this.languageRegex.test(ee)),console.warn('WARNING: `MarianTokenizer` is not yet supported by Hugging Face\'s "fast" tokenizers library. Therefore, you may experience slightly inaccurate results.')}_encode_text($){if($===null)return null;const[U,...ee]=$.trim().split(this.languageRegex);if(ee.length===0)return super._encode_text(U);if(ee.length===2){const[se,Me]=ee;return this.supported_language_codes.includes(se)||console.warn(`Unsupported language code "${se}" detected, which may lead to unexpected behavior. Should be one of: ${JSON.stringify(this.supported_language_codes)}`),(0,n.mergeArrays)([se],super._encode_text(Me))}}}class Nr extends ft{}class ze extends ft{}class Ue extends ft{}class nt extends ft{}class Kt extends ft{}class js extends ft{constructor($,U){super($,U),this.decoder=new kt({})}}class As extends ft{}class Ns extends ft{}class Nn{static TOKENIZER_CLASS_MAPPING={T5Tokenizer:et,DistilBertTokenizer:G,CamembertTokenizer:Z,DebertaTokenizer:yt,DebertaV2Tokenizer:Es,BertTokenizer:qt,HerbertTokenizer:C,ConvBertTokenizer:q,RoFormerTokenizer:R,XLMTokenizer:ce,ElectraTokenizer:ye,MobileBertTokenizer:Ps,SqueezeBertTokenizer:Gr,AlbertTokenizer:Ts,GPT2Tokenizer:ut,BartTokenizer:He,MBartTokenizer:Mt,MBart50Tokenizer:qe,RobertaTokenizer:Pt,WhisperTokenizer:Rs,CodeGenTokenizer:ks,CLIPTokenizer:$s,SiglipTokenizer:Is,MarianTokenizer:as,BloomTokenizer:It,NllbTokenizer:Kr,M2M100Tokenizer:os,LlamaTokenizer:pr,CodeLlamaTokenizer:ir,XLMRobertaTokenizer:Tr,MPNetTokenizer:Cs,FalconTokenizer:Dr,GPTNeoXTokenizer:Ss,EsmTokenizer:Lr,Wav2Vec2CTCTokenizer:Nr,BlenderbotTokenizer:ze,BlenderbotSmallTokenizer:Ue,SpeechT5Tokenizer:nt,NougatTokenizer:Kt,VitsTokenizer:js,Qwen2Tokenizer:zr,GemmaTokenizer:ns,Grok1Tokenizer:wr,CohereTokenizer:As,MgpstrTokenizer:Ns,PreTrainedTokenizer:ft};static async from_pretrained($,{progress_callback:U=null,config:ee=null,cache_dir:se=null,local_files_only:Me=!1,revision:$e="main",legacy:Xe=null}={}){const[Je,Ye]=await d($,{progress_callback:U,config:ee,cache_dir:se,local_files_only:Me,revision:$e,legacy:Xe}),Ke=Ye.tokenizer_class?.replace(/Fast$/,"")??"PreTrainedTokenizer";let $t=this.TOKENIZER_CLASS_MAPPING[Ke];return $t||(console.warn(`Unknown tokenizer class "${Ke}", attempting to construct from base class.`),$t=ft),new $t(Je,Ye)}}}),"./src/utils/audio.js":((e,r,t)=>{t.r(r),t.d(r,{RawAudio:()=>W,hamming:()=>u,hanning:()=>d,mel_filter_bank:()=>k,read_audio:()=>c,spectrogram:()=>S,window_function:()=>O});var s=t("./src/utils/hub.js"),n=t("./src/utils/maths.js"),o=t("./src/utils/core.js"),i=t("./src/env.js"),a=t("./src/utils/tensor.js"),l=t("?7992");async function c(B,Y){if(typeof AudioContext>"u")throw Error("Unable to load audio from path/URL since `AudioContext` is not available in your environment. Instead, audio data should be passed directly to the pipeline/processor. For more information and some example code, see https://huggingface.co/docs/transformers.js/guides/node-audio-processing.");const X=await(await(0,s.getFile)(B)).arrayBuffer(),J=new AudioContext({sampleRate:Y});typeof Y>"u"&&console.warn(`No sampling rate provided, using default of ${J.sampleRate}Hz.`);const re=await J.decodeAudioData(X);let ne;if(re.numberOfChannels===2){const le=Math.sqrt(2),pe=re.getChannelData(0),oe=re.getChannelData(1);ne=new Float32Array(pe.length);for(let K=0;K2595*Math.log10(1+B/700),kaldi:B=>1127*Math.log(1+B/700),slaney:(B,Y=1e3,X=15,J=27/Math.log(6.4))=>B>=Y?X+Math.log(B/Y)*J:3*B/200};function _(B,Y="htk"){const X=f[Y];if(!X)throw new Error('mel_scale should be one of "htk", "slaney" or "kaldi".');return typeof B=="number"?X(B):B.map(J=>X(J))}const y={htk:B=>700*(10**(B/2595)-1),kaldi:B=>700*(Math.exp(B/1127)-1),slaney:(B,Y=1e3,X=15,J=Math.log(6.4)/27)=>B>=X?Y*Math.exp(J*(B-X)):200*B/3};function I(B,Y="htk"){const X=y[Y];if(!X)throw new Error('mel_scale should be one of "htk", "slaney" or "kaldi".');return typeof B=="number"?X(B):B.map(J=>X(J))}function w(B,Y){const X=Float64Array.from({length:Y.length-1},(le,pe)=>Y[pe+1]-Y[pe]),J=Array.from({length:B.length},()=>new Array(Y.length));for(let le=0;lenew Array(B.length));for(let le=0;leB+J*ne)}function k(B,Y,X,J,re,ne=null,le="htk",pe=!1){if(ne!==null&&ne!=="slaney")throw new Error('norm must be one of null or "slaney"');if(B<2)throw new Error(`Require num_frequency_bins: ${B} >= 2`);if(X>J)throw new Error(`Require min_frequency: ${X} <= max_frequency: ${J}`);const oe=_(X,le),K=_(J,le),j=v(oe,K,Y+2);let D=I(j,le),te;if(pe){const Ae=re/((B-1)*2);te=_(Float64Array.from({length:B},(ke,Ve)=>Ve*Ae),le),D=j}else te=v(0,Math.floor(re/2),B);const he=w(te,D);if(ne!==null&&ne==="slaney")for(let Ae=0;Aere)throw Error(`frame_length (${X}) may not be larger than fft_length (${re})`);if(xe!==X)throw new Error(`Length of the window (${xe}) must equal frame_length (${X})`);if(J<=0)throw new Error("hop_length must be greater than zero");if(ne===null&&D!==null)throw new Error("You have provided `mel_filters` but `power` is `None`. Mel spectrogram computation is not yet supported for complex-valued spectrogram. Specify `power` to fix this issue.");if(!j)throw new Error("`preemphasis_htk_flavor=false` is not currently supported.");if(le)switch(pe){case"reflect":{const je=Math.floor((re-1)/2)+1;B=T(B,je,je);break}case"constant":{const je=Math.floor(re/2),Oe=new B.constructor(B.length+2*je);Oe.set(B,je),B=Oe;break}default:throw new Error(`pad_mode="${pe}" not implemented yet.`)}let Ce=Math.floor(1+Math.floor((B.length-X)/J));Q!==null&&CeCe?de&&(fe=z):fe=De=z);const Pe=new n.FFT(re),We=new Float64Array(re),Fe=new Float64Array(Pe.outputBufferSize),tt=new Float32Array(ge*fe);for(let je=0;je=1;--ht)We[ht]-=K*We[ht-1];We[0]*=1-K}for(let ht=0;htMath.pow(pe,.85));break;default:throw new Error(`Unknown window type ${Y}.`)}if(X&&(le=le.subarray(0,B)),J===null)return le;if(B>J)throw new Error(`Length of the window (${B}) may not be larger than frame_length (${J})`);return le}function F(B,Y){let X=44;const J=new ArrayBuffer(X+B.length*4),re=new DataView(J);H(re,0,"RIFF"),re.setUint32(4,36+B.length*4,!0),H(re,8,"WAVE"),H(re,12,"fmt "),re.setUint32(16,16,!0),re.setUint16(20,3,!0),re.setUint16(22,1,!0),re.setUint32(24,Y,!0),re.setUint32(28,Y*4,!0),re.setUint16(32,4,!0),re.setUint16(34,32,!0),H(re,36,"data"),re.setUint32(40,B.length*4,!0);for(let ne=0;ne{let ne=await re.arrayBuffer();l.writeFileSync(J,Buffer.from(ne))};else throw new Error("Unable to save because filesystem is disabled in this environment.");await X(Y,this.toBlob())}}}),"./src/utils/constants.js":((e,r,t)=>{t.r(r),t.d(r,{CHAT_TEMPLATE_NAME:()=>l,CONFIG_NAME:()=>n,FEATURE_EXTRACTOR_NAME:()=>o,GENERATION_CONFIG_NAME:()=>c,GITHUB_ISSUE_URL:()=>s,IMAGE_PROCESSOR_NAME:()=>i,PROCESSOR_NAME:()=>a});const s="https://github.com/huggingface/transformers.js/issues/new/choose",n="config.json",o="preprocessor_config.json",i=o,a="processor_config.json",l="chat_template.jinja",c="generation_config.json"}),"./src/utils/core.js":((e,r,t)=>{t.r(r),t.d(r,{calculateDimensions:()=>c,calculateReflectOffset:()=>f,count:()=>w,dispatchCallback:()=>s,escapeRegExp:()=>o,isIntegralNumber:()=>a,isNullishDimension:()=>l,isTypedArray:()=>i,len:()=>I,mergeArrays:()=>d,pick:()=>y,pop:()=>p,product:()=>u,reverseDictionary:()=>n,saveBlob:()=>_});function s(v,k){v&&v(k)}function n(v){return Object.fromEntries(Object.entries(v).map(([k,T])=>[T,k]))}function o(v){return v.replace(/[.*+?^${}()|[\]\\]/g,"\\$&")}function i(v){return v?.prototype?.__proto__?.constructor?.name==="TypedArray"}function a(v){return Number.isInteger(v)||typeof v=="bigint"}function l(v){return v==null||v===-1}function c(v){const k=[];let T=v;for(;Array.isArray(T);)k.push(T.length),T=T[0];return k}function p(v,k,T=void 0){const b=v[k];if(b!==void 0)return delete v[k],b;if(T===void 0)throw Error(`Key ${k} does not exist in object.`);return T}function d(...v){return Array.prototype.concat.apply([],v)}function u(...v){return v.reduce((k,T)=>k.flatMap(b=>T.map(P=>[b,P])))}function f(v,k){return Math.abs((v+k)%(2*k)-k)}function _(v,k){const T=URL.createObjectURL(k),b=document.createElement("a");b.href=T,b.download=v,b.click(),b.remove(),URL.revokeObjectURL(T)}function y(v,k){return Object.assign({},...k.map(T=>{if(v[T]!==void 0)return{[T]:v[T]}}))}function I(v){let k=0;for(const T of v)++k;return k}function w(v,k){let T=0;for(const b of v)b===k&&++T;return T}}),"./src/utils/data-structures.js":((e,r,t)=>{t.r(r),t.d(r,{CharTrie:()=>n,DictionarySplitter:()=>l,LRUCache:()=>c,PriorityQueue:()=>s,TokenLattice:()=>i});class s{constructor(d=(f,_)=>f>_,u=1/0){this._heap=[],this._comparator=d,this._maxSize=u}get size(){return this._heap.length}isEmpty(){return this.size===0}peek(){return this._heap[0]}push(...d){return this.extend(d)}extend(d){for(const u of d)if(this.size0&&this._swap(0,u),this._heap.pop(),this._siftDown(),d}replace(d){const u=this.peek();return this._heap[0]=d,this._siftDown(),u}_parent(d){return(d+1>>>1)-1}_left(d){return(d<<1)+1}_right(d){return d+1<<1}_greater(d,u){return this._comparator(this._heap[d],this._heap[u])}_swap(d,u){const f=this._heap[d];this._heap[d]=this._heap[u],this._heap[u]=f}_siftUp(){this._siftUpFrom(this.size-1)}_siftUpFrom(d){for(;d>0&&this._greater(d,this._parent(d));)this._swap(d,this._parent(d)),d=this._parent(d)}_siftDown(){let d=0;for(;this._left(d)[]),this.endNodes=Array.from({length:this.len+1},()=>[]);const _=new a(this.bosTokenId,0,0,0,0),y=new a(this.eosTokenId,1,this.len,0,0);this.nodes.push(_.clone()),this.nodes.push(y.clone()),this.beginNodes[this.len].push(y),this.endNodes[0].push(_)}insert(d,u,f,_){const y=this.nodes.length,I=new a(_,y,d,u,f);this.beginNodes[d].push(I),this.endNodes[d+u].push(I),this.nodes.push(I)}viterbi(){const d=this.len;let u=0;for(;u<=d;){if(this.beginNodes[u].length==0)return[];for(let w of this.beginNodes[u]){w.prev=null;let v=0,k=null;for(let T of this.endNodes[u]){const b=T.backtraceScore+w.score;(k===null||b>v)&&(k=T.clone(),v=b)}if(k!==null)w.prev=k,w.backtraceScore=v;else return[]}++u}const f=[],y=this.beginNodes[d][0].prev;if(y===null)return[];let I=y.clone();for(;I.prev!==null;)f.push(I.clone()),I=I.clone().prev.clone();return f.reverse(),f}piece(d){return this.chars.slice(d.pos,d.pos+d.length).join("")}tokens(){return this.viterbi().map(u=>this.piece(u))}tokenIds(){return this.viterbi().map(u=>u.tokenId)}}class a{constructor(d,u,f,_,y){this.tokenId=d,this.nodeId=u,this.pos=f,this.length=_,this.score=y,this.prev=null,this.backtraceScore=0}clone(){const d=new a(this.tokenId,this.nodeId,this.pos,this.length,this.score);return d.prev=this.prev,d.backtraceScore=this.backtraceScore,d}}class l{constructor(d){this.trie=this._buildTrie(d)}_buildTrie(d){const u=Object.create(null);for(const f of d){let _=u;for(let y=0;y_&&u.push(d.slice(_,y)),u.push(w),y+=w.length,_=y):++y}return _this.capacity&&this.cache.delete(this.cache.keys().next().value)}clear(){this.cache.clear()}}}),"./src/utils/devices.js":((e,r,t)=>{t.r(r),t.d(r,{DEVICE_TYPES:()=>s});const s=Object.freeze({auto:"auto",gpu:"gpu",cpu:"cpu",wasm:"wasm",webgpu:"webgpu",cuda:"cuda",dml:"dml",webnn:"webnn","webnn-npu":"webnn-npu","webnn-gpu":"webnn-gpu","webnn-cpu":"webnn-cpu"})}),"./src/utils/dtypes.js":((e,r,t)=>{t.r(r),t.d(r,{DATA_TYPES:()=>i,DEFAULT_DEVICE_DTYPE_MAPPING:()=>a,DEFAULT_DTYPE_SUFFIX_MAPPING:()=>l,isWebGpuFp16Supported:()=>o});var s=t("./src/env.js"),n=t("./src/utils/devices.js");const o=(function(){let c;return async function(){if(c===void 0)if(!s.apis.IS_WEBGPU_AVAILABLE)c=!1;else try{c=(await navigator.gpu.requestAdapter()).features.has("shader-f16")}catch{c=!1}return c}})(),i=Object.freeze({auto:"auto",fp32:"fp32",fp16:"fp16",q8:"q8",int8:"int8",uint8:"uint8",q4:"q4",bnb4:"bnb4",q4f16:"q4f16"}),a=Object.freeze({[n.DEVICE_TYPES.wasm]:i.q8}),l=Object.freeze({[i.fp32]:"",[i.fp16]:"_fp16",[i.int8]:"_int8",[i.uint8]:"_uint8",[i.q8]:"_quantized",[i.q4]:"_q4",[i.q4f16]:"_q4f16",[i.bnb4]:"_bnb4"})}),"./src/utils/generic.js":((e,r,t)=>{t.r(r),t.d(r,{Callable:()=>s});const s=class{constructor(){let n=function(...o){return n._call(...o)};return Object.setPrototypeOf(n,new.target.prototype)}_call(...n){throw Error("Must implement _call method in subclass")}}}),"./src/utils/hub.js":((e,r,t)=>{t.r(r),t.d(r,{MAX_EXTERNAL_DATA_CHUNKS:()=>a,getFile:()=>f,getModelFile:()=>v,getModelJSON:()=>T,getModelText:()=>k});var s=t("?7992"),n=t("?5af5"),o=t("./src/env.js"),i=t("./src/utils/core.js");const a=100,l={txt:"text/plain",html:"text/html",css:"text/css",js:"text/javascript",json:"application/json",png:"image/png",jpg:"image/jpeg",jpeg:"image/jpeg",gif:"image/gif"};class c{constructor(S){if(this.filePath=S,this.headers=new Headers,this.exists=s.existsSync(S),this.exists){this.status=200,this.statusText="OK";let O=s.statSync(S);this.headers.set("content-length",O.size.toString()),this.updateContentType();const F=s.createReadStream(S);this.body=new ReadableStream({start(H){F.on("data",W=>H.enqueue(W)),F.on("end",()=>H.close()),F.on("error",W=>H.error(W))},cancel(){F.destroy()}})}else this.status=404,this.statusText="Not Found",this.body=null}updateContentType(){const S=this.filePath.toString().split(".").pop().toLowerCase();this.headers.set("content-type",l[S]??"application/octet-stream")}clone(){let S=new c(this.filePath);return S.exists=this.exists,S.status=this.status,S.statusText=this.statusText,S.headers=new Headers(this.headers),S}async arrayBuffer(){return(await s.promises.readFile(this.filePath)).buffer}async blob(){const S=await s.promises.readFile(this.filePath);return new Blob([S],{type:this.headers.get("content-type")})}async text(){return await s.promises.readFile(this.filePath,"utf8")}async json(){return JSON.parse(await this.text())}}function p(x,S=null,O=null){let F;try{F=new URL(x)}catch{return!1}return!(S&&!S.includes(F.protocol)||O&&!O.includes(F.hostname))}const d=/^(\b[\w\-.]+\b\/)?\b[\w\-.]{1,96}\b$/;function u(x){return!(!d.test(x)||x.includes("..")||x.includes("--")||x.endsWith(".git")||x.endsWith(".ipynb"))}async function f(x){if(o.env.useFS&&!p(x,["http:","https:","blob:"]))return new c(x instanceof URL?x.protocol==="file:"?x.pathname:x.toString():x);if(typeof process<"u"&&process?.release?.name==="node"){const S=!!pc?.TESTING_REMOTELY,O=o.env.version,F=new Headers;if(F.set("User-Agent",`transformers.js/${O}; is_ci/${S};`),p(x,["http:","https:"],["huggingface.co","hf.co"])){const W=pc?.HF_TOKEN??pc?.HF_ACCESS_TOKEN;W&&F.set("Authorization",`Bearer ${W}`)}return fetch(x,{headers:F})}else return fetch(x)}const _={400:"Bad request error occurred while trying to load file",401:"Unauthorized access to file",403:"Forbidden access to file",404:"Could not locate file",408:"Request timeout error occurred while trying to load file",500:"Internal server error error occurred while trying to load file",502:"Bad gateway error occurred while trying to load file",503:"Service unavailable error occurred while trying to load file",504:"Gateway timeout error occurred while trying to load file"};function y(x,S,O){if(!O)return null;const F=_[x]??`Error (${x}) occurred while trying to load file`;throw Error(`${F}: "${S}".`)}class I{constructor(S){this.path=S}async match(S){let O=n.join(this.path,S),F=new c(O);if(F.exists)return F}async put(S,O,F=void 0){let H=n.join(this.path,S);try{const W=O.headers.get("Content-Length"),B=parseInt(W??"0");let Y=0;await s.promises.mkdir(n.dirname(H),{recursive:!0});const X=s.createWriteStream(H),J=O.body.getReader();for(;;){const{done:re,value:ne}=await J.read();if(re)break;await new Promise((pe,oe)=>{X.write(ne,K=>{if(K){oe(K);return}pe()})}),Y+=ne.length;const le=B?Y/B*100:0;F?.({progress:le,loaded:Y,total:B})}X.close()}catch(W){try{await s.promises.unlink(H)}catch{}throw W}}}async function w(x,...S){for(let O of S)try{let F=await x.match(O);if(F)return F}catch{continue}}async function v(x,S,O=!0,F={},H=!1){if(!o.env.allowLocalModels){if(F.local_files_only)throw Error("Invalid configuration detected: local models are disabled (`env.allowLocalModels=false`) but you have requested to only use local models (`local_files_only=true`).");if(!o.env.allowRemoteModels)throw Error("Invalid configuration detected: both local and remote models are disabled. Fix by setting `env.allowLocalModels` or `env.allowRemoteModels` to `true`.")}(0,i.dispatchCallback)(F.progress_callback,{status:"initiate",name:x,file:S});let W;if(!W&&o.env.useCustomCache){if(!o.env.customCache)throw Error("`env.useCustomCache=true`, but `env.customCache` is not defined.");if(!o.env.customCache.match||!o.env.customCache.put)throw new Error("`env.customCache` must be an object which implements the `match` and `put` functions of the Web Cache API. For more information, see https://developer.mozilla.org/en-US/docs/Web/API/Cache");W=o.env.customCache}if(!W&&o.env.useBrowserCache){if(typeof caches>"u")throw Error("Browser cache is not available in this environment.");try{W=await caches.open("transformers-cache")}catch(te){console.warn("An error occurred while opening the browser cache:",te)}}if(!W&&o.env.useFSCache){if(!o.apis.IS_FS_AVAILABLE)throw Error("File System Cache is not available in this environment.");W=new I(F.cache_dir??o.env.cacheDir)}const B=F.revision??"main",Y=P(x,S),X=u(x),J=X?P(o.env.localModelPath,Y):Y,re=P(o.env.remoteHost,o.env.remotePathTemplate.replaceAll("{model}",x).replaceAll("{revision}",encodeURIComponent(B)),S);let ne;const le=W instanceof I?B==="main"?Y:P(x,B,S):re;let pe=!1,oe;W&&(oe=await w(W,J,le));const K=oe!==void 0;if(oe===void 0){if(o.env.allowLocalModels)if(p(Y,["http:","https:"])){if(F.local_files_only)throw new Error(`\`local_files_only=true\`, but attempted to load a remote file from: ${Y}.`);if(!o.env.allowRemoteModels)throw new Error(`\`env.allowRemoteModels=false\`, but attempted to load a remote file from: ${Y}.`)}else try{oe=await f(J),ne=J}catch(he){console.warn(`Unable to load from local path "${J}": "${he}"`)}if(oe===void 0||oe.status===404){if(F.local_files_only||!o.env.allowRemoteModels){if(O)throw Error(`\`local_files_only=true\` or \`env.allowRemoteModels=false\` and file was not found locally at "${J}".`);return null}if(!X)throw Error(`Local file missing at "${J}" and download aborted due to invalid model ID "${x}".`);if(oe=await f(re),oe.status!==200)return y(oe.status,re,O);ne=le}pe=W&&typeof Response<"u"&&oe instanceof Response&&oe.status===200}(0,i.dispatchCallback)(F.progress_callback,{status:"download",name:x,file:S});let j;if(!(o.apis.IS_NODE_ENV&&H)){let te;F.progress_callback?K&&typeof navigator<"u"&&/firefox/i.test(navigator.userAgent)?(te=new Uint8Array(await oe.arrayBuffer()),(0,i.dispatchCallback)(F.progress_callback,{status:"progress",name:x,file:S,progress:100,loaded:te.length,total:te.length})):te=await b(oe,he=>{(0,i.dispatchCallback)(F.progress_callback,{status:"progress",name:x,file:S,...he})}):te=new Uint8Array(await oe.arrayBuffer()),j=te}if(pe&&ne&&await W.match(ne)===void 0)if(j)await W.put(ne,new Response(j,{headers:oe.headers})).catch(te=>{console.warn(`Unable to add response to browser cache: ${te}.`)});else{const te=F.progress_callback?he=>(0,i.dispatchCallback)(F.progress_callback,{status:"progress",name:x,file:S,...he}):void 0;await W.put(ne,oe,te)}if((0,i.dispatchCallback)(F.progress_callback,{status:"done",name:x,file:S}),j){if(!o.apis.IS_NODE_ENV&&H)throw new Error("Cannot return path in a browser environment.");return j}if(oe instanceof c)return oe.filePath;const D=await W?.match(ne);if(D instanceof c)return D.filePath;if(D instanceof Response)return new Uint8Array(await D.arrayBuffer());if(typeof D=="string")return D;throw new Error("Unable to get model file path or buffer.")}async function k(x,S,O=!0,F={}){const H=await v(x,S,O,F,!1);return H===null?null:new TextDecoder("utf-8").decode(H)}async function T(x,S,O=!0,F={}){const H=await k(x,S,O,F);return H===null?{}:JSON.parse(H)}async function b(x,S){const O=x.headers.get("Content-Length");O===null&&console.warn("Unable to determine content-length from response headers. Will expand buffer when needed.");let F=parseInt(O??"0"),H=new Uint8Array(F),W=0;const B=x.body.getReader();async function Y(){const{done:X,value:J}=await B.read();if(X)return;const re=W+J.length;if(re>F){F=re;const le=new Uint8Array(F);le.set(H),H=le}H.set(J,W),W=re;const ne=W/F*100;return S({progress:ne,loaded:W,total:F}),Y()}return await Y(),H}function P(...x){return x=x.map((S,O)=>(O&&(S=S.replace(new RegExp("^/"),"")),O!==x.length-1&&(S=S.replace(new RegExp("/$"),"")),S)),x.join("/")}}),"./src/utils/image.js":((e,r,t)=>{t.r(r),t.d(r,{RawImage:()=>_,load_image:()=>y});var s=t("./src/utils/core.js"),n=t("./src/utils/hub.js"),o=t("./src/env.js"),i=t("./src/utils/tensor.js"),a=t("?2b25");let l,c,p;const d=o.apis.IS_BROWSER_ENV||o.apis.IS_WEBWORKER_ENV;if(d)l=(I,w)=>{if(!self.OffscreenCanvas)throw new Error("OffscreenCanvas not supported by this browser.");return new self.OffscreenCanvas(I,w)},p=self.createImageBitmap,c=self.ImageData;else if(a)p=async I=>{const v=(await I.metadata()).channels,{data:k,info:T}=await I.rotate().raw().toBuffer({resolveWithObject:!0}),b=new _(new Uint8ClampedArray(k),T.width,T.height,T.channels);return v!==void 0&&v!==T.channels&&b.convert(v),b};else throw new Error("Unable to load image processing library.");const u={0:"nearest",1:"lanczos",2:"bilinear",3:"bicubic",4:"box",5:"hamming"},f=new Map([["png","image/png"],["jpg","image/jpeg"],["jpeg","image/jpeg"],["gif","image/gif"]]);class _{constructor(w,v,k,T){this.data=w,this.width=v,this.height=k,this.channels=T}get size(){return[this.width,this.height]}static async read(w){if(w instanceof _)return w;if(typeof w=="string"||w instanceof URL)return await this.fromURL(w);if(w instanceof Blob)return await this.fromBlob(w);if(typeof HTMLCanvasElement<"u"&&w instanceof HTMLCanvasElement||typeof OffscreenCanvas<"u"&&w instanceof OffscreenCanvas)return this.fromCanvas(w);throw new Error(`Unsupported input type: ${typeof w}`)}static fromCanvas(w){if(!d)throw new Error("fromCanvas() is only supported in browser environments.");const k=w.getContext("2d").getImageData(0,0,w.width,w.height).data;return new _(k,w.width,w.height,4)}static async fromURL(w){const v=await(0,n.getFile)(w);if(v.status!==200)throw new Error(`Unable to read image from "${w}" (${v.status} ${v.statusText})`);const k=await v.blob();return this.fromBlob(k)}static async fromBlob(w){if(d){const v=await p(w),k=l(v.width,v.height).getContext("2d");return k.drawImage(v,0,0),new this(k.getImageData(0,0,v.width,v.height).data,v.width,v.height,4)}else{const v=a(await w.arrayBuffer());return await p(v)}}static fromTensor(w,v="CHW"){if(w.dims.length!==3)throw new Error(`Tensor should have 3 dimensions, but has ${w.dims.length} dimensions.`);if(v==="CHW")w=w.transpose(1,2,0);else if(v!=="HWC")throw new Error(`Unsupported channel format: ${v}`);if(!(w.data instanceof Uint8ClampedArray||w.data instanceof Uint8Array))throw new Error(`Unsupported tensor type: ${w.type}`);switch(w.dims[2]){case 1:case 2:case 3:case 4:return new _(w.data,w.dims[1],w.dims[0],w.dims[2]);default:throw new Error(`Unsupported number of channels: ${w.dims[2]}`)}}grayscale(){if(this.channels===1)return this;const w=new Uint8ClampedArray(this.width*this.height*1);switch(this.channels){case 3:case 4:for(let v=0,k=0;v=0?S=k:F=-k,T>=0?O=T:H=-T,x.drawImage(P,S,O,w,v,F,H,w,v),new _(x.getImageData(0,0,w,v).data,w,v,4).convert(b)}else{let b=this.toSharp();if(k>=0&&T>=0)b=b.extract({left:Math.floor(k),top:Math.floor(T),width:w,height:v});else if(k<=0&&T<=0){const P=Math.floor(-T),x=Math.floor(-k);b=b.extend({top:P,left:x,right:w-this.width-x,bottom:v-this.height-P})}else{let P=[0,0],x=0;T<0?(P[0]=Math.floor(-T),P[1]=v-this.height-P[0]):x=Math.floor(T);let S=[0,0],O=0;k<0?(S[0]=Math.floor(-k),S[1]=w-this.width-S[0]):O=Math.floor(k),b=b.extend({top:P[0],bottom:P[1],left:S[0],right:S[1]}).extract({left:O,top:x,width:w,height:v})}return await p(b)}}async toBlob(w="image/png",v=1){if(!d)throw new Error("toBlob() is only supported in browser environments.");return await this.toCanvas().convertToBlob({type:w,quality:v})}toTensor(w="CHW"){let v=new i.Tensor("uint8",new Uint8Array(this.data),[this.height,this.width,this.channels]);if(w!=="HWC")if(w==="CHW")v=v.permute(2,0,1);else throw new Error(`Unsupported channel format: ${w}`);return v}toCanvas(){if(!d)throw new Error("toCanvas() is only supported in browser environments.");const w=this.clone().rgba(),v=l(w.width,w.height),k=new c(w.data,w.width,w.height);return v.getContext("2d").putImageData(k,0,0),v}split(){const{data:w,width:v,height:k,channels:T}=this,b=w.constructor,P=w.length/T,x=Array.from({length:T},()=>new b(P));for(let S=0;Snew _(S,v,k,1))}_update(w,v,k,T=null){return this.data=w,this.width=v,this.height=k,T!==null&&(this.channels=T),this}clone(){return new _(this.data.slice(),this.width,this.height,this.channels)}convert(w){if(this.channels===w)return this;switch(w){case 1:this.grayscale();break;case 3:this.rgb();break;case 4:this.rgba();break;default:throw new Error(`Conversion failed due to unsupported number of channels: ${this.channels}`)}return this}async save(w){if(d){if(o.apis.IS_WEBWORKER_ENV)throw new Error("Unable to save an image from a Web Worker.");const v=w.split(".").pop().toLowerCase(),k=f.get(v)??"image/png",T=await this.toBlob(k);(0,s.saveBlob)(w,T)}else{if(o.apis.IS_FS_AVAILABLE)return await this.toSharp().toFile(w);throw new Error("Unable to save the image because filesystem is disabled in this environment.")}}toSharp(){if(d)throw new Error("toSharp() is only supported in server-side environments.");return a(this.data,{raw:{width:this.width,height:this.height,channels:this.channels}})}}const y=_.read.bind(_)}),"./src/utils/maths.js":((e,r,t)=>{t.r(r),t.d(r,{FFT:()=>y,bankers_round:()=>v,cos_sim:()=>l,dot:()=>a,dynamic_time_warping:()=>k,interpolate_data:()=>s,log_softmax:()=>i,magnitude:()=>c,max:()=>d,medianFilter:()=>I,min:()=>p,permute_data:()=>n,round:()=>w,softmax:()=>o});function s(T,[b,P,x],[S,O],F="bilinear",H=!1){const W=O/x,B=S/P,Y=new T.constructor(S*O*b),X=P*x,J=S*O;for(let re=0;re=0;--H)S[H]=W,x[H]=b[P[H]],W*=x[H];const O=P.map((H,W)=>S[P.indexOf(W)]),F=new T.constructor(T.length);for(let H=0;H=0;--B)W+=Y%b[B]*O[B],Y=Math.floor(Y/b[B]);F[W]=T[H]}return[F,x]}function o(T){const b=d(T)[0],P=T.map(O=>Math.exp(O-b)),x=P.reduce((O,F)=>O+F,0);return P.map(O=>O/x)}function i(T){const b=d(T)[0];let P=0;for(let O=0;OO-b-x)}function a(T,b){let P=0;for(let x=0;xb+P*P,0))}function p(T){if(T.length===0)throw Error("Array must not be empty");let b=T[0],P=0;for(let x=1;xb&&(b=T[x],P=x);return[b,P]}function u(T){return T>0&&(T&T-1)===0}class f{constructor(b){if(this.size=b|0,this.size<=1||!u(this.size))throw new Error("FFT size must be a power of two larger than 1");this._csize=b<<1,this.table=new Float64Array(this.size*2);for(let x=0;xx;x<<=1)++P;this._width=P%2===0?P-1:P,this._bitrev=new Int32Array(1<>>S&3)<>>1);for(let S=0;S>>1]=b[S];return x}toComplexArray(b,P){const x=P||this.createComplexArray();for(let S=0;S>>1],x[S+1]=0;return x}transform(b,P){if(b===P)throw new Error("Input and output buffers must be different");this._transform4(b,P,1)}realTransform(b,P){if(b===P)throw new Error("Input and output buffers must be different");this._realTransform4(b,P,1)}inverseTransform(b,P){if(b===P)throw new Error("Input and output buffers must be different");this._transform4(b,P,-1);for(let x=0;x>=2;F>=2;F>>=2){H=S/F<<1;const J=H>>>2;for(W=0;W>>1,F>>>1)}else for(W=0,B=0;W>>1,F>>>1,x)}const X=this.table;for(F>>=2;F>=2;F>>=2){H=S/F<<1;const re=H>>>1,ne=re>>>1,le=ne>>>1;for(W=0;W>>1;for(let re=2;re>1;++Y){const X=(Y+1-b)**2/2,J=Math.sqrt(W**2+B**2)**X,re=X*Math.atan2(B,W),ne=2*Y;O[ne]=J*Math.cos(re),O[ne+1]=J*Math.sin(re),F[ne]=O[ne],F[ne+1]=-O[ne+1]}this._slicedChirpBuffer=O.subarray(P,x),this._f=new f(S>>1),this._f.transform(this._chirpBuffer,F)}_transform(b,P,x){const S=this._buffer1,O=this._buffer2,F=this._outBuffer1,H=this._outBuffer2,W=this._chirpBuffer,B=this._slicedChirpBuffer,Y=this._a;if(x)for(let X=0;X>1,ne=P[re];S[X]=ne*B[X],S[J]=ne*B[J]}else for(let X=0;X=T.length&&(W=2*(T.length-1)-W),x[F++]=T[W]}x.sort(),P[O]=x[S]}return P}function w(T,b){const P=Math.pow(10,b);return Math.round(T*P)/P}function v(T){const b=Math.round(T);return Math.abs(T)%1===.5?b%2===0?b:b-1:b}function k(T){const b=T.length,P=T[0].length,x=[b+1,P+1],S=Array.from({length:x[0]},()=>Array(x[1]).fill(1/0));S[0][0]=0;const O=Array.from({length:x[0]},()=>Array(x[1]).fill(-1));for(let Y=1;Y0||H>0;)switch(W.push(F-1),B.push(H-1),O[F][H]){case 0:--F,--H;break;case 1:--F;break;case 2:--H;break;default:throw new Error(`Internal error in dynamic time warping. Unexpected trace[${F}, ${H}]. Please file a bug report.`)}return W.reverse(),B.reverse(),[W,B]}}),"./src/utils/tensor.js":((e,r,t)=>{t.r(r),t.d(r,{DataTypeMap:()=>i,Tensor:()=>a,cat:()=>P,full:()=>B,full_like:()=>Y,interpolate:()=>p,interpolate_4d:()=>d,layer_norm:()=>v,matmul:()=>u,mean:()=>F,mean_pooling:()=>w,ones:()=>X,ones_like:()=>J,permute:()=>c,quantize_embeddings:()=>oe,rand:()=>le,randn:()=>pe,rfft:()=>f,slice:()=>I,stack:()=>x,std_mean:()=>O,topk:()=>_,zeros:()=>re,zeros_like:()=>ne});var s=t("./src/utils/maths.js"),n=t("./src/backends/onnx.js"),o=t("./src/ops/registry.js");const i=Object.freeze({float32:Float32Array,float16:typeof Float16Array<"u"?Float16Array:Uint16Array,float64:Float64Array,string:Array,int8:Int8Array,uint8:Uint8Array,int16:Int16Array,uint16:Uint16Array,int32:Int32Array,uint32:Uint32Array,int64:BigInt64Array,uint64:BigUint64Array,bool:Uint8Array,uint4:Uint8Array,int4:Int8Array});class a{get dims(){return this.ort_tensor.dims}set dims(j){this.ort_tensor.dims=j}get type(){return this.ort_tensor.type}get data(){return this.ort_tensor.data}get size(){return this.ort_tensor.size}get location(){return this.ort_tensor.location}ort_tensor;constructor(...j){return(0,n.isONNXTensor)(j[0])?this.ort_tensor=j[0]:this.ort_tensor=new n.Tensor(j[0],j[1],j[2]),new Proxy(this,{get:(D,te)=>{if(typeof te=="string"){let he=Number(te);if(Number.isInteger(he))return D._getitem(he)}return D[te]},set:(D,te,he)=>D[te]=he})}dispose(){this.ort_tensor.dispose()}*[Symbol.iterator](){const[j,...D]=this.dims;if(D.length>0){const te=D.reduce((he,Ae)=>he*Ae);for(let he=0;he0){const he=te.reduce((Ae,ke)=>Ae*ke);return this._subarray(j,he,te)}else return new a(this.type,[this.data[j]],te)}indexOf(j){const D=this.data;for(let te=0;teve)throw new Error(`Invalid slice: ${de}`);const xe=[Math.max(be,0),Math.min(ve,this.dims[z])];te.push(xe),D.push(xe[1]-xe[0])}else throw new Error(`Invalid slice: ${de}`)}const he=te.map(([z,de])=>de-z),Ae=he.reduce((z,de)=>z*de),ke=this.data,Ve=new ke.constructor(Ae),Te=this.stride();let Q=!0;for(let z=1;z=0;--be){const xe=he[be];de+=(ve%xe+te[be][0])*Te[be],ve=Math.floor(ve/xe)}Ve[z]=ke[de]}return new a(this.type,Ve,D)}permute(...j){return c(this,j)}transpose(...j){return this.permute(...j)}sum(j=null,D=!1){return this.norm(1,j,D)}norm(j="fro",D=null,te=!1){if(j==="fro")j=2;else if(typeof j=="string")throw Error(`Unsupported norm: ${j}`);const he=this.data,Ae=(Q,z)=>Q+z**j;if(D===null){const Q=he.reduce(Ae,0)**(1/j);return new a(this.type,[Q],[])}const[ke,Ve,Te]=S(Ae,this,D,te);if(j!==1)for(let Q=0;Q=0;--Te){const de=this.dims[Te];if(Te!==D){const be=Q%de;Ve+=be*z,z*=this.dims[Te]}Q=Math.floor(Q/de)}he[ke]/=Ae[Ve]}return this}normalize(j=2,D=1){return this.clone().normalize_(j,D)}stride(){return H(this.dims)}squeeze(j=null){return new a(this.type,this.data,k(this.dims,j))}squeeze_(j=null){return this.dims=k(this.dims,j),this}unsqueeze(j=null){return new a(this.type,this.data,T(this.dims,j))}unsqueeze_(j=null){return this.dims=T(this.dims,j),this}flatten_(j=0,D=-1){D=(D+this.dims.length)%this.dims.length;let te=this.dims.slice(0,j),he=this.dims.slice(j,D+1),Ae=this.dims.slice(D+1);return this.dims=[...te,he.reduce((ke,Ve)=>ke*Ve,1),...Ae],this}flatten(j=0,D=-1){return this.clone().flatten_(j,D)}view(...j){let D=-1;for(let he=0;heVe!==D?Ae*ke:Ae,1);j[D]=te.length/he}return new a(this.type,te,j)}neg_(){const j=this.data;for(let D=0;Dj?1:0;return new a("bool",D,this.dims)}lt(j){const D=new Uint8Array(this.data.length),te=this.data;for(let he=0;heMath.min(ke,Ve),this,j,D,1/0);return new a(te,he,Ae)}max(j=null,D=!1){if(j===null){const ke=(0,s.max)(this.data)[0];return new a(this.type,[ke],[])}const[te,he,Ae]=S((ke,Ve)=>Math.max(ke,Ve),this,j,D,-1/0);return new a(te,he,Ae)}argmin(j=null,D=!1){if(j!==null)throw new Error("`dim !== null` not yet implemented.");const te=(0,s.min)(this.data)[1];return new a("int64",[BigInt(te)],[])}argmax(j=null,D=!1){if(j!==null)throw new Error("`dim !== null` not yet implemented.");const te=(0,s.max)(this.data)[1];return new a("int64",[BigInt(te)],[])}to(j){if(this.type===j)return this;if(!i.hasOwnProperty(j))throw new Error(`Unsupported type: ${j}`);let D;const te=["int64","uint64"].includes(this.type),he=["int64","uint64"].includes(j);return te&&!he?D=Number:!te&&he&&(["float16","float32","float64"].includes(this.type)?D=Ae=>BigInt(Math.floor(Ae)):D=BigInt),new a(j,i[j].from(this.data,D),this.dims)}}function l(K,j){const D=K.length,te=j.reduce((Ae,ke)=>Ae*ke);if(D!==te)throw Error(`cannot reshape array of size ${D} into shape (${j})`);let he=K;for(let Ae=j.length-1;Ae>=0;Ae--)he=he.reduce((ke,Ve)=>{let Te=ke[ke.length-1];return Te.lengthnew a("int64",K,[K.length]);async function I(K,j,D,te,he){return await(await o.TensorOpRegistry.slice)({x:K,s:y(j),e:y(D),a:y(te),t:y(he??new Array(te.length).fill(1))})}function w(K,j){const D=K.data,te=j.data,he=[K.dims[0],K.dims[2]],Ae=new D.constructor(he[0]*he[1]),[ke,Ve,Te]=K.dims;let Q=0;for(let z=0;zD!==1):typeof j=="number"?K[j]===1&&K.splice(j,1):Array.isArray(j)&&(K=K.filter((D,te)=>D!==1||!j.includes(te))),K}function T(K,j){return j=b(j,K.length+1),K=K.slice(),K.splice(j,0,1),K}function b(K,j,D=null,te=!0){if(K<-j||K>=j){if(te)throw new Error(`IndexError: index ${K} is out of bounds for dimension${D===null?"":" "+D} with size ${j}`);return K<-j?0:j}return K<0&&(K=(K%j+j)%j),K}function P(K,j=0){j=b(j,K[0].dims.length);const D=K[0].dims.slice();D[j]=K.reduce((ke,Ve)=>ke+Ve.dims[j],0);const te=D.reduce((ke,Ve)=>ke*Ve,1),he=new K[0].data.constructor(te),Ae=K[0].type;if(j===0){let ke=0;for(const Ve of K){const Te=Ve.data;he.set(Te,ke),ke+=Te.length}}else{let ke=0;for(let Ve=0;Ve=0;--be){const Ce=Q[be];let ge=ve%Ce;be===j&&(ge+=ke),de+=ge*xe,xe*=D[be],ve=Math.floor(ve/Ce)}he[de]=Te[z]}ke+=Q[j]}}return new a(Ae,he,D)}function x(K,j=0){return P(K.map(D=>D.unsqueeze(j)),j)}function S(K,j,D=null,te=!1,he=null){const Ae=j.data,ke=j.dims;D=b(D,ke.length);const Ve=ke.slice();Ve[D]=1;const Te=new Ae.constructor(Ae.length/ke[D]);he!==null&&Te.fill(he);for(let Q=0;Q=0;--de){const xe=ke[de];if(de!==D){const Ce=be%xe;z+=Ce*ve,ve*=Ve[de]}be=Math.floor(be/xe)}Te[z]=K(Te[z],Ae[Q],Q,z)}return te||Ve.splice(D,1),[j.type,Te,Ve]}function O(K,j=null,D=1,te=!1){const he=K.data,Ae=K.dims;if(j===null){const ve=he.reduce((De,fe)=>De+fe,0)/he.length,xe=Math.sqrt(he.reduce((De,fe)=>De+(fe-ve)**2,0)/(he.length-D)),Ce=new a(K.type,[ve],[]);return[new a(K.type,[xe],[]),Ce]}j=b(j,Ae.length);const ke=F(K,j,te),Ve=ke.data,[Te,Q,z]=S((be,ve,xe,Ce)=>be+(ve-Ve[Ce])**2,K,j,te);for(let be=0;beQ+z,0);return new a(K.type,[Te/he.length],[])}j=b(j,te.length);const[Ae,ke,Ve]=S((Te,Q)=>Te+Q,K,j,D);if(te[j]!==1)for(let Te=0;Te=0;--D)j[D]=te,te*=K[D];return j}function W(K,j,D,te){const he=K.reduce((Ae,ke)=>Ae*ke,1);return new a(D,new te(he).fill(j),K)}function B(K,j){let D,te;if(typeof j=="number")D="float32",te=Float32Array;else if(typeof j=="bigint")D="int64",te=BigInt64Array;else if(typeof j=="boolean")D="bool",te=Uint8Array;else throw new Error(`Unsupported data type: ${typeof j}`);return W(K,j,D,te)}function Y(K,j){return B(K.dims,j)}function X(K){return W(K,1n,"int64",BigInt64Array)}function J(K){return X(K.dims)}function re(K){return W(K,0n,"int64",BigInt64Array)}function ne(K){return re(K.dims)}function le(K){const j=K.reduce((D,te)=>D*te,1);return new a("float32",Float32Array.from({length:j},()=>Math.random()),K)}function pe(K){const j=K.reduce((te,he)=>te*he,1);function D(){const te=1-Math.random(),he=1-Math.random();return Math.sqrt(-2*Math.log(te))*Math.cos(2*Math.PI*he)}return new a("float32",Float32Array.from({length:j},()=>D()),K)}function oe(K,j){if(K.dims.length!==2)throw new Error("The tensor must have 2 dimensions");if(K.dims.at(-1)%8!==0)throw new Error("The last dimension of the tensor must be a multiple of 8");if(!["binary","ubinary"].includes(j))throw new Error("The precision must be either 'binary' or 'ubinary'");const D=j==="binary",te=D?"int8":"uint8",he=D?Int8Array:Uint8Array,Ae=K.data,ke=new he(Ae.length/8);for(let Ve=0;Ve0?1:0,Q=Math.floor(Ve/8),z=Ve%8;ke[Q]|=Te<<7-z,D&&z===0&&(ke[Q]-=128)}return new a(te,ke,[K.dims[0],K.dims[1]/8])}}),"./src/utils/video.js":((e,r,t)=>{t.r(r),t.d(r,{RawVideo:()=>i,RawVideoFrame:()=>o,load_video:()=>a});var s=t("./src/utils/image.js"),n=t("./src/env.js");class o{constructor(c,p){this.image=c,this.timestamp=p}}class i{constructor(c,p){c.length>0&&c[0]instanceof s.RawImage&&(c=c.map((d,u)=>new o(d,(u+1)/(c.length+1)*p))),this.frames=c,this.duration=p}get width(){return this.frames[0].image.width}get height(){return this.frames[0].image.height}get fps(){return this.frames.length/this.duration}}async function a(l,{num_frames:c=null,fps:p=null}={}){if(!n.apis.IS_BROWSER_ENV)throw new Error("`load_video` is currently only supported in browser environments.");if(c==null&&p==null)throw new Error("Either num_frames or fps must be provided.");const d=[],u=document.createElement("video");if(u.crossOrigin="anonymous",u.muted=!0,typeof l=="string")u.src=l;else if(l instanceof Blob)u.src=URL.createObjectURL(l);else if(l instanceof HTMLVideoElement)u.src=l.src;else throw new Error("Invalid URL or video element provided.");if(await new Promise(k=>u.onloadedmetadata=k),u.seekable.start(0)===u.seekable.end(0)){const T=await(await fetch(u.src)).blob();u.src=URL.createObjectURL(T),await new Promise(b=>u.onloadedmetadata=b)}const f=u.duration;let _,y;c!=null?(_=c,y=c===1?0:f/(c-1)):(y=1/p,_=Math.floor(f/y));let I=[];for(let k=0;k<_;++k)I.push(c===1?f/2:k*y);const w=document.createElement("canvas");w.width=u.videoWidth,w.height=u.videoHeight;const v=w.getContext("2d",{willReadFrequently:!0});for(const k of I){u.currentTime=k,await new Promise(x=>{u.onseeked=x}),v.drawImage(u,0,0,w.width,w.height);const T=v.getImageData(0,0,w.width,w.height),b=new s.RawImage(T.data,w.width,w.height,4),P=new o(b,k);d.push(P)}return u.remove(),new i(d,f)}})},lM={};function jt(e){var r=lM[e];if(r!==void 0)return r.exports;var t=lM[e]={exports:{}};return mT[e](t,t.exports,jt),t.exports}(()=>{var e=Object.getPrototypeOf?t=>Object.getPrototypeOf(t):t=>t.__proto__,r;jt.t=function(t,s){if(s&1&&(t=this(t)),s&8||typeof t=="object"&&t&&(s&4&&t.__esModule||s&16&&typeof t.then=="function"))return t;var n=Object.create(null);jt.r(n);var o={};r=r||[null,e({}),e([]),e(e)];for(var i=s&2&&t;typeof i=="object"&&!~r.indexOf(i);i=e(i))Object.getOwnPropertyNames(i).forEach(a=>o[a]=()=>t[a]);return o.default=()=>t,jt.d(n,o),n}})(),jt.d=(e,r)=>{for(var t in r)jt.o(r,t)&&!jt.o(e,t)&&Object.defineProperty(e,t,{enumerable:!0,get:r[t]})},jt.o=(e,r)=>Object.prototype.hasOwnProperty.call(e,r),jt.r=e=>{typeof Symbol<"u"&&Symbol.toStringTag&&Object.defineProperty(e,Symbol.toStringTag,{value:"Module"}),Object.defineProperty(e,"__esModule",{value:!0})};var m={};(()=>{jt.r(m),jt.d(m,{ASTFeatureExtractor:()=>d.ASTFeatureExtractor,ASTForAudioClassification:()=>t.ASTForAudioClassification,ASTModel:()=>t.ASTModel,ASTPreTrainedModel:()=>t.ASTPreTrainedModel,AlbertForMaskedLM:()=>t.AlbertForMaskedLM,AlbertForQuestionAnswering:()=>t.AlbertForQuestionAnswering,AlbertForSequenceClassification:()=>t.AlbertForSequenceClassification,AlbertModel:()=>t.AlbertModel,AlbertPreTrainedModel:()=>t.AlbertPreTrainedModel,AlbertTokenizer:()=>s.AlbertTokenizer,ArceeForCausalLM:()=>t.ArceeForCausalLM,ArceeModel:()=>t.ArceeModel,ArceePreTrainedModel:()=>t.ArceePreTrainedModel,AudioClassificationPipeline:()=>r.AudioClassificationPipeline,AutoConfig:()=>n.AutoConfig,AutoFeatureExtractor:()=>u.AutoFeatureExtractor,AutoImageProcessor:()=>y.AutoImageProcessor,AutoModel:()=>t.AutoModel,AutoModelForAudioClassification:()=>t.AutoModelForAudioClassification,AutoModelForAudioFrameClassification:()=>t.AutoModelForAudioFrameClassification,AutoModelForAudioTextToText:()=>t.AutoModelForAudioTextToText,AutoModelForCTC:()=>t.AutoModelForCTC,AutoModelForCausalLM:()=>t.AutoModelForCausalLM,AutoModelForDepthEstimation:()=>t.AutoModelForDepthEstimation,AutoModelForDocumentQuestionAnswering:()=>t.AutoModelForDocumentQuestionAnswering,AutoModelForImageClassification:()=>t.AutoModelForImageClassification,AutoModelForImageFeatureExtraction:()=>t.AutoModelForImageFeatureExtraction,AutoModelForImageMatting:()=>t.AutoModelForImageMatting,AutoModelForImageSegmentation:()=>t.AutoModelForImageSegmentation,AutoModelForImageTextToText:()=>t.AutoModelForImageTextToText,AutoModelForImageToImage:()=>t.AutoModelForImageToImage,AutoModelForMaskGeneration:()=>t.AutoModelForMaskGeneration,AutoModelForMaskedLM:()=>t.AutoModelForMaskedLM,AutoModelForNormalEstimation:()=>t.AutoModelForNormalEstimation,AutoModelForObjectDetection:()=>t.AutoModelForObjectDetection,AutoModelForPoseEstimation:()=>t.AutoModelForPoseEstimation,AutoModelForQuestionAnswering:()=>t.AutoModelForQuestionAnswering,AutoModelForSemanticSegmentation:()=>t.AutoModelForSemanticSegmentation,AutoModelForSeq2SeqLM:()=>t.AutoModelForSeq2SeqLM,AutoModelForSequenceClassification:()=>t.AutoModelForSequenceClassification,AutoModelForSpeechSeq2Seq:()=>t.AutoModelForSpeechSeq2Seq,AutoModelForTextToSpectrogram:()=>t.AutoModelForTextToSpectrogram,AutoModelForTextToWaveform:()=>t.AutoModelForTextToWaveform,AutoModelForTokenClassification:()=>t.AutoModelForTokenClassification,AutoModelForUniversalSegmentation:()=>t.AutoModelForUniversalSegmentation,AutoModelForVision2Seq:()=>t.AutoModelForVision2Seq,AutoModelForXVector:()=>t.AutoModelForXVector,AutoModelForZeroShotObjectDetection:()=>t.AutoModelForZeroShotObjectDetection,AutoProcessor:()=>v.AutoProcessor,AutoTokenizer:()=>s.AutoTokenizer,AutomaticSpeechRecognitionPipeline:()=>r.AutomaticSpeechRecognitionPipeline,BackgroundRemovalPipeline:()=>r.BackgroundRemovalPipeline,BartForConditionalGeneration:()=>t.BartForConditionalGeneration,BartForSequenceClassification:()=>t.BartForSequenceClassification,BartModel:()=>t.BartModel,BartPretrainedModel:()=>t.BartPretrainedModel,BartTokenizer:()=>s.BartTokenizer,BaseModelOutput:()=>t.BaseModelOutput,BaseStreamer:()=>k.BaseStreamer,BeitFeatureExtractor:()=>_.BeitFeatureExtractor,BeitForImageClassification:()=>t.BeitForImageClassification,BeitModel:()=>t.BeitModel,BeitPreTrainedModel:()=>t.BeitPreTrainedModel,BertForMaskedLM:()=>t.BertForMaskedLM,BertForQuestionAnswering:()=>t.BertForQuestionAnswering,BertForSequenceClassification:()=>t.BertForSequenceClassification,BertForTokenClassification:()=>t.BertForTokenClassification,BertModel:()=>t.BertModel,BertPreTrainedModel:()=>t.BertPreTrainedModel,BertTokenizer:()=>s.BertTokenizer,BitImageProcessor:()=>_.BitImageProcessor,BlenderbotForConditionalGeneration:()=>t.BlenderbotForConditionalGeneration,BlenderbotModel:()=>t.BlenderbotModel,BlenderbotPreTrainedModel:()=>t.BlenderbotPreTrainedModel,BlenderbotSmallForConditionalGeneration:()=>t.BlenderbotSmallForConditionalGeneration,BlenderbotSmallModel:()=>t.BlenderbotSmallModel,BlenderbotSmallPreTrainedModel:()=>t.BlenderbotSmallPreTrainedModel,BlenderbotSmallTokenizer:()=>s.BlenderbotSmallTokenizer,BlenderbotTokenizer:()=>s.BlenderbotTokenizer,BloomForCausalLM:()=>t.BloomForCausalLM,BloomModel:()=>t.BloomModel,BloomPreTrainedModel:()=>t.BloomPreTrainedModel,BloomTokenizer:()=>s.BloomTokenizer,CLIPFeatureExtractor:()=>_.CLIPFeatureExtractor,CLIPImageProcessor:()=>_.CLIPImageProcessor,CLIPModel:()=>t.CLIPModel,CLIPPreTrainedModel:()=>t.CLIPPreTrainedModel,CLIPSegForImageSegmentation:()=>t.CLIPSegForImageSegmentation,CLIPSegModel:()=>t.CLIPSegModel,CLIPSegPreTrainedModel:()=>t.CLIPSegPreTrainedModel,CLIPTextModel:()=>t.CLIPTextModel,CLIPTextModelWithProjection:()=>t.CLIPTextModelWithProjection,CLIPTokenizer:()=>s.CLIPTokenizer,CLIPVisionModel:()=>t.CLIPVisionModel,CLIPVisionModelWithProjection:()=>t.CLIPVisionModelWithProjection,CamembertForMaskedLM:()=>t.CamembertForMaskedLM,CamembertForQuestionAnswering:()=>t.CamembertForQuestionAnswering,CamembertForSequenceClassification:()=>t.CamembertForSequenceClassification,CamembertForTokenClassification:()=>t.CamembertForTokenClassification,CamembertModel:()=>t.CamembertModel,CamembertPreTrainedModel:()=>t.CamembertPreTrainedModel,CamembertTokenizer:()=>s.CamembertTokenizer,CausalLMOutput:()=>t.CausalLMOutput,CausalLMOutputWithPast:()=>t.CausalLMOutputWithPast,ChineseCLIPFeatureExtractor:()=>_.ChineseCLIPFeatureExtractor,ChineseCLIPModel:()=>t.ChineseCLIPModel,ChineseCLIPPreTrainedModel:()=>t.ChineseCLIPPreTrainedModel,ClapAudioModelWithProjection:()=>t.ClapAudioModelWithProjection,ClapFeatureExtractor:()=>d.ClapFeatureExtractor,ClapModel:()=>t.ClapModel,ClapPreTrainedModel:()=>t.ClapPreTrainedModel,ClapTextModelWithProjection:()=>t.ClapTextModelWithProjection,ClassifierFreeGuidanceLogitsProcessor:()=>b.ClassifierFreeGuidanceLogitsProcessor,CodeGenForCausalLM:()=>t.CodeGenForCausalLM,CodeGenModel:()=>t.CodeGenModel,CodeGenPreTrainedModel:()=>t.CodeGenPreTrainedModel,CodeGenTokenizer:()=>s.CodeGenTokenizer,CodeLlamaTokenizer:()=>s.CodeLlamaTokenizer,CohereForCausalLM:()=>t.CohereForCausalLM,CohereModel:()=>t.CohereModel,CoherePreTrainedModel:()=>t.CoherePreTrainedModel,CohereTokenizer:()=>s.CohereTokenizer,ConvBertForMaskedLM:()=>t.ConvBertForMaskedLM,ConvBertForQuestionAnswering:()=>t.ConvBertForQuestionAnswering,ConvBertForSequenceClassification:()=>t.ConvBertForSequenceClassification,ConvBertForTokenClassification:()=>t.ConvBertForTokenClassification,ConvBertModel:()=>t.ConvBertModel,ConvBertPreTrainedModel:()=>t.ConvBertPreTrainedModel,ConvBertTokenizer:()=>s.ConvBertTokenizer,ConvNextFeatureExtractor:()=>_.ConvNextFeatureExtractor,ConvNextForImageClassification:()=>t.ConvNextForImageClassification,ConvNextImageProcessor:()=>_.ConvNextImageProcessor,ConvNextModel:()=>t.ConvNextModel,ConvNextPreTrainedModel:()=>t.ConvNextPreTrainedModel,ConvNextV2ForImageClassification:()=>t.ConvNextV2ForImageClassification,ConvNextV2Model:()=>t.ConvNextV2Model,ConvNextV2PreTrainedModel:()=>t.ConvNextV2PreTrainedModel,DFineForObjectDetection:()=>t.DFineForObjectDetection,DFineModel:()=>t.DFineModel,DFinePreTrainedModel:()=>t.DFinePreTrainedModel,DINOv3ConvNextModel:()=>t.DINOv3ConvNextModel,DINOv3ConvNextPreTrainedModel:()=>t.DINOv3ConvNextPreTrainedModel,DINOv3ViTImageProcessor:()=>_.DINOv3ViTImageProcessor,DINOv3ViTModel:()=>t.DINOv3ViTModel,DINOv3ViTPreTrainedModel:()=>t.DINOv3ViTPreTrainedModel,DPTFeatureExtractor:()=>_.DPTFeatureExtractor,DPTForDepthEstimation:()=>t.DPTForDepthEstimation,DPTImageProcessor:()=>_.DPTImageProcessor,DPTModel:()=>t.DPTModel,DPTPreTrainedModel:()=>t.DPTPreTrainedModel,DacDecoderModel:()=>t.DacDecoderModel,DacDecoderOutput:()=>t.DacDecoderOutput,DacEncoderModel:()=>t.DacEncoderModel,DacEncoderOutput:()=>t.DacEncoderOutput,DacFeatureExtractor:()=>d.DacFeatureExtractor,DacModel:()=>t.DacModel,DacPreTrainedModel:()=>t.DacPreTrainedModel,DataTypeMap:()=>l.DataTypeMap,DebertaForMaskedLM:()=>t.DebertaForMaskedLM,DebertaForQuestionAnswering:()=>t.DebertaForQuestionAnswering,DebertaForSequenceClassification:()=>t.DebertaForSequenceClassification,DebertaForTokenClassification:()=>t.DebertaForTokenClassification,DebertaModel:()=>t.DebertaModel,DebertaPreTrainedModel:()=>t.DebertaPreTrainedModel,DebertaTokenizer:()=>s.DebertaTokenizer,DebertaV2ForMaskedLM:()=>t.DebertaV2ForMaskedLM,DebertaV2ForQuestionAnswering:()=>t.DebertaV2ForQuestionAnswering,DebertaV2ForSequenceClassification:()=>t.DebertaV2ForSequenceClassification,DebertaV2ForTokenClassification:()=>t.DebertaV2ForTokenClassification,DebertaV2Model:()=>t.DebertaV2Model,DebertaV2PreTrainedModel:()=>t.DebertaV2PreTrainedModel,DebertaV2Tokenizer:()=>s.DebertaV2Tokenizer,DecisionTransformerModel:()=>t.DecisionTransformerModel,DecisionTransformerPreTrainedModel:()=>t.DecisionTransformerPreTrainedModel,DeiTFeatureExtractor:()=>_.DeiTFeatureExtractor,DeiTForImageClassification:()=>t.DeiTForImageClassification,DeiTImageProcessor:()=>_.DeiTImageProcessor,DeiTModel:()=>t.DeiTModel,DeiTPreTrainedModel:()=>t.DeiTPreTrainedModel,DepthAnythingForDepthEstimation:()=>t.DepthAnythingForDepthEstimation,DepthAnythingPreTrainedModel:()=>t.DepthAnythingPreTrainedModel,DepthEstimationPipeline:()=>r.DepthEstimationPipeline,DepthProForDepthEstimation:()=>t.DepthProForDepthEstimation,DepthProPreTrainedModel:()=>t.DepthProPreTrainedModel,DetrFeatureExtractor:()=>_.DetrFeatureExtractor,DetrForObjectDetection:()=>t.DetrForObjectDetection,DetrForSegmentation:()=>t.DetrForSegmentation,DetrImageProcessor:()=>_.DetrImageProcessor,DetrModel:()=>t.DetrModel,DetrObjectDetectionOutput:()=>t.DetrObjectDetectionOutput,DetrPreTrainedModel:()=>t.DetrPreTrainedModel,DetrSegmentationOutput:()=>t.DetrSegmentationOutput,Dinov2ForImageClassification:()=>t.Dinov2ForImageClassification,Dinov2Model:()=>t.Dinov2Model,Dinov2PreTrainedModel:()=>t.Dinov2PreTrainedModel,Dinov2WithRegistersForImageClassification:()=>t.Dinov2WithRegistersForImageClassification,Dinov2WithRegistersModel:()=>t.Dinov2WithRegistersModel,Dinov2WithRegistersPreTrainedModel:()=>t.Dinov2WithRegistersPreTrainedModel,DistilBertForMaskedLM:()=>t.DistilBertForMaskedLM,DistilBertForQuestionAnswering:()=>t.DistilBertForQuestionAnswering,DistilBertForSequenceClassification:()=>t.DistilBertForSequenceClassification,DistilBertForTokenClassification:()=>t.DistilBertForTokenClassification,DistilBertModel:()=>t.DistilBertModel,DistilBertPreTrainedModel:()=>t.DistilBertPreTrainedModel,DistilBertTokenizer:()=>s.DistilBertTokenizer,DocumentQuestionAnsweringPipeline:()=>r.DocumentQuestionAnsweringPipeline,DonutFeatureExtractor:()=>_.DonutFeatureExtractor,DonutImageProcessor:()=>_.DonutImageProcessor,DonutSwinModel:()=>t.DonutSwinModel,DonutSwinPreTrainedModel:()=>t.DonutSwinPreTrainedModel,EdgeTamModel:()=>t.EdgeTamModel,EfficientNetForImageClassification:()=>t.EfficientNetForImageClassification,EfficientNetImageProcessor:()=>_.EfficientNetImageProcessor,EfficientNetModel:()=>t.EfficientNetModel,EfficientNetPreTrainedModel:()=>t.EfficientNetPreTrainedModel,ElectraForMaskedLM:()=>t.ElectraForMaskedLM,ElectraForQuestionAnswering:()=>t.ElectraForQuestionAnswering,ElectraForSequenceClassification:()=>t.ElectraForSequenceClassification,ElectraForTokenClassification:()=>t.ElectraForTokenClassification,ElectraModel:()=>t.ElectraModel,ElectraPreTrainedModel:()=>t.ElectraPreTrainedModel,ElectraTokenizer:()=>s.ElectraTokenizer,EncodecFeatureExtractor:()=>d.EncodecFeatureExtractor,EosTokenCriteria:()=>T.EosTokenCriteria,Ernie4_5ForCausalLM:()=>t.Ernie4_5ForCausalLM,Ernie4_5Model:()=>t.Ernie4_5Model,Ernie4_5PreTrainedModel:()=>t.Ernie4_5PreTrainedModel,EsmForMaskedLM:()=>t.EsmForMaskedLM,EsmForSequenceClassification:()=>t.EsmForSequenceClassification,EsmForTokenClassification:()=>t.EsmForTokenClassification,EsmModel:()=>t.EsmModel,EsmPreTrainedModel:()=>t.EsmPreTrainedModel,EsmTokenizer:()=>s.EsmTokenizer,ExaoneForCausalLM:()=>t.ExaoneForCausalLM,ExaoneModel:()=>t.ExaoneModel,ExaonePreTrainedModel:()=>t.ExaonePreTrainedModel,FFT:()=>c.FFT,FalconForCausalLM:()=>t.FalconForCausalLM,FalconModel:()=>t.FalconModel,FalconPreTrainedModel:()=>t.FalconPreTrainedModel,FalconTokenizer:()=>s.FalconTokenizer,FastViTForImageClassification:()=>t.FastViTForImageClassification,FastViTModel:()=>t.FastViTModel,FastViTPreTrainedModel:()=>t.FastViTPreTrainedModel,FeatureExtractionPipeline:()=>r.FeatureExtractionPipeline,FeatureExtractor:()=>p.FeatureExtractor,FillMaskPipeline:()=>r.FillMaskPipeline,Florence2ForConditionalGeneration:()=>t.Florence2ForConditionalGeneration,Florence2PreTrainedModel:()=>t.Florence2PreTrainedModel,Florence2Processor:()=>w.Florence2Processor,ForcedBOSTokenLogitsProcessor:()=>b.ForcedBOSTokenLogitsProcessor,ForcedEOSTokenLogitsProcessor:()=>b.ForcedEOSTokenLogitsProcessor,GLPNFeatureExtractor:()=>_.GLPNFeatureExtractor,GLPNForDepthEstimation:()=>t.GLPNForDepthEstimation,GLPNModel:()=>t.GLPNModel,GLPNPreTrainedModel:()=>t.GLPNPreTrainedModel,GPT2LMHeadModel:()=>t.GPT2LMHeadModel,GPT2Model:()=>t.GPT2Model,GPT2PreTrainedModel:()=>t.GPT2PreTrainedModel,GPT2Tokenizer:()=>s.GPT2Tokenizer,GPTBigCodeForCausalLM:()=>t.GPTBigCodeForCausalLM,GPTBigCodeModel:()=>t.GPTBigCodeModel,GPTBigCodePreTrainedModel:()=>t.GPTBigCodePreTrainedModel,GPTJForCausalLM:()=>t.GPTJForCausalLM,GPTJModel:()=>t.GPTJModel,GPTJPreTrainedModel:()=>t.GPTJPreTrainedModel,GPTNeoForCausalLM:()=>t.GPTNeoForCausalLM,GPTNeoModel:()=>t.GPTNeoModel,GPTNeoPreTrainedModel:()=>t.GPTNeoPreTrainedModel,GPTNeoXForCausalLM:()=>t.GPTNeoXForCausalLM,GPTNeoXModel:()=>t.GPTNeoXModel,GPTNeoXPreTrainedModel:()=>t.GPTNeoXPreTrainedModel,GPTNeoXTokenizer:()=>s.GPTNeoXTokenizer,Gemma2ForCausalLM:()=>t.Gemma2ForCausalLM,Gemma2Model:()=>t.Gemma2Model,Gemma2PreTrainedModel:()=>t.Gemma2PreTrainedModel,Gemma3ForCausalLM:()=>t.Gemma3ForCausalLM,Gemma3Model:()=>t.Gemma3Model,Gemma3PreTrainedModel:()=>t.Gemma3PreTrainedModel,Gemma3nAudioFeatureExtractor:()=>d.Gemma3nAudioFeatureExtractor,Gemma3nForConditionalGeneration:()=>t.Gemma3nForConditionalGeneration,Gemma3nPreTrainedModel:()=>t.Gemma3nPreTrainedModel,Gemma3nProcessor:()=>w.Gemma3nProcessor,GemmaForCausalLM:()=>t.GemmaForCausalLM,GemmaModel:()=>t.GemmaModel,GemmaPreTrainedModel:()=>t.GemmaPreTrainedModel,GemmaTokenizer:()=>s.GemmaTokenizer,GlmForCausalLM:()=>t.GlmForCausalLM,GlmModel:()=>t.GlmModel,GlmPreTrainedModel:()=>t.GlmPreTrainedModel,GraniteForCausalLM:()=>t.GraniteForCausalLM,GraniteModel:()=>t.GraniteModel,GraniteMoeHybridForCausalLM:()=>t.GraniteMoeHybridForCausalLM,GraniteMoeHybridModel:()=>t.GraniteMoeHybridModel,GraniteMoeHybridPreTrainedModel:()=>t.GraniteMoeHybridPreTrainedModel,GranitePreTrainedModel:()=>t.GranitePreTrainedModel,Grok1Tokenizer:()=>s.Grok1Tokenizer,GroundingDinoForObjectDetection:()=>t.GroundingDinoForObjectDetection,GroundingDinoImageProcessor:()=>_.GroundingDinoImageProcessor,GroundingDinoPreTrainedModel:()=>t.GroundingDinoPreTrainedModel,GroundingDinoProcessor:()=>w.GroundingDinoProcessor,GroupViTModel:()=>t.GroupViTModel,GroupViTPreTrainedModel:()=>t.GroupViTPreTrainedModel,HeliumForCausalLM:()=>t.HeliumForCausalLM,HeliumModel:()=>t.HeliumModel,HeliumPreTrainedModel:()=>t.HeliumPreTrainedModel,HerbertTokenizer:()=>s.HerbertTokenizer,HieraForImageClassification:()=>t.HieraForImageClassification,HieraModel:()=>t.HieraModel,HieraPreTrainedModel:()=>t.HieraPreTrainedModel,HubertForCTC:()=>t.HubertForCTC,HubertForSequenceClassification:()=>t.HubertForSequenceClassification,HubertModel:()=>t.HubertModel,HubertPreTrainedModel:()=>t.HubertPreTrainedModel,IJepaForImageClassification:()=>t.IJepaForImageClassification,IJepaModel:()=>t.IJepaModel,IJepaPreTrainedModel:()=>t.IJepaPreTrainedModel,Idefics3ForConditionalGeneration:()=>t.Idefics3ForConditionalGeneration,Idefics3ImageProcessor:()=>_.Idefics3ImageProcessor,Idefics3PreTrainedModel:()=>t.Idefics3PreTrainedModel,Idefics3Processor:()=>w.Idefics3Processor,ImageClassificationPipeline:()=>r.ImageClassificationPipeline,ImageFeatureExtractionPipeline:()=>r.ImageFeatureExtractionPipeline,ImageFeatureExtractor:()=>d.ImageFeatureExtractor,ImageMattingOutput:()=>t.ImageMattingOutput,ImageProcessor:()=>f.ImageProcessor,ImageSegmentationPipeline:()=>r.ImageSegmentationPipeline,ImageToImagePipeline:()=>r.ImageToImagePipeline,ImageToTextPipeline:()=>r.ImageToTextPipeline,InterruptableStoppingCriteria:()=>T.InterruptableStoppingCriteria,JAISLMHeadModel:()=>t.JAISLMHeadModel,JAISModel:()=>t.JAISModel,JAISPreTrainedModel:()=>t.JAISPreTrainedModel,JinaCLIPImageProcessor:()=>_.JinaCLIPImageProcessor,JinaCLIPModel:()=>t.JinaCLIPModel,JinaCLIPPreTrainedModel:()=>t.JinaCLIPPreTrainedModel,JinaCLIPProcessor:()=>w.JinaCLIPProcessor,JinaCLIPTextModel:()=>t.JinaCLIPTextModel,JinaCLIPVisionModel:()=>t.JinaCLIPVisionModel,Lfm2ForCausalLM:()=>t.Lfm2ForCausalLM,Lfm2Model:()=>t.Lfm2Model,Lfm2PreTrainedModel:()=>t.Lfm2PreTrainedModel,LiteWhisperForConditionalGeneration:()=>t.LiteWhisperForConditionalGeneration,Llama4ForCausalLM:()=>t.Llama4ForCausalLM,Llama4PreTrainedModel:()=>t.Llama4PreTrainedModel,LlamaForCausalLM:()=>t.LlamaForCausalLM,LlamaModel:()=>t.LlamaModel,LlamaPreTrainedModel:()=>t.LlamaPreTrainedModel,LlamaTokenizer:()=>s.LlamaTokenizer,LlavaForConditionalGeneration:()=>t.LlavaForConditionalGeneration,LlavaOnevisionForConditionalGeneration:()=>t.LlavaOnevisionForConditionalGeneration,LlavaOnevisionImageProcessor:()=>_.LlavaOnevisionImageProcessor,LlavaPreTrainedModel:()=>t.LlavaPreTrainedModel,LlavaProcessor:()=>w.LlavaProcessor,LlavaQwen2ForCausalLM:()=>t.LlavaQwen2ForCausalLM,LogitsProcessor:()=>b.LogitsProcessor,LogitsProcessorList:()=>b.LogitsProcessorList,LogitsWarper:()=>b.LogitsWarper,LongT5ForConditionalGeneration:()=>t.LongT5ForConditionalGeneration,LongT5Model:()=>t.LongT5Model,LongT5PreTrainedModel:()=>t.LongT5PreTrainedModel,M2M100ForConditionalGeneration:()=>t.M2M100ForConditionalGeneration,M2M100Model:()=>t.M2M100Model,M2M100PreTrainedModel:()=>t.M2M100PreTrainedModel,M2M100Tokenizer:()=>s.M2M100Tokenizer,MBart50Tokenizer:()=>s.MBart50Tokenizer,MBartForCausalLM:()=>t.MBartForCausalLM,MBartForConditionalGeneration:()=>t.MBartForConditionalGeneration,MBartForSequenceClassification:()=>t.MBartForSequenceClassification,MBartModel:()=>t.MBartModel,MBartPreTrainedModel:()=>t.MBartPreTrainedModel,MBartTokenizer:()=>s.MBartTokenizer,MPNetForMaskedLM:()=>t.MPNetForMaskedLM,MPNetForQuestionAnswering:()=>t.MPNetForQuestionAnswering,MPNetForSequenceClassification:()=>t.MPNetForSequenceClassification,MPNetForTokenClassification:()=>t.MPNetForTokenClassification,MPNetModel:()=>t.MPNetModel,MPNetPreTrainedModel:()=>t.MPNetPreTrainedModel,MPNetTokenizer:()=>s.MPNetTokenizer,MT5ForConditionalGeneration:()=>t.MT5ForConditionalGeneration,MT5Model:()=>t.MT5Model,MT5PreTrainedModel:()=>t.MT5PreTrainedModel,MarianMTModel:()=>t.MarianMTModel,MarianModel:()=>t.MarianModel,MarianPreTrainedModel:()=>t.MarianPreTrainedModel,MarianTokenizer:()=>s.MarianTokenizer,Mask2FormerImageProcessor:()=>_.Mask2FormerImageProcessor,MaskFormerFeatureExtractor:()=>_.MaskFormerFeatureExtractor,MaskFormerForInstanceSegmentation:()=>t.MaskFormerForInstanceSegmentation,MaskFormerImageProcessor:()=>_.MaskFormerImageProcessor,MaskFormerModel:()=>t.MaskFormerModel,MaskFormerPreTrainedModel:()=>t.MaskFormerPreTrainedModel,MaskedLMOutput:()=>t.MaskedLMOutput,MaxLengthCriteria:()=>T.MaxLengthCriteria,Metric3DForDepthEstimation:()=>t.Metric3DForDepthEstimation,Metric3DPreTrainedModel:()=>t.Metric3DPreTrainedModel,Metric3Dv2ForDepthEstimation:()=>t.Metric3Dv2ForDepthEstimation,Metric3Dv2PreTrainedModel:()=>t.Metric3Dv2PreTrainedModel,MgpstrForSceneTextRecognition:()=>t.MgpstrForSceneTextRecognition,MgpstrModelOutput:()=>t.MgpstrModelOutput,MgpstrPreTrainedModel:()=>t.MgpstrPreTrainedModel,MgpstrProcessor:()=>w.MgpstrProcessor,MgpstrTokenizer:()=>s.MgpstrTokenizer,MimiDecoderModel:()=>t.MimiDecoderModel,MimiDecoderOutput:()=>t.MimiDecoderOutput,MimiEncoderModel:()=>t.MimiEncoderModel,MimiEncoderOutput:()=>t.MimiEncoderOutput,MimiModel:()=>t.MimiModel,MimiPreTrainedModel:()=>t.MimiPreTrainedModel,MinLengthLogitsProcessor:()=>b.MinLengthLogitsProcessor,MinNewTokensLengthLogitsProcessor:()=>b.MinNewTokensLengthLogitsProcessor,Ministral3ForCausalLM:()=>t.Ministral3ForCausalLM,Ministral3Model:()=>t.Ministral3Model,Ministral3PreTrainedModel:()=>t.Ministral3PreTrainedModel,MinistralForCausalLM:()=>t.MinistralForCausalLM,MinistralModel:()=>t.MinistralModel,MinistralPreTrainedModel:()=>t.MinistralPreTrainedModel,Mistral3ForConditionalGeneration:()=>t.Mistral3ForConditionalGeneration,MistralForCausalLM:()=>t.MistralForCausalLM,MistralModel:()=>t.MistralModel,MistralPreTrainedModel:()=>t.MistralPreTrainedModel,MobileBertForMaskedLM:()=>t.MobileBertForMaskedLM,MobileBertForQuestionAnswering:()=>t.MobileBertForQuestionAnswering,MobileBertForSequenceClassification:()=>t.MobileBertForSequenceClassification,MobileBertModel:()=>t.MobileBertModel,MobileBertPreTrainedModel:()=>t.MobileBertPreTrainedModel,MobileBertTokenizer:()=>s.MobileBertTokenizer,MobileLLMForCausalLM:()=>t.MobileLLMForCausalLM,MobileLLMModel:()=>t.MobileLLMModel,MobileLLMPreTrainedModel:()=>t.MobileLLMPreTrainedModel,MobileNetV1FeatureExtractor:()=>_.MobileNetV1FeatureExtractor,MobileNetV1ForImageClassification:()=>t.MobileNetV1ForImageClassification,MobileNetV1ForSemanticSegmentation:()=>t.MobileNetV1ForSemanticSegmentation,MobileNetV1ImageProcessor:()=>_.MobileNetV1ImageProcessor,MobileNetV1Model:()=>t.MobileNetV1Model,MobileNetV1PreTrainedModel:()=>t.MobileNetV1PreTrainedModel,MobileNetV2FeatureExtractor:()=>_.MobileNetV2FeatureExtractor,MobileNetV2ForImageClassification:()=>t.MobileNetV2ForImageClassification,MobileNetV2ForSemanticSegmentation:()=>t.MobileNetV2ForSemanticSegmentation,MobileNetV2ImageProcessor:()=>_.MobileNetV2ImageProcessor,MobileNetV2Model:()=>t.MobileNetV2Model,MobileNetV2PreTrainedModel:()=>t.MobileNetV2PreTrainedModel,MobileNetV3FeatureExtractor:()=>_.MobileNetV3FeatureExtractor,MobileNetV3ForImageClassification:()=>t.MobileNetV3ForImageClassification,MobileNetV3ForSemanticSegmentation:()=>t.MobileNetV3ForSemanticSegmentation,MobileNetV3ImageProcessor:()=>_.MobileNetV3ImageProcessor,MobileNetV3Model:()=>t.MobileNetV3Model,MobileNetV3PreTrainedModel:()=>t.MobileNetV3PreTrainedModel,MobileNetV4FeatureExtractor:()=>_.MobileNetV4FeatureExtractor,MobileNetV4ForImageClassification:()=>t.MobileNetV4ForImageClassification,MobileNetV4ForSemanticSegmentation:()=>t.MobileNetV4ForSemanticSegmentation,MobileNetV4ImageProcessor:()=>_.MobileNetV4ImageProcessor,MobileNetV4Model:()=>t.MobileNetV4Model,MobileNetV4PreTrainedModel:()=>t.MobileNetV4PreTrainedModel,MobileViTFeatureExtractor:()=>_.MobileViTFeatureExtractor,MobileViTForImageClassification:()=>t.MobileViTForImageClassification,MobileViTImageProcessor:()=>_.MobileViTImageProcessor,MobileViTModel:()=>t.MobileViTModel,MobileViTPreTrainedModel:()=>t.MobileViTPreTrainedModel,MobileViTV2ForImageClassification:()=>t.MobileViTV2ForImageClassification,MobileViTV2Model:()=>t.MobileViTV2Model,MobileViTV2PreTrainedModel:()=>t.MobileViTV2PreTrainedModel,ModelOutput:()=>t.ModelOutput,ModernBertDecoderForCausalLM:()=>t.ModernBertDecoderForCausalLM,ModernBertDecoderModel:()=>t.ModernBertDecoderModel,ModernBertDecoderPreTrainedModel:()=>t.ModernBertDecoderPreTrainedModel,ModernBertForMaskedLM:()=>t.ModernBertForMaskedLM,ModernBertForSequenceClassification:()=>t.ModernBertForSequenceClassification,ModernBertForTokenClassification:()=>t.ModernBertForTokenClassification,ModernBertModel:()=>t.ModernBertModel,ModernBertPreTrainedModel:()=>t.ModernBertPreTrainedModel,Moondream1ForConditionalGeneration:()=>t.Moondream1ForConditionalGeneration,MoonshineFeatureExtractor:()=>d.MoonshineFeatureExtractor,MoonshineForConditionalGeneration:()=>t.MoonshineForConditionalGeneration,MoonshineModel:()=>t.MoonshineModel,MoonshinePreTrainedModel:()=>t.MoonshinePreTrainedModel,MoonshineProcessor:()=>w.MoonshineProcessor,MptForCausalLM:()=>t.MptForCausalLM,MptModel:()=>t.MptModel,MptPreTrainedModel:()=>t.MptPreTrainedModel,MultiModalityCausalLM:()=>t.MultiModalityCausalLM,MultiModalityPreTrainedModel:()=>t.MultiModalityPreTrainedModel,MusicgenForCausalLM:()=>t.MusicgenForCausalLM,MusicgenForConditionalGeneration:()=>t.MusicgenForConditionalGeneration,MusicgenModel:()=>t.MusicgenModel,MusicgenPreTrainedModel:()=>t.MusicgenPreTrainedModel,NanoChatForCausalLM:()=>t.NanoChatForCausalLM,NanoChatModel:()=>t.NanoChatModel,NanoChatPreTrainedModel:()=>t.NanoChatPreTrainedModel,NeoBertForMaskedLM:()=>t.NeoBertForMaskedLM,NeoBertForQuestionAnswering:()=>t.NeoBertForQuestionAnswering,NeoBertForSequenceClassification:()=>t.NeoBertForSequenceClassification,NeoBertForTokenClassification:()=>t.NeoBertForTokenClassification,NeoBertModel:()=>t.NeoBertModel,NeoBertPreTrainedModel:()=>t.NeoBertPreTrainedModel,NllbTokenizer:()=>s.NllbTokenizer,NoBadWordsLogitsProcessor:()=>b.NoBadWordsLogitsProcessor,NoRepeatNGramLogitsProcessor:()=>b.NoRepeatNGramLogitsProcessor,NomicBertModel:()=>t.NomicBertModel,NomicBertPreTrainedModel:()=>t.NomicBertPreTrainedModel,NougatImageProcessor:()=>_.NougatImageProcessor,NougatTokenizer:()=>s.NougatTokenizer,OPTForCausalLM:()=>t.OPTForCausalLM,OPTModel:()=>t.OPTModel,OPTPreTrainedModel:()=>t.OPTPreTrainedModel,ObjectDetectionPipeline:()=>r.ObjectDetectionPipeline,Olmo2ForCausalLM:()=>t.Olmo2ForCausalLM,Olmo2Model:()=>t.Olmo2Model,Olmo2PreTrainedModel:()=>t.Olmo2PreTrainedModel,OlmoForCausalLM:()=>t.OlmoForCausalLM,OlmoModel:()=>t.OlmoModel,OlmoPreTrainedModel:()=>t.OlmoPreTrainedModel,OpenELMForCausalLM:()=>t.OpenELMForCausalLM,OpenELMModel:()=>t.OpenELMModel,OpenELMPreTrainedModel:()=>t.OpenELMPreTrainedModel,OwlViTFeatureExtractor:()=>_.OwlViTFeatureExtractor,OwlViTForObjectDetection:()=>t.OwlViTForObjectDetection,OwlViTImageProcessor:()=>_.OwlViTImageProcessor,OwlViTModel:()=>t.OwlViTModel,OwlViTPreTrainedModel:()=>t.OwlViTPreTrainedModel,OwlViTProcessor:()=>w.OwlViTProcessor,Owlv2ForObjectDetection:()=>t.Owlv2ForObjectDetection,Owlv2ImageProcessor:()=>_.Owlv2ImageProcessor,Owlv2Model:()=>t.Owlv2Model,Owlv2PreTrainedModel:()=>t.Owlv2PreTrainedModel,PaliGemmaForConditionalGeneration:()=>t.PaliGemmaForConditionalGeneration,PaliGemmaPreTrainedModel:()=>t.PaliGemmaPreTrainedModel,PaliGemmaProcessor:()=>w.PaliGemmaProcessor,ParakeetFeatureExtractor:()=>d.ParakeetFeatureExtractor,ParakeetForCTC:()=>t.ParakeetForCTC,ParakeetPreTrainedModel:()=>t.ParakeetPreTrainedModel,PatchTSMixerForPrediction:()=>t.PatchTSMixerForPrediction,PatchTSMixerModel:()=>t.PatchTSMixerModel,PatchTSMixerPreTrainedModel:()=>t.PatchTSMixerPreTrainedModel,PatchTSTForPrediction:()=>t.PatchTSTForPrediction,PatchTSTModel:()=>t.PatchTSTModel,PatchTSTPreTrainedModel:()=>t.PatchTSTPreTrainedModel,Phi3ForCausalLM:()=>t.Phi3ForCausalLM,Phi3Model:()=>t.Phi3Model,Phi3PreTrainedModel:()=>t.Phi3PreTrainedModel,Phi3VForCausalLM:()=>t.Phi3VForCausalLM,Phi3VImageProcessor:()=>_.Phi3VImageProcessor,Phi3VPreTrainedModel:()=>t.Phi3VPreTrainedModel,Phi3VProcessor:()=>w.Phi3VProcessor,PhiForCausalLM:()=>t.PhiForCausalLM,PhiModel:()=>t.PhiModel,PhiPreTrainedModel:()=>t.PhiPreTrainedModel,Pipeline:()=>r.Pipeline,PixtralImageProcessor:()=>_.PixtralImageProcessor,PixtralProcessor:()=>w.PixtralProcessor,PreTrainedModel:()=>t.PreTrainedModel,PreTrainedTokenizer:()=>s.PreTrainedTokenizer,PretrainedConfig:()=>n.PretrainedConfig,PretrainedMixin:()=>t.PretrainedMixin,Processor:()=>I.Processor,PvtForImageClassification:()=>t.PvtForImageClassification,PvtImageProcessor:()=>_.PvtImageProcessor,PvtModel:()=>t.PvtModel,PvtPreTrainedModel:()=>t.PvtPreTrainedModel,PyAnnoteFeatureExtractor:()=>d.PyAnnoteFeatureExtractor,PyAnnoteForAudioFrameClassification:()=>t.PyAnnoteForAudioFrameClassification,PyAnnoteModel:()=>t.PyAnnoteModel,PyAnnotePreTrainedModel:()=>t.PyAnnotePreTrainedModel,PyAnnoteProcessor:()=>w.PyAnnoteProcessor,QuestionAnsweringModelOutput:()=>t.QuestionAnsweringModelOutput,QuestionAnsweringPipeline:()=>r.QuestionAnsweringPipeline,Qwen2ForCausalLM:()=>t.Qwen2ForCausalLM,Qwen2Model:()=>t.Qwen2Model,Qwen2PreTrainedModel:()=>t.Qwen2PreTrainedModel,Qwen2Tokenizer:()=>s.Qwen2Tokenizer,Qwen2VLForConditionalGeneration:()=>t.Qwen2VLForConditionalGeneration,Qwen2VLImageProcessor:()=>_.Qwen2VLImageProcessor,Qwen2VLPreTrainedModel:()=>t.Qwen2VLPreTrainedModel,Qwen2VLProcessor:()=>w.Qwen2VLProcessor,Qwen3ForCausalLM:()=>t.Qwen3ForCausalLM,Qwen3Model:()=>t.Qwen3Model,Qwen3PreTrainedModel:()=>t.Qwen3PreTrainedModel,RFDetrForObjectDetection:()=>t.RFDetrForObjectDetection,RFDetrModel:()=>t.RFDetrModel,RFDetrObjectDetectionOutput:()=>t.RFDetrObjectDetectionOutput,RFDetrPreTrainedModel:()=>t.RFDetrPreTrainedModel,RTDetrForObjectDetection:()=>t.RTDetrForObjectDetection,RTDetrImageProcessor:()=>_.RTDetrImageProcessor,RTDetrModel:()=>t.RTDetrModel,RTDetrObjectDetectionOutput:()=>t.RTDetrObjectDetectionOutput,RTDetrPreTrainedModel:()=>t.RTDetrPreTrainedModel,RTDetrV2ForObjectDetection:()=>t.RTDetrV2ForObjectDetection,RTDetrV2Model:()=>t.RTDetrV2Model,RTDetrV2ObjectDetectionOutput:()=>t.RTDetrV2ObjectDetectionOutput,RTDetrV2PreTrainedModel:()=>t.RTDetrV2PreTrainedModel,RawAudio:()=>o.RawAudio,RawImage:()=>i.RawImage,RawVideo:()=>a.RawVideo,RawVideoFrame:()=>a.RawVideoFrame,RepetitionPenaltyLogitsProcessor:()=>b.RepetitionPenaltyLogitsProcessor,ResNetForImageClassification:()=>t.ResNetForImageClassification,ResNetModel:()=>t.ResNetModel,ResNetPreTrainedModel:()=>t.ResNetPreTrainedModel,RoFormerForMaskedLM:()=>t.RoFormerForMaskedLM,RoFormerForQuestionAnswering:()=>t.RoFormerForQuestionAnswering,RoFormerForSequenceClassification:()=>t.RoFormerForSequenceClassification,RoFormerForTokenClassification:()=>t.RoFormerForTokenClassification,RoFormerModel:()=>t.RoFormerModel,RoFormerPreTrainedModel:()=>t.RoFormerPreTrainedModel,RoFormerTokenizer:()=>s.RoFormerTokenizer,RobertaForMaskedLM:()=>t.RobertaForMaskedLM,RobertaForQuestionAnswering:()=>t.RobertaForQuestionAnswering,RobertaForSequenceClassification:()=>t.RobertaForSequenceClassification,RobertaForTokenClassification:()=>t.RobertaForTokenClassification,RobertaModel:()=>t.RobertaModel,RobertaPreTrainedModel:()=>t.RobertaPreTrainedModel,RobertaTokenizer:()=>s.RobertaTokenizer,Sam2ImageProcessor:()=>_.Sam2ImageProcessor,Sam2ImageSegmentationOutput:()=>t.Sam2ImageSegmentationOutput,Sam2Model:()=>t.Sam2Model,Sam2PreTrainedModel:()=>t.Sam2PreTrainedModel,Sam2Processor:()=>w.Sam2Processor,Sam2VideoProcessor:()=>w.Sam2VideoProcessor,Sam3ImageProcessor:()=>_.Sam3ImageProcessor,Sam3TrackerModel:()=>t.Sam3TrackerModel,SamImageProcessor:()=>_.SamImageProcessor,SamImageSegmentationOutput:()=>t.SamImageSegmentationOutput,SamModel:()=>t.SamModel,SamPreTrainedModel:()=>t.SamPreTrainedModel,SamProcessor:()=>w.SamProcessor,SapiensForDepthEstimation:()=>t.SapiensForDepthEstimation,SapiensForNormalEstimation:()=>t.SapiensForNormalEstimation,SapiensForSemanticSegmentation:()=>t.SapiensForSemanticSegmentation,SapiensPreTrainedModel:()=>t.SapiensPreTrainedModel,SeamlessM4TFeatureExtractor:()=>d.SeamlessM4TFeatureExtractor,SegformerFeatureExtractor:()=>_.SegformerFeatureExtractor,SegformerForImageClassification:()=>t.SegformerForImageClassification,SegformerForSemanticSegmentation:()=>t.SegformerForSemanticSegmentation,SegformerImageProcessor:()=>_.SegformerImageProcessor,SegformerModel:()=>t.SegformerModel,SegformerPreTrainedModel:()=>t.SegformerPreTrainedModel,Seq2SeqLMOutput:()=>t.Seq2SeqLMOutput,SequenceClassifierOutput:()=>t.SequenceClassifierOutput,SiglipImageProcessor:()=>_.SiglipImageProcessor,SiglipModel:()=>t.SiglipModel,SiglipPreTrainedModel:()=>t.SiglipPreTrainedModel,SiglipTextModel:()=>t.SiglipTextModel,SiglipTokenizer:()=>s.SiglipTokenizer,SiglipVisionModel:()=>t.SiglipVisionModel,SmolLM3ForCausalLM:()=>t.SmolLM3ForCausalLM,SmolLM3Model:()=>t.SmolLM3Model,SmolLM3PreTrainedModel:()=>t.SmolLM3PreTrainedModel,SmolVLMForConditionalGeneration:()=>t.SmolVLMForConditionalGeneration,SmolVLMImageProcessor:()=>_.SmolVLMImageProcessor,SmolVLMProcessor:()=>w.SmolVLMProcessor,SnacDecoderModel:()=>t.SnacDecoderModel,SnacEncoderModel:()=>t.SnacEncoderModel,SnacFeatureExtractor:()=>d.SnacFeatureExtractor,SnacModel:()=>t.SnacModel,SnacPreTrainedModel:()=>t.SnacPreTrainedModel,SpeechT5FeatureExtractor:()=>d.SpeechT5FeatureExtractor,SpeechT5ForSpeechToText:()=>t.SpeechT5ForSpeechToText,SpeechT5ForTextToSpeech:()=>t.SpeechT5ForTextToSpeech,SpeechT5HifiGan:()=>t.SpeechT5HifiGan,SpeechT5Model:()=>t.SpeechT5Model,SpeechT5PreTrainedModel:()=>t.SpeechT5PreTrainedModel,SpeechT5Processor:()=>w.SpeechT5Processor,SpeechT5Tokenizer:()=>s.SpeechT5Tokenizer,SqueezeBertForMaskedLM:()=>t.SqueezeBertForMaskedLM,SqueezeBertForQuestionAnswering:()=>t.SqueezeBertForQuestionAnswering,SqueezeBertForSequenceClassification:()=>t.SqueezeBertForSequenceClassification,SqueezeBertModel:()=>t.SqueezeBertModel,SqueezeBertPreTrainedModel:()=>t.SqueezeBertPreTrainedModel,SqueezeBertTokenizer:()=>s.SqueezeBertTokenizer,StableLmForCausalLM:()=>t.StableLmForCausalLM,StableLmModel:()=>t.StableLmModel,StableLmPreTrainedModel:()=>t.StableLmPreTrainedModel,Starcoder2ForCausalLM:()=>t.Starcoder2ForCausalLM,Starcoder2Model:()=>t.Starcoder2Model,Starcoder2PreTrainedModel:()=>t.Starcoder2PreTrainedModel,StoppingCriteria:()=>T.StoppingCriteria,StoppingCriteriaList:()=>T.StoppingCriteriaList,StyleTextToSpeech2Model:()=>t.StyleTextToSpeech2Model,StyleTextToSpeech2PreTrainedModel:()=>t.StyleTextToSpeech2PreTrainedModel,SummarizationPipeline:()=>r.SummarizationPipeline,SupertonicForConditionalGeneration:()=>t.SupertonicForConditionalGeneration,SupertonicPreTrainedModel:()=>t.SupertonicPreTrainedModel,SuppressTokensAtBeginLogitsProcessor:()=>b.SuppressTokensAtBeginLogitsProcessor,Swin2SRForImageSuperResolution:()=>t.Swin2SRForImageSuperResolution,Swin2SRImageProcessor:()=>_.Swin2SRImageProcessor,Swin2SRModel:()=>t.Swin2SRModel,Swin2SRPreTrainedModel:()=>t.Swin2SRPreTrainedModel,SwinForImageClassification:()=>t.SwinForImageClassification,SwinForSemanticSegmentation:()=>t.SwinForSemanticSegmentation,SwinModel:()=>t.SwinModel,SwinPreTrainedModel:()=>t.SwinPreTrainedModel,T5ForConditionalGeneration:()=>t.T5ForConditionalGeneration,T5Model:()=>t.T5Model,T5PreTrainedModel:()=>t.T5PreTrainedModel,T5Tokenizer:()=>s.T5Tokenizer,TableTransformerForObjectDetection:()=>t.TableTransformerForObjectDetection,TableTransformerModel:()=>t.TableTransformerModel,TableTransformerObjectDetectionOutput:()=>t.TableTransformerObjectDetectionOutput,TableTransformerPreTrainedModel:()=>t.TableTransformerPreTrainedModel,TemperatureLogitsWarper:()=>b.TemperatureLogitsWarper,Tensor:()=>l.Tensor,Text2TextGenerationPipeline:()=>r.Text2TextGenerationPipeline,TextClassificationPipeline:()=>r.TextClassificationPipeline,TextGenerationPipeline:()=>r.TextGenerationPipeline,TextStreamer:()=>k.TextStreamer,TextToAudioPipeline:()=>r.TextToAudioPipeline,TokenClassificationPipeline:()=>r.TokenClassificationPipeline,TokenClassifierOutput:()=>t.TokenClassifierOutput,TokenizerModel:()=>s.TokenizerModel,TopKLogitsWarper:()=>b.TopKLogitsWarper,TopPLogitsWarper:()=>b.TopPLogitsWarper,TrOCRForCausalLM:()=>t.TrOCRForCausalLM,TrOCRPreTrainedModel:()=>t.TrOCRPreTrainedModel,TranslationPipeline:()=>r.TranslationPipeline,UltravoxModel:()=>t.UltravoxModel,UltravoxPreTrainedModel:()=>t.UltravoxPreTrainedModel,UltravoxProcessor:()=>w.UltravoxProcessor,UniSpeechForCTC:()=>t.UniSpeechForCTC,UniSpeechForSequenceClassification:()=>t.UniSpeechForSequenceClassification,UniSpeechModel:()=>t.UniSpeechModel,UniSpeechPreTrainedModel:()=>t.UniSpeechPreTrainedModel,UniSpeechSatForAudioFrameClassification:()=>t.UniSpeechSatForAudioFrameClassification,UniSpeechSatForCTC:()=>t.UniSpeechSatForCTC,UniSpeechSatForSequenceClassification:()=>t.UniSpeechSatForSequenceClassification,UniSpeechSatModel:()=>t.UniSpeechSatModel,UniSpeechSatPreTrainedModel:()=>t.UniSpeechSatPreTrainedModel,VLChatProcessor:()=>w.VLChatProcessor,VLMImageProcessor:()=>_.VLMImageProcessor,VaultGemmaForCausalLM:()=>t.VaultGemmaForCausalLM,VaultGemmaModel:()=>t.VaultGemmaModel,VaultGemmaPreTrainedModel:()=>t.VaultGemmaPreTrainedModel,ViTFeatureExtractor:()=>_.ViTFeatureExtractor,ViTForImageClassification:()=>t.ViTForImageClassification,ViTImageProcessor:()=>_.ViTImageProcessor,ViTMAEModel:()=>t.ViTMAEModel,ViTMAEPreTrainedModel:()=>t.ViTMAEPreTrainedModel,ViTMSNForImageClassification:()=>t.ViTMSNForImageClassification,ViTMSNModel:()=>t.ViTMSNModel,ViTMSNPreTrainedModel:()=>t.ViTMSNPreTrainedModel,ViTModel:()=>t.ViTModel,ViTPreTrainedModel:()=>t.ViTPreTrainedModel,VisionEncoderDecoderModel:()=>t.VisionEncoderDecoderModel,VitMatteForImageMatting:()=>t.VitMatteForImageMatting,VitMatteImageProcessor:()=>_.VitMatteImageProcessor,VitMattePreTrainedModel:()=>t.VitMattePreTrainedModel,VitPoseForPoseEstimation:()=>t.VitPoseForPoseEstimation,VitPoseImageProcessor:()=>_.VitPoseImageProcessor,VitPosePreTrainedModel:()=>t.VitPosePreTrainedModel,VitsModel:()=>t.VitsModel,VitsModelOutput:()=>t.VitsModelOutput,VitsPreTrainedModel:()=>t.VitsPreTrainedModel,VitsTokenizer:()=>s.VitsTokenizer,VoxtralForConditionalGeneration:()=>t.VoxtralForConditionalGeneration,VoxtralProcessor:()=>w.VoxtralProcessor,Wav2Vec2BertForCTC:()=>t.Wav2Vec2BertForCTC,Wav2Vec2BertForSequenceClassification:()=>t.Wav2Vec2BertForSequenceClassification,Wav2Vec2BertModel:()=>t.Wav2Vec2BertModel,Wav2Vec2BertPreTrainedModel:()=>t.Wav2Vec2BertPreTrainedModel,Wav2Vec2CTCTokenizer:()=>s.Wav2Vec2CTCTokenizer,Wav2Vec2FeatureExtractor:()=>d.Wav2Vec2FeatureExtractor,Wav2Vec2ForAudioFrameClassification:()=>t.Wav2Vec2ForAudioFrameClassification,Wav2Vec2ForCTC:()=>t.Wav2Vec2ForCTC,Wav2Vec2ForSequenceClassification:()=>t.Wav2Vec2ForSequenceClassification,Wav2Vec2Model:()=>t.Wav2Vec2Model,Wav2Vec2PreTrainedModel:()=>t.Wav2Vec2PreTrainedModel,Wav2Vec2Processor:()=>w.Wav2Vec2Processor,Wav2Vec2ProcessorWithLM:()=>w.Wav2Vec2ProcessorWithLM,WavLMForAudioFrameClassification:()=>t.WavLMForAudioFrameClassification,WavLMForCTC:()=>t.WavLMForCTC,WavLMForSequenceClassification:()=>t.WavLMForSequenceClassification,WavLMForXVector:()=>t.WavLMForXVector,WavLMModel:()=>t.WavLMModel,WavLMPreTrainedModel:()=>t.WavLMPreTrainedModel,WeSpeakerFeatureExtractor:()=>d.WeSpeakerFeatureExtractor,WeSpeakerResNetModel:()=>t.WeSpeakerResNetModel,WeSpeakerResNetPreTrainedModel:()=>t.WeSpeakerResNetPreTrainedModel,WhisperFeatureExtractor:()=>d.WhisperFeatureExtractor,WhisperForConditionalGeneration:()=>t.WhisperForConditionalGeneration,WhisperModel:()=>t.WhisperModel,WhisperPreTrainedModel:()=>t.WhisperPreTrainedModel,WhisperProcessor:()=>w.WhisperProcessor,WhisperTextStreamer:()=>k.WhisperTextStreamer,WhisperTimeStampLogitsProcessor:()=>b.WhisperTimeStampLogitsProcessor,WhisperTokenizer:()=>s.WhisperTokenizer,XLMForQuestionAnswering:()=>t.XLMForQuestionAnswering,XLMForSequenceClassification:()=>t.XLMForSequenceClassification,XLMForTokenClassification:()=>t.XLMForTokenClassification,XLMModel:()=>t.XLMModel,XLMPreTrainedModel:()=>t.XLMPreTrainedModel,XLMRobertaForMaskedLM:()=>t.XLMRobertaForMaskedLM,XLMRobertaForQuestionAnswering:()=>t.XLMRobertaForQuestionAnswering,XLMRobertaForSequenceClassification:()=>t.XLMRobertaForSequenceClassification,XLMRobertaForTokenClassification:()=>t.XLMRobertaForTokenClassification,XLMRobertaModel:()=>t.XLMRobertaModel,XLMRobertaPreTrainedModel:()=>t.XLMRobertaPreTrainedModel,XLMRobertaTokenizer:()=>s.XLMRobertaTokenizer,XLMTokenizer:()=>s.XLMTokenizer,XLMWithLMHeadModel:()=>t.XLMWithLMHeadModel,XVectorOutput:()=>t.XVectorOutput,YolosFeatureExtractor:()=>_.YolosFeatureExtractor,YolosForObjectDetection:()=>t.YolosForObjectDetection,YolosImageProcessor:()=>_.YolosImageProcessor,YolosModel:()=>t.YolosModel,YolosObjectDetectionOutput:()=>t.YolosObjectDetectionOutput,YolosPreTrainedModel:()=>t.YolosPreTrainedModel,ZeroShotAudioClassificationPipeline:()=>r.ZeroShotAudioClassificationPipeline,ZeroShotClassificationPipeline:()=>r.ZeroShotClassificationPipeline,ZeroShotImageClassificationPipeline:()=>r.ZeroShotImageClassificationPipeline,ZeroShotObjectDetectionPipeline:()=>r.ZeroShotObjectDetectionPipeline,bankers_round:()=>c.bankers_round,cat:()=>l.cat,cos_sim:()=>c.cos_sim,dot:()=>c.dot,dynamic_time_warping:()=>c.dynamic_time_warping,env:()=>e.env,full:()=>l.full,full_like:()=>l.full_like,getCacheShapes:()=>n.getCacheShapes,hamming:()=>o.hamming,hanning:()=>o.hanning,interpolate:()=>l.interpolate,interpolate_4d:()=>l.interpolate_4d,interpolate_data:()=>c.interpolate_data,is_chinese_char:()=>s.is_chinese_char,layer_norm:()=>l.layer_norm,load_image:()=>i.load_image,load_video:()=>a.load_video,log_softmax:()=>c.log_softmax,magnitude:()=>c.magnitude,matmul:()=>l.matmul,max:()=>c.max,mean:()=>l.mean,mean_pooling:()=>l.mean_pooling,medianFilter:()=>c.medianFilter,mel_filter_bank:()=>o.mel_filter_bank,min:()=>c.min,ones:()=>l.ones,ones_like:()=>l.ones_like,permute:()=>l.permute,permute_data:()=>c.permute_data,pipeline:()=>r.pipeline,quantize_embeddings:()=>l.quantize_embeddings,rand:()=>l.rand,randn:()=>l.randn,read_audio:()=>o.read_audio,rfft:()=>l.rfft,round:()=>c.round,slice:()=>l.slice,softmax:()=>c.softmax,spectrogram:()=>o.spectrogram,stack:()=>l.stack,std_mean:()=>l.std_mean,topk:()=>l.topk,window_function:()=>o.window_function,zeros:()=>l.zeros,zeros_like:()=>l.zeros_like});var e=jt("./src/env.js"),r=jt("./src/pipelines.js"),t=jt("./src/models.js"),s=jt("./src/tokenizers.js"),n=jt("./src/configs.js"),o=jt("./src/utils/audio.js"),i=jt("./src/utils/image.js"),a=jt("./src/utils/video.js"),l=jt("./src/utils/tensor.js"),c=jt("./src/utils/maths.js"),p=jt("./src/base/feature_extraction_utils.js"),d=jt("./src/models/feature_extractors.js"),u=jt("./src/models/auto/feature_extraction_auto.js"),f=jt("./src/base/image_processors_utils.js"),_=jt("./src/models/image_processors.js"),y=jt("./src/models/auto/image_processing_auto.js"),I=jt("./src/base/processing_utils.js"),w=jt("./src/models/processors.js"),v=jt("./src/models/auto/processing_auto.js"),k=jt("./src/generation/streamers.js"),T=jt("./src/generation/stopping_criteria.js"),b=jt("./src/generation/logits_process.js")})(),m.ASTFeatureExtractor,m.ASTForAudioClassification,m.ASTModel,m.ASTPreTrainedModel,m.AlbertForMaskedLM,m.AlbertForQuestionAnswering,m.AlbertForSequenceClassification,m.AlbertModel,m.AlbertPreTrainedModel,m.AlbertTokenizer,m.ArceeForCausalLM,m.ArceeModel,m.ArceePreTrainedModel,m.AudioClassificationPipeline,m.AutoConfig,m.AutoFeatureExtractor,m.AutoImageProcessor,m.AutoModel,m.AutoModelForAudioClassification,m.AutoModelForAudioFrameClassification,m.AutoModelForAudioTextToText,m.AutoModelForCTC;var hT=m.AutoModelForCausalLM;m.AutoModelForDepthEstimation,m.AutoModelForDocumentQuestionAnswering,m.AutoModelForImageClassification,m.AutoModelForImageFeatureExtraction,m.AutoModelForImageMatting,m.AutoModelForImageSegmentation,m.AutoModelForImageTextToText,m.AutoModelForImageToImage,m.AutoModelForMaskGeneration,m.AutoModelForMaskedLM,m.AutoModelForNormalEstimation,m.AutoModelForObjectDetection,m.AutoModelForPoseEstimation,m.AutoModelForQuestionAnswering,m.AutoModelForSemanticSegmentation,m.AutoModelForSeq2SeqLM,m.AutoModelForSequenceClassification,m.AutoModelForSpeechSeq2Seq,m.AutoModelForTextToSpectrogram,m.AutoModelForTextToWaveform,m.AutoModelForTokenClassification,m.AutoModelForUniversalSegmentation,m.AutoModelForVision2Seq,m.AutoModelForXVector,m.AutoModelForZeroShotObjectDetection,m.AutoProcessor;var _T=m.AutoTokenizer;m.AutomaticSpeechRecognitionPipeline,m.BackgroundRemovalPipeline,m.BartForConditionalGeneration,m.BartForSequenceClassification,m.BartModel,m.BartPretrainedModel,m.BartTokenizer,m.BaseModelOutput,m.BaseStreamer,m.BeitFeatureExtractor,m.BeitForImageClassification,m.BeitModel,m.BeitPreTrainedModel,m.BertForMaskedLM,m.BertForQuestionAnswering,m.BertForSequenceClassification,m.BertForTokenClassification,m.BertModel,m.BertPreTrainedModel,m.BertTokenizer,m.BitImageProcessor,m.BlenderbotForConditionalGeneration,m.BlenderbotModel,m.BlenderbotPreTrainedModel,m.BlenderbotSmallForConditionalGeneration,m.BlenderbotSmallModel,m.BlenderbotSmallPreTrainedModel,m.BlenderbotSmallTokenizer,m.BlenderbotTokenizer,m.BloomForCausalLM,m.BloomModel,m.BloomPreTrainedModel,m.BloomTokenizer,m.CLIPFeatureExtractor,m.CLIPImageProcessor,m.CLIPModel,m.CLIPPreTrainedModel,m.CLIPSegForImageSegmentation,m.CLIPSegModel,m.CLIPSegPreTrainedModel,m.CLIPTextModel,m.CLIPTextModelWithProjection,m.CLIPTokenizer,m.CLIPVisionModel,m.CLIPVisionModelWithProjection,m.CamembertForMaskedLM,m.CamembertForQuestionAnswering,m.CamembertForSequenceClassification,m.CamembertForTokenClassification,m.CamembertModel,m.CamembertPreTrainedModel,m.CamembertTokenizer,m.CausalLMOutput,m.CausalLMOutputWithPast,m.ChineseCLIPFeatureExtractor,m.ChineseCLIPModel,m.ChineseCLIPPreTrainedModel,m.ClapAudioModelWithProjection,m.ClapFeatureExtractor,m.ClapModel,m.ClapPreTrainedModel,m.ClapTextModelWithProjection,m.ClassifierFreeGuidanceLogitsProcessor,m.CodeGenForCausalLM,m.CodeGenModel,m.CodeGenPreTrainedModel,m.CodeGenTokenizer,m.CodeLlamaTokenizer,m.CohereForCausalLM,m.CohereModel,m.CoherePreTrainedModel,m.CohereTokenizer,m.ConvBertForMaskedLM,m.ConvBertForQuestionAnswering,m.ConvBertForSequenceClassification,m.ConvBertForTokenClassification,m.ConvBertModel,m.ConvBertPreTrainedModel,m.ConvBertTokenizer,m.ConvNextFeatureExtractor,m.ConvNextForImageClassification,m.ConvNextImageProcessor,m.ConvNextModel,m.ConvNextPreTrainedModel,m.ConvNextV2ForImageClassification,m.ConvNextV2Model,m.ConvNextV2PreTrainedModel,m.DFineForObjectDetection,m.DFineModel,m.DFinePreTrainedModel,m.DINOv3ConvNextModel,m.DINOv3ConvNextPreTrainedModel,m.DINOv3ViTImageProcessor,m.DINOv3ViTModel,m.DINOv3ViTPreTrainedModel,m.DPTFeatureExtractor,m.DPTForDepthEstimation,m.DPTImageProcessor,m.DPTModel,m.DPTPreTrainedModel,m.DacDecoderModel,m.DacDecoderOutput,m.DacEncoderModel,m.DacEncoderOutput,m.DacFeatureExtractor,m.DacModel,m.DacPreTrainedModel,m.DataTypeMap,m.DebertaForMaskedLM,m.DebertaForQuestionAnswering,m.DebertaForSequenceClassification,m.DebertaForTokenClassification,m.DebertaModel,m.DebertaPreTrainedModel,m.DebertaTokenizer,m.DebertaV2ForMaskedLM,m.DebertaV2ForQuestionAnswering,m.DebertaV2ForSequenceClassification,m.DebertaV2ForTokenClassification,m.DebertaV2Model,m.DebertaV2PreTrainedModel,m.DebertaV2Tokenizer,m.DecisionTransformerModel,m.DecisionTransformerPreTrainedModel,m.DeiTFeatureExtractor,m.DeiTForImageClassification,m.DeiTImageProcessor,m.DeiTModel,m.DeiTPreTrainedModel,m.DepthAnythingForDepthEstimation,m.DepthAnythingPreTrainedModel,m.DepthEstimationPipeline,m.DepthProForDepthEstimation,m.DepthProPreTrainedModel,m.DetrFeatureExtractor,m.DetrForObjectDetection,m.DetrForSegmentation,m.DetrImageProcessor,m.DetrModel,m.DetrObjectDetectionOutput,m.DetrPreTrainedModel,m.DetrSegmentationOutput,m.Dinov2ForImageClassification,m.Dinov2Model,m.Dinov2PreTrainedModel,m.Dinov2WithRegistersForImageClassification,m.Dinov2WithRegistersModel,m.Dinov2WithRegistersPreTrainedModel,m.DistilBertForMaskedLM,m.DistilBertForQuestionAnswering,m.DistilBertForSequenceClassification,m.DistilBertForTokenClassification,m.DistilBertModel,m.DistilBertPreTrainedModel,m.DistilBertTokenizer,m.DocumentQuestionAnsweringPipeline,m.DonutFeatureExtractor,m.DonutImageProcessor,m.DonutSwinModel,m.DonutSwinPreTrainedModel,m.EdgeTamModel,m.EfficientNetForImageClassification,m.EfficientNetImageProcessor,m.EfficientNetModel,m.EfficientNetPreTrainedModel,m.ElectraForMaskedLM,m.ElectraForQuestionAnswering,m.ElectraForSequenceClassification,m.ElectraForTokenClassification,m.ElectraModel,m.ElectraPreTrainedModel,m.ElectraTokenizer,m.EncodecFeatureExtractor,m.EosTokenCriteria,m.Ernie4_5ForCausalLM,m.Ernie4_5Model,m.Ernie4_5PreTrainedModel,m.EsmForMaskedLM,m.EsmForSequenceClassification,m.EsmForTokenClassification,m.EsmModel,m.EsmPreTrainedModel,m.EsmTokenizer,m.ExaoneForCausalLM,m.ExaoneModel,m.ExaonePreTrainedModel,m.FFT,m.FalconForCausalLM,m.FalconModel,m.FalconPreTrainedModel,m.FalconTokenizer,m.FastViTForImageClassification,m.FastViTModel,m.FastViTPreTrainedModel,m.FeatureExtractionPipeline,m.FeatureExtractor,m.FillMaskPipeline,m.Florence2ForConditionalGeneration,m.Florence2PreTrainedModel,m.Florence2Processor,m.ForcedBOSTokenLogitsProcessor,m.ForcedEOSTokenLogitsProcessor,m.GLPNFeatureExtractor,m.GLPNForDepthEstimation,m.GLPNModel,m.GLPNPreTrainedModel,m.GPT2LMHeadModel,m.GPT2Model,m.GPT2PreTrainedModel,m.GPT2Tokenizer,m.GPTBigCodeForCausalLM,m.GPTBigCodeModel,m.GPTBigCodePreTrainedModel,m.GPTJForCausalLM,m.GPTJModel,m.GPTJPreTrainedModel,m.GPTNeoForCausalLM,m.GPTNeoModel,m.GPTNeoPreTrainedModel,m.GPTNeoXForCausalLM,m.GPTNeoXModel,m.GPTNeoXPreTrainedModel,m.GPTNeoXTokenizer,m.Gemma2ForCausalLM,m.Gemma2Model,m.Gemma2PreTrainedModel,m.Gemma3ForCausalLM,m.Gemma3Model,m.Gemma3PreTrainedModel,m.Gemma3nAudioFeatureExtractor,m.Gemma3nForConditionalGeneration,m.Gemma3nPreTrainedModel,m.Gemma3nProcessor,m.GemmaForCausalLM,m.GemmaModel,m.GemmaPreTrainedModel,m.GemmaTokenizer,m.GlmForCausalLM,m.GlmModel,m.GlmPreTrainedModel,m.GraniteForCausalLM,m.GraniteModel,m.GraniteMoeHybridForCausalLM,m.GraniteMoeHybridModel,m.GraniteMoeHybridPreTrainedModel,m.GranitePreTrainedModel,m.Grok1Tokenizer,m.GroundingDinoForObjectDetection,m.GroundingDinoImageProcessor,m.GroundingDinoPreTrainedModel,m.GroundingDinoProcessor,m.GroupViTModel,m.GroupViTPreTrainedModel,m.HeliumForCausalLM,m.HeliumModel,m.HeliumPreTrainedModel,m.HerbertTokenizer,m.HieraForImageClassification,m.HieraModel,m.HieraPreTrainedModel,m.HubertForCTC,m.HubertForSequenceClassification,m.HubertModel,m.HubertPreTrainedModel,m.IJepaForImageClassification,m.IJepaModel,m.IJepaPreTrainedModel,m.Idefics3ForConditionalGeneration,m.Idefics3ImageProcessor,m.Idefics3PreTrainedModel,m.Idefics3Processor,m.ImageClassificationPipeline,m.ImageFeatureExtractionPipeline,m.ImageFeatureExtractor,m.ImageMattingOutput,m.ImageProcessor,m.ImageSegmentationPipeline,m.ImageToImagePipeline,m.ImageToTextPipeline,m.InterruptableStoppingCriteria,m.JAISLMHeadModel,m.JAISModel,m.JAISPreTrainedModel,m.JinaCLIPImageProcessor,m.JinaCLIPModel,m.JinaCLIPPreTrainedModel,m.JinaCLIPProcessor,m.JinaCLIPTextModel,m.JinaCLIPVisionModel,m.Lfm2ForCausalLM,m.Lfm2Model,m.Lfm2PreTrainedModel,m.LiteWhisperForConditionalGeneration,m.Llama4ForCausalLM,m.Llama4PreTrainedModel,m.LlamaForCausalLM,m.LlamaModel,m.LlamaPreTrainedModel,m.LlamaTokenizer,m.LlavaForConditionalGeneration,m.LlavaOnevisionForConditionalGeneration,m.LlavaOnevisionImageProcessor,m.LlavaPreTrainedModel,m.LlavaProcessor,m.LlavaQwen2ForCausalLM,m.LogitsProcessor,m.LogitsProcessorList,m.LogitsWarper,m.LongT5ForConditionalGeneration,m.LongT5Model,m.LongT5PreTrainedModel,m.M2M100ForConditionalGeneration,m.M2M100Model,m.M2M100PreTrainedModel,m.M2M100Tokenizer,m.MBart50Tokenizer,m.MBartForCausalLM,m.MBartForConditionalGeneration,m.MBartForSequenceClassification,m.MBartModel,m.MBartPreTrainedModel,m.MBartTokenizer,m.MPNetForMaskedLM,m.MPNetForQuestionAnswering,m.MPNetForSequenceClassification,m.MPNetForTokenClassification,m.MPNetModel,m.MPNetPreTrainedModel,m.MPNetTokenizer,m.MT5ForConditionalGeneration,m.MT5Model,m.MT5PreTrainedModel,m.MarianMTModel,m.MarianModel,m.MarianPreTrainedModel,m.MarianTokenizer,m.Mask2FormerImageProcessor,m.MaskFormerFeatureExtractor,m.MaskFormerForInstanceSegmentation,m.MaskFormerImageProcessor,m.MaskFormerModel,m.MaskFormerPreTrainedModel,m.MaskedLMOutput,m.MaxLengthCriteria,m.Metric3DForDepthEstimation,m.Metric3DPreTrainedModel,m.Metric3Dv2ForDepthEstimation,m.Metric3Dv2PreTrainedModel,m.MgpstrForSceneTextRecognition,m.MgpstrModelOutput,m.MgpstrPreTrainedModel,m.MgpstrProcessor,m.MgpstrTokenizer,m.MimiDecoderModel,m.MimiDecoderOutput,m.MimiEncoderModel,m.MimiEncoderOutput,m.MimiModel,m.MimiPreTrainedModel,m.MinLengthLogitsProcessor,m.MinNewTokensLengthLogitsProcessor,m.Ministral3ForCausalLM,m.Ministral3Model,m.Ministral3PreTrainedModel,m.MinistralForCausalLM,m.MinistralModel,m.MinistralPreTrainedModel,m.Mistral3ForConditionalGeneration,m.MistralForCausalLM,m.MistralModel,m.MistralPreTrainedModel,m.MobileBertForMaskedLM,m.MobileBertForQuestionAnswering,m.MobileBertForSequenceClassification,m.MobileBertModel,m.MobileBertPreTrainedModel,m.MobileBertTokenizer,m.MobileLLMForCausalLM,m.MobileLLMModel,m.MobileLLMPreTrainedModel,m.MobileNetV1FeatureExtractor,m.MobileNetV1ForImageClassification,m.MobileNetV1ForSemanticSegmentation,m.MobileNetV1ImageProcessor,m.MobileNetV1Model,m.MobileNetV1PreTrainedModel,m.MobileNetV2FeatureExtractor,m.MobileNetV2ForImageClassification,m.MobileNetV2ForSemanticSegmentation,m.MobileNetV2ImageProcessor,m.MobileNetV2Model,m.MobileNetV2PreTrainedModel,m.MobileNetV3FeatureExtractor,m.MobileNetV3ForImageClassification,m.MobileNetV3ForSemanticSegmentation,m.MobileNetV3ImageProcessor,m.MobileNetV3Model,m.MobileNetV3PreTrainedModel,m.MobileNetV4FeatureExtractor,m.MobileNetV4ForImageClassification,m.MobileNetV4ForSemanticSegmentation,m.MobileNetV4ImageProcessor,m.MobileNetV4Model,m.MobileNetV4PreTrainedModel,m.MobileViTFeatureExtractor,m.MobileViTForImageClassification,m.MobileViTImageProcessor,m.MobileViTModel,m.MobileViTPreTrainedModel,m.MobileViTV2ForImageClassification,m.MobileViTV2Model,m.MobileViTV2PreTrainedModel,m.ModelOutput,m.ModernBertDecoderForCausalLM,m.ModernBertDecoderModel,m.ModernBertDecoderPreTrainedModel,m.ModernBertForMaskedLM,m.ModernBertForSequenceClassification,m.ModernBertForTokenClassification,m.ModernBertModel,m.ModernBertPreTrainedModel,m.Moondream1ForConditionalGeneration,m.MoonshineFeatureExtractor,m.MoonshineForConditionalGeneration,m.MoonshineModel,m.MoonshinePreTrainedModel,m.MoonshineProcessor,m.MptForCausalLM,m.MptModel,m.MptPreTrainedModel,m.MultiModalityCausalLM,m.MultiModalityPreTrainedModel,m.MusicgenForCausalLM,m.MusicgenForConditionalGeneration,m.MusicgenModel,m.MusicgenPreTrainedModel,m.NanoChatForCausalLM,m.NanoChatModel,m.NanoChatPreTrainedModel,m.NeoBertForMaskedLM,m.NeoBertForQuestionAnswering,m.NeoBertForSequenceClassification,m.NeoBertForTokenClassification,m.NeoBertModel,m.NeoBertPreTrainedModel,m.NllbTokenizer,m.NoBadWordsLogitsProcessor,m.NoRepeatNGramLogitsProcessor,m.NomicBertModel,m.NomicBertPreTrainedModel,m.NougatImageProcessor,m.NougatTokenizer,m.OPTForCausalLM,m.OPTModel,m.OPTPreTrainedModel,m.ObjectDetectionPipeline,m.Olmo2ForCausalLM,m.Olmo2Model,m.Olmo2PreTrainedModel,m.OlmoForCausalLM,m.OlmoModel,m.OlmoPreTrainedModel,m.OpenELMForCausalLM,m.OpenELMModel,m.OpenELMPreTrainedModel,m.OwlViTFeatureExtractor,m.OwlViTForObjectDetection,m.OwlViTImageProcessor,m.OwlViTModel,m.OwlViTPreTrainedModel,m.OwlViTProcessor,m.Owlv2ForObjectDetection,m.Owlv2ImageProcessor,m.Owlv2Model,m.Owlv2PreTrainedModel,m.PaliGemmaForConditionalGeneration,m.PaliGemmaPreTrainedModel,m.PaliGemmaProcessor,m.ParakeetFeatureExtractor,m.ParakeetForCTC,m.ParakeetPreTrainedModel,m.PatchTSMixerForPrediction,m.PatchTSMixerModel,m.PatchTSMixerPreTrainedModel,m.PatchTSTForPrediction,m.PatchTSTModel,m.PatchTSTPreTrainedModel,m.Phi3ForCausalLM,m.Phi3Model,m.Phi3PreTrainedModel,m.Phi3VForCausalLM,m.Phi3VImageProcessor,m.Phi3VPreTrainedModel,m.Phi3VProcessor,m.PhiForCausalLM,m.PhiModel,m.PhiPreTrainedModel,m.Pipeline,m.PixtralImageProcessor,m.PixtralProcessor,m.PreTrainedModel,m.PreTrainedTokenizer,m.PretrainedConfig,m.PretrainedMixin,m.Processor,m.PvtForImageClassification,m.PvtImageProcessor,m.PvtModel,m.PvtPreTrainedModel,m.PyAnnoteFeatureExtractor,m.PyAnnoteForAudioFrameClassification,m.PyAnnoteModel,m.PyAnnotePreTrainedModel,m.PyAnnoteProcessor,m.QuestionAnsweringModelOutput,m.QuestionAnsweringPipeline,m.Qwen2ForCausalLM,m.Qwen2Model,m.Qwen2PreTrainedModel,m.Qwen2Tokenizer,m.Qwen2VLForConditionalGeneration,m.Qwen2VLImageProcessor,m.Qwen2VLPreTrainedModel,m.Qwen2VLProcessor,m.Qwen3ForCausalLM,m.Qwen3Model,m.Qwen3PreTrainedModel,m.RFDetrForObjectDetection,m.RFDetrModel,m.RFDetrObjectDetectionOutput,m.RFDetrPreTrainedModel,m.RTDetrForObjectDetection,m.RTDetrImageProcessor,m.RTDetrModel,m.RTDetrObjectDetectionOutput,m.RTDetrPreTrainedModel,m.RTDetrV2ForObjectDetection,m.RTDetrV2Model,m.RTDetrV2ObjectDetectionOutput,m.RTDetrV2PreTrainedModel,m.RawAudio,m.RawImage,m.RawVideo,m.RawVideoFrame,m.RepetitionPenaltyLogitsProcessor,m.ResNetForImageClassification,m.ResNetModel,m.ResNetPreTrainedModel,m.RoFormerForMaskedLM,m.RoFormerForQuestionAnswering,m.RoFormerForSequenceClassification,m.RoFormerForTokenClassification,m.RoFormerModel,m.RoFormerPreTrainedModel,m.RoFormerTokenizer,m.RobertaForMaskedLM,m.RobertaForQuestionAnswering,m.RobertaForSequenceClassification,m.RobertaForTokenClassification,m.RobertaModel,m.RobertaPreTrainedModel,m.RobertaTokenizer,m.Sam2ImageProcessor,m.Sam2ImageSegmentationOutput,m.Sam2Model,m.Sam2PreTrainedModel,m.Sam2Processor,m.Sam2VideoProcessor,m.Sam3ImageProcessor,m.Sam3TrackerModel,m.SamImageProcessor,m.SamImageSegmentationOutput,m.SamModel,m.SamPreTrainedModel,m.SamProcessor,m.SapiensForDepthEstimation,m.SapiensForNormalEstimation,m.SapiensForSemanticSegmentation,m.SapiensPreTrainedModel,m.SeamlessM4TFeatureExtractor,m.SegformerFeatureExtractor,m.SegformerForImageClassification,m.SegformerForSemanticSegmentation,m.SegformerImageProcessor,m.SegformerModel,m.SegformerPreTrainedModel,m.Seq2SeqLMOutput,m.SequenceClassifierOutput,m.SiglipImageProcessor,m.SiglipModel,m.SiglipPreTrainedModel,m.SiglipTextModel,m.SiglipTokenizer,m.SiglipVisionModel,m.SmolLM3ForCausalLM,m.SmolLM3Model,m.SmolLM3PreTrainedModel,m.SmolVLMForConditionalGeneration,m.SmolVLMImageProcessor,m.SmolVLMProcessor,m.SnacDecoderModel,m.SnacEncoderModel,m.SnacFeatureExtractor,m.SnacModel,m.SnacPreTrainedModel,m.SpeechT5FeatureExtractor,m.SpeechT5ForSpeechToText,m.SpeechT5ForTextToSpeech,m.SpeechT5HifiGan,m.SpeechT5Model,m.SpeechT5PreTrainedModel,m.SpeechT5Processor,m.SpeechT5Tokenizer,m.SqueezeBertForMaskedLM,m.SqueezeBertForQuestionAnswering,m.SqueezeBertForSequenceClassification,m.SqueezeBertModel,m.SqueezeBertPreTrainedModel,m.SqueezeBertTokenizer,m.StableLmForCausalLM,m.StableLmModel,m.StableLmPreTrainedModel,m.Starcoder2ForCausalLM,m.Starcoder2Model,m.Starcoder2PreTrainedModel,m.StoppingCriteria,m.StoppingCriteriaList,m.StyleTextToSpeech2Model,m.StyleTextToSpeech2PreTrainedModel,m.SummarizationPipeline,m.SupertonicForConditionalGeneration,m.SupertonicPreTrainedModel,m.SuppressTokensAtBeginLogitsProcessor,m.Swin2SRForImageSuperResolution,m.Swin2SRImageProcessor,m.Swin2SRModel,m.Swin2SRPreTrainedModel,m.SwinForImageClassification,m.SwinForSemanticSegmentation,m.SwinModel,m.SwinPreTrainedModel,m.T5ForConditionalGeneration,m.T5Model,m.T5PreTrainedModel,m.T5Tokenizer,m.TableTransformerForObjectDetection,m.TableTransformerModel,m.TableTransformerObjectDetectionOutput,m.TableTransformerPreTrainedModel,m.TemperatureLogitsWarper,m.Tensor,m.Text2TextGenerationPipeline,m.TextClassificationPipeline,m.TextGenerationPipeline,m.TextStreamer,m.TextToAudioPipeline,m.TokenClassificationPipeline,m.TokenClassifierOutput,m.TokenizerModel,m.TopKLogitsWarper,m.TopPLogitsWarper,m.TrOCRForCausalLM,m.TrOCRPreTrainedModel,m.TranslationPipeline,m.UltravoxModel,m.UltravoxPreTrainedModel,m.UltravoxProcessor,m.UniSpeechForCTC,m.UniSpeechForSequenceClassification,m.UniSpeechModel,m.UniSpeechPreTrainedModel,m.UniSpeechSatForAudioFrameClassification,m.UniSpeechSatForCTC,m.UniSpeechSatForSequenceClassification,m.UniSpeechSatModel,m.UniSpeechSatPreTrainedModel,m.VLChatProcessor,m.VLMImageProcessor,m.VaultGemmaForCausalLM,m.VaultGemmaModel,m.VaultGemmaPreTrainedModel,m.ViTFeatureExtractor,m.ViTForImageClassification,m.ViTImageProcessor,m.ViTMAEModel,m.ViTMAEPreTrainedModel,m.ViTMSNForImageClassification,m.ViTMSNModel,m.ViTMSNPreTrainedModel,m.ViTModel,m.ViTPreTrainedModel,m.VisionEncoderDecoderModel,m.VitMatteForImageMatting,m.VitMatteImageProcessor,m.VitMattePreTrainedModel,m.VitPoseForPoseEstimation,m.VitPoseImageProcessor,m.VitPosePreTrainedModel,m.VitsModel,m.VitsModelOutput,m.VitsPreTrainedModel,m.VitsTokenizer,m.VoxtralForConditionalGeneration,m.VoxtralProcessor,m.Wav2Vec2BertForCTC,m.Wav2Vec2BertForSequenceClassification,m.Wav2Vec2BertModel,m.Wav2Vec2BertPreTrainedModel,m.Wav2Vec2CTCTokenizer,m.Wav2Vec2FeatureExtractor,m.Wav2Vec2ForAudioFrameClassification,m.Wav2Vec2ForCTC,m.Wav2Vec2ForSequenceClassification,m.Wav2Vec2Model,m.Wav2Vec2PreTrainedModel,m.Wav2Vec2Processor,m.Wav2Vec2ProcessorWithLM,m.WavLMForAudioFrameClassification,m.WavLMForCTC,m.WavLMForSequenceClassification,m.WavLMForXVector,m.WavLMModel,m.WavLMPreTrainedModel,m.WeSpeakerFeatureExtractor,m.WeSpeakerResNetModel,m.WeSpeakerResNetPreTrainedModel,m.WhisperFeatureExtractor,m.WhisperForConditionalGeneration,m.WhisperModel,m.WhisperPreTrainedModel,m.WhisperProcessor,m.WhisperTextStreamer,m.WhisperTimeStampLogitsProcessor,m.WhisperTokenizer,m.XLMForQuestionAnswering,m.XLMForSequenceClassification,m.XLMForTokenClassification,m.XLMModel,m.XLMPreTrainedModel,m.XLMRobertaForMaskedLM,m.XLMRobertaForQuestionAnswering,m.XLMRobertaForSequenceClassification,m.XLMRobertaForTokenClassification,m.XLMRobertaModel,m.XLMRobertaPreTrainedModel,m.XLMRobertaTokenizer,m.XLMTokenizer,m.XLMWithLMHeadModel,m.XVectorOutput,m.YolosFeatureExtractor,m.YolosForObjectDetection,m.YolosImageProcessor,m.YolosModel,m.YolosObjectDetectionOutput,m.YolosPreTrainedModel,m.ZeroShotAudioClassificationPipeline,m.ZeroShotClassificationPipeline,m.ZeroShotImageClassificationPipeline,m.ZeroShotObjectDetectionPipeline,m.bankers_round,m.cat,m.cos_sim,m.dot,m.dynamic_time_warping,m.env,m.full,m.full_like,m.getCacheShapes,m.hamming,m.hanning,m.interpolate,m.interpolate_4d,m.interpolate_data,m.is_chinese_char,m.layer_norm,m.load_image,m.load_video,m.log_softmax,m.magnitude,m.matmul,m.max,m.mean,m.mean_pooling,m.medianFilter,m.mel_filter_bank,m.min,m.ones,m.ones_like,m.permute,m.permute_data,m.pipeline,m.quantize_embeddings,m.rand,m.randn,m.read_audio,m.rfft,m.round,m.slice,m.softmax,m.spectrogram,m.stack,m.std_mean,m.topk,m.window_function,m.zeros,m.zeros_like;const cM=Symbol("Comlink.proxy"),fT=Symbol("Comlink.endpoint"),gT=Symbol("Comlink.releaseProxy"),mc=Symbol("Comlink.finalizer"),ya=Symbol("Comlink.thrown"),uM=e=>typeof e=="object"&&e!==null||typeof e=="function",MT={canHandle:e=>uM(e)&&e[cM],serialize(e){const{port1:r,port2:t}=new MessageChannel;return hc(e,r),[t,[t]]},deserialize(e){return e.start(),vT(e)}},wT={canHandle:e=>uM(e)&&ya in e,serialize({value:e}){let r;return e instanceof Error?r={isError:!0,value:{message:e.message,name:e.name,stack:e.stack}}:r={isError:!1,value:e},[r,[]]},deserialize(e){throw e.isError?Object.assign(new Error(e.value.message),e.value):e.value}},dM=new Map([["proxy",MT],["throw",wT]]);function bT(e,r){for(const t of e)if(r===t||t==="*"||t instanceof RegExp&&t.test(r))return!0;return!1}function hc(e,r=globalThis,t=["*"]){r.addEventListener("message",function s(n){if(!n||!n.data)return;if(!bT(t,n.origin)){console.warn(`Invalid origin '${n.origin}' for comlink proxy`);return}const{id:o,type:i,path:a}=Object.assign({path:[]},n.data),l=(n.data.argumentList||[]).map(wn);let c;try{const p=a.slice(0,-1).reduce((u,f)=>u[f],e),d=a.reduce((u,f)=>u[f],e);switch(i){case"GET":c=d;break;case"SET":p[a.slice(-1)[0]]=wn(n.data.value),c=!0;break;case"APPLY":c=d.apply(p,l);break;case"CONSTRUCT":{const u=new d(...l);c=CT(u)}break;case"ENDPOINT":{const{port1:u,port2:f}=new MessageChannel;hc(e,f),c=ET(u,[u])}break;case"RELEASE":c=void 0;break;default:return}}catch(p){c={value:p,[ya]:0}}Promise.resolve(c).catch(p=>({value:p,[ya]:0})).then(p=>{const[d,u]=Pa(p);r.postMessage(Object.assign(Object.assign({},d),{id:o}),u),i==="RELEASE"&&(r.removeEventListener("message",s),pM(r),mc in e&&typeof e[mc]=="function"&&e[mc]())}).catch(p=>{const[d,u]=Pa({value:new TypeError("Unserializable return value"),[ya]:0});r.postMessage(Object.assign(Object.assign({},d),{id:o}),u)})}),r.start&&r.start()}function yT(e){return e.constructor.name==="MessagePort"}function pM(e){yT(e)&&e.close()}function vT(e,r){const t=new Map;return e.addEventListener("message",function(n){const{data:o}=n;if(!o||!o.id)return;const i=t.get(o.id);if(i)try{i(o)}finally{t.delete(o.id)}}),_c(e,t,[],r)}function va(e){if(e)throw new Error("Proxy has been released and is not useable")}function mM(e){return jn(e,new Map,{type:"RELEASE"}).then(()=>{pM(e)})}const xa=new WeakMap,Ta="FinalizationRegistry"in globalThis&&new FinalizationRegistry(e=>{const r=(xa.get(e)||0)-1;xa.set(e,r),r===0&&mM(e)});function xT(e,r){const t=(xa.get(r)||0)+1;xa.set(r,t),Ta&&Ta.register(e,r,e)}function TT(e){Ta&&Ta.unregister(e)}function _c(e,r,t=[],s=function(){}){let n=!1;const o=new Proxy(s,{get(i,a){if(va(n),a===gT)return()=>{TT(o),mM(e),r.clear(),n=!0};if(a==="then"){if(t.length===0)return{then:()=>o};const l=jn(e,r,{type:"GET",path:t.map(c=>c.toString())}).then(wn);return l.then.bind(l)}return _c(e,r,[...t,a])},set(i,a,l){va(n);const[c,p]=Pa(l);return jn(e,r,{type:"SET",path:[...t,a].map(d=>d.toString()),value:c},p).then(wn)},apply(i,a,l){va(n);const c=t[t.length-1];if(c===fT)return jn(e,r,{type:"ENDPOINT"}).then(wn);if(c==="bind")return _c(e,r,t.slice(0,-1));const[p,d]=hM(l);return jn(e,r,{type:"APPLY",path:t.map(u=>u.toString()),argumentList:p},d).then(wn)},construct(i,a){va(n);const[l,c]=hM(a);return jn(e,r,{type:"CONSTRUCT",path:t.map(p=>p.toString()),argumentList:l},c).then(wn)}});return xT(o,e),o}function PT(e){return Array.prototype.concat.apply([],e)}function hM(e){const r=e.map(Pa);return[r.map(t=>t[0]),PT(r.map(t=>t[1]))]}const _M=new WeakMap;function ET(e,r){return _M.set(e,r),e}function CT(e){return Object.assign(e,{[cM]:!0})}function Pa(e){for(const[r,t]of dM)if(t.canHandle(e)){const[s,n]=t.serialize(e);return[{type:"HANDLER",name:r,value:s},n]}return[{type:"RAW",value:e},_M.get(e)||[]]}function wn(e){switch(e.type){case"HANDLER":return dM.get(e.name).deserialize(e.value);case"RAW":return e.value}}function jn(e,r,t,s){return new Promise(n=>{const o=ST();r.set(o,n),e.start&&e.start(),e.postMessage(Object.assign({id:o},t),s)})}function ST(){return new Array(4).fill(0).map(()=>Math.floor(Math.random()*Number.MAX_SAFE_INTEGER).toString(16)).join("-")}const fM="onnx-community/functiongemma-270m-it-ONNX";let gM=null;function MM(e){return gM??=(async()=>{console.log("[Worker] Loading model...");const[r,t]=await Promise.all([_T.from_pretrained(fM,{progress_callback:e}),hT.from_pretrained(fM,{dtype:"q4",device:"webgpu",progress_callback:e})]);return console.log("[Worker] Model loaded!"),{model:t,tokenizer:r}})(),gM}hc({async initModel(e){console.log("[Worker] initModel called"),await MM(e),console.log("[Worker] initModel completed")},async generate(e,r){console.log("[Worker] ========== GENERATE =========="),console.log("[Worker] Messages:",JSON.stringify(e,null,2));const{model:t,tokenizer:s}=await MM(),n=s.apply_chat_template(e,{tools:r,tokenize:!0,add_generation_prompt:!0,return_dict:!0});console.log("[Worker] Input token count:",n.input_ids.dims[1]);const o=await t.generate({...n,max_new_tokens:512}),i=n.input_ids.dims[1],a=s.decode(o.slice(0,[i,null]),{skip_special_tokens:!1});return console.log("[Worker] Raw output:",a),a}})})();