{"0": "cpmant/modeling_cpmant.py:CpmAntLayerNorm", "1": "cpmant/modeling_cpmant.py:CpmAntAttention", "2": "cpmant/modeling_cpmant.py:CpmAntSelfAttentionBlock", "3": "cpmant/modeling_cpmant.py:CpmAntDenseGatedACT", "4": "cpmant/modeling_cpmant.py:CpmAntFeedForward", "5": "cpmant/modeling_cpmant.py:CpmAntFFNBlock", "6": "cpmant/modeling_cpmant.py:CpmAntTransformerBlock", "7": "cpmant/modeling_cpmant.py:CpmAntEncoder", "8": "cpmant/modeling_cpmant.py:CpmAntIntermediate", "9": "cpmant/modeling_cpmant.py:CpmAntSegmentPositionEmbedding", "10": "cpmant/modeling_cpmant.py:CpmAntOutput", "11": "cpmant/modeling_cpmant.py:CpmAntPreTrainedModel", "12": "cpmant/modeling_cpmant.py:CpmAntModel", "13": "cpmant/modeling_cpmant.py:CpmAntForCausalLM", "14": "mistral3/modeling_mistral3.py:Mistral3RMSNorm", "15": "mistral3/modeling_mistral3.py:Mistral3PatchMerger", "16": "mistral3/modeling_mistral3.py:Mistral3MultiModalProjector", "17": "mistral3/modeling_mistral3.py:Mistral3CausalLMOutputWithPast", "18": "mistral3/modeling_mistral3.py:Mistral3ModelOutputWithPast", "19": "mistral3/modeling_mistral3.py:Mistral3PreTrainedModel", "20": "mistral3/modeling_mistral3.py:Mistral3Model", "21": "mistral3/modeling_mistral3.py:Mistral3ForConditionalGeneration", "22": "persimmon/modeling_persimmon.py:PersimmonRotaryEmbedding", "23": "persimmon/modeling_persimmon.py:rotate_half", "24": "persimmon/modeling_persimmon.py:apply_rotary_pos_emb", "25": "persimmon/modeling_persimmon.py:PersimmonMLP", "26": "persimmon/modeling_persimmon.py:eager_attention_forward", "27": "persimmon/modeling_persimmon.py:PersimmonAttention", "28": "persimmon/modeling_persimmon.py:PersimmonDecoderLayer", "29": "persimmon/modeling_persimmon.py:PersimmonPreTrainedModel", "30": "persimmon/modeling_persimmon.py:PersimmonModel", "31": "persimmon/modeling_persimmon.py:PersimmonForCausalLM", "32": "persimmon/modeling_persimmon.py:PersimmonForSequenceClassification", "33": "persimmon/modeling_persimmon.py:PersimmonForTokenClassification", "34": "sam3_video/modeling_sam3_video.py:_load_cv_utils_kernel_once", "35": "sam3_video/modeling_sam3_video.py:Sam3VideoInferenceCache", "36": "sam3_video/modeling_sam3_video.py:Sam3VideoInferenceSession", "37": "sam3_video/modeling_sam3_video.py:Sam3VideoSegmentationOutput", "38": "sam3_video/modeling_sam3_video.py:Sam3VideoPreTrainedModel", "39": "sam3_video/modeling_sam3_video.py:Sam3VideoModel", "40": "sam3_video/modeling_sam3_video.py:fast_diag_box_iou", "41": "sam3_video/modeling_sam3_video.py:mask_iou", "42": "sam3_video/modeling_sam3_video.py:nms_masks", "43": "sam3_video/modeling_sam3_video.py:fill_holes_in_mask_scores", "44": "sam3_video/modeling_sam3_video.py:_get_connected_components_with_padding", "45": "autoformer/modeling_autoformer.py:AutoFormerDecoderOutput", "46": "autoformer/modeling_autoformer.py:AutoformerModelOutput", "47": "autoformer/modeling_autoformer.py:AutoformerFeatureEmbedder", "48": "autoformer/modeling_autoformer.py:AutoformerStdScaler", "49": "autoformer/modeling_autoformer.py:AutoformerMeanScaler", "50": "autoformer/modeling_autoformer.py:AutoformerNOPScaler", "51": "autoformer/modeling_autoformer.py:weighted_average", "52": "autoformer/modeling_autoformer.py:nll", "53": "autoformer/modeling_autoformer.py:AutoformerSinusoidalPositionalEmbedding", "54": "autoformer/modeling_autoformer.py:AutoformerValueEmbedding", "55": "autoformer/modeling_autoformer.py:AutoformerSeriesDecompositionLayer", "56": "autoformer/modeling_autoformer.py:AutoformerLayernorm", "57": "autoformer/modeling_autoformer.py:AutoformerAttention", "58": "autoformer/modeling_autoformer.py:AutoformerEncoderLayer", "59": "autoformer/modeling_autoformer.py:AutoformerDecoderLayer", "60": "autoformer/modeling_autoformer.py:AutoformerPreTrainedModel", "61": "autoformer/modeling_autoformer.py:AutoformerEncoder", "62": "autoformer/modeling_autoformer.py:AutoformerDecoder", "63": "autoformer/modeling_autoformer.py:AutoformerModel", "64": "autoformer/modeling_autoformer.py:AutoformerForPrediction", "65": "dia/modeling_dia.py:DiaPreTrainedModel", "66": "dia/modeling_dia.py:DiaMultiChannelEmbedding", "67": "dia/modeling_dia.py:DiaMLP", "68": "dia/modeling_dia.py:DiaRMSNorm", "69": "dia/modeling_dia.py:DiaRotaryEmbedding", "70": "dia/modeling_dia.py:rotate_half", "71": "dia/modeling_dia.py:apply_rotary_pos_emb", "72": "dia/modeling_dia.py:repeat_kv", "73": "dia/modeling_dia.py:eager_attention_forward", "74": "dia/modeling_dia.py:DiaSelfAttention", "75": "dia/modeling_dia.py:DiaCrossAttention", "76": "dia/modeling_dia.py:DiaEncoderLayer", "77": "dia/modeling_dia.py:DiaEncoder", "78": "dia/modeling_dia.py:DiaDecoderLayer", "79": "dia/modeling_dia.py:DiaDecoder", "80": "dia/modeling_dia.py:DiaModel", "81": "dia/modeling_dia.py:DiaForConditionalGeneration", "82": "florence2/modeling_florence2.py:drop_path", "83": "florence2/modeling_florence2.py:Florence2VisionDropPath", "84": "florence2/modeling_florence2.py:Florence2VisionLearnedAbsolutePositionEmbedding2D", "85": "florence2/modeling_florence2.py:Florence2VisionPositionalEmbeddingCosine1D", "86": "florence2/modeling_florence2.py:Florence2VisionMLP", "87": "florence2/modeling_florence2.py:Florence2VisionConvEmbed", "88": "florence2/modeling_florence2.py:eager_attention_forward", "89": "florence2/modeling_florence2.py:Florence2VisionChannelAttention", "90": "florence2/modeling_florence2.py:Florence2VisionChannelBlock", "91": "florence2/modeling_florence2.py:Florence2VisionWindowAttention", "92": "florence2/modeling_florence2.py:Florence2VisionSpatialBlock", "93": "florence2/modeling_florence2.py:Florence2VisionBlock", "94": "florence2/modeling_florence2.py:Florence2VisionPreTrainedModel", "95": "florence2/modeling_florence2.py:Florence2VisionBackbone", "96": "florence2/modeling_florence2.py:Florence2MultiModalProjector", "97": "florence2/modeling_florence2.py:Florence2Seq2SeqModelOutput", "98": "florence2/modeling_florence2.py:Florence2Seq2SeqLMOutput", "99": "florence2/modeling_florence2.py:Florence2PreTrainedModel", "100": "florence2/modeling_florence2.py:Florence2Model", "101": "florence2/modeling_florence2.py:shift_tokens_right", "102": "florence2/modeling_florence2.py:Florence2ForConditionalGeneration", "103": "vit/modeling_vit.py:ViTEmbeddings", "104": "vit/modeling_vit.py:ViTPatchEmbeddings", "105": "vit/modeling_vit.py:eager_attention_forward", "106": "vit/modeling_vit.py:ViTSelfAttention", "107": "vit/modeling_vit.py:ViTSelfOutput", "108": "vit/modeling_vit.py:ViTAttention", "109": "vit/modeling_vit.py:ViTIntermediate", "110": "vit/modeling_vit.py:ViTOutput", "111": "vit/modeling_vit.py:ViTLayer", "112": "vit/modeling_vit.py:ViTEncoder", "113": "vit/modeling_vit.py:ViTPreTrainedModel", "114": "vit/modeling_vit.py:ViTModel", "115": "vit/modeling_vit.py:ViTPooler", "116": "vit/modeling_vit.py:ViTForMaskedImageModeling", "117": "vit/modeling_vit.py:ViTForImageClassification", "118": "longformer/modeling_longformer.py:LongformerBaseModelOutput", "119": "longformer/modeling_longformer.py:LongformerBaseModelOutputWithPooling", "120": "longformer/modeling_longformer.py:LongformerMaskedLMOutput", "121": "longformer/modeling_longformer.py:LongformerQuestionAnsweringModelOutput", "122": "longformer/modeling_longformer.py:LongformerSequenceClassifierOutput", "123": "longformer/modeling_longformer.py:LongformerMultipleChoiceModelOutput", "124": "longformer/modeling_longformer.py:LongformerTokenClassifierOutput", "125": "longformer/modeling_longformer.py:_get_question_end_index", "126": "longformer/modeling_longformer.py:_compute_global_attention_mask", "127": "longformer/modeling_longformer.py:create_position_ids_from_input_ids", "128": "longformer/modeling_longformer.py:LongformerEmbeddings", "129": "longformer/modeling_longformer.py:LongformerSelfAttention", "130": "longformer/modeling_longformer.py:LongformerSelfOutput", "131": "longformer/modeling_longformer.py:LongformerAttention", "132": "longformer/modeling_longformer.py:LongformerIntermediate", "133": "longformer/modeling_longformer.py:LongformerOutput", "134": "longformer/modeling_longformer.py:LongformerLayer", "135": "longformer/modeling_longformer.py:LongformerEncoder", "136": "longformer/modeling_longformer.py:LongformerPooler", "137": "longformer/modeling_longformer.py:LongformerLMHead", "138": "longformer/modeling_longformer.py:LongformerPreTrainedModel", "139": "longformer/modeling_longformer.py:LongformerModel", "140": "longformer/modeling_longformer.py:LongformerForMaskedLM", "141": "longformer/modeling_longformer.py:LongformerForSequenceClassification", "142": "longformer/modeling_longformer.py:LongformerClassificationHead", "143": "longformer/modeling_longformer.py:LongformerForQuestionAnswering", "144": "longformer/modeling_longformer.py:LongformerForTokenClassification", "145": "longformer/modeling_longformer.py:LongformerForMultipleChoice", "146": "olmo3/modeling_olmo3.py:Olmo3RMSNorm", "147": "olmo3/modeling_olmo3.py:repeat_kv", "148": "olmo3/modeling_olmo3.py:eager_attention_forward", "149": "olmo3/modeling_olmo3.py:apply_rotary_pos_emb", "150": "olmo3/modeling_olmo3.py:rotate_half", "151": "olmo3/modeling_olmo3.py:Olmo3Attention", "152": "olmo3/modeling_olmo3.py:Olmo3MLP", "153": "olmo3/modeling_olmo3.py:Olmo3DecoderLayer", "154": "olmo3/modeling_olmo3.py:Olmo3RotaryEmbedding", "155": "olmo3/modeling_olmo3.py:Olmo3PreTrainedModel", "156": "olmo3/modeling_olmo3.py:Olmo3Model", "157": "olmo3/modeling_olmo3.py:Olmo3ForCausalLM", "158": "reformer/modeling_reformer.py:ReformerDynamicCache", "159": "reformer/modeling_reformer.py:_stable_argsort", "160": "reformer/modeling_reformer.py:_get_least_common_mult_chunk_len", "161": "reformer/modeling_reformer.py:_get_min_chunk_len", "162": "reformer/modeling_reformer.py:AxialPositionEmbeddings", "163": "reformer/modeling_reformer.py:PositionEmbeddings", "164": "reformer/modeling_reformer.py:ReformerEmbeddings", "165": "reformer/modeling_reformer.py:EfficientAttentionMixin", "166": "reformer/modeling_reformer.py:LSHSelfAttention", "167": "reformer/modeling_reformer.py:ReverseSort", "168": "reformer/modeling_reformer.py:LocalSelfAttention", "169": "reformer/modeling_reformer.py:ReformerSelfOutput", "170": "reformer/modeling_reformer.py:ReformerAttention", "171": "reformer/modeling_reformer.py:ReformerFeedForwardDense", "172": "reformer/modeling_reformer.py:ReformerFeedForwardOutput", "173": "reformer/modeling_reformer.py:ChunkReformerFeedForward", "174": "reformer/modeling_reformer.py:ReformerLayer", "175": "reformer/modeling_reformer.py:_ReversibleFunction", "176": "reformer/modeling_reformer.py:ReformerEncoder", "177": "reformer/modeling_reformer.py:ReformerOnlyLMHead", "178": "reformer/modeling_reformer.py:ReformerPreTrainedModel", "179": "reformer/modeling_reformer.py:ReformerModelOutput", "180": "reformer/modeling_reformer.py:ReformerModelWithLMHeadOutput", "181": "reformer/modeling_reformer.py:ReformerModel", "182": "reformer/modeling_reformer.py:ReformerModelWithLMHead", "183": "reformer/modeling_reformer.py:ReformerForMaskedLM", "184": "reformer/modeling_reformer.py:ReformerForSequenceClassification", "185": "reformer/modeling_reformer.py:ReformerClassificationHead", "186": "reformer/modeling_reformer.py:ReformerForQuestionAnswering", "187": "visual_bert/modeling_visual_bert.py:VisualBertEmbeddings", "188": "visual_bert/modeling_visual_bert.py:VisualBertSelfAttention", "189": "visual_bert/modeling_visual_bert.py:VisualBertSelfOutput", "190": "visual_bert/modeling_visual_bert.py:VisualBertAttention", "191": "visual_bert/modeling_visual_bert.py:VisualBertIntermediate", "192": "visual_bert/modeling_visual_bert.py:VisualBertOutput", "193": "visual_bert/modeling_visual_bert.py:VisualBertLayer", "194": "visual_bert/modeling_visual_bert.py:VisualBertEncoder", "195": "visual_bert/modeling_visual_bert.py:VisualBertPooler", "196": "visual_bert/modeling_visual_bert.py:VisualBertPredictionHeadTransform", "197": "visual_bert/modeling_visual_bert.py:VisualBertLMPredictionHead", "198": "visual_bert/modeling_visual_bert.py:VisualBertPreTrainingHeads", "199": "visual_bert/modeling_visual_bert.py:VisualBertPreTrainedModel", "200": "visual_bert/modeling_visual_bert.py:VisualBertForPreTrainingOutput", "201": "visual_bert/modeling_visual_bert.py:VisualBertModel", "202": "visual_bert/modeling_visual_bert.py:VisualBertForPreTraining", "203": "visual_bert/modeling_visual_bert.py:VisualBertForMultipleChoice", "204": "visual_bert/modeling_visual_bert.py:VisualBertForQuestionAnswering", "205": "visual_bert/modeling_visual_bert.py:VisualBertForVisualReasoning", "206": "visual_bert/modeling_visual_bert.py:VisualBertRegionToPhraseAttention", "207": "visual_bert/modeling_visual_bert.py:VisualBertForRegionToPhraseAlignment", "208": "blenderbot_small/modeling_blenderbot_small.py:shift_tokens_right", "209": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallLearnedPositionalEmbedding", "210": "blenderbot_small/modeling_blenderbot_small.py:eager_attention_forward", "211": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallAttention", "212": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallEncoderLayer", "213": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallDecoderLayer", "214": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallPreTrainedModel", "215": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallEncoder", "216": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallDecoder", "217": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallModel", "218": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallForConditionalGeneration", "219": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallDecoderWrapper", "220": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallForCausalLM", "221": "dots1/modeling_dots1.py:Dots1RMSNorm", "222": "dots1/modeling_dots1.py:Dots1RotaryEmbedding", "223": "dots1/modeling_dots1.py:rotate_half", "224": "dots1/modeling_dots1.py:apply_rotary_pos_emb", "225": "dots1/modeling_dots1.py:repeat_kv", "226": "dots1/modeling_dots1.py:eager_attention_forward", "227": "dots1/modeling_dots1.py:Dots1Attention", "228": "dots1/modeling_dots1.py:Dots1MLP", "229": "dots1/modeling_dots1.py:Dots1TopkRouter", "230": "dots1/modeling_dots1.py:Dots1NaiveMoe", "231": "dots1/modeling_dots1.py:Dots1MoE", "232": "dots1/modeling_dots1.py:Dots1DecoderLayer", "233": "dots1/modeling_dots1.py:Dots1PreTrainedModel", "234": "dots1/modeling_dots1.py:Dots1Model", "235": "dots1/modeling_dots1.py:Dots1ForCausalLM", "236": "depth_anything/modeling_depth_anything.py:DepthAnythingReassembleLayer", "237": "depth_anything/modeling_depth_anything.py:DepthAnythingReassembleStage", "238": "depth_anything/modeling_depth_anything.py:DepthAnythingPreActResidualLayer", "239": "depth_anything/modeling_depth_anything.py:DepthAnythingFeatureFusionLayer", "240": "depth_anything/modeling_depth_anything.py:DepthAnythingFeatureFusionStage", "241": "depth_anything/modeling_depth_anything.py:DepthAnythingPreTrainedModel", "242": "depth_anything/modeling_depth_anything.py:DepthAnythingNeck", "243": "depth_anything/modeling_depth_anything.py:DepthAnythingDepthEstimationHead", "244": "depth_anything/modeling_depth_anything.py:DepthAnythingForDepthEstimation", "245": "swiftformer/modeling_swiftformer.py:SwiftFormerPatchEmbedding", "246": "swiftformer/modeling_swiftformer.py:drop_path", "247": "swiftformer/modeling_swiftformer.py:SwiftFormerDropPath", "248": "swiftformer/modeling_swiftformer.py:SwiftFormerEmbeddings", "249": "swiftformer/modeling_swiftformer.py:SwiftFormerConvEncoder", "250": "swiftformer/modeling_swiftformer.py:SwiftFormerMlp", "251": "swiftformer/modeling_swiftformer.py:SwiftFormerEfficientAdditiveAttention", "252": "swiftformer/modeling_swiftformer.py:SwiftFormerLocalRepresentation", "253": "swiftformer/modeling_swiftformer.py:SwiftFormerEncoderBlock", "254": "swiftformer/modeling_swiftformer.py:SwiftFormerStage", "255": "swiftformer/modeling_swiftformer.py:SwiftFormerEncoder", "256": "swiftformer/modeling_swiftformer.py:SwiftFormerPreTrainedModel", "257": "swiftformer/modeling_swiftformer.py:SwiftFormerModel", "258": "swiftformer/modeling_swiftformer.py:SwiftFormerForImageClassification", "259": "moshi/modeling_moshi.py:MoshiConditionalGenerationGenerateOutput", "260": "moshi/modeling_moshi.py:MoshiCausalLMOutputWithPast", "261": "moshi/modeling_moshi.py:MoshiConditionalGenerationOutputWithPast", "262": "moshi/modeling_moshi.py:MoshiUnconditionalInput", "263": "moshi/modeling_moshi.py:MoshiRMSNorm", "264": "moshi/modeling_moshi.py:MoshiFlexibleLinear", "265": "moshi/modeling_moshi.py:MoshiLinear", "266": "moshi/modeling_moshi.py:MoshiRotaryEmbedding", "267": "moshi/modeling_moshi.py:rotate_half", "268": "moshi/modeling_moshi.py:apply_rotary_pos_emb", "269": "moshi/modeling_moshi.py:MoshiGatingMLP", "270": "moshi/modeling_moshi.py:repeat_kv", "271": "moshi/modeling_moshi.py:MoshiAttention", "272": "moshi/modeling_moshi.py:MoshiFlashAttention2", "273": "moshi/modeling_moshi.py:MoshiSdpaAttention", "274": "moshi/modeling_moshi.py:MoshiDecoderLayer", "275": "moshi/modeling_moshi.py:MoshiPreTrainedModel", "276": "moshi/modeling_moshi.py:MoshiDepthDecoder", "277": "moshi/modeling_moshi.py:MoshiModel", "278": "moshi/modeling_moshi.py:MoshiForCausalLM", "279": "moshi/modeling_moshi.py:MoshiForConditionalGeneration", "280": "luke/modeling_luke.py:BaseLukeModelOutputWithPooling", "281": "luke/modeling_luke.py:BaseLukeModelOutput", "282": "luke/modeling_luke.py:LukeMaskedLMOutput", "283": "luke/modeling_luke.py:EntityClassificationOutput", "284": "luke/modeling_luke.py:EntityPairClassificationOutput", "285": "luke/modeling_luke.py:EntitySpanClassificationOutput", "286": "luke/modeling_luke.py:LukeSequenceClassifierOutput", "287": "luke/modeling_luke.py:LukeTokenClassifierOutput", "288": "luke/modeling_luke.py:LukeQuestionAnsweringModelOutput", "289": "luke/modeling_luke.py:LukeMultipleChoiceModelOutput", "290": "luke/modeling_luke.py:LukeEmbeddings", "291": "luke/modeling_luke.py:LukeEntityEmbeddings", "292": "luke/modeling_luke.py:LukeSelfAttention", "293": "luke/modeling_luke.py:LukeSelfOutput", "294": "luke/modeling_luke.py:LukeAttention", "295": "luke/modeling_luke.py:LukeIntermediate", "296": "luke/modeling_luke.py:LukeOutput", "297": "luke/modeling_luke.py:LukeLayer", "298": "luke/modeling_luke.py:LukeEncoder", "299": "luke/modeling_luke.py:LukePooler", "300": "luke/modeling_luke.py:EntityPredictionHeadTransform", "301": "luke/modeling_luke.py:EntityPredictionHead", "302": "luke/modeling_luke.py:LukePreTrainedModel", "303": "luke/modeling_luke.py:LukeModel", "304": "luke/modeling_luke.py:create_position_ids_from_input_ids", "305": "luke/modeling_luke.py:LukeLMHead", "306": "luke/modeling_luke.py:LukeForMaskedLM", "307": "luke/modeling_luke.py:LukeForEntityClassification", "308": "luke/modeling_luke.py:LukeForEntityPairClassification", "309": "luke/modeling_luke.py:LukeForEntitySpanClassification", "310": "luke/modeling_luke.py:LukeForSequenceClassification", "311": "luke/modeling_luke.py:LukeForTokenClassification", "312": "luke/modeling_luke.py:LukeForQuestionAnswering", "313": "luke/modeling_luke.py:LukeForMultipleChoice", "314": "encoder_decoder/modeling_encoder_decoder.py:shift_tokens_right", "315": "encoder_decoder/modeling_encoder_decoder.py:EncoderDecoderModel", "316": "nystromformer/modeling_nystromformer.py:NystromformerEmbeddings", "317": "nystromformer/modeling_nystromformer.py:NystromformerSelfAttention", "318": "nystromformer/modeling_nystromformer.py:NystromformerSelfOutput", "319": "nystromformer/modeling_nystromformer.py:NystromformerAttention", "320": "nystromformer/modeling_nystromformer.py:NystromformerIntermediate", "321": "nystromformer/modeling_nystromformer.py:NystromformerOutput", "322": "nystromformer/modeling_nystromformer.py:NystromformerLayer", "323": "nystromformer/modeling_nystromformer.py:NystromformerEncoder", "324": "nystromformer/modeling_nystromformer.py:NystromformerPredictionHeadTransform", "325": "nystromformer/modeling_nystromformer.py:NystromformerLMPredictionHead", "326": "nystromformer/modeling_nystromformer.py:NystromformerOnlyMLMHead", "327": "nystromformer/modeling_nystromformer.py:NystromformerPreTrainedModel", "328": "nystromformer/modeling_nystromformer.py:NystromformerModel", "329": "nystromformer/modeling_nystromformer.py:NystromformerForMaskedLM", "330": "nystromformer/modeling_nystromformer.py:NystromformerClassificationHead", "331": "nystromformer/modeling_nystromformer.py:NystromformerForSequenceClassification", "332": "nystromformer/modeling_nystromformer.py:NystromformerForMultipleChoice", "333": "nystromformer/modeling_nystromformer.py:NystromformerForTokenClassification", "334": "nystromformer/modeling_nystromformer.py:NystromformerForQuestionAnswering", "335": "informer/modeling_informer.py:InformerFeatureEmbedder", "336": "informer/modeling_informer.py:InformerStdScaler", "337": "informer/modeling_informer.py:InformerMeanScaler", "338": "informer/modeling_informer.py:InformerNOPScaler", "339": "informer/modeling_informer.py:InformerSinusoidalPositionalEmbedding", "340": "informer/modeling_informer.py:InformerValueEmbedding", "341": "informer/modeling_informer.py:InformerPreTrainedModel", "342": "informer/modeling_informer.py:eager_attention_forward", "343": "informer/modeling_informer.py:InformerAttention", "344": "informer/modeling_informer.py:InformerProbSparseAttention", "345": "informer/modeling_informer.py:InformerConvLayer", "346": "informer/modeling_informer.py:InformerEncoderLayer", "347": "informer/modeling_informer.py:InformerDecoderLayer", "348": "informer/modeling_informer.py:InformerEncoder", "349": "informer/modeling_informer.py:InformerDecoder", "350": "informer/modeling_informer.py:InformerModel", "351": "informer/modeling_informer.py:weighted_average", "352": "informer/modeling_informer.py:nll", "353": "informer/modeling_informer.py:InformerForPrediction", "354": "sew/modeling_sew.py:SEWNoLayerNormConvLayer", "355": "sew/modeling_sew.py:SEWLayerNormConvLayer", "356": "sew/modeling_sew.py:SEWGroupNormConvLayer", "357": "sew/modeling_sew.py:SEWPositionalConvEmbedding", "358": "sew/modeling_sew.py:SEWSamePadLayer", "359": "sew/modeling_sew.py:SEWUpsampling", "360": "sew/modeling_sew.py:SEWFeatureEncoder", "361": "sew/modeling_sew.py:eager_attention_forward", "362": "sew/modeling_sew.py:SEWAttention", "363": "sew/modeling_sew.py:SEWFeedForward", "364": "sew/modeling_sew.py:SEWEncoderLayer", "365": "sew/modeling_sew.py:SEWEncoder", "366": "sew/modeling_sew.py:SEWPreTrainedModel", "367": "sew/modeling_sew.py:_compute_mask_indices", "368": "sew/modeling_sew.py:SEWModel", "369": "sew/modeling_sew.py:SEWForCTC", "370": "sew/modeling_sew.py:SEWForSequenceClassification", "371": "fsmt/modeling_fsmt.py:invert_mask", "372": "fsmt/modeling_fsmt.py:triu_onnx", "373": "fsmt/modeling_fsmt.py:_prepare_fsmt_decoder_inputs", "374": "fsmt/modeling_fsmt.py:PretrainedFSMTModel", "375": "fsmt/modeling_fsmt.py:_make_linear_from_emb", "376": "fsmt/modeling_fsmt.py:_check_shapes", "377": "fsmt/modeling_fsmt.py:shift_tokens_right", "378": "fsmt/modeling_fsmt.py:make_padding_mask", "379": "fsmt/modeling_fsmt.py:EncoderLayer", "380": "fsmt/modeling_fsmt.py:FSMTEncoder", "381": "fsmt/modeling_fsmt.py:DecoderLayer", "382": "fsmt/modeling_fsmt.py:FSMTDecoder", "383": "fsmt/modeling_fsmt.py:_reorder_buffer", "384": "fsmt/modeling_fsmt.py:Attention", "385": "fsmt/modeling_fsmt.py:fill_with_neg_inf", "386": "fsmt/modeling_fsmt.py:_get_shape", "387": "fsmt/modeling_fsmt.py:FSMTModel", "388": "fsmt/modeling_fsmt.py:FSMTForConditionalGeneration", "389": "fsmt/modeling_fsmt.py:SinusoidalPositionalEmbedding", "390": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRKeypointMatchingOutput", "391": "efficientloftr/modeling_efficientloftr.py:compute_embeddings", "392": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRRotaryEmbedding", "393": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRConvNormLayer", "394": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRRepVGGBlock", "395": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRRepVGGStage", "396": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRepVGG", "397": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRAggregationLayer", "398": "efficientloftr/modeling_efficientloftr.py:rotate_half", "399": "efficientloftr/modeling_efficientloftr.py:apply_rotary_pos_emb", "400": "efficientloftr/modeling_efficientloftr.py:repeat_kv", "401": "efficientloftr/modeling_efficientloftr.py:eager_attention_forward", "402": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRAttention", "403": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRMLP", "404": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRAggregatedAttention", "405": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRLocalFeatureTransformerLayer", "406": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRLocalFeatureTransformer", "407": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTROutConvBlock", "408": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRFineFusionLayer", "409": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRPreTrainedModel", "410": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRModel", "411": "efficientloftr/modeling_efficientloftr.py:mask_border", "412": "efficientloftr/modeling_efficientloftr.py:create_meshgrid", "413": "efficientloftr/modeling_efficientloftr.py:spatial_expectation2d", "414": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRForKeypointMatching", "415": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormEmbeddings", "416": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:eager_attention_forward", "417": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormSelfAttention", "418": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormCrossAttention", "419": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormSelfOutput", "420": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormAttention", "421": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormIntermediate", "422": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormOutput", "423": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormLayer", "424": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormEncoder", "425": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormPooler", "426": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormPreTrainedModel", "427": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormModel", "428": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormForCausalLM", "429": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormForMaskedLM", "430": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormLMHead", "431": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormForSequenceClassification", "432": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormForMultipleChoice", "433": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormForTokenClassification", "434": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormClassificationHead", "435": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormForQuestionAnswering", "436": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:SinusoidsPositionEmbedding", "437": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoePreTrainedModel", "438": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:_get_feat_extract_output_lengths", "439": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoePreTrainedModelForConditionalGeneration", "440": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:repeat_kv", "441": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:eager_attention_forward", "442": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeAudioAttention", "443": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeAudioEncoderLayer", "444": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeAudioEncoder", "445": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:rotate_half", "446": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:apply_rotary_pos_emb_vision", "447": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeVisionAttention", "448": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeVisionPatchMerger", "449": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeVisionRotaryEmbedding", "450": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeVisionMLP", "451": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeVisionPatchEmbed", "452": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeVisionBlock", "453": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeVisionEncoder", "454": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextRotaryEmbedding", "455": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextExperts", "456": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextTopKRouter", "457": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextSparseMoeBlock", "458": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextRMSNorm", "459": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:apply_rotary_pos_emb", "460": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextAttention", "461": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextMLP", "462": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextDecoderLayer", "463": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextPreTrainedModel", "464": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTextRMSNorm", "465": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextModel", "466": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerCausalLMOutputWithPast", "467": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:load_balancing_loss_func", "468": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerForConditionalGeneration", "469": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerResizeMLP", "470": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerCodePredictorOutputWithPast", "471": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeRMSNorm", "472": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerCodePredictorAttention", "473": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeMLP", "474": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerCodePredictorDecoderLayer", "475": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeRotaryEmbedding", "476": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerCodePredictorModel", "477": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerCodePredictorModelForConditionalGeneration", "478": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerOutputWithPast", "479": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerRotaryEmbedding", "480": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerTextMLP", "481": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerTextExperts", "482": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerTextTopKRouter", "483": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerTextSparseMoeBlock", "484": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerDecoderLayer", "485": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerModel", "486": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerForConditionalGeneration", "487": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCausalConvNet", "488": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCausalTransConvNet", "489": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeConvNeXtBlock", "490": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavAttention", "491": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavMlp", "492": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavRMSNorm", "493": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavLayerScale", "494": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavTransformerLayer", "495": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavTransformerModel", "496": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:SnakeBeta", "497": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavDecoderResidualUnit", "498": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavDecoderBlock", "499": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2Wav", "500": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeForConditionalGeneration", "501": "detr/modeling_detr.py:DetrDecoderOutput", "502": "detr/modeling_detr.py:DetrModelOutput", "503": "detr/modeling_detr.py:DetrObjectDetectionOutput", "504": "detr/modeling_detr.py:DetrSegmentationOutput", "505": "detr/modeling_detr.py:DetrFrozenBatchNorm2d", "506": "detr/modeling_detr.py:replace_batch_norm", "507": "detr/modeling_detr.py:DetrConvEncoder", "508": "detr/modeling_detr.py:DetrConvModel", "509": "detr/modeling_detr.py:DetrSinePositionEmbedding", "510": "detr/modeling_detr.py:DetrLearnedPositionEmbedding", "511": "detr/modeling_detr.py:build_position_encoding", "512": "detr/modeling_detr.py:DetrAttention", "513": "detr/modeling_detr.py:DetrEncoderLayer", "514": "detr/modeling_detr.py:DetrDecoderLayer", "515": "detr/modeling_detr.py:DetrPreTrainedModel", "516": "detr/modeling_detr.py:DetrEncoder", "517": "detr/modeling_detr.py:DetrDecoder", "518": "detr/modeling_detr.py:DetrModel", "519": "detr/modeling_detr.py:DetrMLPPredictionHead", "520": "detr/modeling_detr.py:DetrForObjectDetection", "521": "detr/modeling_detr.py:DetrForSegmentation", "522": "detr/modeling_detr.py:_expand", "523": "detr/modeling_detr.py:DetrMaskHeadSmallConv", "524": "detr/modeling_detr.py:DetrMHAttentionMap", "525": "sam/modeling_sam.py:SamVisionEncoderOutput", "526": "sam/modeling_sam.py:SamImageSegmentationOutput", "527": "sam/modeling_sam.py:SamPatchEmbeddings", "528": "sam/modeling_sam.py:SamMLPBlock", "529": "sam/modeling_sam.py:SamLayerNorm", "530": "sam/modeling_sam.py:eager_attention_forward", "531": "sam/modeling_sam.py:SamAttention", "532": "sam/modeling_sam.py:SamTwoWayAttentionBlock", "533": "sam/modeling_sam.py:SamTwoWayTransformer", "534": "sam/modeling_sam.py:SamFeedForward", "535": "sam/modeling_sam.py:SamMaskDecoder", "536": "sam/modeling_sam.py:SamPositionalEmbedding", "537": "sam/modeling_sam.py:SamMaskEmbedding", "538": "sam/modeling_sam.py:SamPromptEncoder", "539": "sam/modeling_sam.py:SamVisionAttention", "540": "sam/modeling_sam.py:SamVisionSdpaAttention", "541": "sam/modeling_sam.py:SamVisionLayer", "542": "sam/modeling_sam.py:SamVisionNeck", "543": "sam/modeling_sam.py:SamPreTrainedModel", "544": "sam/modeling_sam.py:SamVisionEncoder", "545": "sam/modeling_sam.py:SamVisionModel", "546": "sam/modeling_sam.py:SamModel", "547": "owlvit/modeling_owlvit.py:contrastive_loss", "548": "owlvit/modeling_owlvit.py:owlvit_loss", "549": "owlvit/modeling_owlvit.py:OwlViTOutput", "550": "owlvit/modeling_owlvit.py:_upcast", "551": "owlvit/modeling_owlvit.py:box_area", "552": "owlvit/modeling_owlvit.py:box_iou", "553": "owlvit/modeling_owlvit.py:generalized_box_iou", "554": "owlvit/modeling_owlvit.py:OwlViTObjectDetectionOutput", "555": "owlvit/modeling_owlvit.py:OwlViTImageGuidedObjectDetectionOutput", "556": "owlvit/modeling_owlvit.py:OwlViTVisionEmbeddings", "557": "owlvit/modeling_owlvit.py:OwlViTTextEmbeddings", "558": "owlvit/modeling_owlvit.py:OwlViTAttention", "559": "owlvit/modeling_owlvit.py:OwlViTMLP", "560": "owlvit/modeling_owlvit.py:OwlViTEncoderLayer", "561": "owlvit/modeling_owlvit.py:OwlViTPreTrainedModel", "562": "owlvit/modeling_owlvit.py:OwlViTEncoder", "563": "owlvit/modeling_owlvit.py:OwlViTTextTransformer", "564": "owlvit/modeling_owlvit.py:OwlViTTextModel", "565": "owlvit/modeling_owlvit.py:OwlViTVisionTransformer", "566": "owlvit/modeling_owlvit.py:OwlViTVisionModel", "567": "owlvit/modeling_owlvit.py:OwlViTModel", "568": "owlvit/modeling_owlvit.py:OwlViTBoxPredictionHead", "569": "owlvit/modeling_owlvit.py:OwlViTClassPredictionHead", "570": "owlvit/modeling_owlvit.py:OwlViTForObjectDetection", "571": "apertus/modeling_apertus.py:ApertusMLP", "572": "apertus/modeling_apertus.py:ApertusRMSNorm", "573": "apertus/modeling_apertus.py:ApertusRotaryEmbedding", "574": "apertus/modeling_apertus.py:rotate_half", "575": "apertus/modeling_apertus.py:apply_rotary_pos_emb", "576": "apertus/modeling_apertus.py:repeat_kv", "577": "apertus/modeling_apertus.py:eager_attention_forward", "578": "apertus/modeling_apertus.py:ApertusAttention", "579": "apertus/modeling_apertus.py:ApertusDecoderLayer", "580": "apertus/modeling_apertus.py:ApertusPreTrainedModel", "581": "apertus/modeling_apertus.py:ApertusModel", "582": "apertus/modeling_apertus.py:ApertusForCausalLM", "583": "apertus/modeling_apertus.py:ApertusForTokenClassification", "584": "focalnet/modeling_focalnet.py:FocalNetEncoderOutput", "585": "focalnet/modeling_focalnet.py:FocalNetModelOutput", "586": "focalnet/modeling_focalnet.py:FocalNetMaskedImageModelingOutput", "587": "focalnet/modeling_focalnet.py:FocalNetImageClassifierOutput", "588": "focalnet/modeling_focalnet.py:FocalNetEmbeddings", "589": "focalnet/modeling_focalnet.py:FocalNetPatchEmbeddings", "590": "focalnet/modeling_focalnet.py:drop_path", "591": "focalnet/modeling_focalnet.py:FocalNetDropPath", "592": "focalnet/modeling_focalnet.py:FocalNetModulation", "593": "focalnet/modeling_focalnet.py:FocalNetMlp", "594": "focalnet/modeling_focalnet.py:FocalNetLayer", "595": "focalnet/modeling_focalnet.py:FocalNetStage", "596": "focalnet/modeling_focalnet.py:FocalNetEncoder", "597": "focalnet/modeling_focalnet.py:FocalNetPreTrainedModel", "598": "focalnet/modeling_focalnet.py:FocalNetModel", "599": "focalnet/modeling_focalnet.py:FocalNetForMaskedImageModeling", "600": "focalnet/modeling_focalnet.py:FocalNetForImageClassification", "601": "focalnet/modeling_focalnet.py:FocalNetBackbone", "602": "albert/modeling_albert.py:AlbertEmbeddings", "603": "albert/modeling_albert.py:eager_attention_forward", "604": "albert/modeling_albert.py:AlbertAttention", "605": "albert/modeling_albert.py:AlbertLayer", "606": "albert/modeling_albert.py:AlbertLayerGroup", "607": "albert/modeling_albert.py:AlbertTransformer", "608": "albert/modeling_albert.py:AlbertPreTrainedModel", "609": "albert/modeling_albert.py:AlbertForPreTrainingOutput", "610": "albert/modeling_albert.py:AlbertModel", "611": "albert/modeling_albert.py:AlbertForPreTraining", "612": "albert/modeling_albert.py:AlbertMLMHead", "613": "albert/modeling_albert.py:AlbertSOPHead", "614": "albert/modeling_albert.py:AlbertForMaskedLM", "615": "albert/modeling_albert.py:AlbertForSequenceClassification", "616": "albert/modeling_albert.py:AlbertForTokenClassification", "617": "albert/modeling_albert.py:AlbertForQuestionAnswering", "618": "albert/modeling_albert.py:AlbertForMultipleChoice", "619": "electra/modeling_electra.py:ElectraEmbeddings", "620": "electra/modeling_electra.py:eager_attention_forward", "621": "electra/modeling_electra.py:ElectraSelfAttention", "622": "electra/modeling_electra.py:ElectraCrossAttention", "623": "electra/modeling_electra.py:ElectraSelfOutput", "624": "electra/modeling_electra.py:ElectraAttention", "625": "electra/modeling_electra.py:ElectraIntermediate", "626": "electra/modeling_electra.py:ElectraOutput", "627": "electra/modeling_electra.py:ElectraLayer", "628": "electra/modeling_electra.py:ElectraEncoder", "629": "electra/modeling_electra.py:ElectraDiscriminatorPredictions", "630": "electra/modeling_electra.py:ElectraGeneratorPredictions", "631": "electra/modeling_electra.py:ElectraPreTrainedModel", "632": "electra/modeling_electra.py:ElectraForPreTrainingOutput", "633": "electra/modeling_electra.py:ElectraModel", "634": "electra/modeling_electra.py:ElectraClassificationHead", "635": "electra/modeling_electra.py:ElectraSequenceSummary", "636": "electra/modeling_electra.py:ElectraForSequenceClassification", "637": "electra/modeling_electra.py:ElectraForPreTraining", "638": "electra/modeling_electra.py:ElectraForMaskedLM", "639": "electra/modeling_electra.py:ElectraForTokenClassification", "640": "electra/modeling_electra.py:ElectraForQuestionAnswering", "641": "electra/modeling_electra.py:ElectraForMultipleChoice", "642": "electra/modeling_electra.py:ElectraForCausalLM", "643": "wavlm/modeling_wavlm.py:WavLMSamePadLayer", "644": "wavlm/modeling_wavlm.py:WavLMPositionalConvEmbedding", "645": "wavlm/modeling_wavlm.py:WavLMFeatureProjection", "646": "wavlm/modeling_wavlm.py:WavLMAttention", "647": "wavlm/modeling_wavlm.py:WavLMFeedForward", "648": "wavlm/modeling_wavlm.py:WavLMEncoderLayer", "649": "wavlm/modeling_wavlm.py:WavLMEncoderLayerStableLayerNorm", "650": "wavlm/modeling_wavlm.py:WavLMEncoder", "651": "wavlm/modeling_wavlm.py:WavLMEncoderStableLayerNorm", "652": "wavlm/modeling_wavlm.py:WavLMGumbelVectorQuantizer", "653": "wavlm/modeling_wavlm.py:WavLMPreTrainedModel", "654": "wavlm/modeling_wavlm.py:WavLMNoLayerNormConvLayer", "655": "wavlm/modeling_wavlm.py:WavLMLayerNormConvLayer", "656": "wavlm/modeling_wavlm.py:WavLMGroupNormConvLayer", "657": "wavlm/modeling_wavlm.py:WavLMFeatureEncoder", "658": "wavlm/modeling_wavlm.py:WavLMAdapterLayer", "659": "wavlm/modeling_wavlm.py:WavLMAdapter", "660": "wavlm/modeling_wavlm.py:_compute_mask_indices", "661": "wavlm/modeling_wavlm.py:WavLMModel", "662": "wavlm/modeling_wavlm.py:WavLMForCTC", "663": "wavlm/modeling_wavlm.py:WavLMForSequenceClassification", "664": "wavlm/modeling_wavlm.py:WavLMForAudioFrameClassification", "665": "wavlm/modeling_wavlm.py:AMSoftmaxLoss", "666": "wavlm/modeling_wavlm.py:TDNNLayer", "667": "wavlm/modeling_wavlm.py:WavLMForXVector", "668": "llava_next_video/modeling_llava_next_video.py:LlavaNextVideoModelOutputWithPast", "669": "llava_next_video/modeling_llava_next_video.py:LlavaNextVideoCausalLMOutputWithPast", "670": "llava_next_video/modeling_llava_next_video.py:LlavaNextVideoPooler", "671": "llava_next_video/modeling_llava_next_video.py:LlavaNextVideoMultiModalProjector", "672": "llava_next_video/modeling_llava_next_video.py:LlavaNextVideoPreTrainedModel", "673": "llava_next_video/modeling_llava_next_video.py:get_anyres_image_grid_shape", "674": "llava_next_video/modeling_llava_next_video.py:image_size_to_num_patches", "675": "llava_next_video/modeling_llava_next_video.py:unpad_image", "676": "llava_next_video/modeling_llava_next_video.py:LlavaNextVideoModel", "677": "llava_next_video/modeling_llava_next_video.py:LlavaNextVideoForConditionalGeneration", "678": "sam2_video/modeling_sam2_video.py:Sam2VideoInferenceCache", "679": "sam2_video/modeling_sam2_video.py:Sam2VideoInferenceSession", "680": "sam2_video/modeling_sam2_video.py:Sam2VideoLayerNorm", "681": "sam2_video/modeling_sam2_video.py:Sam2VideoPositionEmbeddingSine", "682": "sam2_video/modeling_sam2_video.py:eager_attention_forward", "683": "sam2_video/modeling_sam2_video.py:Sam2VideoAttention", "684": "sam2_video/modeling_sam2_video.py:Sam2VideoTwoWayAttentionBlock", "685": "sam2_video/modeling_sam2_video.py:Sam2VideoFeedForward", "686": "sam2_video/modeling_sam2_video.py:Sam2VideoImageSegmentationOutput", "687": "sam2_video/modeling_sam2_video.py:Sam2VideoSegmentationOutput", "688": "sam2_video/modeling_sam2_video.py:Sam2VideoPreTrainedModel", "689": "sam2_video/modeling_sam2_video.py:Sam2VideoVisionRotaryEmbedding", "690": "sam2_video/modeling_sam2_video.py:rotate_pairwise", "691": "sam2_video/modeling_sam2_video.py:apply_rotary_pos_emb_2d", "692": "sam2_video/modeling_sam2_video.py:Sam2VideoRoPEAttention", "693": "sam2_video/modeling_sam2_video.py:Sam2VideoMemoryAttentionLayer", "694": "sam2_video/modeling_sam2_video.py:Sam2VideoMemoryAttention", "695": "sam2_video/modeling_sam2_video.py:Sam2VideoMemoryFuserCXBlock", "696": "sam2_video/modeling_sam2_video.py:Sam2VideoMemoryFuser", "697": "sam2_video/modeling_sam2_video.py:Sam2VideoMaskDownSamplerLayer", "698": "sam2_video/modeling_sam2_video.py:Sam2VideoMaskDownSampler", "699": "sam2_video/modeling_sam2_video.py:Sam2VideoMemoryEncoder", "700": "sam2_video/modeling_sam2_video.py:Sam2VideoPositionalEmbedding", "701": "sam2_video/modeling_sam2_video.py:Sam2VideoVisionEncoderOutput", "702": "sam2_video/modeling_sam2_video.py:Sam2VideoMaskEmbedding", "703": "sam2_video/modeling_sam2_video.py:Sam2VideoPromptEncoder", "704": "sam2_video/modeling_sam2_video.py:Sam2VideoTwoWayTransformer", "705": "sam2_video/modeling_sam2_video.py:Sam2VideoMaskDecoder", "706": "sam2_video/modeling_sam2_video.py:get_1d_sine_pe", "707": "sam2_video/modeling_sam2_video.py:Sam2VideoModel", "708": "canine/modeling_canine.py:CanineModelOutputWithPooling", "709": "canine/modeling_canine.py:CanineEmbeddings", "710": "canine/modeling_canine.py:CharactersToMolecules", "711": "canine/modeling_canine.py:ConvProjection", "712": "canine/modeling_canine.py:CanineSelfAttention", "713": "canine/modeling_canine.py:CanineSelfOutput", "714": "canine/modeling_canine.py:CanineAttention", "715": "canine/modeling_canine.py:CanineIntermediate", "716": "canine/modeling_canine.py:CanineOutput", "717": "canine/modeling_canine.py:CanineLayer", "718": "canine/modeling_canine.py:CanineEncoder", "719": "canine/modeling_canine.py:CaninePooler", "720": "canine/modeling_canine.py:CaninePredictionHeadTransform", "721": "canine/modeling_canine.py:CanineLMPredictionHead", "722": "canine/modeling_canine.py:CanineOnlyMLMHead", "723": "canine/modeling_canine.py:CaninePreTrainedModel", "724": "canine/modeling_canine.py:CanineModel", "725": "canine/modeling_canine.py:CanineForSequenceClassification", "726": "canine/modeling_canine.py:CanineForMultipleChoice", "727": "canine/modeling_canine.py:CanineForTokenClassification", "728": "canine/modeling_canine.py:CanineForQuestionAnswering", "729": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3RMSNorm", "730": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3RotaryEmbedding", "731": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3MLP", "732": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3TopkRouter", "733": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3NaiveMoe", "734": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3MoE", "735": "deepseek_v3/modeling_deepseek_v3.py:rotate_half", "736": "deepseek_v3/modeling_deepseek_v3.py:apply_rotary_pos_emb", "737": "deepseek_v3/modeling_deepseek_v3.py:repeat_kv", "738": "deepseek_v3/modeling_deepseek_v3.py:eager_attention_forward", "739": "deepseek_v3/modeling_deepseek_v3.py:apply_rotary_pos_emb_interleave", "740": "deepseek_v3/modeling_deepseek_v3.py:yarn_get_mscale", "741": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3Attention", "742": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3DecoderLayer", "743": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3PreTrainedModel", "744": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3Model", "745": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3ForCausalLM", "746": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3ForSequenceClassification", "747": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3ForTokenClassification", "748": "olmo/modeling_olmo.py:OlmoLayerNorm", "749": "olmo/modeling_olmo.py:OlmoMLP", "750": "olmo/modeling_olmo.py:OlmoRotaryEmbedding", "751": "olmo/modeling_olmo.py:rotate_half", "752": "olmo/modeling_olmo.py:repeat_kv", "753": "olmo/modeling_olmo.py:eager_attention_forward", "754": "olmo/modeling_olmo.py:apply_rotary_pos_emb", "755": "olmo/modeling_olmo.py:OlmoAttention", "756": "olmo/modeling_olmo.py:OlmoDecoderLayer", "757": "olmo/modeling_olmo.py:OlmoPreTrainedModel", "758": "olmo/modeling_olmo.py:OlmoModel", "759": "olmo/modeling_olmo.py:OlmoForCausalLM", "760": "pix2struct/modeling_pix2struct.py:Pix2StructLayerNorm", "761": "pix2struct/modeling_pix2struct.py:Pix2StructVisionEmbeddings", "762": "pix2struct/modeling_pix2struct.py:Pix2StructVisionAttention", "763": "pix2struct/modeling_pix2struct.py:Pix2StructVisionMlp", "764": "pix2struct/modeling_pix2struct.py:Pix2StructVisionLayer", "765": "pix2struct/modeling_pix2struct.py:Pix2StructVisionEncoder", "766": "pix2struct/modeling_pix2struct.py:Pix2StructPreTrainedModel", "767": "pix2struct/modeling_pix2struct.py:Pix2StructVisionModel", "768": "pix2struct/modeling_pix2struct.py:Pix2StructTextDenseGatedActDense", "769": "pix2struct/modeling_pix2struct.py:Pix2StructTextLayerFF", "770": "pix2struct/modeling_pix2struct.py:Pix2StructTextAttention", "771": "pix2struct/modeling_pix2struct.py:Pix2StructTextLayerSelfAttention", "772": "pix2struct/modeling_pix2struct.py:Pix2StructTextLayerCrossAttention", "773": "pix2struct/modeling_pix2struct.py:Pix2StructTextBlock", "774": "pix2struct/modeling_pix2struct.py:Pix2StructTextModel", "775": "pix2struct/modeling_pix2struct.py:Pix2StructForConditionalGeneration", "776": "sam_hq/modeling_sam_hq.py:SamHQVisionEncoderOutput", "777": "sam_hq/modeling_sam_hq.py:SamHQMMaskDecoderOutputs", "778": "sam_hq/modeling_sam_hq.py:SamHQImageSegmentationOutput", "779": "sam_hq/modeling_sam_hq.py:SamHQVisionAttention", "780": "sam_hq/modeling_sam_hq.py:SamHQMLPBlock", "781": "sam_hq/modeling_sam_hq.py:SamHQVisionSdpaAttention", "782": "sam_hq/modeling_sam_hq.py:SamHQVisionLayer", "783": "sam_hq/modeling_sam_hq.py:SamHQPositionalEmbedding", "784": "sam_hq/modeling_sam_hq.py:SamHQPreTrainedModel", "785": "sam_hq/modeling_sam_hq.py:SamHQPatchEmbeddings", "786": "sam_hq/modeling_sam_hq.py:SamHQVisionNeck", "787": "sam_hq/modeling_sam_hq.py:SamHQVisionEncoder", "788": "sam_hq/modeling_sam_hq.py:SamHQLayerNorm", "789": "sam_hq/modeling_sam_hq.py:eager_attention_forward", "790": "sam_hq/modeling_sam_hq.py:SamHQAttention", "791": "sam_hq/modeling_sam_hq.py:SamHQTwoWayAttentionBlock", "792": "sam_hq/modeling_sam_hq.py:SamHQTwoWayTransformer", "793": "sam_hq/modeling_sam_hq.py:SamHQFeedForward", "794": "sam_hq/modeling_sam_hq.py:SamHQMaskDecoder", "795": "sam_hq/modeling_sam_hq.py:SamHQVisionModel", "796": "sam_hq/modeling_sam_hq.py:SamHQMaskEmbedding", "797": "sam_hq/modeling_sam_hq.py:SamHQPromptEncoder", "798": "sam_hq/modeling_sam_hq.py:SamHQModel", "799": "glm46v/modeling_glm46v.py:Glm46VPreTrainedModel", "800": "glm46v/modeling_glm46v.py:Glm46VModelOutputWithPast", "801": "glm46v/modeling_glm46v.py:Glm46VModel", "802": "glm46v/modeling_glm46v.py:Glm46VCausalLMOutputWithPast", "803": "glm46v/modeling_glm46v.py:Glm46VForConditionalGeneration", "804": "t5/modeling_t5.py:T5LayerNorm", "805": "t5/modeling_t5.py:T5DenseActDense", "806": "t5/modeling_t5.py:T5DenseGatedActDense", "807": "t5/modeling_t5.py:T5LayerFF", "808": "t5/modeling_t5.py:T5Attention", "809": "t5/modeling_t5.py:T5LayerSelfAttention", "810": "t5/modeling_t5.py:T5LayerCrossAttention", "811": "t5/modeling_t5.py:T5Block", "812": "t5/modeling_t5.py:T5ClassificationHead", "813": "t5/modeling_t5.py:T5PreTrainedModel", "814": "t5/modeling_t5.py:T5Stack", "815": "t5/modeling_t5.py:T5Model", "816": "t5/modeling_t5.py:T5ForConditionalGeneration", "817": "t5/modeling_t5.py:T5EncoderModel", "818": "t5/modeling_t5.py:T5ForSequenceClassification", "819": "t5/modeling_t5.py:T5ForTokenClassification", "820": "t5/modeling_t5.py:T5ForQuestionAnswering", "821": "openai/modeling_openai.py:Attention", "822": "openai/modeling_openai.py:MLP", "823": "openai/modeling_openai.py:Block", "824": "openai/modeling_openai.py:OpenAIGPTSequenceSummary", "825": "openai/modeling_openai.py:OpenAIGPTPreTrainedModel", "826": "openai/modeling_openai.py:OpenAIGPTDoubleHeadsModelOutput", "827": "openai/modeling_openai.py:OpenAIGPTModel", "828": "openai/modeling_openai.py:OpenAIGPTLMHeadModel", "829": "openai/modeling_openai.py:OpenAIGPTDoubleHeadsModel", "830": "openai/modeling_openai.py:OpenAIGPTForSequenceClassification", "831": "falcon_h1/modeling_falcon_h1.py:FalconHybridMambaAttentionDynamicCache", "832": "falcon_h1/modeling_falcon_h1.py:FalconH1RotaryEmbedding", "833": "falcon_h1/modeling_falcon_h1.py:rotate_half", "834": "falcon_h1/modeling_falcon_h1.py:apply_rotary_pos_emb", "835": "falcon_h1/modeling_falcon_h1.py:repeat_kv", "836": "falcon_h1/modeling_falcon_h1.py:eager_attention_forward", "837": "falcon_h1/modeling_falcon_h1.py:FalconH1Attention", "838": "falcon_h1/modeling_falcon_h1.py:FalconH1RMSNormGated", "839": "falcon_h1/modeling_falcon_h1.py:pad_tensor_by_size", "840": "falcon_h1/modeling_falcon_h1.py:reshape_into_chunks", "841": "falcon_h1/modeling_falcon_h1.py:segment_sum", "842": "falcon_h1/modeling_falcon_h1.py:apply_mask_to_padding_states", "843": "falcon_h1/modeling_falcon_h1.py:FalconH1Mixer", "844": "falcon_h1/modeling_falcon_h1.py:FalconH1MLP", "845": "falcon_h1/modeling_falcon_h1.py:FalconH1RMSNorm", "846": "falcon_h1/modeling_falcon_h1.py:FalconH1DecoderLayer", "847": "falcon_h1/modeling_falcon_h1.py:compute_mup_vector", "848": "falcon_h1/modeling_falcon_h1.py:FalconH1PreTrainedModel", "849": "falcon_h1/modeling_falcon_h1.py:FalconH1Model", "850": "falcon_h1/modeling_falcon_h1.py:FalconH1ForCausalLM", "851": "video_llama_3/modeling_video_llama_3.py:VideoLlama3VisionRotaryEmbedding", "852": "video_llama_3/modeling_video_llama_3.py:VideoLlama3VisionEmbeddings", "853": "video_llama_3/modeling_video_llama_3.py:VideoLlama3VisionMLP", "854": "video_llama_3/modeling_video_llama_3.py:eager_attention_forward", "855": "video_llama_3/modeling_video_llama_3.py:rotate_half", "856": "video_llama_3/modeling_video_llama_3.py:repeat_kv", "857": "video_llama_3/modeling_video_llama_3.py:apply_rotary_pos_emb_vision", "858": "video_llama_3/modeling_video_llama_3.py:VideoLlama3VisionAttention", "859": "video_llama_3/modeling_video_llama_3.py:VideoLlama3VisionEncoderLayer", "860": "video_llama_3/modeling_video_llama_3.py:VideoLlama3VisionEncoder", "861": "video_llama_3/modeling_video_llama_3.py:VideoLlama3PreTrainedModel", "862": "video_llama_3/modeling_video_llama_3.py:VideoLlama3VisionModel", "863": "video_llama_3/modeling_video_llama_3.py:VideoLlama3Projector", "864": "video_llama_3/modeling_video_llama_3.py:VideoLlama3ModelOutputWithPast", "865": "video_llama_3/modeling_video_llama_3.py:VideoLlama3Model", "866": "video_llama_3/modeling_video_llama_3.py:VideoLlama3CausalLMOutputWithPast", "867": "video_llama_3/modeling_video_llama_3.py:VideoLlama3ForConditionalGeneration", "868": "minimax/modeling_minimax.py:MiniMaxRMSNorm", "869": "minimax/modeling_minimax.py:MiniMaxCache", "870": "minimax/modeling_minimax.py:MiniMaxLightningAttention", "871": "minimax/modeling_minimax.py:MiniMaxRotaryEmbedding", "872": "minimax/modeling_minimax.py:rotate_half", "873": "minimax/modeling_minimax.py:apply_rotary_pos_emb", "874": "minimax/modeling_minimax.py:repeat_kv", "875": "minimax/modeling_minimax.py:eager_attention_forward", "876": "minimax/modeling_minimax.py:MiniMaxAttention", "877": "minimax/modeling_minimax.py:MiniMaxTopKRouter", "878": "minimax/modeling_minimax.py:MiniMaxExperts", "879": "minimax/modeling_minimax.py:MiniMaxSparseMoeBlock", "880": "minimax/modeling_minimax.py:MiniMaxDecoderLayer", "881": "minimax/modeling_minimax.py:MiniMaxPreTrainedModel", "882": "minimax/modeling_minimax.py:MiniMaxModel", "883": "minimax/modeling_minimax.py:load_balancing_loss_func", "884": "minimax/modeling_minimax.py:MiniMaxForCausalLM", "885": "minimax/modeling_minimax.py:MiniMaxForSequenceClassification", "886": "minimax/modeling_minimax.py:MiniMaxForTokenClassification", "887": "minimax/modeling_minimax.py:MiniMaxForQuestionAnswering", "888": "starcoder2/modeling_starcoder2.py:Starcoder2MLP", "889": "starcoder2/modeling_starcoder2.py:rotate_half", "890": "starcoder2/modeling_starcoder2.py:apply_rotary_pos_emb", "891": "starcoder2/modeling_starcoder2.py:repeat_kv", "892": "starcoder2/modeling_starcoder2.py:eager_attention_forward", "893": "starcoder2/modeling_starcoder2.py:Starcoder2Attention", "894": "starcoder2/modeling_starcoder2.py:Starcoder2DecoderLayer", "895": "starcoder2/modeling_starcoder2.py:Starcoder2PreTrainedModel", "896": "starcoder2/modeling_starcoder2.py:Starcoder2RotaryEmbedding", "897": "starcoder2/modeling_starcoder2.py:Starcoder2Model", "898": "starcoder2/modeling_starcoder2.py:Starcoder2ForCausalLM", "899": "starcoder2/modeling_starcoder2.py:Starcoder2ForSequenceClassification", "900": "starcoder2/modeling_starcoder2.py:Starcoder2ForTokenClassification", "901": "videomae/modeling_videomae.py:VideoMAEDecoderOutput", "902": "videomae/modeling_videomae.py:VideoMAEForPreTrainingOutput", "903": "videomae/modeling_videomae.py:get_sinusoid_encoding_table", "904": "videomae/modeling_videomae.py:VideoMAEEmbeddings", "905": "videomae/modeling_videomae.py:VideoMAEPatchEmbeddings", "906": "videomae/modeling_videomae.py:eager_attention_forward", "907": "videomae/modeling_videomae.py:VideoMAESelfAttention", "908": "videomae/modeling_videomae.py:VideoMAESelfOutput", "909": "videomae/modeling_videomae.py:VideoMAEAttention", "910": "videomae/modeling_videomae.py:VideoMAEIntermediate", "911": "videomae/modeling_videomae.py:VideoMAEOutput", "912": "videomae/modeling_videomae.py:VideoMAELayer", "913": "videomae/modeling_videomae.py:VideoMAEEncoder", "914": "videomae/modeling_videomae.py:VideoMAEPreTrainedModel", "915": "videomae/modeling_videomae.py:VideoMAEModel", "916": "videomae/modeling_videomae.py:VideoMAEDecoder", "917": "videomae/modeling_videomae.py:VideoMAEForPreTraining", "918": "videomae/modeling_videomae.py:VideoMAEForVideoClassification", "919": "mobilevitv2/modeling_mobilevitv2.py:make_divisible", "920": "mobilevitv2/modeling_mobilevitv2.py:clip", "921": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2ConvLayer", "922": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2InvertedResidual", "923": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2MobileNetLayer", "924": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2LinearSelfAttention", "925": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2FFN", "926": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2TransformerLayer", "927": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2Transformer", "928": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2Layer", "929": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2Encoder", "930": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2PreTrainedModel", "931": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2Model", "932": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2ForImageClassification", "933": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2ASPPPooling", "934": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2ASPP", "935": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2DeepLabV3", "936": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2ForSemanticSegmentation", "937": "sew_d/modeling_sew_d.py:_compute_mask_indices", "938": "sew_d/modeling_sew_d.py:make_log_bucket_position", "939": "sew_d/modeling_sew_d.py:build_relative_position", "940": "sew_d/modeling_sew_d.py:c2p_dynamic_expand", "941": "sew_d/modeling_sew_d.py:p2c_dynamic_expand", "942": "sew_d/modeling_sew_d.py:pos_dynamic_expand", "943": "sew_d/modeling_sew_d.py:get_mask", "944": "sew_d/modeling_sew_d.py:SEWDNoLayerNormConvLayer", "945": "sew_d/modeling_sew_d.py:SEWDLayerNormConvLayer", "946": "sew_d/modeling_sew_d.py:SEWDGroupNormConvLayer", "947": "sew_d/modeling_sew_d.py:SEWDPositionalConvEmbedding", "948": "sew_d/modeling_sew_d.py:SEWDSamePadLayer", "949": "sew_d/modeling_sew_d.py:SEWDUpsampling", "950": "sew_d/modeling_sew_d.py:SEWDFeatureEncoder", "951": "sew_d/modeling_sew_d.py:ContextPooler", "952": "sew_d/modeling_sew_d.py:XSoftmax", "953": "sew_d/modeling_sew_d.py:DropoutContext", "954": "sew_d/modeling_sew_d.py:XDropout", "955": "sew_d/modeling_sew_d.py:StableDropout", "956": "sew_d/modeling_sew_d.py:SEWDSelfOutput", "957": "sew_d/modeling_sew_d.py:DisentangledSelfAttention", "958": "sew_d/modeling_sew_d.py:SEWDAttention", "959": "sew_d/modeling_sew_d.py:SEWDIntermediate", "960": "sew_d/modeling_sew_d.py:SEWDOutput", "961": "sew_d/modeling_sew_d.py:SEWDLayer", "962": "sew_d/modeling_sew_d.py:ConvLayer", "963": "sew_d/modeling_sew_d.py:SEWDTransformerEncoder", "964": "sew_d/modeling_sew_d.py:SEWDEncoder", "965": "sew_d/modeling_sew_d.py:SEWDPreTrainedModel", "966": "sew_d/modeling_sew_d.py:SEWDModel", "967": "sew_d/modeling_sew_d.py:SEWDForCTC", "968": "sew_d/modeling_sew_d.py:SEWDForSequenceClassification", "969": "gpt_neo/modeling_gpt_neo.py:GPTNeoSelfAttention", "970": "gpt_neo/modeling_gpt_neo.py:GPTNeoFlashAttention2", "971": "gpt_neo/modeling_gpt_neo.py:GPTNeoAttention", "972": "gpt_neo/modeling_gpt_neo.py:GPTNeoMLP", "973": "gpt_neo/modeling_gpt_neo.py:GPTNeoBlock", "974": "gpt_neo/modeling_gpt_neo.py:GPTNeoPreTrainedModel", "975": "gpt_neo/modeling_gpt_neo.py:GPTNeoModel", "976": "gpt_neo/modeling_gpt_neo.py:GPTNeoForCausalLM", "977": "gpt_neo/modeling_gpt_neo.py:GPTNeoForSequenceClassification", "978": "gpt_neo/modeling_gpt_neo.py:GPTNeoForTokenClassification", "979": "gpt_neo/modeling_gpt_neo.py:GPTNeoForQuestionAnswering", "980": "decision_transformer/modeling_decision_transformer.py:eager_attention_forward", "981": "decision_transformer/modeling_decision_transformer.py:DecisionTransformerGPT2Attention", "982": "decision_transformer/modeling_decision_transformer.py:DecisionTransformerGPT2MLP", "983": "decision_transformer/modeling_decision_transformer.py:DecisionTransformerGPT2Block", "984": "decision_transformer/modeling_decision_transformer.py:DecisionTransformerGPT2PreTrainedModel", "985": "decision_transformer/modeling_decision_transformer.py:DecisionTransformerGPT2Model", "986": "decision_transformer/modeling_decision_transformer.py:DecisionTransformerOutput", "987": "decision_transformer/modeling_decision_transformer.py:DecisionTransformerPreTrainedModel", "988": "decision_transformer/modeling_decision_transformer.py:DecisionTransformerModel", "989": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLModelOutputWithPast", "990": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLCausalLMOutputWithPast", "991": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLRotaryEmbedding", "992": "qwen2_vl/modeling_qwen2_vl.py:rotate_half", "993": "qwen2_vl/modeling_qwen2_vl.py:apply_multimodal_rotary_pos_emb", "994": "qwen2_vl/modeling_qwen2_vl.py:apply_rotary_pos_emb_vision", "995": "qwen2_vl/modeling_qwen2_vl.py:VisionRotaryEmbedding", "996": "qwen2_vl/modeling_qwen2_vl.py:PatchEmbed", "997": "qwen2_vl/modeling_qwen2_vl.py:PatchMerger", "998": "qwen2_vl/modeling_qwen2_vl.py:VisionMlp", "999": "qwen2_vl/modeling_qwen2_vl.py:repeat_kv", "1000": "qwen2_vl/modeling_qwen2_vl.py:eager_attention_forward", "1001": "qwen2_vl/modeling_qwen2_vl.py:VisionAttention", "1002": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLVisionBlock", "1003": "qwen2_vl/modeling_qwen2_vl.py:Qwen2MLP", "1004": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLAttention", "1005": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLDecoderLayer", "1006": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLPreTrainedModel", "1007": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VisionTransformerPretrainedModel", "1008": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLTextModel", "1009": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLModel", "1010": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLForConditionalGeneration", "1011": "xlnet/modeling_xlnet.py:XLNetRelativeAttention", "1012": "xlnet/modeling_xlnet.py:XLNetFeedForward", "1013": "xlnet/modeling_xlnet.py:XLNetLayer", "1014": "xlnet/modeling_xlnet.py:XLNetPoolerStartLogits", "1015": "xlnet/modeling_xlnet.py:XLNetPoolerEndLogits", "1016": "xlnet/modeling_xlnet.py:XLNetPoolerAnswerClass", "1017": "xlnet/modeling_xlnet.py:XLNetSequenceSummary", "1018": "xlnet/modeling_xlnet.py:XLNetPreTrainedModel", "1019": "xlnet/modeling_xlnet.py:XLNetModelOutput", "1020": "xlnet/modeling_xlnet.py:XLNetLMHeadModelOutput", "1021": "xlnet/modeling_xlnet.py:XLNetForSequenceClassificationOutput", "1022": "xlnet/modeling_xlnet.py:XLNetForTokenClassificationOutput", "1023": "xlnet/modeling_xlnet.py:XLNetForMultipleChoiceOutput", "1024": "xlnet/modeling_xlnet.py:XLNetForQuestionAnsweringSimpleOutput", "1025": "xlnet/modeling_xlnet.py:XLNetForQuestionAnsweringOutput", "1026": "xlnet/modeling_xlnet.py:XLNetModel", "1027": "xlnet/modeling_xlnet.py:XLNetLMHeadModel", "1028": "xlnet/modeling_xlnet.py:XLNetForSequenceClassification", "1029": "xlnet/modeling_xlnet.py:XLNetForTokenClassification", "1030": "xlnet/modeling_xlnet.py:XLNetForMultipleChoice", "1031": "xlnet/modeling_xlnet.py:XLNetForQuestionAnsweringSimple", "1032": "xlnet/modeling_xlnet.py:XLNetForQuestionAnswering", "1033": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackbonePatchEmbeddings", "1034": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneEmbeddings", "1035": "vitpose_backbone/modeling_vitpose_backbone.py:eager_attention_forward", "1036": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneSelfAttention", "1037": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneSelfOutput", "1038": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneAttention", "1039": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseNaiveMoe", "1040": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneMoeMLP", "1041": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneMLP", "1042": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneLayer", "1043": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneEncoder", "1044": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackbonePreTrainedModel", "1045": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackbone", "1046": "mpt/modeling_mpt.py:build_mpt_alibi_tensor", "1047": "mpt/modeling_mpt.py:MptAttention", "1048": "mpt/modeling_mpt.py:MptMLP", "1049": "mpt/modeling_mpt.py:MptBlock", "1050": "mpt/modeling_mpt.py:MptPreTrainedModel", "1051": "mpt/modeling_mpt.py:MptModel", "1052": "mpt/modeling_mpt.py:MptForCausalLM", "1053": "mpt/modeling_mpt.py:MptForSequenceClassification", "1054": "mpt/modeling_mpt.py:MptForTokenClassification", "1055": "mpt/modeling_mpt.py:MptForQuestionAnswering", "1056": "mamba/modeling_mamba.py:MambaCache", "1057": "mamba/modeling_mamba.py:MambaMixer", "1058": "mamba/modeling_mamba.py:MambaRMSNorm", "1059": "mamba/modeling_mamba.py:MambaBlock", "1060": "mamba/modeling_mamba.py:MambaPreTrainedModel", "1061": "mamba/modeling_mamba.py:MambaOutput", "1062": "mamba/modeling_mamba.py:MambaCausalLMOutput", "1063": "mamba/modeling_mamba.py:MambaModel", "1064": "mamba/modeling_mamba.py:MambaForCausalLM", "1065": "internvl/modeling_internvl.py:InternVLVisionRMSNorm", "1066": "internvl/modeling_internvl.py:eager_attention_forward", "1067": "internvl/modeling_internvl.py:InternVLVisionAttention", "1068": "internvl/modeling_internvl.py:InternVLVisionModelOutputWithPooling", "1069": "internvl/modeling_internvl.py:InternVLVisionPatchEmbeddings", "1070": "internvl/modeling_internvl.py:InternVLVisionEmbeddings", "1071": "internvl/modeling_internvl.py:InternVLVisionMLP", "1072": "internvl/modeling_internvl.py:InternVLVisionLayer", "1073": "internvl/modeling_internvl.py:InternVLVisionEncoder", "1074": "internvl/modeling_internvl.py:InternVLVisionPreTrainedModel", "1075": "internvl/modeling_internvl.py:InternVLVisionModel", "1076": "internvl/modeling_internvl.py:InternVLPreTrainedModel", "1077": "internvl/modeling_internvl.py:InternVLMultiModalProjector", "1078": "internvl/modeling_internvl.py:InternVLModelOutputWithPast", "1079": "internvl/modeling_internvl.py:InternVLModel", "1080": "internvl/modeling_internvl.py:InternVLCausalLMOutputWithPast", "1081": "internvl/modeling_internvl.py:InternVLForConditionalGeneration", "1082": "d_fine/modeling_d_fine.py:multi_scale_deformable_attention_v2", "1083": "d_fine/modeling_d_fine.py:DFineMultiscaleDeformableAttention", "1084": "d_fine/modeling_d_fine.py:DFineGate", "1085": "d_fine/modeling_d_fine.py:DFineMultiheadAttention", "1086": "d_fine/modeling_d_fine.py:DFineDecoderLayer", "1087": "d_fine/modeling_d_fine.py:DFinePreTrainedModel", "1088": "d_fine/modeling_d_fine.py:DFineIntegral", "1089": "d_fine/modeling_d_fine.py:DFineDecoderOutput", "1090": "d_fine/modeling_d_fine.py:inverse_sigmoid", "1091": "d_fine/modeling_d_fine.py:weighting_function", "1092": "d_fine/modeling_d_fine.py:distance2bbox", "1093": "d_fine/modeling_d_fine.py:DFineDecoder", "1094": "d_fine/modeling_d_fine.py:DFineFrozenBatchNorm2d", "1095": "d_fine/modeling_d_fine.py:DFineModelOutput", "1096": "d_fine/modeling_d_fine.py:replace_batch_norm", "1097": "d_fine/modeling_d_fine.py:DFineConvEncoder", "1098": "d_fine/modeling_d_fine.py:get_contrastive_denoising_training_group", "1099": "d_fine/modeling_d_fine.py:DFineModel", "1100": "d_fine/modeling_d_fine.py:DFineObjectDetectionOutput", "1101": "d_fine/modeling_d_fine.py:DFineForObjectDetection", "1102": "d_fine/modeling_d_fine.py:DFineMLPPredictionHead", "1103": "d_fine/modeling_d_fine.py:DFineMLP", "1104": "d_fine/modeling_d_fine.py:DFineLQE", "1105": "d_fine/modeling_d_fine.py:DFineConvNormLayer", "1106": "d_fine/modeling_d_fine.py:DFineRepVggBlock", "1107": "d_fine/modeling_d_fine.py:DFineCSPRepLayer", "1108": "d_fine/modeling_d_fine.py:DFineRepNCSPELAN4", "1109": "d_fine/modeling_d_fine.py:DFineSCDown", "1110": "d_fine/modeling_d_fine.py:DFineEncoderLayer", "1111": "d_fine/modeling_d_fine.py:DFineEncoder", "1112": "d_fine/modeling_d_fine.py:DFineHybridEncoder", "1113": "squeezebert/modeling_squeezebert.py:SqueezeBertEmbeddings", "1114": "squeezebert/modeling_squeezebert.py:MatMulWrapper", "1115": "squeezebert/modeling_squeezebert.py:SqueezeBertLayerNorm", "1116": "squeezebert/modeling_squeezebert.py:ConvDropoutLayerNorm", "1117": "squeezebert/modeling_squeezebert.py:ConvActivation", "1118": "squeezebert/modeling_squeezebert.py:SqueezeBertSelfAttention", "1119": "squeezebert/modeling_squeezebert.py:SqueezeBertModule", "1120": "squeezebert/modeling_squeezebert.py:SqueezeBertEncoder", "1121": "squeezebert/modeling_squeezebert.py:SqueezeBertPooler", "1122": "squeezebert/modeling_squeezebert.py:SqueezeBertPredictionHeadTransform", "1123": "squeezebert/modeling_squeezebert.py:SqueezeBertLMPredictionHead", "1124": "squeezebert/modeling_squeezebert.py:SqueezeBertOnlyMLMHead", "1125": "squeezebert/modeling_squeezebert.py:SqueezeBertPreTrainedModel", "1126": "squeezebert/modeling_squeezebert.py:SqueezeBertModel", "1127": "squeezebert/modeling_squeezebert.py:SqueezeBertForMaskedLM", "1128": "squeezebert/modeling_squeezebert.py:SqueezeBertForSequenceClassification", "1129": "squeezebert/modeling_squeezebert.py:SqueezeBertForMultipleChoice", "1130": "squeezebert/modeling_squeezebert.py:SqueezeBertForTokenClassification", "1131": "squeezebert/modeling_squeezebert.py:SqueezeBertForQuestionAnswering", "1132": "bit/modeling_bit.py:get_padding_value", "1133": "bit/modeling_bit.py:WeightStandardizedConv2d", "1134": "bit/modeling_bit.py:BitGroupNormActivation", "1135": "bit/modeling_bit.py:DynamicPad2d", "1136": "bit/modeling_bit.py:BitMaxPool2d", "1137": "bit/modeling_bit.py:BitEmbeddings", "1138": "bit/modeling_bit.py:drop_path", "1139": "bit/modeling_bit.py:BitDropPath", "1140": "bit/modeling_bit.py:make_div", "1141": "bit/modeling_bit.py:BitPreActivationBottleneckLayer", "1142": "bit/modeling_bit.py:BitBottleneckLayer", "1143": "bit/modeling_bit.py:BitDownsampleConv", "1144": "bit/modeling_bit.py:BitStage", "1145": "bit/modeling_bit.py:BitEncoder", "1146": "bit/modeling_bit.py:BitPreTrainedModel", "1147": "bit/modeling_bit.py:BitModel", "1148": "bit/modeling_bit.py:BitForImageClassification", "1149": "bit/modeling_bit.py:BitBackbone", "1150": "deberta/modeling_deberta.py:DebertaLayerNorm", "1151": "deberta/modeling_deberta.py:DebertaSelfOutput", "1152": "deberta/modeling_deberta.py:build_relative_position", "1153": "deberta/modeling_deberta.py:c2p_dynamic_expand", "1154": "deberta/modeling_deberta.py:p2c_dynamic_expand", "1155": "deberta/modeling_deberta.py:pos_dynamic_expand", "1156": "deberta/modeling_deberta.py:scaled_size_sqrt", "1157": "deberta/modeling_deberta.py:build_rpos", "1158": "deberta/modeling_deberta.py:compute_attention_span", "1159": "deberta/modeling_deberta.py:uneven_size_corrected", "1160": "deberta/modeling_deberta.py:DisentangledSelfAttention", "1161": "deberta/modeling_deberta.py:DebertaEmbeddings", "1162": "deberta/modeling_deberta.py:DebertaAttention", "1163": "deberta/modeling_deberta.py:DebertaIntermediate", "1164": "deberta/modeling_deberta.py:DebertaOutput", "1165": "deberta/modeling_deberta.py:DebertaLayer", "1166": "deberta/modeling_deberta.py:DebertaEncoder", "1167": "deberta/modeling_deberta.py:DebertaPreTrainedModel", "1168": "deberta/modeling_deberta.py:DebertaModel", "1169": "deberta/modeling_deberta.py:LegacyDebertaPredictionHeadTransform", "1170": "deberta/modeling_deberta.py:LegacyDebertaLMPredictionHead", "1171": "deberta/modeling_deberta.py:LegacyDebertaOnlyMLMHead", "1172": "deberta/modeling_deberta.py:DebertaLMPredictionHead", "1173": "deberta/modeling_deberta.py:DebertaOnlyMLMHead", "1174": "deberta/modeling_deberta.py:DebertaForMaskedLM", "1175": "deberta/modeling_deberta.py:ContextPooler", "1176": "deberta/modeling_deberta.py:DebertaForSequenceClassification", "1177": "deberta/modeling_deberta.py:DebertaForTokenClassification", "1178": "deberta/modeling_deberta.py:DebertaForQuestionAnswering", "1179": "bert_generation/modeling_bert_generation.py:BertGenerationSelfOutput", "1180": "bert_generation/modeling_bert_generation.py:eager_attention_forward", "1181": "bert_generation/modeling_bert_generation.py:BertGenerationSelfAttention", "1182": "bert_generation/modeling_bert_generation.py:BertGenerationCrossAttention", "1183": "bert_generation/modeling_bert_generation.py:BertGenerationAttention", "1184": "bert_generation/modeling_bert_generation.py:BertGenerationIntermediate", "1185": "bert_generation/modeling_bert_generation.py:BertGenerationOutput", "1186": "bert_generation/modeling_bert_generation.py:BertGenerationLayer", "1187": "bert_generation/modeling_bert_generation.py:BertEncoder", "1188": "bert_generation/modeling_bert_generation.py:BertGenerationEmbeddings", "1189": "bert_generation/modeling_bert_generation.py:BertGenerationPreTrainedModel", "1190": "bert_generation/modeling_bert_generation.py:BertGenerationEncoder", "1191": "bert_generation/modeling_bert_generation.py:BertGenerationOnlyLMHead", "1192": "bert_generation/modeling_bert_generation.py:BertGenerationDecoder", "1193": "llava_next/modeling_llava_next.py:get_anyres_image_grid_shape", "1194": "llava_next/modeling_llava_next.py:image_size_to_num_patches", "1195": "llava_next/modeling_llava_next.py:unpad_image", "1196": "llava_next/modeling_llava_next.py:LlavaNextModelOutputWithPast", "1197": "llava_next/modeling_llava_next.py:LlavaNextCausalLMOutputWithPast", "1198": "llava_next/modeling_llava_next.py:LlavaNextMultiModalProjector", "1199": "llava_next/modeling_llava_next.py:LlavaNextPreTrainedModel", "1200": "llava_next/modeling_llava_next.py:LlavaNextModel", "1201": "llava_next/modeling_llava_next.py:LlavaNextForConditionalGeneration", "1202": "vit_mae/modeling_vit_mae.py:ViTMAEModelOutput", "1203": "vit_mae/modeling_vit_mae.py:ViTMAEDecoderOutput", "1204": "vit_mae/modeling_vit_mae.py:ViTMAEForPreTrainingOutput", "1205": "vit_mae/modeling_vit_mae.py:get_2d_sincos_pos_embed", "1206": "vit_mae/modeling_vit_mae.py:get_2d_sincos_pos_embed_from_grid", "1207": "vit_mae/modeling_vit_mae.py:get_1d_sincos_pos_embed_from_grid", "1208": "vit_mae/modeling_vit_mae.py:ViTMAEEmbeddings", "1209": "vit_mae/modeling_vit_mae.py:ViTMAEPatchEmbeddings", "1210": "vit_mae/modeling_vit_mae.py:eager_attention_forward", "1211": "vit_mae/modeling_vit_mae.py:ViTMAESelfAttention", "1212": "vit_mae/modeling_vit_mae.py:ViTMAESelfOutput", "1213": "vit_mae/modeling_vit_mae.py:ViTMAEAttention", "1214": "vit_mae/modeling_vit_mae.py:ViTMAEIntermediate", "1215": "vit_mae/modeling_vit_mae.py:ViTMAEOutput", "1216": "vit_mae/modeling_vit_mae.py:ViTMAELayer", "1217": "vit_mae/modeling_vit_mae.py:ViTMAEEncoder", "1218": "vit_mae/modeling_vit_mae.py:ViTMAEPreTrainedModel", "1219": "vit_mae/modeling_vit_mae.py:ViTMAEModel", "1220": "vit_mae/modeling_vit_mae.py:ViTMAEDecoder", "1221": "vit_mae/modeling_vit_mae.py:ViTMAEForPreTraining", "1222": "gptj/modeling_gptj.py:create_sinusoidal_positions", "1223": "gptj/modeling_gptj.py:get_embed_positions", "1224": "gptj/modeling_gptj.py:rotate_every_two", "1225": "gptj/modeling_gptj.py:apply_rotary_pos_emb", "1226": "gptj/modeling_gptj.py:GPTJAttention", "1227": "gptj/modeling_gptj.py:GPTJFlashAttention2", "1228": "gptj/modeling_gptj.py:GPTJMLP", "1229": "gptj/modeling_gptj.py:GPTJBlock", "1230": "gptj/modeling_gptj.py:GPTJPreTrainedModel", "1231": "gptj/modeling_gptj.py:GPTJModel", "1232": "gptj/modeling_gptj.py:GPTJForCausalLM", "1233": "gptj/modeling_gptj.py:GPTJForSequenceClassification", "1234": "gptj/modeling_gptj.py:GPTJForQuestionAnswering", "1235": "bloom/modeling_bloom.py:build_alibi_tensor", "1236": "bloom/modeling_bloom.py:dropout_add", "1237": "bloom/modeling_bloom.py:bloom_gelu_forward", "1238": "bloom/modeling_bloom.py:bloom_gelu_back", "1239": "bloom/modeling_bloom.py:GeLUFunction", "1240": "bloom/modeling_bloom.py:BloomGelu", "1241": "bloom/modeling_bloom.py:BloomAttention", "1242": "bloom/modeling_bloom.py:BloomMLP", "1243": "bloom/modeling_bloom.py:BloomBlock", "1244": "bloom/modeling_bloom.py:BloomPreTrainedModel", "1245": "bloom/modeling_bloom.py:BloomModel", "1246": "bloom/modeling_bloom.py:BloomForCausalLM", "1247": "bloom/modeling_bloom.py:BloomForSequenceClassification", "1248": "bloom/modeling_bloom.py:BloomForTokenClassification", "1249": "bloom/modeling_bloom.py:BloomForQuestionAnswering", "1250": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeRMSNorm", "1251": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeMLP", "1252": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeRotaryEmbedding", "1253": "ernie4_5_moe/modeling_ernie4_5_moe.py:rotate_half", "1254": "ernie4_5_moe/modeling_ernie4_5_moe.py:apply_rotary_pos_emb", "1255": "ernie4_5_moe/modeling_ernie4_5_moe.py:repeat_kv", "1256": "ernie4_5_moe/modeling_ernie4_5_moe.py:eager_attention_forward", "1257": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeAttention", "1258": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeStatics", "1259": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeExperts", "1260": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeTopKRouter", "1261": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeSparseMoeBlock", "1262": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeDecoderLayer", "1263": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoePreTrainedModel", "1264": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeModel", "1265": "ernie4_5_moe/modeling_ernie4_5_moe.py:load_balancing_loss_func", "1266": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeForCausalLM", "1267": "xlstm/modeling_xlstm.py:small_init_method", "1268": "xlstm/modeling_xlstm.py:wang_init_method", "1269": "xlstm/modeling_xlstm.py:xLSTMPreTrainedModel", "1270": "xlstm/modeling_xlstm.py:xLSTMCache", "1271": "xlstm/modeling_xlstm.py:xLSTMOutput", "1272": "xlstm/modeling_xlstm.py:xLSTMModel", "1273": "xlstm/modeling_xlstm.py:xLSTMCausalLMOutput", "1274": "xlstm/modeling_xlstm.py:xLSTMForCausalLM", "1275": "beit/modeling_beit.py:BeitModelOutputWithPooling", "1276": "beit/modeling_beit.py:drop_path", "1277": "beit/modeling_beit.py:BeitDropPath", "1278": "beit/modeling_beit.py:BeitEmbeddings", "1279": "beit/modeling_beit.py:BeitPatchEmbeddings", "1280": "beit/modeling_beit.py:BeitSelfAttention", "1281": "beit/modeling_beit.py:BeitSdpaSelfAttention", "1282": "beit/modeling_beit.py:BeitSelfOutput", "1283": "beit/modeling_beit.py:BeitAttention", "1284": "beit/modeling_beit.py:BeitIntermediate", "1285": "beit/modeling_beit.py:BeitOutput", "1286": "beit/modeling_beit.py:BeitLayer", "1287": "beit/modeling_beit.py:BeitRelativePositionBias", "1288": "beit/modeling_beit.py:BeitEncoder", "1289": "beit/modeling_beit.py:BeitPreTrainedModel", "1290": "beit/modeling_beit.py:BeitModel", "1291": "beit/modeling_beit.py:BeitPooler", "1292": "beit/modeling_beit.py:BeitForMaskedImageModeling", "1293": "beit/modeling_beit.py:BeitForImageClassification", "1294": "beit/modeling_beit.py:BeitConvModule", "1295": "beit/modeling_beit.py:BeitPyramidPoolingBlock", "1296": "beit/modeling_beit.py:BeitPyramidPoolingModule", "1297": "beit/modeling_beit.py:BeitUperHead", "1298": "beit/modeling_beit.py:BeitFCNHead", "1299": "beit/modeling_beit.py:BeitForSemanticSegmentation", "1300": "beit/modeling_beit.py:BeitBackbone", "1301": "glmasr/modeling_glmasr.py:GlmAsrRotaryEmbedding", "1302": "glmasr/modeling_glmasr.py:rotate_half", "1303": "glmasr/modeling_glmasr.py:repeat_kv", "1304": "glmasr/modeling_glmasr.py:eager_attention_forward", "1305": "glmasr/modeling_glmasr.py:apply_rotary_pos_emb", "1306": "glmasr/modeling_glmasr.py:GlmAsrAttention", "1307": "glmasr/modeling_glmasr.py:GlmAsrMLP", "1308": "glmasr/modeling_glmasr.py:GlmAsrEncoderLayer", "1309": "glmasr/modeling_glmasr.py:GlmAsrPreTrainedModel", "1310": "glmasr/modeling_glmasr.py:GlmAsrEncoder", "1311": "glmasr/modeling_glmasr.py:GlmAsrMultiModalProjector", "1312": "glmasr/modeling_glmasr.py:GlmAsrForConditionalGeneration", "1313": "granite/modeling_granite.py:rotate_half", "1314": "granite/modeling_granite.py:apply_rotary_pos_emb", "1315": "granite/modeling_granite.py:repeat_kv", "1316": "granite/modeling_granite.py:eager_attention_forward", "1317": "granite/modeling_granite.py:GraniteAttention", "1318": "granite/modeling_granite.py:GraniteRMSNorm", "1319": "granite/modeling_granite.py:GraniteMLP", "1320": "granite/modeling_granite.py:GraniteDecoderLayer", "1321": "granite/modeling_granite.py:GranitePreTrainedModel", "1322": "granite/modeling_granite.py:GraniteRotaryEmbedding", "1323": "granite/modeling_granite.py:GraniteModel", "1324": "granite/modeling_granite.py:GraniteForCausalLM", "1325": "olmoe/modeling_olmoe.py:OlmoeRMSNorm", "1326": "olmoe/modeling_olmoe.py:OlmoeRotaryEmbedding", "1327": "olmoe/modeling_olmoe.py:OlmoeMLP", "1328": "olmoe/modeling_olmoe.py:rotate_half", "1329": "olmoe/modeling_olmoe.py:apply_rotary_pos_emb", "1330": "olmoe/modeling_olmoe.py:repeat_kv", "1331": "olmoe/modeling_olmoe.py:eager_attention_forward", "1332": "olmoe/modeling_olmoe.py:OlmoeAttention", "1333": "olmoe/modeling_olmoe.py:OlmoeExperts", "1334": "olmoe/modeling_olmoe.py:OlmoeTopKRouter", "1335": "olmoe/modeling_olmoe.py:OlmoeSparseMoeBlock", "1336": "olmoe/modeling_olmoe.py:OlmoeDecoderLayer", "1337": "olmoe/modeling_olmoe.py:OlmoePreTrainedModel", "1338": "olmoe/modeling_olmoe.py:OlmoeModel", "1339": "olmoe/modeling_olmoe.py:load_balancing_loss_func", "1340": "olmoe/modeling_olmoe.py:OlmoeForCausalLM", "1341": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerGatedAttention", "1342": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerBatchNorm", "1343": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerPositionalEncoding", "1344": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerNormLayer", "1345": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerMLP", "1346": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerChannelFeatureMixerBlock", "1347": "patchtsmixer/modeling_patchtsmixer.py:eager_attention_forward", "1348": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerAttention", "1349": "patchtsmixer/modeling_patchtsmixer.py:PatchMixerBlock", "1350": "patchtsmixer/modeling_patchtsmixer.py:FeatureMixerBlock", "1351": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerLayer", "1352": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerBlock", "1353": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForPredictionHead", "1354": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerLinearHead", "1355": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerPreTrainedModel", "1356": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerPretrainHead", "1357": "patchtsmixer/modeling_patchtsmixer.py:random_masking", "1358": "patchtsmixer/modeling_patchtsmixer.py:forecast_masking", "1359": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerPatchify", "1360": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerMasking", "1361": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerStdScaler", "1362": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerMeanScaler", "1363": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerNOPScaler", "1364": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerEncoderOutput", "1365": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerEncoder", "1366": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerModelOutput", "1367": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerModel", "1368": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForPreTrainingOutput", "1369": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForPretraining", "1370": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForPredictionOutput", "1371": "patchtsmixer/modeling_patchtsmixer.py:SamplePatchTSMixerPredictionOutput", "1372": "patchtsmixer/modeling_patchtsmixer.py:SamplePatchTSMixerRegressionOutput", "1373": "patchtsmixer/modeling_patchtsmixer.py:nll", "1374": "patchtsmixer/modeling_patchtsmixer.py:weighted_average", "1375": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForPrediction", "1376": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForTimeSeriesClassificationOutput", "1377": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForTimeSeriesClassification", "1378": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForRegressionOutput", "1379": "patchtsmixer/modeling_patchtsmixer.py:InjectScalerStatistics4D", "1380": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForRegression", "1381": "vilt/modeling_vilt.py:ViltForImagesAndTextClassificationOutput", "1382": "vilt/modeling_vilt.py:ViltEmbeddings", "1383": "vilt/modeling_vilt.py:TextEmbeddings", "1384": "vilt/modeling_vilt.py:ViltPatchEmbeddings", "1385": "vilt/modeling_vilt.py:ViltSelfAttention", "1386": "vilt/modeling_vilt.py:ViltSelfOutput", "1387": "vilt/modeling_vilt.py:ViltAttention", "1388": "vilt/modeling_vilt.py:ViltIntermediate", "1389": "vilt/modeling_vilt.py:ViltOutput", "1390": "vilt/modeling_vilt.py:ViltLayer", "1391": "vilt/modeling_vilt.py:ViltEncoder", "1392": "vilt/modeling_vilt.py:ViltPreTrainedModel", "1393": "vilt/modeling_vilt.py:ViltModel", "1394": "vilt/modeling_vilt.py:ViltPooler", "1395": "vilt/modeling_vilt.py:ViltForMaskedLM", "1396": "vilt/modeling_vilt.py:ViltPredictionHeadTransform", "1397": "vilt/modeling_vilt.py:ViltMLMHead", "1398": "vilt/modeling_vilt.py:ViltForQuestionAnswering", "1399": "vilt/modeling_vilt.py:ViltForImageAndTextRetrieval", "1400": "vilt/modeling_vilt.py:ViltForImagesAndTextClassification", "1401": "vilt/modeling_vilt.py:ViltForTokenClassification", "1402": "mobilenet_v1/modeling_mobilenet_v1.py:apply_tf_padding", "1403": "mobilenet_v1/modeling_mobilenet_v1.py:MobileNetV1ConvLayer", "1404": "mobilenet_v1/modeling_mobilenet_v1.py:MobileNetV1PreTrainedModel", "1405": "mobilenet_v1/modeling_mobilenet_v1.py:MobileNetV1Model", "1406": "mobilenet_v1/modeling_mobilenet_v1.py:MobileNetV1ForImageClassification", "1407": "rembert/modeling_rembert.py:RemBertEmbeddings", "1408": "rembert/modeling_rembert.py:RemBertPooler", "1409": "rembert/modeling_rembert.py:RemBertSelfAttention", "1410": "rembert/modeling_rembert.py:RemBertSelfOutput", "1411": "rembert/modeling_rembert.py:RemBertAttention", "1412": "rembert/modeling_rembert.py:RemBertIntermediate", "1413": "rembert/modeling_rembert.py:RemBertOutput", "1414": "rembert/modeling_rembert.py:RemBertLayer", "1415": "rembert/modeling_rembert.py:RemBertEncoder", "1416": "rembert/modeling_rembert.py:RemBertPredictionHeadTransform", "1417": "rembert/modeling_rembert.py:RemBertLMPredictionHead", "1418": "rembert/modeling_rembert.py:RemBertOnlyMLMHead", "1419": "rembert/modeling_rembert.py:RemBertPreTrainedModel", "1420": "rembert/modeling_rembert.py:RemBertModel", "1421": "rembert/modeling_rembert.py:RemBertForMaskedLM", "1422": "rembert/modeling_rembert.py:RemBertForCausalLM", "1423": "rembert/modeling_rembert.py:RemBertForSequenceClassification", "1424": "rembert/modeling_rembert.py:RemBertForMultipleChoice", "1425": "rembert/modeling_rembert.py:RemBertForTokenClassification", "1426": "rembert/modeling_rembert.py:RemBertForQuestionAnswering", "1427": "t5gemma2/modeling_t5gemma2.py:T5Gemma2RMSNorm", "1428": "t5gemma2/modeling_t5gemma2.py:T5Gemma2MLP", "1429": "t5gemma2/modeling_t5gemma2.py:T5Gemma2RotaryEmbedding", "1430": "t5gemma2/modeling_t5gemma2.py:rotate_half", "1431": "t5gemma2/modeling_t5gemma2.py:apply_rotary_pos_emb", "1432": "t5gemma2/modeling_t5gemma2.py:repeat_kv", "1433": "t5gemma2/modeling_t5gemma2.py:eager_attention_forward", "1434": "t5gemma2/modeling_t5gemma2.py:T5Gemma2SelfAttention", "1435": "t5gemma2/modeling_t5gemma2.py:T5Gemma2MergedAttention", "1436": "t5gemma2/modeling_t5gemma2.py:T5Gemma2EncoderLayer", "1437": "t5gemma2/modeling_t5gemma2.py:T5Gemma2DecoderLayer", "1438": "t5gemma2/modeling_t5gemma2.py:T5Gemma2LMHead", "1439": "t5gemma2/modeling_t5gemma2.py:T5Gemma2ClassificationHead", "1440": "t5gemma2/modeling_t5gemma2.py:T5Gemma2MultiModalProjector", "1441": "t5gemma2/modeling_t5gemma2.py:T5Gemma2TextScaledWordEmbedding", "1442": "t5gemma2/modeling_t5gemma2.py:T5Gemma2PreTrainedModel", "1443": "t5gemma2/modeling_t5gemma2.py:sliding_window_mask_function", "1444": "t5gemma2/modeling_t5gemma2.py:T5Gemma2Encoder", "1445": "t5gemma2/modeling_t5gemma2.py:bidirectional_mask_function", "1446": "t5gemma2/modeling_t5gemma2.py:T5Gemma2Decoder", "1447": "t5gemma2/modeling_t5gemma2.py:T5Gemma2Model", "1448": "t5gemma2/modeling_t5gemma2.py:T5Gemma2ForConditionalGeneration", "1449": "t5gemma2/modeling_t5gemma2.py:T5Gemma2ForSequenceClassification", "1450": "t5gemma2/modeling_t5gemma2.py:T5Gemma2ForTokenClassification", "1451": "imagegpt/modeling_imagegpt.py:ImageGPTLayerNorm", "1452": "imagegpt/modeling_imagegpt.py:ImageGPTAttention", "1453": "imagegpt/modeling_imagegpt.py:ImageGPTMLP", "1454": "imagegpt/modeling_imagegpt.py:ImageGPTBlock", "1455": "imagegpt/modeling_imagegpt.py:ImageGPTPreTrainedModel", "1456": "imagegpt/modeling_imagegpt.py:ImageGPTModel", "1457": "imagegpt/modeling_imagegpt.py:ImageGPTForCausalImageModeling", "1458": "imagegpt/modeling_imagegpt.py:ImageGPTForImageClassification", "1459": "timesformer/modeling_timesformer.py:TimesformerPatchEmbeddings", "1460": "timesformer/modeling_timesformer.py:TimesformerEmbeddings", "1461": "timesformer/modeling_timesformer.py:drop_path", "1462": "timesformer/modeling_timesformer.py:TimeSformerDropPath", "1463": "timesformer/modeling_timesformer.py:TimesformerSelfAttention", "1464": "timesformer/modeling_timesformer.py:TimesformerSelfOutput", "1465": "timesformer/modeling_timesformer.py:TimeSformerAttention", "1466": "timesformer/modeling_timesformer.py:TimesformerIntermediate", "1467": "timesformer/modeling_timesformer.py:TimesformerOutput", "1468": "timesformer/modeling_timesformer.py:TimesformerLayer", "1469": "timesformer/modeling_timesformer.py:TimesformerEncoder", "1470": "timesformer/modeling_timesformer.py:TimesformerPreTrainedModel", "1471": "timesformer/modeling_timesformer.py:TimesformerModel", "1472": "timesformer/modeling_timesformer.py:TimesformerForVideoClassification", "1473": "altclip/modeling_altclip.py:contrastive_loss", "1474": "altclip/modeling_altclip.py:clip_loss", "1475": "altclip/modeling_altclip.py:AltCLIPOutput", "1476": "altclip/modeling_altclip.py:AltRobertaEmbeddings", "1477": "altclip/modeling_altclip.py:AltRobertaSelfAttention", "1478": "altclip/modeling_altclip.py:AltRobertaSelfOutput", "1479": "altclip/modeling_altclip.py:AltRobertaAttention", "1480": "altclip/modeling_altclip.py:AltRobertaIntermediate", "1481": "altclip/modeling_altclip.py:AltRobertaOutput", "1482": "altclip/modeling_altclip.py:AltRobertaLayer", "1483": "altclip/modeling_altclip.py:AltRobertaEncoder", "1484": "altclip/modeling_altclip.py:AltRobertaPooler", "1485": "altclip/modeling_altclip.py:eager_attention_forward", "1486": "altclip/modeling_altclip.py:AltCLIPAttention", "1487": "altclip/modeling_altclip.py:AltCLIPMLP", "1488": "altclip/modeling_altclip.py:AltCLIPEncoderLayer", "1489": "altclip/modeling_altclip.py:AltCLIPEncoder", "1490": "altclip/modeling_altclip.py:AltCLIPVisionEmbeddings", "1491": "altclip/modeling_altclip.py:AltCLIPPreTrainedModel", "1492": "altclip/modeling_altclip.py:AltCLIPVisionTransformer", "1493": "altclip/modeling_altclip.py:AltCLIPVisionModel", "1494": "altclip/modeling_altclip.py:AltRobertaModel", "1495": "altclip/modeling_altclip.py:AltCLIPTextModel", "1496": "altclip/modeling_altclip.py:AltCLIPModel", "1497": "udop/modeling_udop.py:BaseModelOutputWithAttentionMask", "1498": "udop/modeling_udop.py:get_visual_bbox", "1499": "udop/modeling_udop.py:pad_sequence", "1500": "udop/modeling_udop.py:combine_image_text_embeddings", "1501": "udop/modeling_udop.py:UdopPatchEmbeddings", "1502": "udop/modeling_udop.py:UdopPreTrainedModel", "1503": "udop/modeling_udop.py:UdopLayerNorm", "1504": "udop/modeling_udop.py:UdopDenseActDense", "1505": "udop/modeling_udop.py:UdopDenseGatedActDense", "1506": "udop/modeling_udop.py:UdopLayerFF", "1507": "udop/modeling_udop.py:UdopAttention", "1508": "udop/modeling_udop.py:UdopLayerSelfAttention", "1509": "udop/modeling_udop.py:UdopLayerCrossAttention", "1510": "udop/modeling_udop.py:UdopBlock", "1511": "udop/modeling_udop.py:UdopCellEmbeddings", "1512": "udop/modeling_udop.py:RelativePositionBiasBase", "1513": "udop/modeling_udop.py:RelativePositionBias1D", "1514": "udop/modeling_udop.py:RelativePositionBiasHorizontal", "1515": "udop/modeling_udop.py:RelativePositionBiasVertical", "1516": "udop/modeling_udop.py:RelativePositionBiasAggregated", "1517": "udop/modeling_udop.py:create_relative_bias", "1518": "udop/modeling_udop.py:UdopStack", "1519": "udop/modeling_udop.py:UdopModel", "1520": "udop/modeling_udop.py:UdopForConditionalGeneration", "1521": "udop/modeling_udop.py:UdopEncoderModel", "1522": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesFeatureEmbedder", "1523": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesStdScaler", "1524": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesMeanScaler", "1525": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesNOPScaler", "1526": "time_series_transformer/modeling_time_series_transformer.py:nll", "1527": "time_series_transformer/modeling_time_series_transformer.py:weighted_average", "1528": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesSinusoidalPositionalEmbedding", "1529": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesValueEmbedding", "1530": "time_series_transformer/modeling_time_series_transformer.py:eager_attention_forward", "1531": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesTransformerAttention", "1532": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesTransformerEncoderLayer", "1533": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesTransformerDecoderLayer", "1534": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesTransformerPreTrainedModel", "1535": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesTransformerEncoder", "1536": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesTransformerDecoder", "1537": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesTransformerModel", "1538": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesTransformerForPrediction", "1539": "pe_audio_video/modeling_pe_audio_video.py:PeAudioVideoMaskedGroupNorm", "1540": "pe_audio_video/modeling_pe_audio_video.py:PeAudioVideoConvBlock1d", "1541": "pe_audio_video/modeling_pe_audio_video.py:PeAudioVideoResnetBlock1d", "1542": "pe_audio_video/modeling_pe_audio_video.py:PeAudioVideoEncoderPatchEmbedder", "1543": "pe_audio_video/modeling_pe_audio_video.py:PeAudioVideoContrastiveHead", "1544": "pe_audio_video/modeling_pe_audio_video.py:PeAudioVideoEncoderEmbedder", "1545": "pe_audio_video/modeling_pe_audio_video.py:repeat_kv", "1546": "pe_audio_video/modeling_pe_audio_video.py:eager_attention_forward", "1547": "pe_audio_video/modeling_pe_audio_video.py:stack_freqs", "1548": "pe_audio_video/modeling_pe_audio_video.py:apply_rotary_pos_emb", "1549": "pe_audio_video/modeling_pe_audio_video.py:PeAudioVideoEncoderAttention", "1550": "pe_audio_video/modeling_pe_audio_video.py:PeAudioVideoEncoderMLP", "1551": "pe_audio_video/modeling_pe_audio_video.py:PeAudioVideoEncoderLayer", "1552": "pe_audio_video/modeling_pe_audio_video.py:PeAudioVideoEncoderRMSNorm", "1553": "pe_audio_video/modeling_pe_audio_video.py:PeAudioVideoEncoderRotaryEmbedding", "1554": "pe_audio_video/modeling_pe_audio_video.py:PeAudioVideoPreTrainedModel", "1555": "pe_audio_video/modeling_pe_audio_video.py:PeAudioVideoEncoderOutput", "1556": "pe_audio_video/modeling_pe_audio_video.py:PeAudioVideoEncoder", "1557": "pe_audio_video/modeling_pe_audio_video.py:PeAudioVideoOutput", "1558": "pe_audio_video/modeling_pe_audio_video.py:AudioVideoEmbeddings", "1559": "pe_audio_video/modeling_pe_audio_video.py:PeAudioVideoModel", "1560": "gemma2/modeling_gemma2.py:Gemma2RMSNorm", "1561": "gemma2/modeling_gemma2.py:Gemma2MLP", "1562": "gemma2/modeling_gemma2.py:Gemma2RotaryEmbedding", "1563": "gemma2/modeling_gemma2.py:rotate_half", "1564": "gemma2/modeling_gemma2.py:apply_rotary_pos_emb", "1565": "gemma2/modeling_gemma2.py:repeat_kv", "1566": "gemma2/modeling_gemma2.py:eager_attention_forward", "1567": "gemma2/modeling_gemma2.py:Gemma2Attention", "1568": "gemma2/modeling_gemma2.py:Gemma2DecoderLayer", "1569": "gemma2/modeling_gemma2.py:Gemma2PreTrainedModel", "1570": "gemma2/modeling_gemma2.py:Gemma2Model", "1571": "gemma2/modeling_gemma2.py:Gemma2ForCausalLM", "1572": "gemma2/modeling_gemma2.py:Gemma2ForSequenceClassification", "1573": "gemma2/modeling_gemma2.py:Gemma2ForTokenClassification", "1574": "qwen2_audio/modeling_qwen2_audio.py:Qwen2AudioCausalLMOutputWithPast", "1575": "qwen2_audio/modeling_qwen2_audio.py:eager_attention_forward", "1576": "qwen2_audio/modeling_qwen2_audio.py:Qwen2AudioAttention", "1577": "qwen2_audio/modeling_qwen2_audio.py:Qwen2AudioEncoderLayer", "1578": "qwen2_audio/modeling_qwen2_audio.py:Qwen2AudioPreTrainedModel", "1579": "qwen2_audio/modeling_qwen2_audio.py:Qwen2AudioEncoder", "1580": "qwen2_audio/modeling_qwen2_audio.py:Qwen2AudioMultiModalProjector", "1581": "qwen2_audio/modeling_qwen2_audio.py:Qwen2AudioForConditionalGeneration", "1582": "seed_oss/modeling_seed_oss.py:SeedOssRMSNorm", "1583": "seed_oss/modeling_seed_oss.py:SeedOssMLP", "1584": "seed_oss/modeling_seed_oss.py:rotate_half", "1585": "seed_oss/modeling_seed_oss.py:apply_rotary_pos_emb", "1586": "seed_oss/modeling_seed_oss.py:repeat_kv", "1587": "seed_oss/modeling_seed_oss.py:eager_attention_forward", "1588": "seed_oss/modeling_seed_oss.py:SeedOssAttention", "1589": "seed_oss/modeling_seed_oss.py:SeedOssDecoderLayer", "1590": "seed_oss/modeling_seed_oss.py:SeedOssPreTrainedModel", "1591": "seed_oss/modeling_seed_oss.py:SeedOssRotaryEmbedding", "1592": "seed_oss/modeling_seed_oss.py:SeedOssModel", "1593": "seed_oss/modeling_seed_oss.py:SeedOssForCausalLM", "1594": "seed_oss/modeling_seed_oss.py:SeedOssForSequenceClassification", "1595": "seed_oss/modeling_seed_oss.py:SeedOssForTokenClassification", "1596": "seed_oss/modeling_seed_oss.py:SeedOssForQuestionAnswering", "1597": "resnet/modeling_resnet.py:ResNetConvLayer", "1598": "resnet/modeling_resnet.py:ResNetEmbeddings", "1599": "resnet/modeling_resnet.py:ResNetShortCut", "1600": "resnet/modeling_resnet.py:ResNetBasicLayer", "1601": "resnet/modeling_resnet.py:ResNetBottleNeckLayer", "1602": "resnet/modeling_resnet.py:ResNetStage", "1603": "resnet/modeling_resnet.py:ResNetEncoder", "1604": "resnet/modeling_resnet.py:ResNetPreTrainedModel", "1605": "resnet/modeling_resnet.py:ResNetModel", "1606": "resnet/modeling_resnet.py:ResNetForImageClassification", "1607": "resnet/modeling_resnet.py:ResNetBackbone", "1608": "colqwen2/modeling_colqwen2.py:ColQwen2PreTrainedModel", "1609": "colqwen2/modeling_colqwen2.py:ColQwen2ForRetrievalOutput", "1610": "colqwen2/modeling_colqwen2.py:ColQwen2ForRetrieval", "1611": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoInferenceCache", "1612": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoInferenceSession", "1613": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoLayerNorm", "1614": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoPositionEmbeddingSine", "1615": "sam3_tracker_video/modeling_sam3_tracker_video.py:eager_attention_forward", "1616": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoAttention", "1617": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoTwoWayAttentionBlock", "1618": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoFeedForward", "1619": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoImageSegmentationOutput", "1620": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoSegmentationOutput", "1621": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoPreTrainedModel", "1622": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoVisionRotaryEmbedding", "1623": "sam3_tracker_video/modeling_sam3_tracker_video.py:rotate_pairwise", "1624": "sam3_tracker_video/modeling_sam3_tracker_video.py:apply_rotary_pos_emb_2d", "1625": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoRoPEAttention", "1626": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoMemoryAttentionLayer", "1627": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoMemoryAttention", "1628": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoMemoryFuserCXBlock", "1629": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoMemoryFuser", "1630": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoMaskDownSamplerLayer", "1631": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoMaskDownSampler", "1632": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoMemoryEncoder", "1633": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoVisionEncoderOutput", "1634": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoPositionalEmbedding", "1635": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoMaskEmbedding", "1636": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoPromptEncoder", "1637": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoTwoWayTransformer", "1638": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoMaskDecoder", "1639": "sam3_tracker_video/modeling_sam3_tracker_video.py:get_1d_sine_pe", "1640": "sam3_tracker_video/modeling_sam3_tracker_video.py:Sam3TrackerVideoModel", "1641": "donut/modeling_donut_swin.py:DonutSwinEncoderOutput", "1642": "donut/modeling_donut_swin.py:DonutSwinModelOutput", "1643": "donut/modeling_donut_swin.py:DonutSwinImageClassifierOutput", "1644": "donut/modeling_donut_swin.py:window_partition", "1645": "donut/modeling_donut_swin.py:window_reverse", "1646": "donut/modeling_donut_swin.py:DonutSwinEmbeddings", "1647": "donut/modeling_donut_swin.py:DonutSwinPatchEmbeddings", "1648": "donut/modeling_donut_swin.py:DonutSwinPatchMerging", "1649": "donut/modeling_donut_swin.py:drop_path", "1650": "donut/modeling_donut_swin.py:DonutSwinDropPath", "1651": "donut/modeling_donut_swin.py:DonutSwinSelfAttention", "1652": "donut/modeling_donut_swin.py:DonutSwinSelfOutput", "1653": "donut/modeling_donut_swin.py:DonutSwinAttention", "1654": "donut/modeling_donut_swin.py:DonutSwinIntermediate", "1655": "donut/modeling_donut_swin.py:DonutSwinOutput", "1656": "donut/modeling_donut_swin.py:DonutSwinLayer", "1657": "donut/modeling_donut_swin.py:DonutSwinStage", "1658": "donut/modeling_donut_swin.py:DonutSwinEncoder", "1659": "donut/modeling_donut_swin.py:DonutSwinPreTrainedModel", "1660": "donut/modeling_donut_swin.py:DonutSwinModel", "1661": "donut/modeling_donut_swin.py:DonutSwinForImageClassification", "1662": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoVisionEmbeddings", "1663": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerEmbeddings", "1664": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoPreTrainedModel", "1665": "instructblipvideo/modeling_instructblipvideo.py:eager_attention_forward", "1666": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoAttention", "1667": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoMLP", "1668": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoEncoderLayer", "1669": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoEncoder", "1670": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoVisionModel", "1671": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerMultiHeadAttention", "1672": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerSelfOutput", "1673": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerAttention", "1674": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerIntermediate", "1675": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerOutput", "1676": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerLayer", "1677": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerEncoder", "1678": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerModel", "1679": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoForConditionalGenerationModelOutput", "1680": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoModel", "1681": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoForConditionalGeneration", "1682": "cwm/modeling_cwm.py:CwmRotaryEmbedding", "1683": "cwm/modeling_cwm.py:rotate_half", "1684": "cwm/modeling_cwm.py:apply_rotary_pos_emb", "1685": "cwm/modeling_cwm.py:repeat_kv", "1686": "cwm/modeling_cwm.py:eager_attention_forward", "1687": "cwm/modeling_cwm.py:CwmAttention", "1688": "cwm/modeling_cwm.py:CwmRMSNorm", "1689": "cwm/modeling_cwm.py:CwmMLP", "1690": "cwm/modeling_cwm.py:CwmDecoderLayer", "1691": "cwm/modeling_cwm.py:CwmPreTrainedModel", "1692": "cwm/modeling_cwm.py:CwmModelOutputWithPast", "1693": "cwm/modeling_cwm.py:CwmModel", "1694": "cwm/modeling_cwm.py:CwmForCausalLM", "1695": "llava_onevision/modeling_llava_onevision.py:LlavaOnevisionModelOutputWithPast", "1696": "llava_onevision/modeling_llava_onevision.py:LlavaOnevisionCausalLMOutputWithPast", "1697": "llava_onevision/modeling_llava_onevision.py:LlavaOnevisionPreTrainedModel", "1698": "llava_onevision/modeling_llava_onevision.py:LlavaOnevisionMultiModalProjector", "1699": "llava_onevision/modeling_llava_onevision.py:get_anyres_image_grid_shape", "1700": "llava_onevision/modeling_llava_onevision.py:image_size_to_num_patches", "1701": "llava_onevision/modeling_llava_onevision.py:unpad_image", "1702": "llava_onevision/modeling_llava_onevision.py:LlavaOnevisionModel", "1703": "llava_onevision/modeling_llava_onevision.py:LlavaOnevisionForConditionalGeneration", "1704": "mixtral/modeling_mixtral.py:MixtralExperts", "1705": "mixtral/modeling_mixtral.py:MixtralTopKRouter", "1706": "mixtral/modeling_mixtral.py:MixtralSparseMoeBlock", "1707": "mixtral/modeling_mixtral.py:MixtralRMSNorm", "1708": "mixtral/modeling_mixtral.py:MixtralRotaryEmbedding", "1709": "mixtral/modeling_mixtral.py:rotate_half", "1710": "mixtral/modeling_mixtral.py:apply_rotary_pos_emb", "1711": "mixtral/modeling_mixtral.py:repeat_kv", "1712": "mixtral/modeling_mixtral.py:eager_attention_forward", "1713": "mixtral/modeling_mixtral.py:MixtralAttention", "1714": "mixtral/modeling_mixtral.py:MixtralDecoderLayer", "1715": "mixtral/modeling_mixtral.py:MixtralPreTrainedModel", "1716": "mixtral/modeling_mixtral.py:MixtralModel", "1717": "mixtral/modeling_mixtral.py:load_balancing_loss_func", "1718": "mixtral/modeling_mixtral.py:MixtralForCausalLM", "1719": "mixtral/modeling_mixtral.py:MixtralForSequenceClassification", "1720": "mixtral/modeling_mixtral.py:MixtralForTokenClassification", "1721": "mixtral/modeling_mixtral.py:MixtralForQuestionAnswering", "1722": "video_llava/modeling_video_llava.py:VideoLlavaModelOutputWithPast", "1723": "video_llava/modeling_video_llava.py:VideoLlavaCausalLMOutputWithPast", "1724": "video_llava/modeling_video_llava.py:VideoLlavaMultiModalProjector", "1725": "video_llava/modeling_video_llava.py:VideoLlavaPreTrainedModel", "1726": "video_llava/modeling_video_llava.py:VideoLlavaModel", "1727": "video_llava/modeling_video_llava.py:VideoLlavaForConditionalGeneration", "1728": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2PreTrainedModel", "1729": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2LearnableAffineBlock", "1730": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2ConvLayer", "1731": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2ConvLayerLight", "1732": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2Embeddings", "1733": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2BasicLayer", "1734": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2Stage", "1735": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2Encoder", "1736": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2Backbone", "1737": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2ForImageClassification", "1738": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersTop1Router", "1739": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersLayerNorm", "1740": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersDenseActDense", "1741": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersExperts", "1742": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersSparseMLP", "1743": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersLayerFF", "1744": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersAttention", "1745": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersLayerSelfAttention", "1746": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersLayerCrossAttention", "1747": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersBlock", "1748": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersPreTrainedModel", "1749": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersStack", "1750": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersModel", "1751": "switch_transformers/modeling_switch_transformers.py:router_z_loss_func", "1752": "switch_transformers/modeling_switch_transformers.py:load_balancing_loss_func", "1753": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersForConditionalGeneration", "1754": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersEncoderModel", "1755": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextFlexibleLinear", "1756": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextPreTrainedModel", "1757": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextConv1dPaddingCache", "1758": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextEmbeddings", "1759": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextRMSNorm", "1760": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextLinear", "1761": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextRotaryEmbedding", "1762": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextGatingMLP", "1763": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:rotate_half", "1764": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:apply_rotary_pos_emb", "1765": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:repeat_kv", "1766": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextAttention", "1767": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextFlashAttention2", "1768": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextSdpaAttention", "1769": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextDecoderLayer", "1770": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextModel", "1771": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextForConditionalGeneration", "1772": "pe_audio/modeling_pe_audio.py:Snake1d", "1773": "pe_audio/modeling_pe_audio.py:PeAudioDacResidualUnit", "1774": "pe_audio/modeling_pe_audio.py:PeAudioDacEncoderBlock", "1775": "pe_audio/modeling_pe_audio.py:PeAudioDacEncoder", "1776": "pe_audio/modeling_pe_audio.py:PeAudioEncoderEmbedder", "1777": "pe_audio/modeling_pe_audio.py:PeAudioContrastiveHead", "1778": "pe_audio/modeling_pe_audio.py:PeAudioMaskedGroupNorm", "1779": "pe_audio/modeling_pe_audio.py:PeAudioConvBlock1d", "1780": "pe_audio/modeling_pe_audio.py:PeAudioResnetBlock1d", "1781": "pe_audio/modeling_pe_audio.py:PeAudioEncoderPatchEmbedder", "1782": "pe_audio/modeling_pe_audio.py:repeat_kv", "1783": "pe_audio/modeling_pe_audio.py:eager_attention_forward", "1784": "pe_audio/modeling_pe_audio.py:stack_freqs", "1785": "pe_audio/modeling_pe_audio.py:apply_rotary_pos_emb", "1786": "pe_audio/modeling_pe_audio.py:PeAudioEncoderRMSNorm", "1787": "pe_audio/modeling_pe_audio.py:PeAudioEncoderAttention", "1788": "pe_audio/modeling_pe_audio.py:PeAudioEncoderMLP", "1789": "pe_audio/modeling_pe_audio.py:PeAudioEncoderLayer", "1790": "pe_audio/modeling_pe_audio.py:PeAudioPreTrainedModel", "1791": "pe_audio/modeling_pe_audio.py:PeAudioEncoderOutput", "1792": "pe_audio/modeling_pe_audio.py:PeAudioEncoderRotaryEmbedding", "1793": "pe_audio/modeling_pe_audio.py:PeAudioEncoder", "1794": "pe_audio/modeling_pe_audio.py:PeAudioOutput", "1795": "pe_audio/modeling_pe_audio.py:PeAudioModel", "1796": "pe_audio/modeling_pe_audio.py:PeAudioFrameLevelModel", "1797": "bitnet/modeling_bitnet.py:BitNetRMSNorm", "1798": "bitnet/modeling_bitnet.py:BitNetMLP", "1799": "bitnet/modeling_bitnet.py:rotate_half", "1800": "bitnet/modeling_bitnet.py:apply_rotary_pos_emb", "1801": "bitnet/modeling_bitnet.py:repeat_kv", "1802": "bitnet/modeling_bitnet.py:eager_attention_forward", "1803": "bitnet/modeling_bitnet.py:BitNetAttention", "1804": "bitnet/modeling_bitnet.py:BitNetDecoderLayer", "1805": "bitnet/modeling_bitnet.py:BitNetRotaryEmbedding", "1806": "bitnet/modeling_bitnet.py:BitNetPreTrainedModel", "1807": "bitnet/modeling_bitnet.py:BitNetModel", "1808": "bitnet/modeling_bitnet.py:BitNetForCausalLM", "1809": "shieldgemma2/modeling_shieldgemma2.py:ShieldGemma2ImageClassifierOutputWithNoAttention", "1810": "shieldgemma2/modeling_shieldgemma2.py:ShieldGemma2ForImageClassification", "1811": "modernbert/modeling_modernbert.py:ApplyRotaryEmbUnpad", "1812": "modernbert/modeling_modernbert.py:apply_rotary_unpadded", "1813": "modernbert/modeling_modernbert.py:ModernBertUnpaddedRotaryEmbedding", "1814": "modernbert/modeling_modernbert.py:ModernBertEmbeddings", "1815": "modernbert/modeling_modernbert.py:ModernBertMLP", "1816": "modernbert/modeling_modernbert.py:ModernBertRotaryEmbedding", "1817": "modernbert/modeling_modernbert.py:rotate_half", "1818": "modernbert/modeling_modernbert.py:apply_rotary_pos_emb", "1819": "modernbert/modeling_modernbert.py:eager_attention_forward", "1820": "modernbert/modeling_modernbert.py:flash_attention_forward", "1821": "modernbert/modeling_modernbert.py:sdpa_attention_forward", "1822": "modernbert/modeling_modernbert.py:ModernBertAttention", "1823": "modernbert/modeling_modernbert.py:ModernBertEncoderLayer", "1824": "modernbert/modeling_modernbert.py:ModernBertPreTrainedModel", "1825": "modernbert/modeling_modernbert.py:_unpad_modernbert_input", "1826": "modernbert/modeling_modernbert.py:_pad_modernbert_output", "1827": "modernbert/modeling_modernbert.py:ModernBertModel", "1828": "modernbert/modeling_modernbert.py:ModernBertPredictionHead", "1829": "modernbert/modeling_modernbert.py:ModernBertForMaskedLM", "1830": "modernbert/modeling_modernbert.py:ModernBertForSequenceClassification", "1831": "modernbert/modeling_modernbert.py:ModernBertForTokenClassification", "1832": "modernbert/modeling_modernbert.py:ModernBertForQuestionAnswering", "1833": "modernbert/modeling_modernbert.py:ModernBertForMultipleChoice", "1834": "bros/modeling_bros.py:BrosSpadeOutput", "1835": "bros/modeling_bros.py:BrosPositionalEmbedding1D", "1836": "bros/modeling_bros.py:BrosPositionalEmbedding2D", "1837": "bros/modeling_bros.py:BrosBboxEmbeddings", "1838": "bros/modeling_bros.py:BrosTextEmbeddings", "1839": "bros/modeling_bros.py:BrosSelfAttention", "1840": "bros/modeling_bros.py:BrosSelfOutput", "1841": "bros/modeling_bros.py:BrosAttention", "1842": "bros/modeling_bros.py:BrosIntermediate", "1843": "bros/modeling_bros.py:BrosOutput", "1844": "bros/modeling_bros.py:BrosLayer", "1845": "bros/modeling_bros.py:BrosEncoder", "1846": "bros/modeling_bros.py:BrosPooler", "1847": "bros/modeling_bros.py:BrosRelationExtractor", "1848": "bros/modeling_bros.py:BrosPreTrainedModel", "1849": "bros/modeling_bros.py:BrosModel", "1850": "bros/modeling_bros.py:BrosForTokenClassification", "1851": "bros/modeling_bros.py:BrosSpadeEEForTokenClassification", "1852": "bros/modeling_bros.py:BrosSpadeELForTokenClassification", "1853": "diffllama/modeling_diffllama.py:DiffLlamaMLP", "1854": "diffllama/modeling_diffllama.py:DiffLlamaRotaryEmbedding", "1855": "diffllama/modeling_diffllama.py:rotate_half", "1856": "diffllama/modeling_diffllama.py:apply_rotary_pos_emb", "1857": "diffllama/modeling_diffllama.py:repeat_kv", "1858": "diffllama/modeling_diffllama.py:lambda_init_fn", "1859": "diffllama/modeling_diffllama.py:DiffLlamaAttention", "1860": "diffllama/modeling_diffllama.py:DiffLlamaFlashAttention2", "1861": "diffllama/modeling_diffllama.py:DiffLlamaSdpaAttention", "1862": "diffllama/modeling_diffllama.py:DiffLlamaRMSNorm", "1863": "diffllama/modeling_diffllama.py:DiffLlamaDecoderLayer", "1864": "diffllama/modeling_diffllama.py:DiffLlamaPreTrainedModel", "1865": "diffllama/modeling_diffllama.py:DiffLlamaModel", "1866": "diffllama/modeling_diffllama.py:DiffLlamaForCausalLM", "1867": "diffllama/modeling_diffllama.py:DiffLlamaForSequenceClassification", "1868": "diffllama/modeling_diffllama.py:DiffLlamaForQuestionAnswering", "1869": "diffllama/modeling_diffllama.py:DiffLlamaForTokenClassification", "1870": "speech_encoder_decoder/modeling_speech_encoder_decoder.py:shift_tokens_right", "1871": "speech_encoder_decoder/modeling_speech_encoder_decoder.py:SpeechEncoderDecoderModel", "1872": "nemotron/modeling_nemotron.py:_cast_if_autocast_enabled", "1873": "nemotron/modeling_nemotron.py:NemotronLayerNorm1P", "1874": "nemotron/modeling_nemotron.py:NemotronRotaryEmbedding", "1875": "nemotron/modeling_nemotron.py:rotate_half", "1876": "nemotron/modeling_nemotron.py:apply_rotary_pos_emb", "1877": "nemotron/modeling_nemotron.py:NemotronMLP", "1878": "nemotron/modeling_nemotron.py:repeat_kv", "1879": "nemotron/modeling_nemotron.py:NemotronAttention", "1880": "nemotron/modeling_nemotron.py:NemotronFlashAttention2", "1881": "nemotron/modeling_nemotron.py:NemotronSdpaAttention", "1882": "nemotron/modeling_nemotron.py:NemotronDecoderLayer", "1883": "nemotron/modeling_nemotron.py:NemotronPreTrainedModel", "1884": "nemotron/modeling_nemotron.py:NemotronModel", "1885": "nemotron/modeling_nemotron.py:NemotronForCausalLM", "1886": "nemotron/modeling_nemotron.py:NemotronForSequenceClassification", "1887": "nemotron/modeling_nemotron.py:NemotronForQuestionAnswering", "1888": "nemotron/modeling_nemotron.py:NemotronForTokenClassification", "1889": "janus/modeling_janus.py:JanusPreTrainedModel", "1890": "janus/modeling_janus.py:JanusVQVAEOutput", "1891": "janus/modeling_janus.py:JanusBaseModelOutputWithPast", "1892": "janus/modeling_janus.py:JanusCausalLMOutputWithPast", "1893": "janus/modeling_janus.py:JanusVisionEmbeddings", "1894": "janus/modeling_janus.py:repeat_kv", "1895": "janus/modeling_janus.py:eager_attention_forward", "1896": "janus/modeling_janus.py:JanusVisionAttention", "1897": "janus/modeling_janus.py:JanusVisionMLP", "1898": "janus/modeling_janus.py:JanusVisionEncoderLayer", "1899": "janus/modeling_janus.py:JanusVisionEncoder", "1900": "janus/modeling_janus.py:JanusAttention", "1901": "janus/modeling_janus.py:JanusMLP", "1902": "janus/modeling_janus.py:JanusEncoderLayer", "1903": "janus/modeling_janus.py:JanusVisionModel", "1904": "janus/modeling_janus.py:JanusVisionAlignerMLP", "1905": "janus/modeling_janus.py:JanusVQVAEVectorQuantizer", "1906": "janus/modeling_janus.py:JanusVQVAEResnetBlock", "1907": "janus/modeling_janus.py:JanusVQVAEAttnBlock", "1908": "janus/modeling_janus.py:JanusVQVAEConvDownsample", "1909": "janus/modeling_janus.py:JanusVQVAEConvUpsample", "1910": "janus/modeling_janus.py:JanusVQVAEMidBlock", "1911": "janus/modeling_janus.py:JanusVQVAEEncoder", "1912": "janus/modeling_janus.py:JanusVQVAEDecoder", "1913": "janus/modeling_janus.py:JanusVQVAE", "1914": "janus/modeling_janus.py:JanusVQVAEAlignerMLP", "1915": "janus/modeling_janus.py:JanusVQVAEHead", "1916": "janus/modeling_janus.py:JanusModel", "1917": "janus/modeling_janus.py:JanusForConditionalGeneration", "1918": "maskformer/modeling_maskformer.py:DetrDecoderOutput", "1919": "maskformer/modeling_maskformer.py:MaskFormerPixelLevelModuleOutput", "1920": "maskformer/modeling_maskformer.py:MaskFormerPixelDecoderOutput", "1921": "maskformer/modeling_maskformer.py:MaskFormerModelOutput", "1922": "maskformer/modeling_maskformer.py:MaskFormerForInstanceSegmentationOutput", "1923": "maskformer/modeling_maskformer.py:upsample_like", "1924": "maskformer/modeling_maskformer.py:dice_loss", "1925": "maskformer/modeling_maskformer.py:sigmoid_focal_loss", "1926": "maskformer/modeling_maskformer.py:pair_wise_dice_loss", "1927": "maskformer/modeling_maskformer.py:pair_wise_sigmoid_focal_loss", "1928": "maskformer/modeling_maskformer.py:DetrAttention", "1929": "maskformer/modeling_maskformer.py:DetrDecoderLayer", "1930": "maskformer/modeling_maskformer.py:DetrDecoder", "1931": "maskformer/modeling_maskformer.py:MaskFormerHungarianMatcher", "1932": "maskformer/modeling_maskformer.py:MaskFormerLoss", "1933": "maskformer/modeling_maskformer.py:MaskFormerFPNConvLayer", "1934": "maskformer/modeling_maskformer.py:MaskFormerFPNLayer", "1935": "maskformer/modeling_maskformer.py:MaskFormerFPNModel", "1936": "maskformer/modeling_maskformer.py:MaskFormerPixelDecoder", "1937": "maskformer/modeling_maskformer.py:MaskFormerSinePositionEmbedding", "1938": "maskformer/modeling_maskformer.py:PredictionBlock", "1939": "maskformer/modeling_maskformer.py:MaskformerMLPPredictionHead", "1940": "maskformer/modeling_maskformer.py:MaskFormerPixelLevelModule", "1941": "maskformer/modeling_maskformer.py:MaskFormerTransformerModule", "1942": "maskformer/modeling_maskformer.py:MaskFormerPreTrainedModel", "1943": "maskformer/modeling_maskformer.py:MaskFormerModel", "1944": "maskformer/modeling_maskformer.py:MaskFormerForInstanceSegmentation", "1945": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinModelOutputWithPooling", "1946": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinBaseModelOutput", "1947": "maskformer/modeling_maskformer_swin.py:window_partition", "1948": "maskformer/modeling_maskformer_swin.py:window_reverse", "1949": "maskformer/modeling_maskformer_swin.py:drop_path", "1950": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinEmbeddings", "1951": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinPatchEmbeddings", "1952": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinPatchMerging", "1953": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinDropPath", "1954": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinSelfAttention", "1955": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinSelfOutput", "1956": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinAttention", "1957": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinIntermediate", "1958": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinOutput", "1959": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinLayer", "1960": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinStage", "1961": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinEncoder", "1962": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinPreTrainedModel", "1963": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinModel", "1964": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinBackbone", "1965": "deit/modeling_deit.py:DeiTEmbeddings", "1966": "deit/modeling_deit.py:DeiTPatchEmbeddings", "1967": "deit/modeling_deit.py:eager_attention_forward", "1968": "deit/modeling_deit.py:DeiTSelfAttention", "1969": "deit/modeling_deit.py:DeiTSelfOutput", "1970": "deit/modeling_deit.py:DeiTAttention", "1971": "deit/modeling_deit.py:DeiTIntermediate", "1972": "deit/modeling_deit.py:DeiTOutput", "1973": "deit/modeling_deit.py:DeiTLayer", "1974": "deit/modeling_deit.py:DeiTEncoder", "1975": "deit/modeling_deit.py:DeiTPreTrainedModel", "1976": "deit/modeling_deit.py:DeiTModel", "1977": "deit/modeling_deit.py:DeiTPooler", "1978": "deit/modeling_deit.py:DeiTForMaskedImageModeling", "1979": "deit/modeling_deit.py:DeiTForImageClassification", "1980": "deit/modeling_deit.py:DeiTForImageClassificationWithTeacherOutput", "1981": "deit/modeling_deit.py:DeiTForImageClassificationWithTeacher", "1982": "idefics/modeling_idefics.py:IdeficsBaseModelOutputWithPast", "1983": "idefics/modeling_idefics.py:IdeficsCausalLMOutputWithPast", "1984": "idefics/modeling_idefics.py:expand_inputs_for_generation", "1985": "idefics/modeling_idefics.py:freeze_model", "1986": "idefics/modeling_idefics.py:IdeficsDecoupledEmbedding", "1987": "idefics/modeling_idefics.py:IdeficsDecoupledLinear", "1988": "idefics/modeling_idefics.py:IdeficsRMSNorm", "1989": "idefics/modeling_idefics.py:IdeficsEmbedding", "1990": "idefics/modeling_idefics.py:rotate_half", "1991": "idefics/modeling_idefics.py:apply_rotary_pos_emb", "1992": "idefics/modeling_idefics.py:IdeficsMLP", "1993": "idefics/modeling_idefics.py:eager_attention_forward", "1994": "idefics/modeling_idefics.py:IdeficsAttention", "1995": "idefics/modeling_idefics.py:IdeficsDecoderLayer", "1996": "idefics/modeling_idefics.py:IdeficsGatedCrossAttentionLayer", "1997": "idefics/modeling_idefics.py:IdeficsPreTrainedModel", "1998": "idefics/modeling_idefics.py:IdeficsModel", "1999": "idefics/modeling_idefics.py:IdeficsForVisionText2Text", "2000": "perceiver/modeling_perceiver.py:PerceiverModelOutput", "2001": "perceiver/modeling_perceiver.py:PerceiverDecoderOutput", "2002": "perceiver/modeling_perceiver.py:PerceiverMaskedLMOutput", "2003": "perceiver/modeling_perceiver.py:PerceiverClassifierOutput", "2004": "perceiver/modeling_perceiver.py:PerceiverEmbeddings", "2005": "perceiver/modeling_perceiver.py:PerceiverSelfAttention", "2006": "perceiver/modeling_perceiver.py:PerceiverSelfOutput", "2007": "perceiver/modeling_perceiver.py:PerceiverAttention", "2008": "perceiver/modeling_perceiver.py:PerceiverMLP", "2009": "perceiver/modeling_perceiver.py:PerceiverLayer", "2010": "perceiver/modeling_perceiver.py:PerceiverEncoder", "2011": "perceiver/modeling_perceiver.py:PerceiverPreTrainedModel", "2012": "perceiver/modeling_perceiver.py:PerceiverModel", "2013": "perceiver/modeling_perceiver.py:PerceiverForMaskedLM", "2014": "perceiver/modeling_perceiver.py:PerceiverForSequenceClassification", "2015": "perceiver/modeling_perceiver.py:PerceiverForImageClassificationLearned", "2016": "perceiver/modeling_perceiver.py:PerceiverForImageClassificationFourier", "2017": "perceiver/modeling_perceiver.py:PerceiverForImageClassificationConvProcessing", "2018": "perceiver/modeling_perceiver.py:PerceiverForOpticalFlow", "2019": "perceiver/modeling_perceiver.py:PerceiverForMultimodalAutoencoding", "2020": "perceiver/modeling_perceiver.py:build_position_encoding", "2021": "perceiver/modeling_perceiver.py:PerceiverAbstractDecoder", "2022": "perceiver/modeling_perceiver.py:PerceiverProjectionDecoder", "2023": "perceiver/modeling_perceiver.py:PerceiverBasicDecoder", "2024": "perceiver/modeling_perceiver.py:PerceiverClassificationDecoder", "2025": "perceiver/modeling_perceiver.py:PerceiverOpticalFlowDecoder", "2026": "perceiver/modeling_perceiver.py:PerceiverBasicVideoAutoencodingDecoder", "2027": "perceiver/modeling_perceiver.py:restructure", "2028": "perceiver/modeling_perceiver.py:PerceiverMultimodalDecoder", "2029": "perceiver/modeling_perceiver.py:space_to_depth", "2030": "perceiver/modeling_perceiver.py:Conv2dSamePadding", "2031": "perceiver/modeling_perceiver.py:Conv2DDownsample", "2032": "perceiver/modeling_perceiver.py:generate_fourier_features", "2033": "perceiver/modeling_perceiver.py:build_linear_positions", "2034": "perceiver/modeling_perceiver.py:PerceiverAbstractPositionEncoding", "2035": "perceiver/modeling_perceiver.py:PerceiverTrainablePositionEncoding", "2036": "perceiver/modeling_perceiver.py:_check_or_build_spatial_positions", "2037": "perceiver/modeling_perceiver.py:PerceiverFourierPositionEncoding", "2038": "perceiver/modeling_perceiver.py:AbstractPreprocessor", "2039": "perceiver/modeling_perceiver.py:PerceiverTextPreprocessor", "2040": "perceiver/modeling_perceiver.py:PerceiverEmbeddingDecoder", "2041": "perceiver/modeling_perceiver.py:PerceiverMultimodalPostprocessor", "2042": "perceiver/modeling_perceiver.py:PerceiverClassificationPostprocessor", "2043": "perceiver/modeling_perceiver.py:PerceiverAudioPostprocessor", "2044": "perceiver/modeling_perceiver.py:PerceiverProjectionPostprocessor", "2045": "perceiver/modeling_perceiver.py:PerceiverImagePreprocessor", "2046": "perceiver/modeling_perceiver.py:PerceiverOneHotPreprocessor", "2047": "perceiver/modeling_perceiver.py:PerceiverAudioPreprocessor", "2048": "perceiver/modeling_perceiver.py:PerceiverMultimodalPreprocessor", "2049": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersPatchEmbeddings", "2050": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersEmbeddings", "2051": "dinov2_with_registers/modeling_dinov2_with_registers.py:eager_attention_forward", "2052": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersSelfAttention", "2053": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersSelfOutput", "2054": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersAttention", "2055": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersLayerScale", "2056": "dinov2_with_registers/modeling_dinov2_with_registers.py:drop_path", "2057": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersDropPath", "2058": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersMLP", "2059": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersSwiGLUFFN", "2060": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersLayer", "2061": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersEncoder", "2062": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersPreTrainedModel", "2063": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersModel", "2064": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersForImageClassification", "2065": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersBackbone", "2066": "yolos/modeling_yolos.py:YolosObjectDetectionOutput", "2067": "yolos/modeling_yolos.py:YolosEmbeddings", "2068": "yolos/modeling_yolos.py:InterpolateInitialPositionEmbeddings", "2069": "yolos/modeling_yolos.py:InterpolateMidPositionEmbeddings", "2070": "yolos/modeling_yolos.py:YolosPatchEmbeddings", "2071": "yolos/modeling_yolos.py:eager_attention_forward", "2072": "yolos/modeling_yolos.py:YolosSelfAttention", "2073": "yolos/modeling_yolos.py:YolosSelfOutput", "2074": "yolos/modeling_yolos.py:YolosAttention", "2075": "yolos/modeling_yolos.py:YolosIntermediate", "2076": "yolos/modeling_yolos.py:YolosOutput", "2077": "yolos/modeling_yolos.py:YolosLayer", "2078": "yolos/modeling_yolos.py:YolosEncoder", "2079": "yolos/modeling_yolos.py:YolosPreTrainedModel", "2080": "yolos/modeling_yolos.py:YolosModel", "2081": "yolos/modeling_yolos.py:YolosPooler", "2082": "yolos/modeling_yolos.py:YolosMLPPredictionHead", "2083": "yolos/modeling_yolos.py:YolosForObjectDetection", "2084": "vision_text_dual_encoder/modeling_vision_text_dual_encoder.py:contrastive_loss", "2085": "vision_text_dual_encoder/modeling_vision_text_dual_encoder.py:clip_loss", "2086": "vision_text_dual_encoder/modeling_vision_text_dual_encoder.py:VisionTextDualEncoderModel", "2087": "qwen2/modeling_qwen2.py:Qwen2MLP", "2088": "qwen2/modeling_qwen2.py:Qwen2RotaryEmbedding", "2089": "qwen2/modeling_qwen2.py:rotate_half", "2090": "qwen2/modeling_qwen2.py:apply_rotary_pos_emb", "2091": "qwen2/modeling_qwen2.py:repeat_kv", "2092": "qwen2/modeling_qwen2.py:eager_attention_forward", "2093": "qwen2/modeling_qwen2.py:Qwen2Attention", "2094": "qwen2/modeling_qwen2.py:Qwen2RMSNorm", "2095": "qwen2/modeling_qwen2.py:Qwen2DecoderLayer", "2096": "qwen2/modeling_qwen2.py:Qwen2PreTrainedModel", "2097": "qwen2/modeling_qwen2.py:Qwen2Model", "2098": "qwen2/modeling_qwen2.py:Qwen2ForCausalLM", "2099": "qwen2/modeling_qwen2.py:Qwen2ForSequenceClassification", "2100": "qwen2/modeling_qwen2.py:Qwen2ForTokenClassification", "2101": "qwen2/modeling_qwen2.py:Qwen2ForQuestionAnswering", "2102": "glm4/modeling_glm4.py:Glm4MLP", "2103": "glm4/modeling_glm4.py:Glm4DecoderLayer", "2104": "glm4/modeling_glm4.py:repeat_kv", "2105": "glm4/modeling_glm4.py:eager_attention_forward", "2106": "glm4/modeling_glm4.py:rotate_half", "2107": "glm4/modeling_glm4.py:apply_rotary_pos_emb", "2108": "glm4/modeling_glm4.py:Glm4Attention", "2109": "glm4/modeling_glm4.py:Glm4RotaryEmbedding", "2110": "glm4/modeling_glm4.py:Glm4RMSNorm", "2111": "glm4/modeling_glm4.py:Glm4PreTrainedModel", "2112": "glm4/modeling_glm4.py:Glm4Model", "2113": "glm4/modeling_glm4.py:Glm4ForCausalLM", "2114": "glm4/modeling_glm4.py:Glm4ForSequenceClassification", "2115": "glm4/modeling_glm4.py:Glm4ForTokenClassification", "2116": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2Embeddings", "2117": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2SelfAttention", "2118": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2Attention", "2119": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2SelfOutput", "2120": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2Intermediate", "2121": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2Output", "2122": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2Layer", "2123": "layoutlmv2/modeling_layoutlmv2.py:relative_position_bucket", "2124": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2Encoder", "2125": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2PreTrainedModel", "2126": "layoutlmv2/modeling_layoutlmv2.py:my_convert_sync_batchnorm", "2127": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2VisualBackbone", "2128": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2Pooler", "2129": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2Model", "2130": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2ForSequenceClassification", "2131": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2ForTokenClassification", "2132": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2ForQuestionAnswering", "2133": "gpt_bigcode/modeling_gpt_bigcode.py:upcast_masked_softmax", "2134": "gpt_bigcode/modeling_gpt_bigcode.py:upcast_softmax", "2135": "gpt_bigcode/modeling_gpt_bigcode.py:masked_softmax", "2136": "gpt_bigcode/modeling_gpt_bigcode.py:repeat_kv", "2137": "gpt_bigcode/modeling_gpt_bigcode.py:eager_attention_forward", "2138": "gpt_bigcode/modeling_gpt_bigcode.py:GPTBigCodeAttention", "2139": "gpt_bigcode/modeling_gpt_bigcode.py:GPTBigCodeMLP", "2140": "gpt_bigcode/modeling_gpt_bigcode.py:GPTBigCodeBlock", "2141": "gpt_bigcode/modeling_gpt_bigcode.py:GPTBigCodePreTrainedModel", "2142": "gpt_bigcode/modeling_gpt_bigcode.py:GPTBigCodeModel", "2143": "gpt_bigcode/modeling_gpt_bigcode.py:GPTBigCodeForCausalLM", "2144": "gpt_bigcode/modeling_gpt_bigcode.py:GPTBigCodeForSequenceClassification", "2145": "gpt_bigcode/modeling_gpt_bigcode.py:GPTBigCodeForTokenClassification", "2146": "clip/modeling_clip.py:contrastive_loss", "2147": "clip/modeling_clip.py:clip_loss", "2148": "clip/modeling_clip.py:_get_vector_norm", "2149": "clip/modeling_clip.py:CLIPVisionModelOutput", "2150": "clip/modeling_clip.py:CLIPTextModelOutput", "2151": "clip/modeling_clip.py:CLIPOutput", "2152": "clip/modeling_clip.py:CLIPVisionEmbeddings", "2153": "clip/modeling_clip.py:CLIPTextEmbeddings", "2154": "clip/modeling_clip.py:eager_attention_forward", "2155": "clip/modeling_clip.py:CLIPAttention", "2156": "clip/modeling_clip.py:CLIPMLP", "2157": "clip/modeling_clip.py:CLIPEncoderLayer", "2158": "clip/modeling_clip.py:CLIPPreTrainedModel", "2159": "clip/modeling_clip.py:CLIPEncoder", "2160": "clip/modeling_clip.py:CLIPTextTransformer", "2161": "clip/modeling_clip.py:CLIPTextModel", "2162": "clip/modeling_clip.py:CLIPVisionTransformer", "2163": "clip/modeling_clip.py:CLIPVisionModel", "2164": "clip/modeling_clip.py:CLIPModel", "2165": "clip/modeling_clip.py:CLIPTextModelWithProjection", "2166": "clip/modeling_clip.py:CLIPVisionModelWithProjection", "2167": "clip/modeling_clip.py:CLIPForImageClassification", "2168": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2Experts", "2169": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2Moe", "2170": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2MLP", "2171": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2RMSNorm", "2172": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2RotaryEmbedding", "2173": "deepseek_v2/modeling_deepseek_v2.py:repeat_kv", "2174": "deepseek_v2/modeling_deepseek_v2.py:eager_attention_forward", "2175": "deepseek_v2/modeling_deepseek_v2.py:apply_rotary_emb", "2176": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2Attention", "2177": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2DecoderLayer", "2178": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2PreTrainedModel", "2179": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2Model", "2180": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2ForCausalLM", "2181": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2ForSequenceClassification", "2182": "convnext/modeling_convnext.py:drop_path", "2183": "convnext/modeling_convnext.py:ConvNextDropPath", "2184": "convnext/modeling_convnext.py:ConvNextLayerNorm", "2185": "convnext/modeling_convnext.py:ConvNextEmbeddings", "2186": "convnext/modeling_convnext.py:ConvNextLayer", "2187": "convnext/modeling_convnext.py:ConvNextStage", "2188": "convnext/modeling_convnext.py:ConvNextEncoder", "2189": "convnext/modeling_convnext.py:ConvNextPreTrainedModel", "2190": "convnext/modeling_convnext.py:ConvNextModel", "2191": "convnext/modeling_convnext.py:ConvNextForImageClassification", "2192": "convnext/modeling_convnext.py:ConvNextBackbone", "2193": "umt5/modeling_umt5.py:UMT5LayerNorm", "2194": "umt5/modeling_umt5.py:UMT5DenseActDense", "2195": "umt5/modeling_umt5.py:UMT5DenseGatedActDense", "2196": "umt5/modeling_umt5.py:UMT5LayerFF", "2197": "umt5/modeling_umt5.py:UMT5Attention", "2198": "umt5/modeling_umt5.py:UMT5LayerSelfAttention", "2199": "umt5/modeling_umt5.py:UMT5LayerCrossAttention", "2200": "umt5/modeling_umt5.py:UMT5Block", "2201": "umt5/modeling_umt5.py:UMT5ClassificationHead", "2202": "umt5/modeling_umt5.py:UMT5PreTrainedModel", "2203": "umt5/modeling_umt5.py:UMT5Stack", "2204": "umt5/modeling_umt5.py:UMT5Model", "2205": "umt5/modeling_umt5.py:UMT5ForConditionalGeneration", "2206": "umt5/modeling_umt5.py:UMT5EncoderModel", "2207": "umt5/modeling_umt5.py:UMT5ForSequenceClassification", "2208": "umt5/modeling_umt5.py:UMT5ForTokenClassification", "2209": "umt5/modeling_umt5.py:UMT5ForQuestionAnswering", "2210": "aya_vision/modeling_aya_vision.py:AyaVisionMultiModalProjector", "2211": "aya_vision/modeling_aya_vision.py:AyaVisionPreTrainedModel", "2212": "aya_vision/modeling_aya_vision.py:AyaVisionCausalLMOutputWithPast", "2213": "aya_vision/modeling_aya_vision.py:AyaVisionModelOutputWithPast", "2214": "aya_vision/modeling_aya_vision.py:AyaVisionModel", "2215": "aya_vision/modeling_aya_vision.py:AyaVisionForConditionalGeneration", "2216": "olmo2/modeling_olmo2.py:Olmo2RMSNorm", "2217": "olmo2/modeling_olmo2.py:Olmo2RotaryEmbedding", "2218": "olmo2/modeling_olmo2.py:repeat_kv", "2219": "olmo2/modeling_olmo2.py:eager_attention_forward", "2220": "olmo2/modeling_olmo2.py:apply_rotary_pos_emb", "2221": "olmo2/modeling_olmo2.py:rotate_half", "2222": "olmo2/modeling_olmo2.py:Olmo2Attention", "2223": "olmo2/modeling_olmo2.py:Olmo2MLP", "2224": "olmo2/modeling_olmo2.py:Olmo2DecoderLayer", "2225": "olmo2/modeling_olmo2.py:Olmo2PreTrainedModel", "2226": "olmo2/modeling_olmo2.py:Olmo2Model", "2227": "olmo2/modeling_olmo2.py:Olmo2ForCausalLM", "2228": "textnet/modeling_textnet.py:TextNetConvLayer", "2229": "textnet/modeling_textnet.py:TextNetRepConvLayer", "2230": "textnet/modeling_textnet.py:TextNetStage", "2231": "textnet/modeling_textnet.py:TextNetEncoder", "2232": "textnet/modeling_textnet.py:TextNetPreTrainedModel", "2233": "textnet/modeling_textnet.py:TextNetModel", "2234": "textnet/modeling_textnet.py:TextNetForImageClassification", "2235": "textnet/modeling_textnet.py:TextNetBackbone", "2236": "flex_olmo/modeling_flex_olmo.py:FlexOlmoRMSNorm", "2237": "flex_olmo/modeling_flex_olmo.py:FlexOlmoRotaryEmbedding", "2238": "flex_olmo/modeling_flex_olmo.py:FlexOlmoMLP", "2239": "flex_olmo/modeling_flex_olmo.py:repeat_kv", "2240": "flex_olmo/modeling_flex_olmo.py:eager_attention_forward", "2241": "flex_olmo/modeling_flex_olmo.py:apply_rotary_pos_emb", "2242": "flex_olmo/modeling_flex_olmo.py:rotate_half", "2243": "flex_olmo/modeling_flex_olmo.py:FlexOlmoAttention", "2244": "flex_olmo/modeling_flex_olmo.py:FlexOlmoExperts", "2245": "flex_olmo/modeling_flex_olmo.py:FlexOlmoTopKRouter", "2246": "flex_olmo/modeling_flex_olmo.py:FlexOlmoSparseMoeBlock", "2247": "flex_olmo/modeling_flex_olmo.py:FlexOlmoDecoderLayer", "2248": "flex_olmo/modeling_flex_olmo.py:FlexOlmoPreTrainedModel", "2249": "flex_olmo/modeling_flex_olmo.py:FlexOlmoModel", "2250": "flex_olmo/modeling_flex_olmo.py:load_balancing_loss_func", "2251": "flex_olmo/modeling_flex_olmo.py:FlexOlmoForCausalLM", "2252": "distilbert/modeling_distilbert.py:create_sinusoidal_embeddings", "2253": "distilbert/modeling_distilbert.py:_create_sinusoidal_embeddings", "2254": "distilbert/modeling_distilbert.py:Embeddings", "2255": "distilbert/modeling_distilbert.py:eager_attention_forward", "2256": "distilbert/modeling_distilbert.py:DistilBertSelfAttention", "2257": "distilbert/modeling_distilbert.py:FFN", "2258": "distilbert/modeling_distilbert.py:TransformerBlock", "2259": "distilbert/modeling_distilbert.py:Transformer", "2260": "distilbert/modeling_distilbert.py:DistilBertPreTrainedModel", "2261": "distilbert/modeling_distilbert.py:DistilBertModel", "2262": "distilbert/modeling_distilbert.py:DistilBertForMaskedLM", "2263": "distilbert/modeling_distilbert.py:DistilBertForSequenceClassification", "2264": "distilbert/modeling_distilbert.py:DistilBertForQuestionAnswering", "2265": "distilbert/modeling_distilbert.py:DistilBertForTokenClassification", "2266": "distilbert/modeling_distilbert.py:DistilBertForMultipleChoice", "2267": "timm_wrapper/modeling_timm_wrapper.py:TimmWrapperModelOutput", "2268": "timm_wrapper/modeling_timm_wrapper.py:_create_timm_model_with_error_handling", "2269": "timm_wrapper/modeling_timm_wrapper.py:TimmWrapperPreTrainedModel", "2270": "timm_wrapper/modeling_timm_wrapper.py:TimmWrapperModel", "2271": "timm_wrapper/modeling_timm_wrapper.py:TimmWrapperForImageClassification", "2272": "trocr/modeling_trocr.py:TrOCRLearnedPositionalEmbedding", "2273": "trocr/modeling_trocr.py:TrOCRScaledWordEmbedding", "2274": "trocr/modeling_trocr.py:TrOCRSinusoidalPositionalEmbedding", "2275": "trocr/modeling_trocr.py:TrOCRAttention", "2276": "trocr/modeling_trocr.py:TrOCRDecoderLayer", "2277": "trocr/modeling_trocr.py:TrOCRPreTrainedModel", "2278": "trocr/modeling_trocr.py:TrOCRDecoder", "2279": "trocr/modeling_trocr.py:TrOCRDecoderWrapper", "2280": "trocr/modeling_trocr.py:TrOCRForCausalLM", "2281": "phimoe/modeling_phimoe.py:PhimoeRotaryEmbedding", "2282": "phimoe/modeling_phimoe.py:rotate_half", "2283": "phimoe/modeling_phimoe.py:apply_rotary_pos_emb", "2284": "phimoe/modeling_phimoe.py:repeat_kv", "2285": "phimoe/modeling_phimoe.py:eager_attention_forward", "2286": "phimoe/modeling_phimoe.py:PhimoeAttention", "2287": "phimoe/modeling_phimoe.py:PhimoeMultiplier", "2288": "phimoe/modeling_phimoe.py:PhimoeExperts", "2289": "phimoe/modeling_phimoe.py:sparsemixer", "2290": "phimoe/modeling_phimoe.py:PhimoeTopKRouter", "2291": "phimoe/modeling_phimoe.py:PhimoeSparseMoeBlock", "2292": "phimoe/modeling_phimoe.py:PhimoeRMSNorm", "2293": "phimoe/modeling_phimoe.py:PhimoeDecoderLayer", "2294": "phimoe/modeling_phimoe.py:PhimoePreTrainedModel", "2295": "phimoe/modeling_phimoe.py:PhimoeModel", "2296": "phimoe/modeling_phimoe.py:load_balancing_loss_func", "2297": "phimoe/modeling_phimoe.py:PhimoeForCausalLM", "2298": "phimoe/modeling_phimoe.py:PhimoeForSequenceClassification", "2299": "mimi/modeling_mimi.py:MimiOutput", "2300": "mimi/modeling_mimi.py:MimiConv1dPaddingCache", "2301": "mimi/modeling_mimi.py:MimiEncoderOutput", "2302": "mimi/modeling_mimi.py:MimiDecoderOutput", "2303": "mimi/modeling_mimi.py:MimiConv1d", "2304": "mimi/modeling_mimi.py:MimiConvTranspose1d", "2305": "mimi/modeling_mimi.py:MimiResnetBlock", "2306": "mimi/modeling_mimi.py:MimiEncoder", "2307": "mimi/modeling_mimi.py:MimiLayerScale", "2308": "mimi/modeling_mimi.py:MimiRotaryEmbedding", "2309": "mimi/modeling_mimi.py:rotate_half", "2310": "mimi/modeling_mimi.py:apply_rotary_pos_emb", "2311": "mimi/modeling_mimi.py:MimiMLP", "2312": "mimi/modeling_mimi.py:repeat_kv", "2313": "mimi/modeling_mimi.py:MimiAttention", "2314": "mimi/modeling_mimi.py:MimiFlashAttention2", "2315": "mimi/modeling_mimi.py:MimiSdpaAttention", "2316": "mimi/modeling_mimi.py:MimiTransformerLayer", "2317": "mimi/modeling_mimi.py:MimiTransformerModel", "2318": "mimi/modeling_mimi.py:MimiDecoder", "2319": "mimi/modeling_mimi.py:MimiEuclideanCodebook", "2320": "mimi/modeling_mimi.py:MimiVectorQuantization", "2321": "mimi/modeling_mimi.py:MimiResidualVectorQuantizer", "2322": "mimi/modeling_mimi.py:MimiSplitResidualVectorQuantizer", "2323": "mimi/modeling_mimi.py:MimiPreTrainedModel", "2324": "mimi/modeling_mimi.py:MimiModel", "2325": "hiera/modeling_hiera.py:HieraEncoderOutput", "2326": "hiera/modeling_hiera.py:HieraModelOutput", "2327": "hiera/modeling_hiera.py:HieraForImageClassificationOutput", "2328": "hiera/modeling_hiera.py:HieraForPreTrainingOutput", "2329": "hiera/modeling_hiera.py:HieraPatchEmbeddings", "2330": "hiera/modeling_hiera.py:HieraEmbeddings", "2331": "hiera/modeling_hiera.py:HieraMaskUnitAttention", "2332": "hiera/modeling_hiera.py:drop_path", "2333": "hiera/modeling_hiera.py:HieraDropPath", "2334": "hiera/modeling_hiera.py:HieraMlp", "2335": "hiera/modeling_hiera.py:HieraLayer", "2336": "hiera/modeling_hiera.py:HieraStage", "2337": "hiera/modeling_hiera.py:undo_windowing", "2338": "hiera/modeling_hiera.py:HieraEncoder", "2339": "hiera/modeling_hiera.py:unroll", "2340": "hiera/modeling_hiera.py:HieraPreTrainedModel", "2341": "hiera/modeling_hiera.py:HieraPooler", "2342": "hiera/modeling_hiera.py:HieraModel", "2343": "hiera/modeling_hiera.py:HieraDecoder", "2344": "hiera/modeling_hiera.py:HieraMultiScaleHead", "2345": "hiera/modeling_hiera.py:HieraForPreTraining", "2346": "hiera/modeling_hiera.py:HieraForImageClassification", "2347": "hiera/modeling_hiera.py:HieraBackbone", "2348": "gemma3n/modeling_gemma3n.py:Gemma3nModelOutputWithPast", "2349": "gemma3n/modeling_gemma3n.py:Gemma3nCausalLMOutputWithPast", "2350": "gemma3n/modeling_gemma3n.py:Gemma3nRMSNorm", "2351": "gemma3n/modeling_gemma3n.py:Gemma3nAudioRelativePositionEmbedding", "2352": "gemma3n/modeling_gemma3n.py:Gemma3nAudioAttention", "2353": "gemma3n/modeling_gemma3n.py:Gemma3nAudioCumulativeGroupNorm", "2354": "gemma3n/modeling_gemma3n.py:Gemma3nAudioSSCPConvBlock", "2355": "gemma3n/modeling_gemma3n.py:Gemma3nAudioSubSampleConvProjection", "2356": "gemma3n/modeling_gemma3n.py:Gemma3nAudioConformerAttention", "2357": "gemma3n/modeling_gemma3n.py:Gemma3nAudioConformerFeedForward", "2358": "gemma3n/modeling_gemma3n.py:Gemma3nAudioConformerLightConv1d", "2359": "gemma3n/modeling_gemma3n.py:Gemma3nAudioConformerBlock", "2360": "gemma3n/modeling_gemma3n.py:Gemma3nAudioEncoder", "2361": "gemma3n/modeling_gemma3n.py:Gemma3nTextScaledWordEmbedding", "2362": "gemma3n/modeling_gemma3n.py:Gemma3nTextLaurelBlock", "2363": "gemma3n/modeling_gemma3n.py:Gemma3nTextMLP", "2364": "gemma3n/modeling_gemma3n.py:Gemma3nTextAltUp", "2365": "gemma3n/modeling_gemma3n.py:rotate_half", "2366": "gemma3n/modeling_gemma3n.py:repeat_kv", "2367": "gemma3n/modeling_gemma3n.py:eager_attention_forward", "2368": "gemma3n/modeling_gemma3n.py:apply_rotary_pos_emb", "2369": "gemma3n/modeling_gemma3n.py:Gemma3nTextAttention", "2370": "gemma3n/modeling_gemma3n.py:Gemma3nTextDecoderLayer", "2371": "gemma3n/modeling_gemma3n.py:Gemma3nPreTrainedModel", "2372": "gemma3n/modeling_gemma3n.py:Gemma3nRotaryEmbedding", "2373": "gemma3n/modeling_gemma3n.py:Gemma3nTextModel", "2374": "gemma3n/modeling_gemma3n.py:Gemma3nForCausalLM", "2375": "gemma3n/modeling_gemma3n.py:Gemma3nMultimodalEmbedder", "2376": "gemma3n/modeling_gemma3n.py:Gemma3nModel", "2377": "gemma3n/modeling_gemma3n.py:Gemma3nForConditionalGeneration", "2378": "plbart/modeling_plbart.py:PLBartScaledWordEmbedding", "2379": "plbart/modeling_plbart.py:PLBartPreTrainedModel", "2380": "plbart/modeling_plbart.py:PLBartLearnedPositionalEmbedding", "2381": "plbart/modeling_plbart.py:eager_attention_forward", "2382": "plbart/modeling_plbart.py:PLBartAttention", "2383": "plbart/modeling_plbart.py:PLBartEncoderLayer", "2384": "plbart/modeling_plbart.py:PLBartEncoder", "2385": "plbart/modeling_plbart.py:PLBartDecoderLayer", "2386": "plbart/modeling_plbart.py:PLBartDecoder", "2387": "plbart/modeling_plbart.py:shift_tokens_right", "2388": "plbart/modeling_plbart.py:PLBartModel", "2389": "plbart/modeling_plbart.py:PLBartForConditionalGeneration", "2390": "plbart/modeling_plbart.py:PLBartClassificationHead", "2391": "plbart/modeling_plbart.py:PLBartForSequenceClassification", "2392": "plbart/modeling_plbart.py:PLBartDecoderWrapper", "2393": "plbart/modeling_plbart.py:PLBartForCausalLM", "2394": "longt5/modeling_longt5.py:_pad_to_multiple", "2395": "longt5/modeling_longt5.py:_split_into_blocks", "2396": "longt5/modeling_longt5.py:_concatenate_3_blocks", "2397": "longt5/modeling_longt5.py:_make_3block_relative_position_ids", "2398": "longt5/modeling_longt5.py:_mask_local_attention_mask", "2399": "longt5/modeling_longt5.py:_get_local_attention_mask", "2400": "longt5/modeling_longt5.py:_make_global_fixed_block_ids", "2401": "longt5/modeling_longt5.py:_make_side_relative_position_ids", "2402": "longt5/modeling_longt5.py:_create_global_aggregates", "2403": "longt5/modeling_longt5.py:LongT5LayerNorm", "2404": "longt5/modeling_longt5.py:LongT5DenseActDense", "2405": "longt5/modeling_longt5.py:LongT5DenseGatedActDense", "2406": "longt5/modeling_longt5.py:LongT5LayerFF", "2407": "longt5/modeling_longt5.py:LongT5Attention", "2408": "longt5/modeling_longt5.py:LongT5LocalAttention", "2409": "longt5/modeling_longt5.py:LongT5TransientGlobalAttention", "2410": "longt5/modeling_longt5.py:LongT5LayerSelfAttention", "2411": "longt5/modeling_longt5.py:LongT5LayerLocalSelfAttention", "2412": "longt5/modeling_longt5.py:LongT5LayerTransientGlobalSelfAttention", "2413": "longt5/modeling_longt5.py:LongT5LayerCrossAttention", "2414": "longt5/modeling_longt5.py:LongT5Block", "2415": "longt5/modeling_longt5.py:LongT5PreTrainedModel", "2416": "longt5/modeling_longt5.py:LongT5Stack", "2417": "longt5/modeling_longt5.py:LongT5Model", "2418": "longt5/modeling_longt5.py:LongT5ForConditionalGeneration", "2419": "longt5/modeling_longt5.py:LongT5EncoderModel", "2420": "ctrl/modeling_ctrl.py:angle_defn", "2421": "ctrl/modeling_ctrl.py:positional_encoding", "2422": "ctrl/modeling_ctrl.py:scaled_dot_product_attention", "2423": "ctrl/modeling_ctrl.py:MultiHeadAttention", "2424": "ctrl/modeling_ctrl.py:point_wise_feed_forward_network", "2425": "ctrl/modeling_ctrl.py:EncoderLayer", "2426": "ctrl/modeling_ctrl.py:CTRLPreTrainedModel", "2427": "ctrl/modeling_ctrl.py:CTRLModel", "2428": "ctrl/modeling_ctrl.py:CTRLLMHeadModel", "2429": "ctrl/modeling_ctrl.py:CTRLForSequenceClassification", "2430": "idefics2/modeling_idefics2.py:Idefics2BaseModelOutputWithPast", "2431": "idefics2/modeling_idefics2.py:Idefics2CausalLMOutputWithPast", "2432": "idefics2/modeling_idefics2.py:Idefics2VisionEmbeddings", "2433": "idefics2/modeling_idefics2.py:eager_attention_forward", "2434": "idefics2/modeling_idefics2.py:Idefics2VisionAttention", "2435": "idefics2/modeling_idefics2.py:Idefics2VisionMLP", "2436": "idefics2/modeling_idefics2.py:Idefics2MLP", "2437": "idefics2/modeling_idefics2.py:Idefics2MultiheadAttentionPoolingHead", "2438": "idefics2/modeling_idefics2.py:Idefics2EncoderLayer", "2439": "idefics2/modeling_idefics2.py:Idefics2Encoder", "2440": "idefics2/modeling_idefics2.py:Idefics2PreTrainedModel", "2441": "idefics2/modeling_idefics2.py:Idefics2VisionTransformer", "2442": "idefics2/modeling_idefics2.py:repeat_kv", "2443": "idefics2/modeling_idefics2.py:Idefics2RMSNorm", "2444": "idefics2/modeling_idefics2.py:Idefics2PerceiverAttention", "2445": "idefics2/modeling_idefics2.py:Idefics2PerceiverLayer", "2446": "idefics2/modeling_idefics2.py:Idefics2PerceiverResampler", "2447": "idefics2/modeling_idefics2.py:Idefics2Connector", "2448": "idefics2/modeling_idefics2.py:Idefics2Model", "2449": "idefics2/modeling_idefics2.py:Idefics2ForConditionalGeneration", "2450": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatForPreTrainingOutput", "2451": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatSamePadLayer", "2452": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatPositionalConvEmbedding", "2453": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatNoLayerNormConvLayer", "2454": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatLayerNormConvLayer", "2455": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatGroupNormConvLayer", "2456": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatFeatureEncoder", "2457": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatFeatureProjection", "2458": "unispeech_sat/modeling_unispeech_sat.py:eager_attention_forward", "2459": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatAttention", "2460": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatFeedForward", "2461": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatEncoderLayer", "2462": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatEncoder", "2463": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatAttnAdapterLayer", "2464": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatEncoderLayerStableLayerNorm", "2465": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatEncoderStableLayerNorm", "2466": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatGumbelVectorQuantizer", "2467": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatPreTrainedModel", "2468": "unispeech_sat/modeling_unispeech_sat.py:_compute_mask_indices", "2469": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatModel", "2470": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatForPreTraining", "2471": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatForCTC", "2472": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatForSequenceClassification", "2473": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatForAudioFrameClassification", "2474": "unispeech_sat/modeling_unispeech_sat.py:AMSoftmaxLoss", "2475": "unispeech_sat/modeling_unispeech_sat.py:TDNNLayer", "2476": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatForXVector", "2477": "bert/modeling_bert.py:BertEmbeddings", "2478": "bert/modeling_bert.py:eager_attention_forward", "2479": "bert/modeling_bert.py:BertSelfAttention", "2480": "bert/modeling_bert.py:BertCrossAttention", "2481": "bert/modeling_bert.py:BertSelfOutput", "2482": "bert/modeling_bert.py:BertAttention", "2483": "bert/modeling_bert.py:BertIntermediate", "2484": "bert/modeling_bert.py:BertOutput", "2485": "bert/modeling_bert.py:BertLayer", "2486": "bert/modeling_bert.py:BertEncoder", "2487": "bert/modeling_bert.py:BertPooler", "2488": "bert/modeling_bert.py:BertPredictionHeadTransform", "2489": "bert/modeling_bert.py:BertLMPredictionHead", "2490": "bert/modeling_bert.py:BertOnlyMLMHead", "2491": "bert/modeling_bert.py:BertOnlyNSPHead", "2492": "bert/modeling_bert.py:BertPreTrainingHeads", "2493": "bert/modeling_bert.py:BertPreTrainedModel", "2494": "bert/modeling_bert.py:BertForPreTrainingOutput", "2495": "bert/modeling_bert.py:BertModel", "2496": "bert/modeling_bert.py:BertForPreTraining", "2497": "bert/modeling_bert.py:BertLMHeadModel", "2498": "bert/modeling_bert.py:BertForMaskedLM", "2499": "bert/modeling_bert.py:BertForNextSentencePrediction", "2500": "bert/modeling_bert.py:BertForSequenceClassification", "2501": "bert/modeling_bert.py:BertForMultipleChoice", "2502": "bert/modeling_bert.py:BertForTokenClassification", "2503": "bert/modeling_bert.py:BertForQuestionAnswering", "2504": "ijepa/modeling_ijepa.py:IJepaPatchEmbeddings", "2505": "ijepa/modeling_ijepa.py:IJepaEmbeddings", "2506": "ijepa/modeling_ijepa.py:eager_attention_forward", "2507": "ijepa/modeling_ijepa.py:IJepaSelfAttention", "2508": "ijepa/modeling_ijepa.py:IJepaSelfOutput", "2509": "ijepa/modeling_ijepa.py:IJepaAttention", "2510": "ijepa/modeling_ijepa.py:IJepaIntermediate", "2511": "ijepa/modeling_ijepa.py:IJepaOutput", "2512": "ijepa/modeling_ijepa.py:IJepaLayer", "2513": "ijepa/modeling_ijepa.py:IJepaPreTrainedModel", "2514": "ijepa/modeling_ijepa.py:IJepaEncoder", "2515": "ijepa/modeling_ijepa.py:IJepaPooler", "2516": "ijepa/modeling_ijepa.py:IJepaModel", "2517": "ijepa/modeling_ijepa.py:IJepaForImageClassification", "2518": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3PatchEmbeddings", "2519": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3TextEmbeddings", "2520": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3PreTrainedModel", "2521": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3SelfAttention", "2522": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3SelfOutput", "2523": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3Attention", "2524": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3Layer", "2525": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3Encoder", "2526": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3Intermediate", "2527": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3Output", "2528": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3Model", "2529": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3ClassificationHead", "2530": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3ForTokenClassification", "2531": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3ForQuestionAnswering", "2532": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3ForSequenceClassification", "2533": "bamba/modeling_bamba.py:BambaFlashAttentionKwargs", "2534": "bamba/modeling_bamba.py:HybridMambaAttentionDynamicCache", "2535": "bamba/modeling_bamba.py:BambaRotaryEmbedding", "2536": "bamba/modeling_bamba.py:rotate_half", "2537": "bamba/modeling_bamba.py:repeat_kv", "2538": "bamba/modeling_bamba.py:eager_attention_forward", "2539": "bamba/modeling_bamba.py:apply_rotary_pos_emb", "2540": "bamba/modeling_bamba.py:BambaAttention", "2541": "bamba/modeling_bamba.py:BambaRMSNormGated", "2542": "bamba/modeling_bamba.py:pad_tensor_by_size", "2543": "bamba/modeling_bamba.py:reshape_into_chunks", "2544": "bamba/modeling_bamba.py:segment_sum", "2545": "bamba/modeling_bamba.py:apply_mask_to_padding_states", "2546": "bamba/modeling_bamba.py:BambaMixer", "2547": "bamba/modeling_bamba.py:BambaMLP", "2548": "bamba/modeling_bamba.py:BambaRMSNorm", "2549": "bamba/modeling_bamba.py:BambaDecoderLayer", "2550": "bamba/modeling_bamba.py:BambaPreTrainedModel", "2551": "bamba/modeling_bamba.py:BambaModel", "2552": "bamba/modeling_bamba.py:BambaForCausalLM", "2553": "lilt/modeling_lilt.py:LiltTextEmbeddings", "2554": "lilt/modeling_lilt.py:LiltLayoutEmbeddings", "2555": "lilt/modeling_lilt.py:LiltSelfAttention", "2556": "lilt/modeling_lilt.py:LiltSelfOutput", "2557": "lilt/modeling_lilt.py:LiltAttention", "2558": "lilt/modeling_lilt.py:LiltIntermediate", "2559": "lilt/modeling_lilt.py:LiltOutput", "2560": "lilt/modeling_lilt.py:LiltLayer", "2561": "lilt/modeling_lilt.py:LiltEncoder", "2562": "lilt/modeling_lilt.py:LiltPooler", "2563": "lilt/modeling_lilt.py:LiltPreTrainedModel", "2564": "lilt/modeling_lilt.py:LiltModel", "2565": "lilt/modeling_lilt.py:LiltForSequenceClassification", "2566": "lilt/modeling_lilt.py:LiltForTokenClassification", "2567": "lilt/modeling_lilt.py:LiltClassificationHead", "2568": "lilt/modeling_lilt.py:LiltForQuestionAnswering", "2569": "lfm2_vl/modeling_lfm2_vl.py:Lfm2VlMultiModalProjector", "2570": "lfm2_vl/modeling_lfm2_vl.py:Lfm2VlPreTrainedModel", "2571": "lfm2_vl/modeling_lfm2_vl.py:Lfm2VlCausalLMOutputWithPast", "2572": "lfm2_vl/modeling_lfm2_vl.py:Lfm2VlModelOutputWithPast", "2573": "lfm2_vl/modeling_lfm2_vl.py:Lfm2VlModel", "2574": "lfm2_vl/modeling_lfm2_vl.py:Lfm2VlForConditionalGeneration", "2575": "gpt_neox/modeling_gpt_neox.py:GPTNeoXMLP", "2576": "gpt_neox/modeling_gpt_neox.py:GPTNeoXRotaryEmbedding", "2577": "gpt_neox/modeling_gpt_neox.py:rotate_half", "2578": "gpt_neox/modeling_gpt_neox.py:apply_rotary_pos_emb", "2579": "gpt_neox/modeling_gpt_neox.py:eager_attention_forward", "2580": "gpt_neox/modeling_gpt_neox.py:GPTNeoXAttention", "2581": "gpt_neox/modeling_gpt_neox.py:GPTNeoXLayer", "2582": "gpt_neox/modeling_gpt_neox.py:GPTNeoXRMSNorm", "2583": "gpt_neox/modeling_gpt_neox.py:GPTNeoXDecoderLayer", "2584": "gpt_neox/modeling_gpt_neox.py:GPTNeoXPreTrainedModel", "2585": "gpt_neox/modeling_gpt_neox.py:GPTNeoXModel", "2586": "gpt_neox/modeling_gpt_neox.py:GPTNeoXForCausalLM", "2587": "gpt_neox/modeling_gpt_neox.py:GPTNeoXForSequenceClassification", "2588": "gpt_neox/modeling_gpt_neox.py:GPTNeoXForTokenClassification", "2589": "gpt_neox/modeling_gpt_neox.py:GPTNeoXForQuestionAnswering", "2590": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLEmbeddings", "2591": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:eager_attention_forward", "2592": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLSelfAttention", "2593": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLCrossAttention", "2594": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLSelfOutput", "2595": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLAttention", "2596": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLOutput", "2597": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLIntermediate", "2598": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLLayer", "2599": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLEncoder", "2600": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLPreTrainedModel", "2601": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLPooler", "2602": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLModel", "2603": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLLMHead", "2604": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLClassificationHead", "2605": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLForCausalLM", "2606": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLForMaskedLM", "2607": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLForSequenceClassification", "2608": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLForMultipleChoice", "2609": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLForTokenClassification", "2610": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLForQuestionAnswering", "2611": "kosmos2_5/modeling_kosmos2_5.py:_expand_mask", "2612": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5ModelOutput", "2613": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5ForConditionalGenerationModelOutput", "2614": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5LayerNorm", "2615": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5VisionEmbeddings", "2616": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5VisionMlp", "2617": "kosmos2_5/modeling_kosmos2_5.py:eager_attention_forward", "2618": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5VisionAttention", "2619": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5VisionLayer", "2620": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5VisionEncoder", "2621": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5TextSinusoidalPositionalEmbedding", "2622": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5TextFFN", "2623": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5TextAttention", "2624": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5TextBlock", "2625": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5TextTransformer", "2626": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5ImageToTextProjection", "2627": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5PreTrainedModel", "2628": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5VisionModel", "2629": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5TextModel", "2630": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5Model", "2631": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5TextForCausalLM", "2632": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5ForConditionalGeneration", "2633": "timesfm/modeling_timesfm.py:TimesFmOutput", "2634": "timesfm/modeling_timesfm.py:TimesFmOutputForPrediction", "2635": "timesfm/modeling_timesfm.py:TimesFmMLP", "2636": "timesfm/modeling_timesfm.py:TimesFmResidualBlock", "2637": "timesfm/modeling_timesfm.py:TimesFmRMSNorm", "2638": "timesfm/modeling_timesfm.py:TimesFmPositionalEmbedding", "2639": "timesfm/modeling_timesfm.py:simple_eager_attention_forward", "2640": "timesfm/modeling_timesfm.py:TimesFmAttention", "2641": "timesfm/modeling_timesfm.py:TimesFmDecoderLayer", "2642": "timesfm/modeling_timesfm.py:TimesFmPreTrainedModel", "2643": "timesfm/modeling_timesfm.py:TimesFmModel", "2644": "timesfm/modeling_timesfm.py:TimesFmModelForPrediction", "2645": "superpoint/modeling_superpoint.py:remove_keypoints_from_borders", "2646": "superpoint/modeling_superpoint.py:top_k_keypoints", "2647": "superpoint/modeling_superpoint.py:simple_nms", "2648": "superpoint/modeling_superpoint.py:SuperPointKeypointDescriptionOutput", "2649": "superpoint/modeling_superpoint.py:SuperPointConvBlock", "2650": "superpoint/modeling_superpoint.py:SuperPointEncoder", "2651": "superpoint/modeling_superpoint.py:SuperPointInterestPointDecoder", "2652": "superpoint/modeling_superpoint.py:SuperPointDescriptorDecoder", "2653": "superpoint/modeling_superpoint.py:SuperPointPreTrainedModel", "2654": "superpoint/modeling_superpoint.py:SuperPointForKeypointDetection", "2655": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetConvLayer", "2656": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetEmbeddings", "2657": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetShortCut", "2658": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetBasicLayer", "2659": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetBottleNeckLayer", "2660": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetStage", "2661": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetEncoder", "2662": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetPreTrainedModel", "2663": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetBackbone", "2664": "rt_detr/modeling_rt_detr.py:MultiScaleDeformableAttention", "2665": "rt_detr/modeling_rt_detr.py:RTDetrDecoderOutput", "2666": "rt_detr/modeling_rt_detr.py:RTDetrModelOutput", "2667": "rt_detr/modeling_rt_detr.py:RTDetrObjectDetectionOutput", "2668": "rt_detr/modeling_rt_detr.py:_get_clones", "2669": "rt_detr/modeling_rt_detr.py:inverse_sigmoid", "2670": "rt_detr/modeling_rt_detr.py:RTDetrFrozenBatchNorm2d", "2671": "rt_detr/modeling_rt_detr.py:replace_batch_norm", "2672": "rt_detr/modeling_rt_detr.py:get_contrastive_denoising_training_group", "2673": "rt_detr/modeling_rt_detr.py:RTDetrConvEncoder", "2674": "rt_detr/modeling_rt_detr.py:RTDetrConvNormLayer", "2675": "rt_detr/modeling_rt_detr.py:RTDetrEncoderLayer", "2676": "rt_detr/modeling_rt_detr.py:RTDetrRepVggBlock", "2677": "rt_detr/modeling_rt_detr.py:RTDetrCSPRepLayer", "2678": "rt_detr/modeling_rt_detr.py:RTDetrMultiscaleDeformableAttention", "2679": "rt_detr/modeling_rt_detr.py:RTDetrMultiheadAttention", "2680": "rt_detr/modeling_rt_detr.py:RTDetrDecoderLayer", "2681": "rt_detr/modeling_rt_detr.py:RTDetrPreTrainedModel", "2682": "rt_detr/modeling_rt_detr.py:RTDetrEncoder", "2683": "rt_detr/modeling_rt_detr.py:RTDetrHybridEncoder", "2684": "rt_detr/modeling_rt_detr.py:RTDetrDecoder", "2685": "rt_detr/modeling_rt_detr.py:RTDetrMLPPredictionHead", "2686": "rt_detr/modeling_rt_detr.py:RTDetrModel", "2687": "rt_detr/modeling_rt_detr.py:RTDetrForObjectDetection", "2688": "gemma3/modeling_gemma3.py:Gemma3ModelOutputWithPast", "2689": "gemma3/modeling_gemma3.py:Gemma3CausalLMOutputWithPast", "2690": "gemma3/modeling_gemma3.py:Gemma3TextScaledWordEmbedding", "2691": "gemma3/modeling_gemma3.py:Gemma3MLP", "2692": "gemma3/modeling_gemma3.py:Gemma3RMSNorm", "2693": "gemma3/modeling_gemma3.py:Gemma3RotaryEmbedding", "2694": "gemma3/modeling_gemma3.py:rotate_half", "2695": "gemma3/modeling_gemma3.py:apply_rotary_pos_emb", "2696": "gemma3/modeling_gemma3.py:repeat_kv", "2697": "gemma3/modeling_gemma3.py:eager_attention_forward", "2698": "gemma3/modeling_gemma3.py:Gemma3Attention", "2699": "gemma3/modeling_gemma3.py:Gemma3DecoderLayer", "2700": "gemma3/modeling_gemma3.py:Gemma3PreTrainedModel", "2701": "gemma3/modeling_gemma3.py:_bidirectional_window_overlay", "2702": "gemma3/modeling_gemma3.py:Gemma3TextModel", "2703": "gemma3/modeling_gemma3.py:Gemma3ForCausalLM", "2704": "gemma3/modeling_gemma3.py:Gemma3MultiModalProjector", "2705": "gemma3/modeling_gemma3.py:token_type_ids_mask_function", "2706": "gemma3/modeling_gemma3.py:create_causal_mask_mapping", "2707": "gemma3/modeling_gemma3.py:Gemma3Model", "2708": "gemma3/modeling_gemma3.py:Gemma3ForConditionalGeneration", "2709": "gemma3/modeling_gemma3.py:Gemma3ForSequenceClassification", "2710": "gemma3/modeling_gemma3.py:Gemma3TextForSequenceClassification", "2711": "mamba2/modeling_mamba2.py:pad_tensor_by_size", "2712": "mamba2/modeling_mamba2.py:reshape_into_chunks", "2713": "mamba2/modeling_mamba2.py:segment_sum", "2714": "mamba2/modeling_mamba2.py:apply_mask_to_padding_states", "2715": "mamba2/modeling_mamba2.py:Mamba2Cache", "2716": "mamba2/modeling_mamba2.py:MambaRMSNormGated", "2717": "mamba2/modeling_mamba2.py:Mamba2Mixer", "2718": "mamba2/modeling_mamba2.py:Mamba2RMSNorm", "2719": "mamba2/modeling_mamba2.py:Mamba2Block", "2720": "mamba2/modeling_mamba2.py:Mamba2PreTrainedModel", "2721": "mamba2/modeling_mamba2.py:Mamba2Output", "2722": "mamba2/modeling_mamba2.py:Mamba2CausalLMOutput", "2723": "mamba2/modeling_mamba2.py:Mamba2Model", "2724": "mamba2/modeling_mamba2.py:Mamba2ForCausalLM", "2725": "mlcd/modeling_mlcd.py:MLCDMLP", "2726": "mlcd/modeling_mlcd.py:MLCDRotaryEmbedding", "2727": "mlcd/modeling_mlcd.py:MLCDVisionEmbeddings", "2728": "mlcd/modeling_mlcd.py:eager_attention_forward", "2729": "mlcd/modeling_mlcd.py:rotate_half", "2730": "mlcd/modeling_mlcd.py:repeat_kv", "2731": "mlcd/modeling_mlcd.py:apply_rotary_pos_emb_vision", "2732": "mlcd/modeling_mlcd.py:MLCDAttention", "2733": "mlcd/modeling_mlcd.py:MLCDEncoderLayer", "2734": "mlcd/modeling_mlcd.py:MLCDEncoder", "2735": "mlcd/modeling_mlcd.py:MLCDPreTrainedModel", "2736": "mlcd/modeling_mlcd.py:MLCDVisionTransformer", "2737": "mlcd/modeling_mlcd.py:MLCDVisionModel", "2738": "clvp/modeling_clvp.py:contrastive_loss", "2739": "clvp/modeling_clvp.py:clvp_loss", "2740": "clvp/modeling_clvp.py:rotate_half", "2741": "clvp/modeling_clvp.py:apply_rotary_pos_emb", "2742": "clvp/modeling_clvp.py:_pad_extra_bos_eos_tokens", "2743": "clvp/modeling_clvp.py:ClvpEncoderOutput", "2744": "clvp/modeling_clvp.py:ClvpOutput", "2745": "clvp/modeling_clvp.py:ClvpRMSNorm", "2746": "clvp/modeling_clvp.py:ClvpRotaryPositionalEmbedding", "2747": "clvp/modeling_clvp.py:ClvpSelfAttention", "2748": "clvp/modeling_clvp.py:ClvpGatedLinearUnit", "2749": "clvp/modeling_clvp.py:ClvpEncoderMLP", "2750": "clvp/modeling_clvp.py:ClvpEncoderLayer", "2751": "clvp/modeling_clvp.py:ClvpSequenceSummary", "2752": "clvp/modeling_clvp.py:ClvpDecoderMLP", "2753": "clvp/modeling_clvp.py:ClvpDecoderLayer", "2754": "clvp/modeling_clvp.py:ClvpConditioningEncoder", "2755": "clvp/modeling_clvp.py:ClvpPreTrainedModel", "2756": "clvp/modeling_clvp.py:ClvpEncoder", "2757": "clvp/modeling_clvp.py:ClvpDecoder", "2758": "clvp/modeling_clvp.py:ClvpModel", "2759": "clvp/modeling_clvp.py:ClvpForCausalLM", "2760": "clvp/modeling_clvp.py:ClvpModelForConditionalGeneration", "2761": "ovis2/modeling_ovis2.py:Ovis2ModelOutputWithPast", "2762": "ovis2/modeling_ovis2.py:Ovis2CausalLMOutputWithPast", "2763": "ovis2/modeling_ovis2.py:Ovis2RMSNorm", "2764": "ovis2/modeling_ovis2.py:Ovis2VisionMLP", "2765": "ovis2/modeling_ovis2.py:Ovis2VisionEmbeddings", "2766": "ovis2/modeling_ovis2.py:eager_attention_forward", "2767": "ovis2/modeling_ovis2.py:Ovis2VisionAttention", "2768": "ovis2/modeling_ovis2.py:Ovis2MLP", "2769": "ovis2/modeling_ovis2.py:Ovis2Attention", "2770": "ovis2/modeling_ovis2.py:Ovis2VisionEncoderLayer", "2771": "ovis2/modeling_ovis2.py:Ovis2VisionEncoder", "2772": "ovis2/modeling_ovis2.py:Ovis2VisionTransformer", "2773": "ovis2/modeling_ovis2.py:Ovis2VisualEmbeddingTable", "2774": "ovis2/modeling_ovis2.py:Ovis2PreTrainedModel", "2775": "ovis2/modeling_ovis2.py:hard_softmax", "2776": "ovis2/modeling_ovis2.py:Ovis2VisionModel", "2777": "ovis2/modeling_ovis2.py:Ovis2Model", "2778": "ovis2/modeling_ovis2.py:Ovis2ForConditionalGeneration", "2779": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalVisionMLP", "2780": "phi4_multimodal/modeling_phi4_multimodal.py:simple_eager_attention_forward", "2781": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalVisionAttention", "2782": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalVisionEncoderLayer", "2783": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalVisionEncoder", "2784": "phi4_multimodal/modeling_phi4_multimodal.py:variance_scaling_", "2785": "phi4_multimodal/modeling_phi4_multimodal.py:lecun_normal_", "2786": "phi4_multimodal/modeling_phi4_multimodal.py:default_flax_embed_init", "2787": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalVisionPreTrainedModel", "2788": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalVisionEmbeddings", "2789": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalVisionMultiheadAttentionPoolingHead", "2790": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalVisionModel", "2791": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalImageEmbedding", "2792": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioMLP", "2793": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioAttention", "2794": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioDepthWiseSeparableConv1d", "2795": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioGluPointWiseConv", "2796": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioConvModule", "2797": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioConformerEncoderLayer", "2798": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioNemoConvSubsampling", "2799": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioRelativeAttentionBias", "2800": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioMeanVarianceNormLayer", "2801": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioPreTrainedModel", "2802": "phi4_multimodal/modeling_phi4_multimodal.py:unfold_tensor", "2803": "phi4_multimodal/modeling_phi4_multimodal.py:adaptive_enc_mask", "2804": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioModel", "2805": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioEmbedding", "2806": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalRMSNorm", "2807": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalMLP", "2808": "phi4_multimodal/modeling_phi4_multimodal.py:rotate_half", "2809": "phi4_multimodal/modeling_phi4_multimodal.py:repeat_kv", "2810": "phi4_multimodal/modeling_phi4_multimodal.py:eager_attention_forward", "2811": "phi4_multimodal/modeling_phi4_multimodal.py:apply_rotary_pos_emb", "2812": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAttention", "2813": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalDecoderLayer", "2814": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalFeatureEmbedding", "2815": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalPreTrainedModel", "2816": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalRotaryEmbedding", "2817": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalModel", "2818": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalForCausalLM", "2819": "xlm/modeling_xlm.py:create_sinusoidal_embeddings", "2820": "xlm/modeling_xlm.py:get_masks", "2821": "xlm/modeling_xlm.py:XLMSquadHeadOutput", "2822": "xlm/modeling_xlm.py:XLMPoolerStartLogits", "2823": "xlm/modeling_xlm.py:XLMPoolerEndLogits", "2824": "xlm/modeling_xlm.py:XLMPoolerAnswerClass", "2825": "xlm/modeling_xlm.py:XLMSQuADHead", "2826": "xlm/modeling_xlm.py:XLMSequenceSummary", "2827": "xlm/modeling_xlm.py:MultiHeadAttention", "2828": "xlm/modeling_xlm.py:TransformerFFN", "2829": "xlm/modeling_xlm.py:XLMPreTrainedModel", "2830": "xlm/modeling_xlm.py:XLMForQuestionAnsweringOutput", "2831": "xlm/modeling_xlm.py:XLMModel", "2832": "xlm/modeling_xlm.py:XLMPredLayer", "2833": "xlm/modeling_xlm.py:XLMWithLMHeadModel", "2834": "xlm/modeling_xlm.py:XLMForSequenceClassification", "2835": "xlm/modeling_xlm.py:XLMForQuestionAnsweringSimple", "2836": "xlm/modeling_xlm.py:XLMForQuestionAnswering", "2837": "xlm/modeling_xlm.py:XLMForTokenClassification", "2838": "xlm/modeling_xlm.py:XLMForMultipleChoice", "2839": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1RMSNorm", "2840": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1MLP", "2841": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:rotate_half", "2842": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:apply_rotary_pos_emb", "2843": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:repeat_kv", "2844": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:eager_attention_forward", "2845": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1Attention", "2846": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1DecoderLayer", "2847": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1PreTrainedModel", "2848": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1RotaryEmbedding", "2849": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1Model", "2850": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1ForCausalLM", "2851": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1ForSequenceClassification", "2852": "roformer/modeling_roformer.py:RoFormerSinusoidalPositionalEmbedding", "2853": "roformer/modeling_roformer.py:RoFormerEmbeddings", "2854": "roformer/modeling_roformer.py:RoFormerSelfAttention", "2855": "roformer/modeling_roformer.py:RoFormerSelfOutput", "2856": "roformer/modeling_roformer.py:RoFormerAttention", "2857": "roformer/modeling_roformer.py:RoFormerIntermediate", "2858": "roformer/modeling_roformer.py:RoFormerOutput", "2859": "roformer/modeling_roformer.py:RoFormerLayer", "2860": "roformer/modeling_roformer.py:RoFormerEncoder", "2861": "roformer/modeling_roformer.py:RoFormerSequenceSummary", "2862": "roformer/modeling_roformer.py:RoFormerPredictionHeadTransform", "2863": "roformer/modeling_roformer.py:RoFormerLMPredictionHead", "2864": "roformer/modeling_roformer.py:RoFormerOnlyMLMHead", "2865": "roformer/modeling_roformer.py:RoFormerPreTrainedModel", "2866": "roformer/modeling_roformer.py:RoFormerModel", "2867": "roformer/modeling_roformer.py:RoFormerForMaskedLM", "2868": "roformer/modeling_roformer.py:RoFormerForCausalLM", "2869": "roformer/modeling_roformer.py:RoFormerClassificationHead", "2870": "roformer/modeling_roformer.py:RoFormerForSequenceClassification", "2871": "roformer/modeling_roformer.py:RoFormerForMultipleChoice", "2872": "roformer/modeling_roformer.py:RoFormerForTokenClassification", "2873": "roformer/modeling_roformer.py:RoFormerForQuestionAnswering", "2874": "sam3_tracker/modeling_sam3_tracker.py:Sam3TrackerImageSegmentationOutput", "2875": "sam3_tracker/modeling_sam3_tracker.py:Sam3TrackerFeedForward", "2876": "sam3_tracker/modeling_sam3_tracker.py:Sam3TrackerPreTrainedModel", "2877": "sam3_tracker/modeling_sam3_tracker.py:Sam3TrackerPositionalEmbedding", "2878": "sam3_tracker/modeling_sam3_tracker.py:Sam3TrackerMaskEmbedding", "2879": "sam3_tracker/modeling_sam3_tracker.py:Sam3TrackerPromptEncoder", "2880": "sam3_tracker/modeling_sam3_tracker.py:eager_attention_forward", "2881": "sam3_tracker/modeling_sam3_tracker.py:Sam3TrackerAttention", "2882": "sam3_tracker/modeling_sam3_tracker.py:Sam3TrackerTwoWayAttentionBlock", "2883": "sam3_tracker/modeling_sam3_tracker.py:Sam3TrackerTwoWayTransformer", "2884": "sam3_tracker/modeling_sam3_tracker.py:Sam3TrackerLayerNorm", "2885": "sam3_tracker/modeling_sam3_tracker.py:Sam3TrackerMaskDecoder", "2886": "sam3_tracker/modeling_sam3_tracker.py:Sam3TrackerVisionEncoderOutput", "2887": "sam3_tracker/modeling_sam3_tracker.py:Sam3TrackerModel", "2888": "csm/modeling_csm.py:CsmOutputWithPast", "2889": "csm/modeling_csm.py:CsmRMSNorm", "2890": "csm/modeling_csm.py:CsmRotaryEmbedding", "2891": "csm/modeling_csm.py:CsmMLP", "2892": "csm/modeling_csm.py:rotate_half", "2893": "csm/modeling_csm.py:apply_rotary_pos_emb", "2894": "csm/modeling_csm.py:repeat_kv", "2895": "csm/modeling_csm.py:eager_attention_forward", "2896": "csm/modeling_csm.py:CsmAttention", "2897": "csm/modeling_csm.py:CsmDecoderLayer", "2898": "csm/modeling_csm.py:CsmPreTrainedModel", "2899": "csm/modeling_csm.py:CsmDepthDecoderModel", "2900": "csm/modeling_csm.py:CsmCodebooksHead", "2901": "csm/modeling_csm.py:CsmDepthDecoderForCausalLM", "2902": "csm/modeling_csm.py:CsmBackboneModelEmbeddings", "2903": "csm/modeling_csm.py:CsmBackboneModel", "2904": "csm/modeling_csm.py:CsmForConditionalGeneration", "2905": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:GPTNeoXJapanesePreTrainedModel", "2906": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:GPTNeoXJapaneseRotaryEmbedding", "2907": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:rotate_half", "2908": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:apply_rotary_pos_emb", "2909": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:GPTNeoXJapaneseAttention", "2910": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:bias_dropout_add", "2911": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:GPTNeoXJapaneseMLP", "2912": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:GPTNeoXJapaneseLayer", "2913": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:GPTNeoXJapaneseModel", "2914": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:GPTNeoXJapaneseForCausalLM", "2915": "jais2/modeling_jais2.py:Jais2MLP", "2916": "jais2/modeling_jais2.py:rotate_half", "2917": "jais2/modeling_jais2.py:apply_rotary_pos_emb", "2918": "jais2/modeling_jais2.py:repeat_kv", "2919": "jais2/modeling_jais2.py:eager_attention_forward", "2920": "jais2/modeling_jais2.py:Jais2Attention", "2921": "jais2/modeling_jais2.py:Jais2DecoderLayer", "2922": "jais2/modeling_jais2.py:Jais2PreTrainedModel", "2923": "jais2/modeling_jais2.py:Jais2RotaryEmbedding", "2924": "jais2/modeling_jais2.py:Jais2Model", "2925": "jais2/modeling_jais2.py:Jais2ForCausalLM", "2926": "got_ocr2/modeling_got_ocr2.py:GotOcr2MLPBlock", "2927": "got_ocr2/modeling_got_ocr2.py:GotOcr2VisionAttention", "2928": "got_ocr2/modeling_got_ocr2.py:GotOcr2VisionLayer", "2929": "got_ocr2/modeling_got_ocr2.py:GotOcr2PreTrainedModel", "2930": "got_ocr2/modeling_got_ocr2.py:GotOcr2VisionEncoderOutput", "2931": "got_ocr2/modeling_got_ocr2.py:GotOcr2PatchEmbeddings", "2932": "got_ocr2/modeling_got_ocr2.py:GotOcr2LayerNorm", "2933": "got_ocr2/modeling_got_ocr2.py:GotOcr2VisionNeck", "2934": "got_ocr2/modeling_got_ocr2.py:GotOcr2VisionEncoder", "2935": "got_ocr2/modeling_got_ocr2.py:GotOcr2MultiModalProjector", "2936": "got_ocr2/modeling_got_ocr2.py:GotOcr2CausalLMOutputWithPast", "2937": "got_ocr2/modeling_got_ocr2.py:GotOcr2ModelOutputWithPast", "2938": "got_ocr2/modeling_got_ocr2.py:GotOcr2Model", "2939": "got_ocr2/modeling_got_ocr2.py:GotOcr2ForConditionalGeneration", "2940": "oneformer/modeling_oneformer.py:_get_clones", "2941": "oneformer/modeling_oneformer.py:multi_scale_deformable_attention", "2942": "oneformer/modeling_oneformer.py:dice_loss", "2943": "oneformer/modeling_oneformer.py:sigmoid_cross_entropy_loss", "2944": "oneformer/modeling_oneformer.py:pair_wise_dice_loss", "2945": "oneformer/modeling_oneformer.py:pair_wise_sigmoid_cross_entropy_loss", "2946": "oneformer/modeling_oneformer.py:sample_point", "2947": "oneformer/modeling_oneformer.py:OneFormerHungarianMatcher", "2948": "oneformer/modeling_oneformer.py:OneFormerLoss", "2949": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoderOutput", "2950": "oneformer/modeling_oneformer.py:OneFormerPixelDecoderOutput", "2951": "oneformer/modeling_oneformer.py:OneFormerPixelLevelModuleOutput", "2952": "oneformer/modeling_oneformer.py:OneFormerModelOutput", "2953": "oneformer/modeling_oneformer.py:OneFormerForUniversalSegmentationOutput", "2954": "oneformer/modeling_oneformer.py:OneFormerPixelDecoderEncoderMultiscaleDeformableAttention", "2955": "oneformer/modeling_oneformer.py:OneFormerPixelDecoderEncoderLayer", "2956": "oneformer/modeling_oneformer.py:OneFormerPixelDecoderEncoderOnly", "2957": "oneformer/modeling_oneformer.py:OneFormerPixelDecoder", "2958": "oneformer/modeling_oneformer.py:OneFormerPixelLevelModule", "2959": "oneformer/modeling_oneformer.py:OneFormerAttention", "2960": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoderSelfAttentionLayer", "2961": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoderCrossAttentionLayer", "2962": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoderFFNLayer", "2963": "oneformer/modeling_oneformer.py:OneFormerMLPPredictionHead", "2964": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoderLayer", "2965": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoderQueryTransformerDecoder", "2966": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoderQueryTransformerDecoderLayer", "2967": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoderQueryTransformer", "2968": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoder", "2969": "oneformer/modeling_oneformer.py:OneFormerTransformerModule", "2970": "oneformer/modeling_oneformer.py:OneFormerSinePositionEmbedding", "2971": "oneformer/modeling_oneformer.py:PredictionBlock", "2972": "oneformer/modeling_oneformer.py:OneFormerTextMapperAttention", "2973": "oneformer/modeling_oneformer.py:OneFormerTextTransformerDecoderLayer", "2974": "oneformer/modeling_oneformer.py:OneFormerTextContextDecoder", "2975": "oneformer/modeling_oneformer.py:OneFormerTextMLP", "2976": "oneformer/modeling_oneformer.py:OneFormerTextTransformerLayer", "2977": "oneformer/modeling_oneformer.py:OneFormerTextTransformer", "2978": "oneformer/modeling_oneformer.py:OneFormerTextEncoder", "2979": "oneformer/modeling_oneformer.py:OneFormerTextMapper", "2980": "oneformer/modeling_oneformer.py:OneFormerTaskModel", "2981": "oneformer/modeling_oneformer.py:OneFormerPreTrainedModel", "2982": "oneformer/modeling_oneformer.py:OneFormerModel", "2983": "oneformer/modeling_oneformer.py:OneFormerForUniversalSegmentation", "2984": "vit_msn/modeling_vit_msn.py:ViTMSNEmbeddings", "2985": "vit_msn/modeling_vit_msn.py:ViTMSNPatchEmbeddings", "2986": "vit_msn/modeling_vit_msn.py:eager_attention_forward", "2987": "vit_msn/modeling_vit_msn.py:ViTMSNSelfAttention", "2988": "vit_msn/modeling_vit_msn.py:ViTMSNSelfOutput", "2989": "vit_msn/modeling_vit_msn.py:ViTMSNAttention", "2990": "vit_msn/modeling_vit_msn.py:ViTMSNIntermediate", "2991": "vit_msn/modeling_vit_msn.py:ViTMSNOutput", "2992": "vit_msn/modeling_vit_msn.py:ViTMSNLayer", "2993": "vit_msn/modeling_vit_msn.py:ViTMSNEncoder", "2994": "vit_msn/modeling_vit_msn.py:ViTMSNPreTrainedModel", "2995": "vit_msn/modeling_vit_msn.py:ViTMSNModel", "2996": "vit_msn/modeling_vit_msn.py:ViTMSNForImageClassification", "2997": "llava/modeling_llava.py:LlavaModelOutputWithPast", "2998": "llava/modeling_llava.py:LlavaCausalLMOutputWithPast", "2999": "llava/modeling_llava.py:LlavaMultiModalProjector", "3000": "llava/modeling_llava.py:LlavaPreTrainedModel", "3001": "llava/modeling_llava.py:LlavaModel", "3002": "llava/modeling_llava.py:LlavaForConditionalGeneration", "3003": "sam2/modeling_sam2.py:Sam2VisionEncoderOutput", "3004": "sam2/modeling_sam2.py:Sam2ImageSegmentationOutput", "3005": "sam2/modeling_sam2.py:Sam2PatchEmbeddings", "3006": "sam2/modeling_sam2.py:Sam2SinePositionEmbedding", "3007": "sam2/modeling_sam2.py:Sam2VisionNeck", "3008": "sam2/modeling_sam2.py:eager_attention_forward", "3009": "sam2/modeling_sam2.py:do_pool", "3010": "sam2/modeling_sam2.py:Sam2MultiScaleAttention", "3011": "sam2/modeling_sam2.py:Sam2FeedForward", "3012": "sam2/modeling_sam2.py:window_partition", "3013": "sam2/modeling_sam2.py:window_unpartition", "3014": "sam2/modeling_sam2.py:Sam2MultiScaleBlock", "3015": "sam2/modeling_sam2.py:Sam2HieraDetModelOutput", "3016": "sam2/modeling_sam2.py:Sam2PreTrainedModel", "3017": "sam2/modeling_sam2.py:Sam2HieraDetModel", "3018": "sam2/modeling_sam2.py:Sam2VisionModel", "3019": "sam2/modeling_sam2.py:Sam2PositionalEmbedding", "3020": "sam2/modeling_sam2.py:Sam2MaskEmbedding", "3021": "sam2/modeling_sam2.py:Sam2PromptEncoder", "3022": "sam2/modeling_sam2.py:Sam2Attention", "3023": "sam2/modeling_sam2.py:Sam2TwoWayAttentionBlock", "3024": "sam2/modeling_sam2.py:Sam2TwoWayTransformer", "3025": "sam2/modeling_sam2.py:Sam2LayerNorm", "3026": "sam2/modeling_sam2.py:Sam2MaskDecoder", "3027": "sam2/modeling_sam2.py:Sam2Model", "3028": "sam3/modeling_sam3.py:Sam3VisionEncoderOutput", "3029": "sam3/modeling_sam3.py:Sam3GeometryEncoderOutput", "3030": "sam3/modeling_sam3.py:Sam3DETREncoderOutput", "3031": "sam3/modeling_sam3.py:Sam3DETRDecoderOutput", "3032": "sam3/modeling_sam3.py:Sam3MaskDecoderOutput", "3033": "sam3/modeling_sam3.py:Sam3ImageSegmentationOutput", "3034": "sam3/modeling_sam3.py:inverse_sigmoid", "3035": "sam3/modeling_sam3.py:concat_padded_sequences", "3036": "sam3/modeling_sam3.py:box_cxcywh_to_xyxy", "3037": "sam3/modeling_sam3.py:Sam3MLP", "3038": "sam3/modeling_sam3.py:eager_attention_forward", "3039": "sam3/modeling_sam3.py:Sam3Attention", "3040": "sam3/modeling_sam3.py:Sam3ViTRotaryEmbedding", "3041": "sam3/modeling_sam3.py:rotate_pairwise", "3042": "sam3/modeling_sam3.py:apply_rotary_pos_emb_2d", "3043": "sam3/modeling_sam3.py:Sam3ViTRoPEAttention", "3044": "sam3/modeling_sam3.py:Sam3ViTPatchEmbeddings", "3045": "sam3/modeling_sam3.py:Sam3ViTEmbeddings", "3046": "sam3/modeling_sam3.py:window_partition", "3047": "sam3/modeling_sam3.py:window_unpartition", "3048": "sam3/modeling_sam3.py:Sam3ViTLayerScale", "3049": "sam3/modeling_sam3.py:Sam3ViTLayer", "3050": "sam3/modeling_sam3.py:Sam3PreTrainedModel", "3051": "sam3/modeling_sam3.py:Sam3ViTModel", "3052": "sam3/modeling_sam3.py:Sam3SinePositionEmbedding", "3053": "sam3/modeling_sam3.py:Sam3FPNLayer", "3054": "sam3/modeling_sam3.py:Sam3VisionNeck", "3055": "sam3/modeling_sam3.py:Sam3VisionModel", "3056": "sam3/modeling_sam3.py:Sam3GeometryEncoderLayer", "3057": "sam3/modeling_sam3.py:Sam3GeometryEncoder", "3058": "sam3/modeling_sam3.py:Sam3DetrEncoderLayer", "3059": "sam3/modeling_sam3.py:Sam3DetrEncoder", "3060": "sam3/modeling_sam3.py:Sam3DecoderMLP", "3061": "sam3/modeling_sam3.py:Sam3DetrDecoderLayer", "3062": "sam3/modeling_sam3.py:Sam3DetrDecoder", "3063": "sam3/modeling_sam3.py:Sam3DotProductScoring", "3064": "sam3/modeling_sam3.py:Sam3MaskEmbedder", "3065": "sam3/modeling_sam3.py:Sam3PixelDecoder", "3066": "sam3/modeling_sam3.py:Sam3MaskDecoder", "3067": "sam3/modeling_sam3.py:Sam3Model", "3068": "biogpt/modeling_biogpt.py:BioGptLearnedPositionalEmbedding", "3069": "biogpt/modeling_biogpt.py:BioGptScaledWordEmbedding", "3070": "biogpt/modeling_biogpt.py:eager_attention_forward", "3071": "biogpt/modeling_biogpt.py:BioGptAttention", "3072": "biogpt/modeling_biogpt.py:BioGptDecoderLayer", "3073": "biogpt/modeling_biogpt.py:BioGptPreTrainedModel", "3074": "biogpt/modeling_biogpt.py:BioGptModel", "3075": "biogpt/modeling_biogpt.py:BioGptForCausalLM", "3076": "biogpt/modeling_biogpt.py:BioGptForTokenClassification", "3077": "biogpt/modeling_biogpt.py:BioGptForSequenceClassification", "3078": "qwen3_next/modeling_qwen3_next.py:Qwen3NextRMSNormGated", "3079": "qwen3_next/modeling_qwen3_next.py:Qwen3NextDynamicCache", "3080": "qwen3_next/modeling_qwen3_next.py:Qwen3NextRotaryEmbedding", "3081": "qwen3_next/modeling_qwen3_next.py:Qwen3NextRMSNorm", "3082": "qwen3_next/modeling_qwen3_next.py:rotate_half", "3083": "qwen3_next/modeling_qwen3_next.py:apply_rotary_pos_emb", "3084": "qwen3_next/modeling_qwen3_next.py:repeat_kv", "3085": "qwen3_next/modeling_qwen3_next.py:eager_attention_forward", "3086": "qwen3_next/modeling_qwen3_next.py:Qwen3NextAttention", "3087": "qwen3_next/modeling_qwen3_next.py:apply_mask_to_padding_states", "3088": "qwen3_next/modeling_qwen3_next.py:torch_causal_conv1d_update", "3089": "qwen3_next/modeling_qwen3_next.py:l2norm", "3090": "qwen3_next/modeling_qwen3_next.py:torch_chunk_gated_delta_rule", "3091": "qwen3_next/modeling_qwen3_next.py:torch_recurrent_gated_delta_rule", "3092": "qwen3_next/modeling_qwen3_next.py:Qwen3NextGatedDeltaNet", "3093": "qwen3_next/modeling_qwen3_next.py:Qwen3NextMLP", "3094": "qwen3_next/modeling_qwen3_next.py:Qwen3NextExperts", "3095": "qwen3_next/modeling_qwen3_next.py:Qwen3NextTopKRouter", "3096": "qwen3_next/modeling_qwen3_next.py:Qwen3NextSparseMoeBlock", "3097": "qwen3_next/modeling_qwen3_next.py:Qwen3NextDecoderLayer", "3098": "qwen3_next/modeling_qwen3_next.py:Qwen3NextPreTrainedModel", "3099": "qwen3_next/modeling_qwen3_next.py:Qwen3NextModel", "3100": "qwen3_next/modeling_qwen3_next.py:load_balancing_loss_func", "3101": "qwen3_next/modeling_qwen3_next.py:Qwen3NextForCausalLM", "3102": "qwen3_next/modeling_qwen3_next.py:Qwen3NextForSequenceClassification", "3103": "qwen3_next/modeling_qwen3_next.py:Qwen3NextForTokenClassification", "3104": "qwen3_next/modeling_qwen3_next.py:Qwen3NextForQuestionAnswering", "3105": "vitpose/modeling_vitpose.py:VitPoseEstimatorOutput", "3106": "vitpose/modeling_vitpose.py:VitPosePreTrainedModel", "3107": "vitpose/modeling_vitpose.py:flip_back", "3108": "vitpose/modeling_vitpose.py:VitPoseSimpleDecoder", "3109": "vitpose/modeling_vitpose.py:VitPoseClassicDecoder", "3110": "vitpose/modeling_vitpose.py:VitPoseForPoseEstimation", "3111": "superglue/modeling_superglue.py:concat_pairs", "3112": "superglue/modeling_superglue.py:normalize_keypoints", "3113": "superglue/modeling_superglue.py:log_sinkhorn_iterations", "3114": "superglue/modeling_superglue.py:log_optimal_transport", "3115": "superglue/modeling_superglue.py:arange_like", "3116": "superglue/modeling_superglue.py:SuperGlueKeypointMatchingOutput", "3117": "superglue/modeling_superglue.py:SuperGlueMultiLayerPerceptron", "3118": "superglue/modeling_superglue.py:SuperGlueKeypointEncoder", "3119": "superglue/modeling_superglue.py:SuperGlueSelfAttention", "3120": "superglue/modeling_superglue.py:SuperGlueSelfOutput", "3121": "superglue/modeling_superglue.py:SuperGlueAttention", "3122": "superglue/modeling_superglue.py:SuperGlueAttentionalPropagation", "3123": "superglue/modeling_superglue.py:SuperGlueAttentionalGNN", "3124": "superglue/modeling_superglue.py:SuperGlueFinalProjection", "3125": "superglue/modeling_superglue.py:SuperGluePreTrainedModel", "3126": "superglue/modeling_superglue.py:SuperGlueForKeypointMatching", "3127": "colpali/modeling_colpali.py:ColPaliPreTrainedModel", "3128": "colpali/modeling_colpali.py:ColPaliForRetrievalOutput", "3129": "colpali/modeling_colpali.py:ColPaliForRetrieval", "3130": "vivit/modeling_vivit.py:VivitTubeletEmbeddings", "3131": "vivit/modeling_vivit.py:VivitEmbeddings", "3132": "vivit/modeling_vivit.py:eager_attention_forward", "3133": "vivit/modeling_vivit.py:VivitSelfAttention", "3134": "vivit/modeling_vivit.py:VivitSelfOutput", "3135": "vivit/modeling_vivit.py:VivitAttention", "3136": "vivit/modeling_vivit.py:VivitIntermediate", "3137": "vivit/modeling_vivit.py:VivitOutput", "3138": "vivit/modeling_vivit.py:VivitLayer", "3139": "vivit/modeling_vivit.py:VivitEncoder", "3140": "vivit/modeling_vivit.py:VivitPooler", "3141": "vivit/modeling_vivit.py:VivitPreTrainedModel", "3142": "vivit/modeling_vivit.py:VivitModel", "3143": "vivit/modeling_vivit.py:VivitForVideoClassification", "3144": "siglip2/modeling_siglip2.py:Siglip2VisionOutput", "3145": "siglip2/modeling_siglip2.py:Siglip2TextOutput", "3146": "siglip2/modeling_siglip2.py:Siglip2Output", "3147": "siglip2/modeling_siglip2.py:Siglip2VisionEmbeddings", "3148": "siglip2/modeling_siglip2.py:Siglip2TextEmbeddings", "3149": "siglip2/modeling_siglip2.py:eager_attention_forward", "3150": "siglip2/modeling_siglip2.py:Siglip2Attention", "3151": "siglip2/modeling_siglip2.py:Siglip2MLP", "3152": "siglip2/modeling_siglip2.py:Siglip2EncoderLayer", "3153": "siglip2/modeling_siglip2.py:variance_scaling_", "3154": "siglip2/modeling_siglip2.py:lecun_normal_", "3155": "siglip2/modeling_siglip2.py:default_flax_embed_init", "3156": "siglip2/modeling_siglip2.py:Siglip2PreTrainedModel", "3157": "siglip2/modeling_siglip2.py:Siglip2Encoder", "3158": "siglip2/modeling_siglip2.py:Siglip2VisionTransformer", "3159": "siglip2/modeling_siglip2.py:Siglip2TextTransformer", "3160": "siglip2/modeling_siglip2.py:Siglip2TextModel", "3161": "siglip2/modeling_siglip2.py:Siglip2MultiheadAttentionPoolingHead", "3162": "siglip2/modeling_siglip2.py:Siglip2VisionModel", "3163": "siglip2/modeling_siglip2.py:Siglip2Model", "3164": "siglip2/modeling_siglip2.py:Siglip2ForImageClassification", "3165": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeRMSNorm", "3166": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeRotaryEmbedding", "3167": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeMLP", "3168": "qwen2_moe/modeling_qwen2_moe.py:rotate_half", "3169": "qwen2_moe/modeling_qwen2_moe.py:apply_rotary_pos_emb", "3170": "qwen2_moe/modeling_qwen2_moe.py:repeat_kv", "3171": "qwen2_moe/modeling_qwen2_moe.py:eager_attention_forward", "3172": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeAttention", "3173": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeExperts", "3174": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeTopKRouter", "3175": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeSparseMoeBlock", "3176": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeDecoderLayer", "3177": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoePreTrainedModel", "3178": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeModel", "3179": "qwen2_moe/modeling_qwen2_moe.py:load_balancing_loss_func", "3180": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeForCausalLM", "3181": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeForSequenceClassification", "3182": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeForTokenClassification", "3183": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeForQuestionAnswering", "3184": "mra/modeling_mra.py:load_cuda_kernels", "3185": "mra/modeling_mra.py:sparse_max", "3186": "mra/modeling_mra.py:sparse_mask", "3187": "mra/modeling_mra.py:mm_to_sparse", "3188": "mra/modeling_mra.py:sparse_dense_mm", "3189": "mra/modeling_mra.py:transpose_indices", "3190": "mra/modeling_mra.py:MraSampledDenseMatMul", "3191": "mra/modeling_mra.py:MraSparseDenseMatMul", "3192": "mra/modeling_mra.py:MraReduceSum", "3193": "mra/modeling_mra.py:get_low_resolution_logit", "3194": "mra/modeling_mra.py:get_block_idxes", "3195": "mra/modeling_mra.py:mra2_attention", "3196": "mra/modeling_mra.py:MraEmbeddings", "3197": "mra/modeling_mra.py:MraSelfAttention", "3198": "mra/modeling_mra.py:MraSelfOutput", "3199": "mra/modeling_mra.py:MraAttention", "3200": "mra/modeling_mra.py:MraIntermediate", "3201": "mra/modeling_mra.py:MraOutput", "3202": "mra/modeling_mra.py:MraLayer", "3203": "mra/modeling_mra.py:MraEncoder", "3204": "mra/modeling_mra.py:MraPredictionHeadTransform", "3205": "mra/modeling_mra.py:MraLMPredictionHead", "3206": "mra/modeling_mra.py:MraOnlyMLMHead", "3207": "mra/modeling_mra.py:MraPreTrainedModel", "3208": "mra/modeling_mra.py:MraModel", "3209": "mra/modeling_mra.py:MraForMaskedLM", "3210": "mra/modeling_mra.py:MraClassificationHead", "3211": "mra/modeling_mra.py:MraForSequenceClassification", "3212": "mra/modeling_mra.py:MraForMultipleChoice", "3213": "mra/modeling_mra.py:MraForTokenClassification", "3214": "mra/modeling_mra.py:MraForQuestionAnswering", "3215": "paligemma/modeling_paligemma.py:PaligemmaModelOutputWithPast", "3216": "paligemma/modeling_paligemma.py:PaliGemmaCausalLMOutputWithPast", "3217": "paligemma/modeling_paligemma.py:PaliGemmaMultiModalProjector", "3218": "paligemma/modeling_paligemma.py:token_type_ids_mask_function", "3219": "paligemma/modeling_paligemma.py:create_causal_mask_mapping", "3220": "paligemma/modeling_paligemma.py:PaliGemmaPreTrainedModel", "3221": "paligemma/modeling_paligemma.py:PaliGemmaModel", "3222": "paligemma/modeling_paligemma.py:PaliGemmaForConditionalGeneration", "3223": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRProjector", "3224": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRVisionRotaryEmbedding", "3225": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRRotaryEmbedding", "3226": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRMLP", "3227": "paddleocr_vl/modeling_paddleocr_vl.py:repeat_kv", "3228": "paddleocr_vl/modeling_paddleocr_vl.py:eager_attention_forward", "3229": "paddleocr_vl/modeling_paddleocr_vl.py:rotate_half", "3230": "paddleocr_vl/modeling_paddleocr_vl.py:apply_multimodal_rotary_pos_emb", "3231": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRAttention", "3232": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRRMSNorm", "3233": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRDecoderLayer", "3234": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRVLPreTrainedModel", "3235": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRTextModel", "3236": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRVisionModel", "3237": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRVisionEmbeddings", "3238": "paddleocr_vl/modeling_paddleocr_vl.py:apply_rotary_pos_emb_vision", "3239": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRVisionAttention", "3240": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRVisionMLP", "3241": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRVisionEncoderLayer", "3242": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRVisionEncoder", "3243": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRVisionTransformer", "3244": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRVLModelOutputWithPast", "3245": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRVLCausalLMOutputWithPast", "3246": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRVLModel", "3247": "paddleocr_vl/modeling_paddleocr_vl.py:PaddleOCRVLForConditionalGeneration", "3248": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerForPreTrainingOutput", "3249": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerSamePadLayer", "3250": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerPositionalConvEmbedding", "3251": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerRotaryPositionalEmbedding", "3252": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerRelPositionalEmbedding", "3253": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerNoLayerNormConvLayer", "3254": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerLayerNormConvLayer", "3255": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerGroupNormConvLayer", "3256": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerFeatureEncoder", "3257": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerFeatureProjection", "3258": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerFeedForward", "3259": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerConvolutionModule", "3260": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerSelfAttention", "3261": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerEncoderLayer", "3262": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerEncoder", "3263": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerGumbelVectorQuantizer", "3264": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerAdapter", "3265": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerAdapterLayer", "3266": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerPreTrainedModel", "3267": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:_compute_mask_indices", "3268": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerModel", "3269": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerForPreTraining", "3270": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerForCTC", "3271": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerForSequenceClassification", "3272": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerForAudioFrameClassification", "3273": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:AMSoftmaxLoss", "3274": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:TDNNLayer", "3275": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerForXVector", "3276": "lfm2_moe/modeling_lfm2_moe.py:Lfm2MoeRMSNorm", "3277": "lfm2_moe/modeling_lfm2_moe.py:Lfm2MoeRotaryEmbedding", "3278": "lfm2_moe/modeling_lfm2_moe.py:Lfm2MoeMLP", "3279": "lfm2_moe/modeling_lfm2_moe.py:Lfm2MoeExperts", "3280": "lfm2_moe/modeling_lfm2_moe.py:Lfm2MoeSparseMoeBlock", "3281": "lfm2_moe/modeling_lfm2_moe.py:Lfm2MoeHybridConvCache", "3282": "lfm2_moe/modeling_lfm2_moe.py:rotate_half", "3283": "lfm2_moe/modeling_lfm2_moe.py:apply_rotary_pos_emb", "3284": "lfm2_moe/modeling_lfm2_moe.py:repeat_kv", "3285": "lfm2_moe/modeling_lfm2_moe.py:eager_attention_forward", "3286": "lfm2_moe/modeling_lfm2_moe.py:Lfm2MoeAttention", "3287": "lfm2_moe/modeling_lfm2_moe.py:apply_mask_to_padding_states", "3288": "lfm2_moe/modeling_lfm2_moe.py:Lfm2MoeShortConv", "3289": "lfm2_moe/modeling_lfm2_moe.py:Lfm2MoeDecoderLayer", "3290": "lfm2_moe/modeling_lfm2_moe.py:Lfm2MoePreTrainedModel", "3291": "lfm2_moe/modeling_lfm2_moe.py:Lfm2MoeModel", "3292": "lfm2_moe/modeling_lfm2_moe.py:Lfm2MoeForCausalLM", "3293": "swinv2/modeling_swinv2.py:Swinv2EncoderOutput", "3294": "swinv2/modeling_swinv2.py:Swinv2ModelOutput", "3295": "swinv2/modeling_swinv2.py:Swinv2MaskedImageModelingOutput", "3296": "swinv2/modeling_swinv2.py:Swinv2ImageClassifierOutput", "3297": "swinv2/modeling_swinv2.py:window_partition", "3298": "swinv2/modeling_swinv2.py:window_reverse", "3299": "swinv2/modeling_swinv2.py:drop_path", "3300": "swinv2/modeling_swinv2.py:Swinv2DropPath", "3301": "swinv2/modeling_swinv2.py:Swinv2Embeddings", "3302": "swinv2/modeling_swinv2.py:Swinv2PatchEmbeddings", "3303": "swinv2/modeling_swinv2.py:Swinv2PatchMerging", "3304": "swinv2/modeling_swinv2.py:Swinv2SelfAttention", "3305": "swinv2/modeling_swinv2.py:Swinv2SelfOutput", "3306": "swinv2/modeling_swinv2.py:Swinv2Attention", "3307": "swinv2/modeling_swinv2.py:Swinv2Intermediate", "3308": "swinv2/modeling_swinv2.py:Swinv2Output", "3309": "swinv2/modeling_swinv2.py:Swinv2Layer", "3310": "swinv2/modeling_swinv2.py:Swinv2Stage", "3311": "swinv2/modeling_swinv2.py:Swinv2Encoder", "3312": "swinv2/modeling_swinv2.py:Swinv2PreTrainedModel", "3313": "swinv2/modeling_swinv2.py:Swinv2Model", "3314": "swinv2/modeling_swinv2.py:Swinv2ForMaskedImageModeling", "3315": "swinv2/modeling_swinv2.py:Swinv2ForImageClassification", "3316": "swinv2/modeling_swinv2.py:Swinv2Backbone", "3317": "tvp/modeling_tvp.py:TvpVideoGroundingOutput", "3318": "tvp/modeling_tvp.py:TvpLoss", "3319": "tvp/modeling_tvp.py:TvpVisionModel", "3320": "tvp/modeling_tvp.py:TvpVisualInputEmbedding", "3321": "tvp/modeling_tvp.py:TvpTextInputEmbeddings", "3322": "tvp/modeling_tvp.py:TvpAttention", "3323": "tvp/modeling_tvp.py:TvpIntermediate", "3324": "tvp/modeling_tvp.py:TvpOutputLayer", "3325": "tvp/modeling_tvp.py:TvpEncodeLayer", "3326": "tvp/modeling_tvp.py:TvpEncoder", "3327": "tvp/modeling_tvp.py:TvpPooler", "3328": "tvp/modeling_tvp.py:TvpPreTrainedModel", "3329": "tvp/modeling_tvp.py:TvpFrameDownPadPrompter", "3330": "tvp/modeling_tvp.py:TvpFramePadPrompter", "3331": "tvp/modeling_tvp.py:TvpModel", "3332": "tvp/modeling_tvp.py:TvpVideoGroundingHead", "3333": "tvp/modeling_tvp.py:TvpForVideoGrounding", "3334": "splinter/modeling_splinter.py:SplinterEmbeddings", "3335": "splinter/modeling_splinter.py:eager_attention_forward", "3336": "splinter/modeling_splinter.py:SplinterSelfAttention", "3337": "splinter/modeling_splinter.py:SplinterSelfOutput", "3338": "splinter/modeling_splinter.py:SplinterAttention", "3339": "splinter/modeling_splinter.py:SplinterIntermediate", "3340": "splinter/modeling_splinter.py:SplinterOutput", "3341": "splinter/modeling_splinter.py:SplinterLayer", "3342": "splinter/modeling_splinter.py:SplinterEncoder", "3343": "splinter/modeling_splinter.py:SplinterPreTrainedModel", "3344": "splinter/modeling_splinter.py:SplinterModel", "3345": "splinter/modeling_splinter.py:SplinterFullyConnectedLayer", "3346": "splinter/modeling_splinter.py:QuestionAwareSpanSelectionHead", "3347": "splinter/modeling_splinter.py:SplinterForQuestionAnswering", "3348": "splinter/modeling_splinter.py:SplinterForPreTrainingOutput", "3349": "splinter/modeling_splinter.py:SplinterForPreTraining", "3350": "audioflamingo3/modeling_audioflamingo3.py:eager_attention_forward", "3351": "audioflamingo3/modeling_audioflamingo3.py:AudioFlamingo3Attention", "3352": "audioflamingo3/modeling_audioflamingo3.py:AudioFlamingo3EncoderLayer", "3353": "audioflamingo3/modeling_audioflamingo3.py:AudioFlamingo3PreTrainedModel", "3354": "audioflamingo3/modeling_audioflamingo3.py:AudioFlamingo3Encoder", "3355": "audioflamingo3/modeling_audioflamingo3.py:AudioFlamingo3MultiModalProjector", "3356": "audioflamingo3/modeling_audioflamingo3.py:AudioFlamingo3ForConditionalGeneration", "3357": "esm/modeling_esmfold.py:EsmForProteinFoldingOutput", "3358": "esm/modeling_esmfold.py:is_fp16_enabled", "3359": "esm/modeling_esmfold.py:is_deepspeed_initialized", "3360": "esm/modeling_esmfold.py:collate_dense_tensors", "3361": "esm/modeling_esmfold.py:flatten_final_dims", "3362": "esm/modeling_esmfold.py:permute_final_dims", "3363": "esm/modeling_esmfold.py:dict_multimap", "3364": "esm/modeling_esmfold.py:EsmFoldLinear", "3365": "esm/modeling_esmfold.py:EsmFoldLayerNorm", "3366": "esm/modeling_esmfold.py:softmax_no_cast", "3367": "esm/modeling_esmfold.py:EsmFoldAttention", "3368": "esm/modeling_esmfold.py:EsmFoldTriangleAttention", "3369": "esm/modeling_esmfold.py:EsmFoldTriangleMultiplicativeUpdate", "3370": "esm/modeling_esmfold.py:EsmFoldPreTrainedModel", "3371": "esm/modeling_esmfold.py:EsmFoldSelfAttention", "3372": "esm/modeling_esmfold.py:EsmFoldDropout", "3373": "esm/modeling_esmfold.py:EsmFoldSequenceToPair", "3374": "esm/modeling_esmfold.py:EsmFoldPairToSequence", "3375": "esm/modeling_esmfold.py:EsmFoldResidueMLP", "3376": "esm/modeling_esmfold.py:EsmFoldTriangularSelfAttentionBlock", "3377": "esm/modeling_esmfold.py:EsmCategoricalMixture", "3378": "esm/modeling_esmfold.py:categorical_lddt", "3379": "esm/modeling_esmfold.py:get_axial_mask", "3380": "esm/modeling_esmfold.py:EsmFoldRelativePosition", "3381": "esm/modeling_esmfold.py:EsmFoldAngleResnetBlock", "3382": "esm/modeling_esmfold.py:EsmFoldAngleResnet", "3383": "esm/modeling_esmfold.py:EsmFoldInvariantPointAttention", "3384": "esm/modeling_esmfold.py:EsmFoldBackboneUpdate", "3385": "esm/modeling_esmfold.py:EsmFoldStructureModuleTransitionLayer", "3386": "esm/modeling_esmfold.py:EsmFoldStructureModuleTransition", "3387": "esm/modeling_esmfold.py:EsmFoldStructureModule", "3388": "esm/modeling_esmfold.py:EsmFoldingTrunk", "3389": "esm/modeling_esmfold.py:EsmForProteinFolding", "3390": "esm/modeling_esm.py:rotate_half", "3391": "esm/modeling_esm.py:apply_rotary_pos_emb", "3392": "esm/modeling_esm.py:gelu", "3393": "esm/modeling_esm.py:symmetrize", "3394": "esm/modeling_esm.py:average_product_correct", "3395": "esm/modeling_esm.py:RotaryEmbedding", "3396": "esm/modeling_esm.py:EsmContactPredictionHead", "3397": "esm/modeling_esm.py:EsmEmbeddings", "3398": "esm/modeling_esm.py:eager_attention_forward", "3399": "esm/modeling_esm.py:EsmSelfAttention", "3400": "esm/modeling_esm.py:EsmSelfOutput", "3401": "esm/modeling_esm.py:EsmAttention", "3402": "esm/modeling_esm.py:EsmIntermediate", "3403": "esm/modeling_esm.py:EsmOutput", "3404": "esm/modeling_esm.py:EsmLayer", "3405": "esm/modeling_esm.py:EsmEncoder", "3406": "esm/modeling_esm.py:EsmPooler", "3407": "esm/modeling_esm.py:EsmPreTrainedModel", "3408": "esm/modeling_esm.py:EsmModel", "3409": "esm/modeling_esm.py:EsmForMaskedLM", "3410": "esm/modeling_esm.py:EsmLMHead", "3411": "esm/modeling_esm.py:EsmForSequenceClassification", "3412": "esm/modeling_esm.py:EsmForTokenClassification", "3413": "esm/modeling_esm.py:EsmClassificationHead", "3414": "esm/modeling_esm.py:create_position_ids_from_input_ids", "3415": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoLayerNorm", "3416": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoMemoryFuserCXBlock", "3417": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoVisionEncoderOutput", "3418": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoVisionRotaryEmbedding", "3419": "edgetam_video/modeling_edgetam_video.py:eager_attention_forward", "3420": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoAttention", "3421": "edgetam_video/modeling_edgetam_video.py:rotate_pairwise", "3422": "edgetam_video/modeling_edgetam_video.py:apply_rotary_pos_emb_2d_self_attn", "3423": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoRoPESelfAttention", "3424": "edgetam_video/modeling_edgetam_video.py:apply_rotary_pos_emb_2d_cross_attn", "3425": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoRoPECrossAttention", "3426": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoTwoWayAttentionBlock", "3427": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoPositionEmbeddingSine", "3428": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoMemoryFuser", "3429": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoMaskDownSamplerLayer", "3430": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoMaskDownSampler", "3431": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoMemoryEncoder", "3432": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoFeedForward", "3433": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoPositionalEmbedding", "3434": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoPreTrainedModel", "3435": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoInferenceCache", "3436": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoInferenceSession", "3437": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoMemoryAttentionMLP", "3438": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoMemoryAttentionLayer", "3439": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoMemoryAttention", "3440": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoPerceiverMLP", "3441": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoPerceiverAttention", "3442": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoPerceiverEncoderLayer", "3443": "edgetam_video/modeling_edgetam_video.py:window_partition", "3444": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoPerceiverResampler", "3445": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoImageSegmentationOutput", "3446": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoSegmentationOutput", "3447": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoMaskEmbedding", "3448": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoPromptEncoder", "3449": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoTwoWayTransformer", "3450": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoMaskDecoder", "3451": "edgetam_video/modeling_edgetam_video.py:get_1d_sine_pe", "3452": "edgetam_video/modeling_edgetam_video.py:EdgeTamVideoModel", "3453": "xmod/modeling_xmod.py:XmodEmbeddings", "3454": "xmod/modeling_xmod.py:eager_attention_forward", "3455": "xmod/modeling_xmod.py:XmodSelfAttention", "3456": "xmod/modeling_xmod.py:XmodCrossAttention", "3457": "xmod/modeling_xmod.py:XmodSelfOutput", "3458": "xmod/modeling_xmod.py:XmodAttention", "3459": "xmod/modeling_xmod.py:XmodIntermediate", "3460": "xmod/modeling_xmod.py:XmodAdapter", "3461": "xmod/modeling_xmod.py:XmodOutput", "3462": "xmod/modeling_xmod.py:XmodLayer", "3463": "xmod/modeling_xmod.py:XmodEncoder", "3464": "xmod/modeling_xmod.py:XmodPooler", "3465": "xmod/modeling_xmod.py:XmodPreTrainedModel", "3466": "xmod/modeling_xmod.py:XmodModel", "3467": "xmod/modeling_xmod.py:XmodForCausalLM", "3468": "xmod/modeling_xmod.py:XmodForMaskedLM", "3469": "xmod/modeling_xmod.py:XmodLMHead", "3470": "xmod/modeling_xmod.py:XmodForSequenceClassification", "3471": "xmod/modeling_xmod.py:XmodForMultipleChoice", "3472": "xmod/modeling_xmod.py:XmodForTokenClassification", "3473": "xmod/modeling_xmod.py:XmodClassificationHead", "3474": "xmod/modeling_xmod.py:XmodForQuestionAnswering", "3475": "mgp_str/modeling_mgp_str.py:drop_path", "3476": "mgp_str/modeling_mgp_str.py:MgpstrDropPath", "3477": "mgp_str/modeling_mgp_str.py:MgpstrModelOutput", "3478": "mgp_str/modeling_mgp_str.py:MgpstrEmbeddings", "3479": "mgp_str/modeling_mgp_str.py:MgpstrMlp", "3480": "mgp_str/modeling_mgp_str.py:MgpstrAttention", "3481": "mgp_str/modeling_mgp_str.py:MgpstrLayer", "3482": "mgp_str/modeling_mgp_str.py:MgpstrEncoder", "3483": "mgp_str/modeling_mgp_str.py:MgpstrA3Module", "3484": "mgp_str/modeling_mgp_str.py:MgpstrPreTrainedModel", "3485": "mgp_str/modeling_mgp_str.py:MgpstrModel", "3486": "mgp_str/modeling_mgp_str.py:MgpstrForSceneTextRecognition", "3487": "edgetam/modeling_edgetam.py:EdgeTamLayerNorm", "3488": "edgetam/modeling_edgetam.py:EdgeTamVisionEncoderOutput", "3489": "edgetam/modeling_edgetam.py:eager_attention_forward", "3490": "edgetam/modeling_edgetam.py:EdgeTamAttention", "3491": "edgetam/modeling_edgetam.py:EdgeTamTwoWayAttentionBlock", "3492": "edgetam/modeling_edgetam.py:EdgeTamFeedForward", "3493": "edgetam/modeling_edgetam.py:EdgeTamPreTrainedModel", "3494": "edgetam/modeling_edgetam.py:EdgeTamSinePositionEmbedding", "3495": "edgetam/modeling_edgetam.py:EdgeTamVisionNeck", "3496": "edgetam/modeling_edgetam.py:EdgeTamVisionModel", "3497": "edgetam/modeling_edgetam.py:EdgeTamImageSegmentationOutput", "3498": "edgetam/modeling_edgetam.py:EdgeTamPositionalEmbedding", "3499": "edgetam/modeling_edgetam.py:EdgeTamMaskEmbedding", "3500": "edgetam/modeling_edgetam.py:EdgeTamPromptEncoder", "3501": "edgetam/modeling_edgetam.py:EdgeTamTwoWayTransformer", "3502": "edgetam/modeling_edgetam.py:EdgeTamMaskDecoder", "3503": "edgetam/modeling_edgetam.py:EdgeTamModel", "3504": "codegen/modeling_codegen.py:create_sinusoidal_positions", "3505": "codegen/modeling_codegen.py:rotate_every_two", "3506": "codegen/modeling_codegen.py:apply_rotary_pos_emb", "3507": "codegen/modeling_codegen.py:CodeGenAttention", "3508": "codegen/modeling_codegen.py:CodeGenMLP", "3509": "codegen/modeling_codegen.py:CodeGenBlock", "3510": "codegen/modeling_codegen.py:CodeGenPreTrainedModel", "3511": "codegen/modeling_codegen.py:CodeGenModel", "3512": "codegen/modeling_codegen.py:CodeGenForCausalLM", "3513": "zamba/modeling_zamba.py:ZambaRMSNorm", "3514": "zamba/modeling_zamba.py:repeat_kv", "3515": "zamba/modeling_zamba.py:ZambaHybridDynamicCache", "3516": "zamba/modeling_zamba.py:eager_attention_forward", "3517": "zamba/modeling_zamba.py:ZambaAttention", "3518": "zamba/modeling_zamba.py:ZambaMambaMixer", "3519": "zamba/modeling_zamba.py:ZambaMLP", "3520": "zamba/modeling_zamba.py:ZambaAttentionDecoderLayer", "3521": "zamba/modeling_zamba.py:ZambaMambaDecoderLayer", "3522": "zamba/modeling_zamba.py:ZambaHybridLayer", "3523": "zamba/modeling_zamba.py:ZambaPreTrainedModel", "3524": "zamba/modeling_zamba.py:ZambaModel", "3525": "zamba/modeling_zamba.py:ZambaForCausalLM", "3526": "zamba/modeling_zamba.py:ZambaForSequenceClassification", "3527": "speech_to_text/modeling_speech_to_text.py:shift_tokens_right", "3528": "speech_to_text/modeling_speech_to_text.py:Conv1dSubsampler", "3529": "speech_to_text/modeling_speech_to_text.py:Speech2TextSinusoidalPositionalEmbedding", "3530": "speech_to_text/modeling_speech_to_text.py:eager_attention_forward", "3531": "speech_to_text/modeling_speech_to_text.py:Speech2TextAttention", "3532": "speech_to_text/modeling_speech_to_text.py:Speech2TextEncoderLayer", "3533": "speech_to_text/modeling_speech_to_text.py:Speech2TextDecoderLayer", "3534": "speech_to_text/modeling_speech_to_text.py:Speech2TextPreTrainedModel", "3535": "speech_to_text/modeling_speech_to_text.py:Speech2TextEncoder", "3536": "speech_to_text/modeling_speech_to_text.py:Speech2TextDecoder", "3537": "speech_to_text/modeling_speech_to_text.py:Speech2TextModel", "3538": "speech_to_text/modeling_speech_to_text.py:Speech2TextForConditionalGeneration", "3539": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTEmbeddings", "3540": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTPatchEmbeddings", "3541": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:eager_attention_forward", "3542": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTSelfAttention", "3543": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTSelfOutput", "3544": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTAttention", "3545": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTIntermediate", "3546": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTOutput", "3547": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTLayer", "3548": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTEncoder", "3549": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTPreTrainedModel", "3550": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTModel", "3551": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTMLPHead", "3552": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTForAudioClassification", "3553": "seggpt/modeling_seggpt.py:SegGptEncoderOutput", "3554": "seggpt/modeling_seggpt.py:SegGptImageSegmentationOutput", "3555": "seggpt/modeling_seggpt.py:SegGptPatchEmbeddings", "3556": "seggpt/modeling_seggpt.py:SegGptEmbeddings", "3557": "seggpt/modeling_seggpt.py:SegGptAttention", "3558": "seggpt/modeling_seggpt.py:SegGptMlp", "3559": "seggpt/modeling_seggpt.py:drop_path", "3560": "seggpt/modeling_seggpt.py:SegGptDropPath", "3561": "seggpt/modeling_seggpt.py:SegGptLayer", "3562": "seggpt/modeling_seggpt.py:SegGptEncoder", "3563": "seggpt/modeling_seggpt.py:SegGptLayerNorm", "3564": "seggpt/modeling_seggpt.py:SegGptDecoderHead", "3565": "seggpt/modeling_seggpt.py:SegGptDecoder", "3566": "seggpt/modeling_seggpt.py:SegGptPreTrainedModel", "3567": "seggpt/modeling_seggpt.py:SegGptModel", "3568": "seggpt/modeling_seggpt.py:patchify", "3569": "seggpt/modeling_seggpt.py:unpatchify", "3570": "seggpt/modeling_seggpt.py:SegGptLoss", "3571": "seggpt/modeling_seggpt.py:SegGptForImageSegmentation", "3572": "roberta/modeling_roberta.py:RobertaEmbeddings", "3573": "roberta/modeling_roberta.py:eager_attention_forward", "3574": "roberta/modeling_roberta.py:RobertaSelfAttention", "3575": "roberta/modeling_roberta.py:RobertaCrossAttention", "3576": "roberta/modeling_roberta.py:RobertaSelfOutput", "3577": "roberta/modeling_roberta.py:RobertaAttention", "3578": "roberta/modeling_roberta.py:RobertaIntermediate", "3579": "roberta/modeling_roberta.py:RobertaOutput", "3580": "roberta/modeling_roberta.py:RobertaLayer", "3581": "roberta/modeling_roberta.py:RobertaPreTrainedModel", "3582": "roberta/modeling_roberta.py:RobertaEncoder", "3583": "roberta/modeling_roberta.py:RobertaPooler", "3584": "roberta/modeling_roberta.py:RobertaModel", "3585": "roberta/modeling_roberta.py:RobertaForCausalLM", "3586": "roberta/modeling_roberta.py:RobertaForMaskedLM", "3587": "roberta/modeling_roberta.py:RobertaLMHead", "3588": "roberta/modeling_roberta.py:RobertaForSequenceClassification", "3589": "roberta/modeling_roberta.py:RobertaForMultipleChoice", "3590": "roberta/modeling_roberta.py:RobertaForTokenClassification", "3591": "roberta/modeling_roberta.py:RobertaClassificationHead", "3592": "roberta/modeling_roberta.py:RobertaForQuestionAnswering", "3593": "funnel/modeling_funnel.py:FunnelEmbeddings", "3594": "funnel/modeling_funnel.py:FunnelAttentionStructure", "3595": "funnel/modeling_funnel.py:_relative_shift_gather", "3596": "funnel/modeling_funnel.py:FunnelRelMultiheadAttention", "3597": "funnel/modeling_funnel.py:FunnelPositionwiseFFN", "3598": "funnel/modeling_funnel.py:FunnelLayer", "3599": "funnel/modeling_funnel.py:FunnelEncoder", "3600": "funnel/modeling_funnel.py:upsample", "3601": "funnel/modeling_funnel.py:FunnelDecoder", "3602": "funnel/modeling_funnel.py:FunnelDiscriminatorPredictions", "3603": "funnel/modeling_funnel.py:FunnelPreTrainedModel", "3604": "funnel/modeling_funnel.py:FunnelClassificationHead", "3605": "funnel/modeling_funnel.py:FunnelForPreTrainingOutput", "3606": "funnel/modeling_funnel.py:FunnelBaseModel", "3607": "funnel/modeling_funnel.py:FunnelModel", "3608": "funnel/modeling_funnel.py:FunnelForPreTraining", "3609": "funnel/modeling_funnel.py:FunnelForMaskedLM", "3610": "funnel/modeling_funnel.py:FunnelForSequenceClassification", "3611": "funnel/modeling_funnel.py:FunnelForMultipleChoice", "3612": "funnel/modeling_funnel.py:FunnelForTokenClassification", "3613": "funnel/modeling_funnel.py:FunnelForQuestionAnswering", "3614": "qwen3/modeling_qwen3.py:Qwen3RMSNorm", "3615": "qwen3/modeling_qwen3.py:Qwen3MLP", "3616": "qwen3/modeling_qwen3.py:Qwen3RotaryEmbedding", "3617": "qwen3/modeling_qwen3.py:rotate_half", "3618": "qwen3/modeling_qwen3.py:apply_rotary_pos_emb", "3619": "qwen3/modeling_qwen3.py:repeat_kv", "3620": "qwen3/modeling_qwen3.py:eager_attention_forward", "3621": "qwen3/modeling_qwen3.py:Qwen3Attention", "3622": "qwen3/modeling_qwen3.py:Qwen3DecoderLayer", "3623": "qwen3/modeling_qwen3.py:Qwen3PreTrainedModel", "3624": "qwen3/modeling_qwen3.py:Qwen3Model", "3625": "qwen3/modeling_qwen3.py:Qwen3ForCausalLM", "3626": "qwen3/modeling_qwen3.py:Qwen3ForSequenceClassification", "3627": "qwen3/modeling_qwen3.py:Qwen3ForTokenClassification", "3628": "qwen3/modeling_qwen3.py:Qwen3ForQuestionAnswering", "3629": "opt/modeling_opt.py:OPTLearnedPositionalEmbedding", "3630": "opt/modeling_opt.py:eager_attention_forward", "3631": "opt/modeling_opt.py:OPTAttention", "3632": "opt/modeling_opt.py:OPTDecoderLayer", "3633": "opt/modeling_opt.py:OPTPreTrainedModel", "3634": "opt/modeling_opt.py:OPTDecoder", "3635": "opt/modeling_opt.py:OPTModel", "3636": "opt/modeling_opt.py:OPTForCausalLM", "3637": "opt/modeling_opt.py:OPTForSequenceClassification", "3638": "opt/modeling_opt.py:OPTForQuestionAnswering", "3639": "rag/modeling_rag.py:RetrievAugLMMarginOutput", "3640": "rag/modeling_rag.py:RetrievAugLMOutput", "3641": "rag/modeling_rag.py:RagPreTrainedModel", "3642": "rag/modeling_rag.py:RagModel", "3643": "rag/modeling_rag.py:RagSequenceForGeneration", "3644": "rag/modeling_rag.py:RagTokenForGeneration", "3645": "marian/modeling_marian.py:shift_tokens_right", "3646": "marian/modeling_marian.py:MarianSinusoidalPositionalEmbedding", "3647": "marian/modeling_marian.py:eager_attention_forward", "3648": "marian/modeling_marian.py:MarianAttention", "3649": "marian/modeling_marian.py:MarianEncoderLayer", "3650": "marian/modeling_marian.py:MarianDecoderLayer", "3651": "marian/modeling_marian.py:MarianPreTrainedModel", "3652": "marian/modeling_marian.py:MarianEncoder", "3653": "marian/modeling_marian.py:MarianDecoder", "3654": "marian/modeling_marian.py:MarianModel", "3655": "marian/modeling_marian.py:MarianMTModel", "3656": "marian/modeling_marian.py:MarianDecoderWrapper", "3657": "marian/modeling_marian.py:MarianForCausalLM", "3658": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderEmbeddings", "3659": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderMLP", "3660": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderRotaryEmbedding", "3661": "modernbert_decoder/modeling_modernbert_decoder.py:rotate_half", "3662": "modernbert_decoder/modeling_modernbert_decoder.py:apply_rotary_pos_emb", "3663": "modernbert_decoder/modeling_modernbert_decoder.py:eager_attention_forward", "3664": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderAttention", "3665": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderLayer", "3666": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderPredictionHead", "3667": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderPreTrainedModel", "3668": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderModel", "3669": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderForCausalLM", "3670": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderForSequenceClassification", "3671": "bridgetower/modeling_bridgetower.py:BridgeTowerModelOutput", "3672": "bridgetower/modeling_bridgetower.py:BridgeTowerContrastiveOutput", "3673": "bridgetower/modeling_bridgetower.py:BridgeTowerResidualAttention", "3674": "bridgetower/modeling_bridgetower.py:BridgeTowerTransformer", "3675": "bridgetower/modeling_bridgetower.py:BridgeTowerVisionEmbeddings", "3676": "bridgetower/modeling_bridgetower.py:BridgeTowerVisionTransformer", "3677": "bridgetower/modeling_bridgetower.py:BridgeTowerLinkTower", "3678": "bridgetower/modeling_bridgetower.py:BridgeTowerSelfOutput", "3679": "bridgetower/modeling_bridgetower.py:BridgeTowerIntermediate", "3680": "bridgetower/modeling_bridgetower.py:BridgeTowerOutput", "3681": "bridgetower/modeling_bridgetower.py:BridgeTowerPooler", "3682": "bridgetower/modeling_bridgetower.py:eager_attention_forward", "3683": "bridgetower/modeling_bridgetower.py:BridgeTowerSelfAttention", "3684": "bridgetower/modeling_bridgetower.py:BridgeTowerCrossAttention", "3685": "bridgetower/modeling_bridgetower.py:BridgeTowerAttention", "3686": "bridgetower/modeling_bridgetower.py:BridgeTowerBertCrossLayer", "3687": "bridgetower/modeling_bridgetower.py:BridgeTowerTextLayer", "3688": "bridgetower/modeling_bridgetower.py:BridgeTowerTextEncoder", "3689": "bridgetower/modeling_bridgetower.py:BridgeTowerTextEmbeddings", "3690": "bridgetower/modeling_bridgetower.py:BridgeTowerPreTrainedModel", "3691": "bridgetower/modeling_bridgetower.py:BridgeTowerVisionModel", "3692": "bridgetower/modeling_bridgetower.py:BridgeTowerTextModel", "3693": "bridgetower/modeling_bridgetower.py:BridgeTowerModel", "3694": "bridgetower/modeling_bridgetower.py:BridgeTowerPredictionHeadTransform", "3695": "bridgetower/modeling_bridgetower.py:BridgeTowerMLMHead", "3696": "bridgetower/modeling_bridgetower.py:BridgeTowerITMHead", "3697": "bridgetower/modeling_bridgetower.py:BridgeTowerForMaskedLM", "3698": "bridgetower/modeling_bridgetower.py:BridgeTowerForImageAndTextRetrieval", "3699": "bridgetower/modeling_bridgetower.py:BridgeTowerContrastiveHead", "3700": "bridgetower/modeling_bridgetower.py:BridgeTowerForContrastiveLearning", "3701": "nllb_moe/modeling_nllb_moe.py:NllbMoeScaledWordEmbedding", "3702": "nllb_moe/modeling_nllb_moe.py:NllbMoeSinusoidalPositionalEmbedding", "3703": "nllb_moe/modeling_nllb_moe.py:NllbMoeTop2Router", "3704": "nllb_moe/modeling_nllb_moe.py:NllbMoeDenseActDense", "3705": "nllb_moe/modeling_nllb_moe.py:NllbMoeExperts", "3706": "nllb_moe/modeling_nllb_moe.py:NllbMoeSparseMLP", "3707": "nllb_moe/modeling_nllb_moe.py:eager_attention_forward", "3708": "nllb_moe/modeling_nllb_moe.py:NllbMoeAttention", "3709": "nllb_moe/modeling_nllb_moe.py:NllbMoeEncoderLayer", "3710": "nllb_moe/modeling_nllb_moe.py:NllbMoeDecoderLayer", "3711": "nllb_moe/modeling_nllb_moe.py:NllbMoePreTrainedModel", "3712": "nllb_moe/modeling_nllb_moe.py:NllbMoeEncoder", "3713": "nllb_moe/modeling_nllb_moe.py:NllbMoeDecoder", "3714": "nllb_moe/modeling_nllb_moe.py:NllbMoeModel", "3715": "nllb_moe/modeling_nllb_moe.py:load_balancing_loss_func", "3716": "nllb_moe/modeling_nllb_moe.py:shift_tokens_right", "3717": "nllb_moe/modeling_nllb_moe.py:NllbMoeForConditionalGeneration", "3718": "fnet/modeling_fnet.py:_two_dim_matmul", "3719": "fnet/modeling_fnet.py:two_dim_matmul", "3720": "fnet/modeling_fnet.py:fftn", "3721": "fnet/modeling_fnet.py:FNetEmbeddings", "3722": "fnet/modeling_fnet.py:FNetBasicFourierTransform", "3723": "fnet/modeling_fnet.py:FNetBasicOutput", "3724": "fnet/modeling_fnet.py:FNetFourierTransform", "3725": "fnet/modeling_fnet.py:FNetIntermediate", "3726": "fnet/modeling_fnet.py:FNetOutput", "3727": "fnet/modeling_fnet.py:FNetLayer", "3728": "fnet/modeling_fnet.py:FNetEncoder", "3729": "fnet/modeling_fnet.py:FNetPooler", "3730": "fnet/modeling_fnet.py:FNetPredictionHeadTransform", "3731": "fnet/modeling_fnet.py:FNetLMPredictionHead", "3732": "fnet/modeling_fnet.py:FNetOnlyMLMHead", "3733": "fnet/modeling_fnet.py:FNetOnlyNSPHead", "3734": "fnet/modeling_fnet.py:FNetPreTrainingHeads", "3735": "fnet/modeling_fnet.py:FNetPreTrainedModel", "3736": "fnet/modeling_fnet.py:FNetForPreTrainingOutput", "3737": "fnet/modeling_fnet.py:FNetModel", "3738": "fnet/modeling_fnet.py:FNetForPreTraining", "3739": "fnet/modeling_fnet.py:FNetForMaskedLM", "3740": "fnet/modeling_fnet.py:FNetForNextSentencePrediction", "3741": "fnet/modeling_fnet.py:FNetForSequenceClassification", "3742": "fnet/modeling_fnet.py:FNetForMultipleChoice", "3743": "fnet/modeling_fnet.py:FNetForTokenClassification", "3744": "fnet/modeling_fnet.py:FNetForQuestionAnswering", "3745": "musicgen/modeling_musicgen.py:MusicgenUnconditionalInput", "3746": "musicgen/modeling_musicgen.py:shift_tokens_right", "3747": "musicgen/modeling_musicgen.py:MusicgenSinusoidalPositionalEmbedding", "3748": "musicgen/modeling_musicgen.py:eager_attention_forward", "3749": "musicgen/modeling_musicgen.py:MusicgenAttention", "3750": "musicgen/modeling_musicgen.py:MusicgenDecoderLayer", "3751": "musicgen/modeling_musicgen.py:MusicgenPreTrainedModel", "3752": "musicgen/modeling_musicgen.py:MusicgenDecoder", "3753": "musicgen/modeling_musicgen.py:MusicgenModel", "3754": "musicgen/modeling_musicgen.py:MusicgenForCausalLM", "3755": "musicgen/modeling_musicgen.py:MusicgenForConditionalGeneration", "3756": "align/modeling_align.py:AlignVisionModelOutput", "3757": "align/modeling_align.py:AlignTextModelOutput", "3758": "align/modeling_align.py:AlignOutput", "3759": "align/modeling_align.py:contrastive_loss", "3760": "align/modeling_align.py:align_loss", "3761": "align/modeling_align.py:round_filters", "3762": "align/modeling_align.py:correct_pad", "3763": "align/modeling_align.py:AlignVisionEmbeddings", "3764": "align/modeling_align.py:AlignVisionDepthwiseConv2d", "3765": "align/modeling_align.py:AlignVisionExpansionLayer", "3766": "align/modeling_align.py:AlignVisionDepthwiseLayer", "3767": "align/modeling_align.py:AlignVisionSqueezeExciteLayer", "3768": "align/modeling_align.py:AlignVisionFinalBlockLayer", "3769": "align/modeling_align.py:AlignVisionBlock", "3770": "align/modeling_align.py:AlignVisionEncoder", "3771": "align/modeling_align.py:AlignTextEmbeddings", "3772": "align/modeling_align.py:eager_attention_forward", "3773": "align/modeling_align.py:AlignTextSelfAttention", "3774": "align/modeling_align.py:AlignTextSelfOutput", "3775": "align/modeling_align.py:AlignTextAttention", "3776": "align/modeling_align.py:AlignTextIntermediate", "3777": "align/modeling_align.py:AlignTextOutput", "3778": "align/modeling_align.py:AlignTextLayer", "3779": "align/modeling_align.py:AlignTextEncoder", "3780": "align/modeling_align.py:AlignTextPooler", "3781": "align/modeling_align.py:AlignPreTrainedModel", "3782": "align/modeling_align.py:AlignTextModel", "3783": "align/modeling_align.py:AlignVisionModel", "3784": "align/modeling_align.py:AlignModel", "3785": "blip/modeling_blip_text.py:BlipTextEmbeddings", "3786": "blip/modeling_blip_text.py:BlipTextSelfAttention", "3787": "blip/modeling_blip_text.py:BlipTextSelfOutput", "3788": "blip/modeling_blip_text.py:BlipTextAttention", "3789": "blip/modeling_blip_text.py:BlipTextIntermediate", "3790": "blip/modeling_blip_text.py:BlipTextOutput", "3791": "blip/modeling_blip_text.py:BlipTextLayer", "3792": "blip/modeling_blip_text.py:BlipTextEncoder", "3793": "blip/modeling_blip_text.py:BlipTextPooler", "3794": "blip/modeling_blip_text.py:BlipTextPredictionHeadTransform", "3795": "blip/modeling_blip_text.py:BlipTextLMPredictionHead", "3796": "blip/modeling_blip_text.py:BlipTextOnlyMLMHead", "3797": "blip/modeling_blip_text.py:BlipTextPreTrainedModel", "3798": "blip/modeling_blip_text.py:BlipTextModel", "3799": "blip/modeling_blip_text.py:BlipTextLMHeadModel", "3800": "blip/modeling_blip.py:contrastive_loss", "3801": "blip/modeling_blip.py:blip_loss", "3802": "blip/modeling_blip.py:BlipForConditionalGenerationModelOutput", "3803": "blip/modeling_blip.py:BlipTextVisionModelOutput", "3804": "blip/modeling_blip.py:BlipImageTextMatchingModelOutput", "3805": "blip/modeling_blip.py:BlipOutput", "3806": "blip/modeling_blip.py:BlipVisionEmbeddings", "3807": "blip/modeling_blip.py:BlipTextEmbeddings", "3808": "blip/modeling_blip.py:BlipAttention", "3809": "blip/modeling_blip.py:BlipMLP", "3810": "blip/modeling_blip.py:BlipEncoderLayer", "3811": "blip/modeling_blip.py:BlipPreTrainedModel", "3812": "blip/modeling_blip.py:BlipEncoder", "3813": "blip/modeling_blip.py:BlipVisionModel", "3814": "blip/modeling_blip.py:BlipModel", "3815": "blip/modeling_blip.py:BlipForConditionalGeneration", "3816": "blip/modeling_blip.py:BlipForQuestionAnswering", "3817": "blip/modeling_blip.py:BlipForImageTextRetrieval", "3818": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboEncoderOutput", "3819": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboDecoderOutput", "3820": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboObjectDetectionOutput", "3821": "omdet_turbo/modeling_omdet_turbo.py:MultiScaleDeformableAttention", "3822": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboLRUCache", "3823": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboLanguageBackbone", "3824": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboVisionBackbone", "3825": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboMultiscaleDeformableAttention", "3826": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboConvNormLayer", "3827": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboRepVggBlock", "3828": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboCSPRepLayer", "3829": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboMultiheadAttention", "3830": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboEncoderLayer", "3831": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboEncoder", "3832": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboHybridEncoder", "3833": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboMLPWithDropout", "3834": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboMLP", "3835": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboResidualLayer", "3836": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboTaskEncoder", "3837": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboDeformableTransformerDecoderLayer", "3838": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboPreTrainedModel", "3839": "omdet_turbo/modeling_omdet_turbo.py:_cosine_similarity_scaled", "3840": "omdet_turbo/modeling_omdet_turbo.py:get_class_similarity", "3841": "omdet_turbo/modeling_omdet_turbo.py:_inverse_sigmoid", "3842": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboDecoder", "3843": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboForObjectDetection", "3844": "vitmatte/modeling_vitmatte.py:ImageMattingOutput", "3845": "vitmatte/modeling_vitmatte.py:VitMattePreTrainedModel", "3846": "vitmatte/modeling_vitmatte.py:VitMatteBasicConv3x3", "3847": "vitmatte/modeling_vitmatte.py:VitMatteConvStream", "3848": "vitmatte/modeling_vitmatte.py:VitMatteFusionBlock", "3849": "vitmatte/modeling_vitmatte.py:VitMatteHead", "3850": "vitmatte/modeling_vitmatte.py:VitMatteDetailCaptureModule", "3851": "vitmatte/modeling_vitmatte.py:VitMatteForImageMatting", "3852": "table_transformer/modeling_table_transformer.py:TableTransformerDecoderOutput", "3853": "table_transformer/modeling_table_transformer.py:TableTransformerModelOutput", "3854": "table_transformer/modeling_table_transformer.py:TableTransformerObjectDetectionOutput", "3855": "table_transformer/modeling_table_transformer.py:TableTransformerFrozenBatchNorm2d", "3856": "table_transformer/modeling_table_transformer.py:replace_batch_norm", "3857": "table_transformer/modeling_table_transformer.py:TableTransformerConvEncoder", "3858": "table_transformer/modeling_table_transformer.py:TableTransformerConvModel", "3859": "table_transformer/modeling_table_transformer.py:TableTransformerSinePositionEmbedding", "3860": "table_transformer/modeling_table_transformer.py:TableTransformerLearnedPositionEmbedding", "3861": "table_transformer/modeling_table_transformer.py:build_position_encoding", "3862": "table_transformer/modeling_table_transformer.py:TableTransformerAttention", "3863": "table_transformer/modeling_table_transformer.py:TableTransformerEncoderLayer", "3864": "table_transformer/modeling_table_transformer.py:TableTransformerDecoderLayer", "3865": "table_transformer/modeling_table_transformer.py:TableTransformerPreTrainedModel", "3866": "table_transformer/modeling_table_transformer.py:TableTransformerEncoder", "3867": "table_transformer/modeling_table_transformer.py:TableTransformerDecoder", "3868": "table_transformer/modeling_table_transformer.py:TableTransformerModel", "3869": "table_transformer/modeling_table_transformer.py:TableTransformerForObjectDetection", "3870": "table_transformer/modeling_table_transformer.py:TableTransformerMLPPredictionHead", "3871": "granite_speech/modeling_granite_speech.py:GraniteSpeechCausalLMOutputWithPast", "3872": "granite_speech/modeling_granite_speech.py:GraniteSpeechEncoderProjector", "3873": "granite_speech/modeling_granite_speech.py:GraniteSpeechConformerFeedForward", "3874": "granite_speech/modeling_granite_speech.py:GraniteSpeechConformerAttention", "3875": "granite_speech/modeling_granite_speech.py:GraniteSpeechConformerDepthWiseConv1d", "3876": "granite_speech/modeling_granite_speech.py:GraniteSpeechConformerConvModule", "3877": "granite_speech/modeling_granite_speech.py:GraniteSpeechConformerBlock", "3878": "granite_speech/modeling_granite_speech.py:GraniteSpeechCTCEncoder", "3879": "granite_speech/modeling_granite_speech.py:GraniteSpeechPreTrainedModel", "3880": "granite_speech/modeling_granite_speech.py:GraniteSpeechForConditionalGeneration", "3881": "efficientnet/modeling_efficientnet.py:round_filters", "3882": "efficientnet/modeling_efficientnet.py:correct_pad", "3883": "efficientnet/modeling_efficientnet.py:EfficientNetEmbeddings", "3884": "efficientnet/modeling_efficientnet.py:EfficientNetDepthwiseConv2d", "3885": "efficientnet/modeling_efficientnet.py:EfficientNetExpansionLayer", "3886": "efficientnet/modeling_efficientnet.py:EfficientNetDepthwiseLayer", "3887": "efficientnet/modeling_efficientnet.py:EfficientNetSqueezeExciteLayer", "3888": "efficientnet/modeling_efficientnet.py:EfficientNetFinalBlockLayer", "3889": "efficientnet/modeling_efficientnet.py:EfficientNetBlock", "3890": "efficientnet/modeling_efficientnet.py:EfficientNetEncoder", "3891": "efficientnet/modeling_efficientnet.py:EfficientNetPreTrainedModel", "3892": "efficientnet/modeling_efficientnet.py:EfficientNetModel", "3893": "efficientnet/modeling_efficientnet.py:EfficientNetForImageClassification", "3894": "vits/modeling_vits.py:VitsModelOutput", "3895": "vits/modeling_vits.py:VitsTextEncoderOutput", "3896": "vits/modeling_vits.py:fused_add_tanh_sigmoid_multiply", "3897": "vits/modeling_vits.py:_unconstrained_rational_quadratic_spline", "3898": "vits/modeling_vits.py:_rational_quadratic_spline", "3899": "vits/modeling_vits.py:VitsWaveNet", "3900": "vits/modeling_vits.py:VitsPosteriorEncoder", "3901": "vits/modeling_vits.py:HifiGanResidualBlock", "3902": "vits/modeling_vits.py:VitsHifiGan", "3903": "vits/modeling_vits.py:VitsResidualCouplingLayer", "3904": "vits/modeling_vits.py:VitsResidualCouplingBlock", "3905": "vits/modeling_vits.py:VitsDilatedDepthSeparableConv", "3906": "vits/modeling_vits.py:VitsConvFlow", "3907": "vits/modeling_vits.py:VitsElementwiseAffine", "3908": "vits/modeling_vits.py:VitsStochasticDurationPredictor", "3909": "vits/modeling_vits.py:VitsDurationPredictor", "3910": "vits/modeling_vits.py:VitsAttention", "3911": "vits/modeling_vits.py:VitsFeedForward", "3912": "vits/modeling_vits.py:VitsEncoderLayer", "3913": "vits/modeling_vits.py:VitsEncoder", "3914": "vits/modeling_vits.py:VitsTextEncoder", "3915": "vits/modeling_vits.py:VitsPreTrainedModel", "3916": "vits/modeling_vits.py:VitsModel", "3917": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeTextRotaryEmbedding", "3918": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:repeat_kv", "3919": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:eager_attention_forward", "3920": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:rotate_half_text", "3921": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:apply_rotary_pos_emb", "3922": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeTextAttention", "3923": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeRMSNorm", "3924": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeMLP", "3925": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeMoeStatics", "3926": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeMoeTopKRouter", "3927": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeMoeExperts", "3928": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeSparseMoeBlock", "3929": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeMoeBlock", "3930": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeDecoderLayer", "3931": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoePreTrainedModel", "3932": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeTextModel", "3933": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5VLVisionMLP", "3934": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoePatchEmbed", "3935": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeVisionRotaryEmbedding", "3936": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:rotate_half", "3937": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:apply_rotary_pos_emb_vision", "3938": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeVisionAttention", "3939": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeVisionBlock", "3940": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeVisionTransformerPretrainedModel", "3941": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeVisionMLP", "3942": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeVariableResolutionResamplerModel", "3943": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeModel", "3944": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:load_balancing_loss_func", "3945": "ernie4_5_vl_moe/modeling_ernie4_5_vl_moe.py:Ernie4_5_VL_MoeForConditionalGeneration", "3946": "ministral/modeling_ministral.py:MinistralMLP", "3947": "ministral/modeling_ministral.py:rotate_half", "3948": "ministral/modeling_ministral.py:apply_rotary_pos_emb", "3949": "ministral/modeling_ministral.py:repeat_kv", "3950": "ministral/modeling_ministral.py:eager_attention_forward", "3951": "ministral/modeling_ministral.py:MinistralAttention", "3952": "ministral/modeling_ministral.py:MinistralRMSNorm", "3953": "ministral/modeling_ministral.py:MinistralDecoderLayer", "3954": "ministral/modeling_ministral.py:MinistralPreTrainedModel", "3955": "ministral/modeling_ministral.py:MinistralRotaryEmbedding", "3956": "ministral/modeling_ministral.py:MinistralModel", "3957": "ministral/modeling_ministral.py:MinistralForCausalLM", "3958": "ministral/modeling_ministral.py:MinistralForSequenceClassification", "3959": "ministral/modeling_ministral.py:MinistralForTokenClassification", "3960": "ministral/modeling_ministral.py:MinistralForQuestionAnswering", "3961": "qwen2_5_omni/modeling_qwen2_5_omni.py:kaiser_sinc_filter1d", "3962": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniPreTrainedModel", "3963": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniPreTrainedModelForConditionalGeneration", "3964": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniThinkerCausalLMOutputWithPast", "3965": "qwen2_5_omni/modeling_qwen2_5_omni.py:repeat_kv", "3966": "qwen2_5_omni/modeling_qwen2_5_omni.py:eager_attention_forward", "3967": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniAudioAttention", "3968": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniAudioEncoderLayer", "3969": "qwen2_5_omni/modeling_qwen2_5_omni.py:SinusoidsPositionEmbedding", "3970": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniAudioEncoder", "3971": "qwen2_5_omni/modeling_qwen2_5_omni.py:rotate_half", "3972": "qwen2_5_omni/modeling_qwen2_5_omni.py:apply_rotary_pos_emb_vision", "3973": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniVisionAttention", "3974": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniMLP", "3975": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniVisionBlock", "3976": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5_VisionRotaryEmbedding", "3977": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5_VisionPatchEmbed", "3978": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniPatchMerger", "3979": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniVisionEncoder", "3980": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniRotaryEmbedding", "3981": "qwen2_5_omni/modeling_qwen2_5_omni.py:apply_multimodal_rotary_pos_emb", "3982": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniAttention", "3983": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2MLP", "3984": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniDecoderLayer", "3985": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniThinkerTextModel", "3986": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniThinkerForConditionalGeneration", "3987": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniTalkerCausalLMOutputWithPast", "3988": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniTalkerModel", "3989": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniTalkerForConditionalGeneration", "3990": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniDiTRotaryEmbedding", "3991": "qwen2_5_omni/modeling_qwen2_5_omni.py:TimeDelayNetBlock", "3992": "qwen2_5_omni/modeling_qwen2_5_omni.py:Res2NetBlock", "3993": "qwen2_5_omni/modeling_qwen2_5_omni.py:SqueezeExcitationBlock", "3994": "qwen2_5_omni/modeling_qwen2_5_omni.py:AttentiveStatisticsPooling", "3995": "qwen2_5_omni/modeling_qwen2_5_omni.py:SqueezeExcitationRes2NetBlock", "3996": "qwen2_5_omni/modeling_qwen2_5_omni.py:ECAPA_TimeDelayNet", "3997": "qwen2_5_omni/modeling_qwen2_5_omni.py:DiTInputEmbedding", "3998": "qwen2_5_omni/modeling_qwen2_5_omni.py:DiTCodecEmbedding", "3999": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5_OmniAdaLayerNormZero", "4000": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5_OmniAdaLayerNormZero_Final", "4001": "qwen2_5_omni/modeling_qwen2_5_omni.py:DiTMLP", "4002": "qwen2_5_omni/modeling_qwen2_5_omni.py:apply_rotary_pos_emb", "4003": "qwen2_5_omni/modeling_qwen2_5_omni.py:DiTAttention", "4004": "qwen2_5_omni/modeling_qwen2_5_omni.py:SinusPositionEmbedding", "4005": "qwen2_5_omni/modeling_qwen2_5_omni.py:DiTTimestepEmbedding", "4006": "qwen2_5_omni/modeling_qwen2_5_omni.py:DiTDecoderLayer", "4007": "qwen2_5_omni/modeling_qwen2_5_omni.py:SnakeBeta", "4008": "qwen2_5_omni/modeling_qwen2_5_omni.py:UpSample1d", "4009": "qwen2_5_omni/modeling_qwen2_5_omni.py:DownSample1d", "4010": "qwen2_5_omni/modeling_qwen2_5_omni.py:TorchActivation1d", "4011": "qwen2_5_omni/modeling_qwen2_5_omni.py:AMPBlock", "4012": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniToken2WavBigVGANModel", "4013": "qwen2_5_omni/modeling_qwen2_5_omni.py:RungeKutta4ODESolver", "4014": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniToken2WavDiTModel", "4015": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniToken2WavModel", "4016": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniForConditionalGeneration", "4017": "pegasus/modeling_pegasus.py:shift_tokens_right", "4018": "pegasus/modeling_pegasus.py:PegasusSinusoidalPositionalEmbedding", "4019": "pegasus/modeling_pegasus.py:eager_attention_forward", "4020": "pegasus/modeling_pegasus.py:PegasusAttention", "4021": "pegasus/modeling_pegasus.py:PegasusEncoderLayer", "4022": "pegasus/modeling_pegasus.py:PegasusDecoderLayer", "4023": "pegasus/modeling_pegasus.py:PegasusPreTrainedModel", "4024": "pegasus/modeling_pegasus.py:PegasusEncoder", "4025": "pegasus/modeling_pegasus.py:PegasusDecoder", "4026": "pegasus/modeling_pegasus.py:PegasusModel", "4027": "pegasus/modeling_pegasus.py:PegasusForConditionalGeneration", "4028": "pegasus/modeling_pegasus.py:PegasusDecoderWrapper", "4029": "pegasus/modeling_pegasus.py:PegasusForCausalLM", "4030": "blt/modeling_blt.py:BltMLP", "4031": "blt/modeling_blt.py:BltRMSNorm", "4032": "blt/modeling_blt.py:BltRotaryEmbedding", "4033": "blt/modeling_blt.py:BltTransformerLayer", "4034": "blt/modeling_blt.py:repeat_kv", "4035": "blt/modeling_blt.py:eager_attention_forward", "4036": "blt/modeling_blt.py:rotate_half", "4037": "blt/modeling_blt.py:apply_rotary_pos_emb", "4038": "blt/modeling_blt.py:BltSelfAttention", "4039": "blt/modeling_blt.py:BltCrossAttention", "4040": "blt/modeling_blt.py:BltPreTrainedModel", "4041": "blt/modeling_blt.py:BltLocalEncoder", "4042": "blt/modeling_blt.py:BltLocalDecoder", "4043": "blt/modeling_blt.py:BltGlobalTransformer", "4044": "blt/modeling_blt.py:process_patch_lengths", "4045": "blt/modeling_blt.py:BltPatcher", "4046": "blt/modeling_blt.py:rolling_polynomial_hash", "4047": "blt/modeling_blt.py:byte_group_hash_function", "4048": "blt/modeling_blt.py:compute_hash_embeddings", "4049": "blt/modeling_blt.py:_prepare_patch_cross_attention_mask", "4050": "blt/modeling_blt.py:BltModel", "4051": "blt/modeling_blt.py:BltForCausalLM", "4052": "stablelm/modeling_stablelm.py:StableLmRotaryEmbedding", "4053": "stablelm/modeling_stablelm.py:rotate_half", "4054": "stablelm/modeling_stablelm.py:apply_rotary_pos_emb", "4055": "stablelm/modeling_stablelm.py:StableLmMLP", "4056": "stablelm/modeling_stablelm.py:StableLmLayerNormPerHead", "4057": "stablelm/modeling_stablelm.py:repeat_kv", "4058": "stablelm/modeling_stablelm.py:eager_attention_forward", "4059": "stablelm/modeling_stablelm.py:StableLmAttention", "4060": "stablelm/modeling_stablelm.py:StableLmDecoderLayer", "4061": "stablelm/modeling_stablelm.py:StableLmPreTrainedModel", "4062": "stablelm/modeling_stablelm.py:StableLmModel", "4063": "stablelm/modeling_stablelm.py:StableLmForCausalLM", "4064": "stablelm/modeling_stablelm.py:StableLmForSequenceClassification", "4065": "stablelm/modeling_stablelm.py:StableLmForTokenClassification", "4066": "dinov2/modeling_dinov2.py:Dinov2Embeddings", "4067": "dinov2/modeling_dinov2.py:Dinov2PatchEmbeddings", "4068": "dinov2/modeling_dinov2.py:eager_attention_forward", "4069": "dinov2/modeling_dinov2.py:Dinov2SelfAttention", "4070": "dinov2/modeling_dinov2.py:Dinov2SelfOutput", "4071": "dinov2/modeling_dinov2.py:Dinov2Attention", "4072": "dinov2/modeling_dinov2.py:Dinov2LayerScale", "4073": "dinov2/modeling_dinov2.py:drop_path", "4074": "dinov2/modeling_dinov2.py:Dinov2DropPath", "4075": "dinov2/modeling_dinov2.py:Dinov2MLP", "4076": "dinov2/modeling_dinov2.py:Dinov2SwiGLUFFN", "4077": "dinov2/modeling_dinov2.py:Dinov2Layer", "4078": "dinov2/modeling_dinov2.py:Dinov2Encoder", "4079": "dinov2/modeling_dinov2.py:Dinov2PreTrainedModel", "4080": "dinov2/modeling_dinov2.py:Dinov2Model", "4081": "dinov2/modeling_dinov2.py:Dinov2ForImageClassification", "4082": "dinov2/modeling_dinov2.py:Dinov2Backbone", "4083": "glm4v/modeling_glm4v.py:Glm4vRMSNorm", "4084": "glm4v/modeling_glm4v.py:Glm4VisionMlp", "4085": "glm4v/modeling_glm4v.py:Glm4vVisionPatchEmbed", "4086": "glm4v/modeling_glm4v.py:Glm4vVisionRotaryEmbedding", "4087": "glm4v/modeling_glm4v.py:Glm4vVisionPatchMerger", "4088": "glm4v/modeling_glm4v.py:Glm4vVisionEmbeddings", "4089": "glm4v/modeling_glm4v.py:rotate_half", "4090": "glm4v/modeling_glm4v.py:apply_rotary_pos_emb_vision", "4091": "glm4v/modeling_glm4v.py:repeat_kv", "4092": "glm4v/modeling_glm4v.py:eager_attention_forward", "4093": "glm4v/modeling_glm4v.py:Glm4vVisionAttention", "4094": "glm4v/modeling_glm4v.py:Glm4vVisionBlock", "4095": "glm4v/modeling_glm4v.py:Glm4vTextRotaryEmbedding", "4096": "glm4v/modeling_glm4v.py:rotate_half_llm", "4097": "glm4v/modeling_glm4v.py:apply_multimodal_rotary_pos_emb", "4098": "glm4v/modeling_glm4v.py:Glm4vTextAttention", "4099": "glm4v/modeling_glm4v.py:Glm4vTextMLP", "4100": "glm4v/modeling_glm4v.py:Glm4vTextDecoderLayer", "4101": "glm4v/modeling_glm4v.py:Glm4vModelOutputWithPast", "4102": "glm4v/modeling_glm4v.py:Glm4vPreTrainedModel", "4103": "glm4v/modeling_glm4v.py:Glm4vVisionModel", "4104": "glm4v/modeling_glm4v.py:Glm4vTextModel", "4105": "glm4v/modeling_glm4v.py:Glm4vModel", "4106": "glm4v/modeling_glm4v.py:Glm4vCausalLMOutputWithPast", "4107": "glm4v/modeling_glm4v.py:Glm4vForConditionalGeneration", "4108": "idefics3/modeling_idefics3.py:Idefics3BaseModelOutputWithPast", "4109": "idefics3/modeling_idefics3.py:Idefics3CausalLMOutputWithPast", "4110": "idefics3/modeling_idefics3.py:Idefics3VisionEmbeddings", "4111": "idefics3/modeling_idefics3.py:eager_attention_forward", "4112": "idefics3/modeling_idefics3.py:Idefics3VisionAttention", "4113": "idefics3/modeling_idefics3.py:Idefics3VisionMLP", "4114": "idefics3/modeling_idefics3.py:Idefics3SimpleMLP", "4115": "idefics3/modeling_idefics3.py:Idefics3EncoderLayer", "4116": "idefics3/modeling_idefics3.py:Idefics3Encoder", "4117": "idefics3/modeling_idefics3.py:repeat_kv", "4118": "idefics3/modeling_idefics3.py:Idefics3RMSNorm", "4119": "idefics3/modeling_idefics3.py:Idefics3Connector", "4120": "idefics3/modeling_idefics3.py:Idefics3PreTrainedModel", "4121": "idefics3/modeling_idefics3.py:Idefics3VisionTransformer", "4122": "idefics3/modeling_idefics3.py:Idefics3Model", "4123": "idefics3/modeling_idefics3.py:Idefics3ForConditionalGeneration", "4124": "cohere2/modeling_cohere2.py:Cohere2RotaryEmbedding", "4125": "cohere2/modeling_cohere2.py:Cohere2LayerNorm", "4126": "cohere2/modeling_cohere2.py:repeat_kv", "4127": "cohere2/modeling_cohere2.py:eager_attention_forward", "4128": "cohere2/modeling_cohere2.py:rotate_half", "4129": "cohere2/modeling_cohere2.py:apply_rotary_pos_emb", "4130": "cohere2/modeling_cohere2.py:Cohere2Attention", "4131": "cohere2/modeling_cohere2.py:Cohere2MLP", "4132": "cohere2/modeling_cohere2.py:Cohere2DecoderLayer", "4133": "cohere2/modeling_cohere2.py:Cohere2PreTrainedModel", "4134": "cohere2/modeling_cohere2.py:Cohere2Model", "4135": "cohere2/modeling_cohere2.py:Cohere2ForCausalLM", "4136": "segformer/modeling_segformer.py:SegFormerImageClassifierOutput", "4137": "segformer/modeling_segformer.py:drop_path", "4138": "segformer/modeling_segformer.py:SegformerDropPath", "4139": "segformer/modeling_segformer.py:SegformerOverlapPatchEmbeddings", "4140": "segformer/modeling_segformer.py:SegformerEfficientSelfAttention", "4141": "segformer/modeling_segformer.py:SegformerSelfOutput", "4142": "segformer/modeling_segformer.py:SegformerAttention", "4143": "segformer/modeling_segformer.py:SegformerDWConv", "4144": "segformer/modeling_segformer.py:SegformerMixFFN", "4145": "segformer/modeling_segformer.py:SegformerLayer", "4146": "segformer/modeling_segformer.py:SegformerEncoder", "4147": "segformer/modeling_segformer.py:SegformerPreTrainedModel", "4148": "segformer/modeling_segformer.py:SegformerModel", "4149": "segformer/modeling_segformer.py:SegformerForImageClassification", "4150": "segformer/modeling_segformer.py:SegformerMLP", "4151": "segformer/modeling_segformer.py:SegformerDecodeHead", "4152": "segformer/modeling_segformer.py:SegformerForSemanticSegmentation", "4153": "vipllava/modeling_vipllava.py:VipLlavaModelOutputWithPast", "4154": "vipllava/modeling_vipllava.py:VipLlavaCausalLMOutputWithPast", "4155": "vipllava/modeling_vipllava.py:VipLlavaMultiModalProjector", "4156": "vipllava/modeling_vipllava.py:VipLlavaPreTrainedModel", "4157": "vipllava/modeling_vipllava.py:VipLlavaModel", "4158": "vipllava/modeling_vipllava.py:VipLlavaForConditionalGeneration", "4159": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoContrastiveEmbedding", "4160": "mm_grounding_dino/modeling_mm_grounding_dino.py:MultiScaleDeformableAttention", "4161": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoLearnedPositionEmbedding", "4162": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoMultiscaleDeformableAttention", "4163": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoBiMultiHeadAttention", "4164": "mm_grounding_dino/modeling_mm_grounding_dino.py:drop_path", "4165": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoDropPath", "4166": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoFusionLayer", "4167": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoPreTrainedModel", "4168": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoFrozenBatchNorm2d", "4169": "mm_grounding_dino/modeling_mm_grounding_dino.py:replace_batch_norm", "4170": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoConvEncoder", "4171": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoConvModel", "4172": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoEncoderOutput", "4173": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoMultiheadAttention", "4174": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoTextEnhancerLayer", "4175": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoDeformableLayer", "4176": "mm_grounding_dino/modeling_mm_grounding_dino.py:get_sine_pos_embed", "4177": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoEncoderLayer", "4178": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoEncoder", "4179": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoDecoderOutput", "4180": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoDecoderLayer", "4181": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoDecoder", "4182": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoModelOutput", "4183": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoSinePositionEmbedding", "4184": "mm_grounding_dino/modeling_mm_grounding_dino.py:build_position_encoding", "4185": "mm_grounding_dino/modeling_mm_grounding_dino.py:generate_masks_with_special_tokens_and_transfer_map", "4186": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoModel", "4187": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoMLPPredictionHead", "4188": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoObjectDetectionOutput", "4189": "mm_grounding_dino/modeling_mm_grounding_dino.py:build_label_maps", "4190": "mm_grounding_dino/modeling_mm_grounding_dino.py:build_text_mask", "4191": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoForObjectDetection", "4192": "chinese_clip/modeling_chinese_clip.py:contrastive_loss", "4193": "chinese_clip/modeling_chinese_clip.py:chinese_clip_loss", "4194": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPOutput", "4195": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextEmbeddings", "4196": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPVisionEmbeddings", "4197": "chinese_clip/modeling_chinese_clip.py:eager_attention_forward", "4198": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextSelfAttention", "4199": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextSelfOutput", "4200": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextAttention", "4201": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPVisionAttention", "4202": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextIntermediate", "4203": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextOutput", "4204": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPVisionMLP", "4205": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextLayer", "4206": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPVisionLayer", "4207": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextPooler", "4208": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPPreTrainedModel", "4209": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextEncoder", "4210": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPVisionEncoder", "4211": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPVisionTransformer", "4212": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextModel", "4213": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPVisionModel", "4214": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPModel", "4215": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2ForPreTrainingOutput", "4216": "wav2vec2/modeling_wav2vec2.py:_compute_mask_indices", "4217": "wav2vec2/modeling_wav2vec2.py:_sample_negative_indices", "4218": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2NoLayerNormConvLayer", "4219": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2LayerNormConvLayer", "4220": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2GroupNormConvLayer", "4221": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2PositionalConvEmbedding", "4222": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2SamePadLayer", "4223": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2FeatureEncoder", "4224": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2FeatureProjection", "4225": "wav2vec2/modeling_wav2vec2.py:eager_attention_forward", "4226": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2Attention", "4227": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2FeedForward", "4228": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2EncoderLayer", "4229": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2EncoderLayerStableLayerNorm", "4230": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2Encoder", "4231": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2EncoderStableLayerNorm", "4232": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2GumbelVectorQuantizer", "4233": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2Adapter", "4234": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2AdapterLayer", "4235": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2AttnAdapterLayer", "4236": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2PreTrainedModel", "4237": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2Model", "4238": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2ForPreTraining", "4239": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2ForMaskedLM", "4240": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2ForCTC", "4241": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2ForSequenceClassification", "4242": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2ForAudioFrameClassification", "4243": "wav2vec2/modeling_wav2vec2.py:AMSoftmaxLoss", "4244": "wav2vec2/modeling_wav2vec2.py:TDNNLayer", "4245": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2ForXVector", "4246": "granitemoe/modeling_granitemoe.py:GraniteMoeRMSNorm", "4247": "granitemoe/modeling_granitemoe.py:GraniteMoeRotaryEmbedding", "4248": "granitemoe/modeling_granitemoe.py:GraniteMoeParallelExperts", "4249": "granitemoe/modeling_granitemoe.py:GraniteMoeTopKGating", "4250": "granitemoe/modeling_granitemoe.py:GraniteMoeMoE", "4251": "granitemoe/modeling_granitemoe.py:rotate_half", "4252": "granitemoe/modeling_granitemoe.py:apply_rotary_pos_emb", "4253": "granitemoe/modeling_granitemoe.py:repeat_kv", "4254": "granitemoe/modeling_granitemoe.py:eager_attention_forward", "4255": "granitemoe/modeling_granitemoe.py:GraniteMoeAttention", "4256": "granitemoe/modeling_granitemoe.py:GraniteMoeDecoderLayer", "4257": "granitemoe/modeling_granitemoe.py:GraniteMoePreTrainedModel", "4258": "granitemoe/modeling_granitemoe.py:GraniteMoeModel", "4259": "granitemoe/modeling_granitemoe.py:load_balancing_loss_func", "4260": "granitemoe/modeling_granitemoe.py:GraniteMoeForCausalLM", "4261": "convnextv2/modeling_convnextv2.py:drop_path", "4262": "convnextv2/modeling_convnextv2.py:ConvNextV2DropPath", "4263": "convnextv2/modeling_convnextv2.py:ConvNextV2GRN", "4264": "convnextv2/modeling_convnextv2.py:ConvNextV2LayerNorm", "4265": "convnextv2/modeling_convnextv2.py:ConvNextV2Embeddings", "4266": "convnextv2/modeling_convnextv2.py:ConvNextV2Layer", "4267": "convnextv2/modeling_convnextv2.py:ConvNextV2Stage", "4268": "convnextv2/modeling_convnextv2.py:ConvNextV2Encoder", "4269": "convnextv2/modeling_convnextv2.py:ConvNextV2PreTrainedModel", "4270": "convnextv2/modeling_convnextv2.py:ConvNextV2Model", "4271": "convnextv2/modeling_convnextv2.py:ConvNextV2ForImageClassification", "4272": "convnextv2/modeling_convnextv2.py:ConvNextV2Backbone", "4273": "pixtral/modeling_pixtral.py:position_ids_in_meshgrid", "4274": "pixtral/modeling_pixtral.py:PixtralRotaryEmbedding", "4275": "pixtral/modeling_pixtral.py:rotate_half", "4276": "pixtral/modeling_pixtral.py:apply_rotary_pos_emb", "4277": "pixtral/modeling_pixtral.py:eager_attention_forward", "4278": "pixtral/modeling_pixtral.py:PixtralAttention", "4279": "pixtral/modeling_pixtral.py:PixtralMLP", "4280": "pixtral/modeling_pixtral.py:PixtralRMSNorm", "4281": "pixtral/modeling_pixtral.py:PixtralAttentionLayer", "4282": "pixtral/modeling_pixtral.py:PixtralTransformer", "4283": "pixtral/modeling_pixtral.py:PixtralPreTrainedModel", "4284": "pixtral/modeling_pixtral.py:generate_block_attention_mask", "4285": "pixtral/modeling_pixtral.py:PixtralVisionModel", "4286": "univnet/modeling_univnet.py:UnivNetModelOutput", "4287": "univnet/modeling_univnet.py:UnivNetKernelPredictorResidualBlock", "4288": "univnet/modeling_univnet.py:UnivNetKernelPredictor", "4289": "univnet/modeling_univnet.py:UnivNetLvcResidualBlock", "4290": "univnet/modeling_univnet.py:UnivNetLvcBlock", "4291": "univnet/modeling_univnet.py:UnivNetModel", "4292": "markuplm/modeling_markuplm.py:XPathEmbeddings", "4293": "markuplm/modeling_markuplm.py:MarkupLMEmbeddings", "4294": "markuplm/modeling_markuplm.py:MarkupLMSelfOutput", "4295": "markuplm/modeling_markuplm.py:MarkupLMIntermediate", "4296": "markuplm/modeling_markuplm.py:MarkupLMOutput", "4297": "markuplm/modeling_markuplm.py:MarkupLMPooler", "4298": "markuplm/modeling_markuplm.py:MarkupLMPredictionHeadTransform", "4299": "markuplm/modeling_markuplm.py:MarkupLMLMPredictionHead", "4300": "markuplm/modeling_markuplm.py:MarkupLMOnlyMLMHead", "4301": "markuplm/modeling_markuplm.py:eager_attention_forward", "4302": "markuplm/modeling_markuplm.py:MarkupLMSelfAttention", "4303": "markuplm/modeling_markuplm.py:MarkupLMAttention", "4304": "markuplm/modeling_markuplm.py:MarkupLMLayer", "4305": "markuplm/modeling_markuplm.py:MarkupLMEncoder", "4306": "markuplm/modeling_markuplm.py:MarkupLMPreTrainedModel", "4307": "markuplm/modeling_markuplm.py:MarkupLMModel", "4308": "markuplm/modeling_markuplm.py:MarkupLMForQuestionAnswering", "4309": "markuplm/modeling_markuplm.py:MarkupLMForTokenClassification", "4310": "markuplm/modeling_markuplm.py:MarkupLMForSequenceClassification", "4311": "layoutlm/modeling_layoutlm.py:LayoutLMEmbeddings", "4312": "layoutlm/modeling_layoutlm.py:eager_attention_forward", "4313": "layoutlm/modeling_layoutlm.py:LayoutLMSelfAttention", "4314": "layoutlm/modeling_layoutlm.py:LayoutLMSelfOutput", "4315": "layoutlm/modeling_layoutlm.py:LayoutLMAttention", "4316": "layoutlm/modeling_layoutlm.py:LayoutLMIntermediate", "4317": "layoutlm/modeling_layoutlm.py:LayoutLMOutput", "4318": "layoutlm/modeling_layoutlm.py:LayoutLMLayer", "4319": "layoutlm/modeling_layoutlm.py:LayoutLMEncoder", "4320": "layoutlm/modeling_layoutlm.py:LayoutLMPooler", "4321": "layoutlm/modeling_layoutlm.py:LayoutLMPredictionHeadTransform", "4322": "layoutlm/modeling_layoutlm.py:LayoutLMLMPredictionHead", "4323": "layoutlm/modeling_layoutlm.py:LayoutLMOnlyMLMHead", "4324": "layoutlm/modeling_layoutlm.py:LayoutLMPreTrainedModel", "4325": "layoutlm/modeling_layoutlm.py:LayoutLMModel", "4326": "layoutlm/modeling_layoutlm.py:LayoutLMForMaskedLM", "4327": "layoutlm/modeling_layoutlm.py:LayoutLMForSequenceClassification", "4328": "layoutlm/modeling_layoutlm.py:LayoutLMForTokenClassification", "4329": "layoutlm/modeling_layoutlm.py:LayoutLMForQuestionAnswering", "4330": "pvt/modeling_pvt.py:drop_path", "4331": "pvt/modeling_pvt.py:PvtDropPath", "4332": "pvt/modeling_pvt.py:PvtPatchEmbeddings", "4333": "pvt/modeling_pvt.py:PvtSelfOutput", "4334": "pvt/modeling_pvt.py:PvtEfficientSelfAttention", "4335": "pvt/modeling_pvt.py:PvtAttention", "4336": "pvt/modeling_pvt.py:PvtFFN", "4337": "pvt/modeling_pvt.py:PvtLayer", "4338": "pvt/modeling_pvt.py:PvtEncoder", "4339": "pvt/modeling_pvt.py:PvtPreTrainedModel", "4340": "pvt/modeling_pvt.py:PvtModel", "4341": "pvt/modeling_pvt.py:PvtForImageClassification", "4342": "moonshine/modeling_moonshine.py:MoonshineEncoderMLP", "4343": "moonshine/modeling_moonshine.py:MoonshineDecoderMLP", "4344": "moonshine/modeling_moonshine.py:MoonshineRotaryEmbedding", "4345": "moonshine/modeling_moonshine.py:repeat_kv", "4346": "moonshine/modeling_moonshine.py:eager_attention_forward", "4347": "moonshine/modeling_moonshine.py:rotate_half", "4348": "moonshine/modeling_moonshine.py:apply_rotary_pos_emb", "4349": "moonshine/modeling_moonshine.py:MoonshineAttention", "4350": "moonshine/modeling_moonshine.py:MoonshineEncoderLayer", "4351": "moonshine/modeling_moonshine.py:MoonshineDecoderLayer", "4352": "moonshine/modeling_moonshine.py:MoonshinePreTrainedModel", "4353": "moonshine/modeling_moonshine.py:MoonshineEncoder", "4354": "moonshine/modeling_moonshine.py:MoonshineDecoder", "4355": "moonshine/modeling_moonshine.py:_compute_mask_indices", "4356": "moonshine/modeling_moonshine.py:MoonshineModel", "4357": "moonshine/modeling_moonshine.py:shift_tokens_right", "4358": "moonshine/modeling_moonshine.py:MoonshineForConditionalGeneration", "4359": "rt_detr_v2/modeling_rt_detr_v2.py:multi_scale_deformable_attention_v2", "4360": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2MultiscaleDeformableAttention", "4361": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2MultiheadAttention", "4362": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2DecoderLayer", "4363": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2PreTrainedModel", "4364": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2DecoderOutput", "4365": "rt_detr_v2/modeling_rt_detr_v2.py:inverse_sigmoid", "4366": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2Decoder", "4367": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2ModelOutput", "4368": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2FrozenBatchNorm2d", "4369": "rt_detr_v2/modeling_rt_detr_v2.py:replace_batch_norm", "4370": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2ConvEncoder", "4371": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2ConvNormLayer", "4372": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2EncoderLayer", "4373": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2RepVggBlock", "4374": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2CSPRepLayer", "4375": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2Encoder", "4376": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2HybridEncoder", "4377": "rt_detr_v2/modeling_rt_detr_v2.py:get_contrastive_denoising_training_group", "4378": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2Model", "4379": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2MLPPredictionHead", "4380": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2ObjectDetectionOutput", "4381": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2ForObjectDetection", "4382": "nanochat/modeling_nanochat.py:NanoChatRMSNorm", "4383": "nanochat/modeling_nanochat.py:NanoChatRotaryEmbedding", "4384": "nanochat/modeling_nanochat.py:apply_rotary_pos_emb", "4385": "nanochat/modeling_nanochat.py:repeat_kv", "4386": "nanochat/modeling_nanochat.py:eager_attention_forward", "4387": "nanochat/modeling_nanochat.py:rotate_half", "4388": "nanochat/modeling_nanochat.py:NanoChatAttention", "4389": "nanochat/modeling_nanochat.py:NanoChatMLP", "4390": "nanochat/modeling_nanochat.py:NanoChatDecoderLayer", "4391": "nanochat/modeling_nanochat.py:NanoChatPreTrainedModel", "4392": "nanochat/modeling_nanochat.py:NanoChatModel", "4393": "nanochat/modeling_nanochat.py:NanoChatForCausalLM", "4394": "clap/modeling_clap.py:interpolate", "4395": "clap/modeling_clap.py:window_partition", "4396": "clap/modeling_clap.py:window_reverse", "4397": "clap/modeling_clap.py:contrastive_loss", "4398": "clap/modeling_clap.py:ClapTextModelOutput", "4399": "clap/modeling_clap.py:ClapAudioModelOutput", "4400": "clap/modeling_clap.py:ClapOutput", "4401": "clap/modeling_clap.py:ClapDropPath", "4402": "clap/modeling_clap.py:ClapAudioAFFBlock", "4403": "clap/modeling_clap.py:ClapAudioPatchEmbed", "4404": "clap/modeling_clap.py:ClapAudioSelfAttention", "4405": "clap/modeling_clap.py:ClapAudioSelfOutput", "4406": "clap/modeling_clap.py:ClapAudioAttention", "4407": "clap/modeling_clap.py:ClapAudioIntermediate", "4408": "clap/modeling_clap.py:ClapAudioOutput", "4409": "clap/modeling_clap.py:ClapAudioLayer", "4410": "clap/modeling_clap.py:ClapAudioStage", "4411": "clap/modeling_clap.py:ClapAudioPatchMerging", "4412": "clap/modeling_clap.py:ClapAudioEncoder", "4413": "clap/modeling_clap.py:ClapProjectionLayer", "4414": "clap/modeling_clap.py:ClapTextEmbeddings", "4415": "clap/modeling_clap.py:eager_attention_forward", "4416": "clap/modeling_clap.py:ClapTextSelfAttention", "4417": "clap/modeling_clap.py:ClapTextSelfOutput", "4418": "clap/modeling_clap.py:ClapTextAttention", "4419": "clap/modeling_clap.py:ClapTextIntermediate", "4420": "clap/modeling_clap.py:ClapTextOutput", "4421": "clap/modeling_clap.py:ClapTextLayer", "4422": "clap/modeling_clap.py:ClapTextEncoder", "4423": "clap/modeling_clap.py:ClapTextPooler", "4424": "clap/modeling_clap.py:ClapPreTrainedModel", "4425": "clap/modeling_clap.py:ClapAudioModel", "4426": "clap/modeling_clap.py:ClapTextModel", "4427": "clap/modeling_clap.py:ClapModel", "4428": "clap/modeling_clap.py:ClapTextModelWithProjection", "4429": "clap/modeling_clap.py:ClapAudioModelWithProjection", "4430": "poolformer/modeling_poolformer.py:drop_path", "4431": "poolformer/modeling_poolformer.py:PoolFormerDropPath", "4432": "poolformer/modeling_poolformer.py:PoolFormerEmbeddings", "4433": "poolformer/modeling_poolformer.py:PoolFormerGroupNorm", "4434": "poolformer/modeling_poolformer.py:PoolFormerPooling", "4435": "poolformer/modeling_poolformer.py:PoolFormerOutput", "4436": "poolformer/modeling_poolformer.py:PoolFormerLayer", "4437": "poolformer/modeling_poolformer.py:PoolFormerEncoder", "4438": "poolformer/modeling_poolformer.py:PoolFormerPreTrainedModel", "4439": "poolformer/modeling_poolformer.py:PoolFormerModel", "4440": "poolformer/modeling_poolformer.py:PoolFormerFinalPooler", "4441": "poolformer/modeling_poolformer.py:PoolFormerForImageClassification", "4442": "ernie/modeling_ernie.py:ErnieEmbeddings", "4443": "ernie/modeling_ernie.py:eager_attention_forward", "4444": "ernie/modeling_ernie.py:ErnieSelfAttention", "4445": "ernie/modeling_ernie.py:ErnieCrossAttention", "4446": "ernie/modeling_ernie.py:ErnieSelfOutput", "4447": "ernie/modeling_ernie.py:ErnieAttention", "4448": "ernie/modeling_ernie.py:ErnieIntermediate", "4449": "ernie/modeling_ernie.py:ErnieOutput", "4450": "ernie/modeling_ernie.py:ErnieLayer", "4451": "ernie/modeling_ernie.py:ErniePooler", "4452": "ernie/modeling_ernie.py:ErniePredictionHeadTransform", "4453": "ernie/modeling_ernie.py:ErnieLMPredictionHead", "4454": "ernie/modeling_ernie.py:ErnieEncoder", "4455": "ernie/modeling_ernie.py:ErniePreTrainedModel", "4456": "ernie/modeling_ernie.py:ErnieModel", "4457": "ernie/modeling_ernie.py:ErnieForPreTrainingOutput", "4458": "ernie/modeling_ernie.py:ErniePreTrainingHeads", "4459": "ernie/modeling_ernie.py:ErnieForPreTraining", "4460": "ernie/modeling_ernie.py:ErnieOnlyMLMHead", "4461": "ernie/modeling_ernie.py:ErnieForCausalLM", "4462": "ernie/modeling_ernie.py:ErnieForMaskedLM", "4463": "ernie/modeling_ernie.py:ErnieOnlyNSPHead", "4464": "ernie/modeling_ernie.py:ErnieForNextSentencePrediction", "4465": "ernie/modeling_ernie.py:ErnieForSequenceClassification", "4466": "ernie/modeling_ernie.py:ErnieForMultipleChoice", "4467": "ernie/modeling_ernie.py:ErnieForTokenClassification", "4468": "ernie/modeling_ernie.py:ErnieForQuestionAnswering", "4469": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertRotaryPositionalEmbedding", "4470": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertRelPositionalEmbedding", "4471": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertFeatureProjection", "4472": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertFeedForward", "4473": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertConvolutionModule", "4474": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertSelfAttention", "4475": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertEncoderLayer", "4476": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertEncoder", "4477": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertAdapter", "4478": "wav2vec2_bert/modeling_wav2vec2_bert.py:_compute_new_attention_mask", "4479": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertAdapterLayer", "4480": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertPreTrainedModel", "4481": "wav2vec2_bert/modeling_wav2vec2_bert.py:_compute_mask_indices", "4482": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertModel", "4483": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertForCTC", "4484": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertForSequenceClassification", "4485": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertForAudioFrameClassification", "4486": "wav2vec2_bert/modeling_wav2vec2_bert.py:AMSoftmaxLoss", "4487": "wav2vec2_bert/modeling_wav2vec2_bert.py:TDNNLayer", "4488": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertForXVector", "4489": "auto/modeling_auto.py:AutoModelForMaskGeneration", "4490": "auto/modeling_auto.py:AutoModelForKeypointDetection", "4491": "auto/modeling_auto.py:AutoModelForKeypointMatching", "4492": "auto/modeling_auto.py:AutoModelForTextEncoding", "4493": "auto/modeling_auto.py:AutoModelForImageToImage", "4494": "auto/modeling_auto.py:AutoModel", "4495": "auto/modeling_auto.py:AutoModelForPreTraining", "4496": "auto/modeling_auto.py:_AutoModelWithLMHead", "4497": "auto/modeling_auto.py:AutoModelForCausalLM", "4498": "auto/modeling_auto.py:AutoModelForMaskedLM", "4499": "auto/modeling_auto.py:AutoModelForSeq2SeqLM", "4500": "auto/modeling_auto.py:AutoModelForSequenceClassification", "4501": "auto/modeling_auto.py:AutoModelForQuestionAnswering", "4502": "auto/modeling_auto.py:AutoModelForTableQuestionAnswering", "4503": "auto/modeling_auto.py:AutoModelForVisualQuestionAnswering", "4504": "auto/modeling_auto.py:AutoModelForDocumentQuestionAnswering", "4505": "auto/modeling_auto.py:AutoModelForTokenClassification", "4506": "auto/modeling_auto.py:AutoModelForMultipleChoice", "4507": "auto/modeling_auto.py:AutoModelForNextSentencePrediction", "4508": "auto/modeling_auto.py:AutoModelForImageClassification", "4509": "auto/modeling_auto.py:AutoModelForZeroShotImageClassification", "4510": "auto/modeling_auto.py:AutoModelForImageSegmentation", "4511": "auto/modeling_auto.py:AutoModelForSemanticSegmentation", "4512": "auto/modeling_auto.py:AutoModelForTimeSeriesPrediction", "4513": "auto/modeling_auto.py:AutoModelForUniversalSegmentation", "4514": "auto/modeling_auto.py:AutoModelForInstanceSegmentation", "4515": "auto/modeling_auto.py:AutoModelForObjectDetection", "4516": "auto/modeling_auto.py:AutoModelForZeroShotObjectDetection", "4517": "auto/modeling_auto.py:AutoModelForDepthEstimation", "4518": "auto/modeling_auto.py:AutoModelForVideoClassification", "4519": "auto/modeling_auto.py:_AutoModelForVision2Seq", "4520": "auto/modeling_auto.py:AutoModelForImageTextToText", "4521": "auto/modeling_auto.py:AutoModelForMultimodalLM", "4522": "auto/modeling_auto.py:AutoModelForAudioClassification", "4523": "auto/modeling_auto.py:AutoModelForCTC", "4524": "auto/modeling_auto.py:AutoModelForSpeechSeq2Seq", "4525": "auto/modeling_auto.py:AutoModelForAudioFrameClassification", "4526": "auto/modeling_auto.py:AutoModelForAudioXVector", "4527": "auto/modeling_auto.py:AutoModelForTextToSpectrogram", "4528": "auto/modeling_auto.py:AutoModelForTextToWaveform", "4529": "auto/modeling_auto.py:AutoBackbone", "4530": "auto/modeling_auto.py:AutoModelForMaskedImageModeling", "4531": "auto/modeling_auto.py:AutoModelForAudioTokenization", "4532": "auto/modeling_auto.py:AutoModelWithLMHead", "4533": "auto/modeling_auto.py:AutoModelForVision2Seq", "4534": "cvt/modeling_cvt.py:BaseModelOutputWithCLSToken", "4535": "cvt/modeling_cvt.py:drop_path", "4536": "cvt/modeling_cvt.py:CvtDropPath", "4537": "cvt/modeling_cvt.py:CvtEmbeddings", "4538": "cvt/modeling_cvt.py:CvtConvEmbeddings", "4539": "cvt/modeling_cvt.py:CvtSelfAttentionConvProjection", "4540": "cvt/modeling_cvt.py:CvtSelfAttentionLinearProjection", "4541": "cvt/modeling_cvt.py:CvtSelfAttentionProjection", "4542": "cvt/modeling_cvt.py:CvtSelfAttention", "4543": "cvt/modeling_cvt.py:CvtSelfOutput", "4544": "cvt/modeling_cvt.py:CvtAttention", "4545": "cvt/modeling_cvt.py:CvtIntermediate", "4546": "cvt/modeling_cvt.py:CvtOutput", "4547": "cvt/modeling_cvt.py:CvtLayer", "4548": "cvt/modeling_cvt.py:CvtStage", "4549": "cvt/modeling_cvt.py:CvtEncoder", "4550": "cvt/modeling_cvt.py:CvtPreTrainedModel", "4551": "cvt/modeling_cvt.py:CvtModel", "4552": "cvt/modeling_cvt.py:CvtForImageClassification", "4553": "flava/modeling_flava.py:FlavaModelOutput", "4554": "flava/modeling_flava.py:FlavaLosses", "4555": "flava/modeling_flava.py:FlavaForPreTrainingOutput", "4556": "flava/modeling_flava.py:FlavaImageEmbeddings", "4557": "flava/modeling_flava.py:PatchEmbeddings", "4558": "flava/modeling_flava.py:FlavaTextEmbeddings", "4559": "flava/modeling_flava.py:FlavaSelfAttention", "4560": "flava/modeling_flava.py:FlavaSelfOutput", "4561": "flava/modeling_flava.py:FlavaAttention", "4562": "flava/modeling_flava.py:FlavaIntermediate", "4563": "flava/modeling_flava.py:FlavaOutput", "4564": "flava/modeling_flava.py:FlavaLayer", "4565": "flava/modeling_flava.py:FlavaEncoder", "4566": "flava/modeling_flava.py:FlavaPooler", "4567": "flava/modeling_flava.py:FlavaPreTrainedModel", "4568": "flava/modeling_flava.py:FlavaImageModel", "4569": "flava/modeling_flava.py:FlavaTextModel", "4570": "flava/modeling_flava.py:FlavaMultimodalModel", "4571": "flava/modeling_flava.py:FlavaModel", "4572": "flava/modeling_flava.py:FlavaImageCodebookResPath", "4573": "flava/modeling_flava.py:FlavaImageCodebookBlock", "4574": "flava/modeling_flava.py:FlavaImageCodebookLayerGroup", "4575": "flava/modeling_flava.py:FlavaImageCodebook", "4576": "flava/modeling_flava.py:FlavaPredictionHeadTransform", "4577": "flava/modeling_flava.py:FlavaMaskedPredictionHead", "4578": "flava/modeling_flava.py:FlavaITMHead", "4579": "flava/modeling_flava.py:FlavaGlobalContrastiveHead", "4580": "flava/modeling_flava.py:FlavaForPreTraining", "4581": "ministral3/modeling_ministral3.py:rotate_half", "4582": "ministral3/modeling_ministral3.py:apply_rotary_pos_emb", "4583": "ministral3/modeling_ministral3.py:repeat_kv", "4584": "ministral3/modeling_ministral3.py:eager_attention_forward", "4585": "ministral3/modeling_ministral3.py:_get_llama_4_attn_scale", "4586": "ministral3/modeling_ministral3.py:Ministral3Attention", "4587": "ministral3/modeling_ministral3.py:Ministral3MLP", "4588": "ministral3/modeling_ministral3.py:Ministral3RMSNorm", "4589": "ministral3/modeling_ministral3.py:Ministral3DecoderLayer", "4590": "ministral3/modeling_ministral3.py:Ministral3PreTrainedModel", "4591": "ministral3/modeling_ministral3.py:Ministral3RotaryEmbedding", "4592": "ministral3/modeling_ministral3.py:Ministral3Model", "4593": "ministral3/modeling_ministral3.py:Ministral3ForCausalLM", "4594": "ministral3/modeling_ministral3.py:Ministral3ForTokenClassification", "4595": "ministral3/modeling_ministral3.py:Ministral3ForSequenceClassification", "4596": "ministral3/modeling_ministral3.py:Ministral3ForQuestionAnswering", "4597": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLMLP", "4598": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VisionPatchEmbed", "4599": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VisionRotaryEmbedding", "4600": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLPatchMerger", "4601": "qwen2_5_vl/modeling_qwen2_5_vl.py:rotate_half", "4602": "qwen2_5_vl/modeling_qwen2_5_vl.py:apply_rotary_pos_emb_vision", "4603": "qwen2_5_vl/modeling_qwen2_5_vl.py:repeat_kv", "4604": "qwen2_5_vl/modeling_qwen2_5_vl.py:eager_attention_forward", "4605": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLVisionAttention", "4606": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLVisionBlock", "4607": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLPreTrainedModel", "4608": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VisionTransformerPretrainedModel", "4609": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLModelOutputWithPast", "4610": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLRotaryEmbedding", "4611": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2MLP", "4612": "qwen2_5_vl/modeling_qwen2_5_vl.py:apply_multimodal_rotary_pos_emb", "4613": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLAttention", "4614": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLDecoderLayer", "4615": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLTextModel", "4616": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLModel", "4617": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLCausalLMOutputWithPast", "4618": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLForConditionalGeneration", "4619": "groupvit/modeling_groupvit.py:contrastive_loss", "4620": "groupvit/modeling_groupvit.py:groupvit_loss", "4621": "groupvit/modeling_groupvit.py:hard_softmax", "4622": "groupvit/modeling_groupvit.py:gumbel_softmax", "4623": "groupvit/modeling_groupvit.py:resize_attention_map", "4624": "groupvit/modeling_groupvit.py:get_grouping_from_attentions", "4625": "groupvit/modeling_groupvit.py:GroupViTCrossAttentionLayer", "4626": "groupvit/modeling_groupvit.py:GroupViTAssignAttention", "4627": "groupvit/modeling_groupvit.py:GroupViTTokenAssign", "4628": "groupvit/modeling_groupvit.py:GroupViTModelOutput", "4629": "groupvit/modeling_groupvit.py:GroupViTPatchEmbeddings", "4630": "groupvit/modeling_groupvit.py:GroupViTVisionEmbeddings", "4631": "groupvit/modeling_groupvit.py:GroupViTTextEmbeddings", "4632": "groupvit/modeling_groupvit.py:GroupViTStage", "4633": "groupvit/modeling_groupvit.py:GroupViTMLP", "4634": "groupvit/modeling_groupvit.py:GroupViTMixerMLP", "4635": "groupvit/modeling_groupvit.py:GroupViTAttention", "4636": "groupvit/modeling_groupvit.py:GroupViTEncoderLayer", "4637": "groupvit/modeling_groupvit.py:GroupViTPreTrainedModel", "4638": "groupvit/modeling_groupvit.py:GroupViTVisionEncoder", "4639": "groupvit/modeling_groupvit.py:GroupViTTextEncoder", "4640": "groupvit/modeling_groupvit.py:GroupViTTextTransformer", "4641": "groupvit/modeling_groupvit.py:GroupViTTextModel", "4642": "groupvit/modeling_groupvit.py:GroupViTVisionTransformer", "4643": "groupvit/modeling_groupvit.py:GroupViTVisionModel", "4644": "groupvit/modeling_groupvit.py:GroupViTModel", "4645": "voxtral/modeling_voxtral.py:eager_attention_forward", "4646": "voxtral/modeling_voxtral.py:VoxtralAttention", "4647": "voxtral/modeling_voxtral.py:VoxtralEncoderLayer", "4648": "voxtral/modeling_voxtral.py:VoxtralPreTrainedModel", "4649": "voxtral/modeling_voxtral.py:VoxtralEncoder", "4650": "voxtral/modeling_voxtral.py:VoxtralMultiModalProjector", "4651": "voxtral/modeling_voxtral.py:VoxtralForConditionalGeneration", "4652": "afmoe/modeling_afmoe.py:AfmoeRotaryEmbedding", "4653": "afmoe/modeling_afmoe.py:AfmoeRMSNorm", "4654": "afmoe/modeling_afmoe.py:AfmoeMLP", "4655": "afmoe/modeling_afmoe.py:AfmoeTokenChoiceRouter", "4656": "afmoe/modeling_afmoe.py:AfmoeExperts", "4657": "afmoe/modeling_afmoe.py:AfmoeMoE", "4658": "afmoe/modeling_afmoe.py:rotate_half", "4659": "afmoe/modeling_afmoe.py:apply_rotary_pos_emb", "4660": "afmoe/modeling_afmoe.py:repeat_kv", "4661": "afmoe/modeling_afmoe.py:eager_attention_forward", "4662": "afmoe/modeling_afmoe.py:AfmoeAttention", "4663": "afmoe/modeling_afmoe.py:AfmoeDecoderLayer", "4664": "afmoe/modeling_afmoe.py:AfmoePreTrainedModel", "4665": "afmoe/modeling_afmoe.py:AfmoeModel", "4666": "afmoe/modeling_afmoe.py:AfmoeForCausalLM", "4667": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaRMSNorm", "4668": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaRotaryEmbedding", "4669": "recurrent_gemma/modeling_recurrent_gemma.py:rotate_half", "4670": "recurrent_gemma/modeling_recurrent_gemma.py:apply_rotary_pos_emb", "4671": "recurrent_gemma/modeling_recurrent_gemma.py:repeat_kv", "4672": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaSdpaAttention", "4673": "recurrent_gemma/modeling_recurrent_gemma.py:SqrtBoundDerivative", "4674": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaRglru", "4675": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaRecurrentBlock", "4676": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaMlp", "4677": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaDecoderLayer", "4678": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaPreTrainedModel", "4679": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaModel", "4680": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaForCausalLM", "4681": "led/modeling_led.py:shift_tokens_right", "4682": "led/modeling_led.py:_prepare_4d_attention_mask_inverted", "4683": "led/modeling_led.py:LEDLearnedPositionalEmbedding", "4684": "led/modeling_led.py:LEDEncoderSelfAttention", "4685": "led/modeling_led.py:LEDEncoderAttention", "4686": "led/modeling_led.py:LEDDecoderAttention", "4687": "led/modeling_led.py:LEDEncoderLayer", "4688": "led/modeling_led.py:LEDDecoderLayer", "4689": "led/modeling_led.py:LEDClassificationHead", "4690": "led/modeling_led.py:LEDPreTrainedModel", "4691": "led/modeling_led.py:LEDEncoderBaseModelOutput", "4692": "led/modeling_led.py:LEDSeq2SeqModelOutput", "4693": "led/modeling_led.py:LEDSeq2SeqLMOutput", "4694": "led/modeling_led.py:LEDSeq2SeqSequenceClassifierOutput", "4695": "led/modeling_led.py:LEDSeq2SeqQuestionAnsweringModelOutput", "4696": "led/modeling_led.py:LEDEncoder", "4697": "led/modeling_led.py:LEDDecoder", "4698": "led/modeling_led.py:LEDModel", "4699": "led/modeling_led.py:LEDForConditionalGeneration", "4700": "led/modeling_led.py:LEDForSequenceClassification", "4701": "led/modeling_led.py:LEDForQuestionAnswering", "4702": "megatron_bert/modeling_megatron_bert.py:MegatronBertEmbeddings", "4703": "megatron_bert/modeling_megatron_bert.py:MegatronBertSelfAttention", "4704": "megatron_bert/modeling_megatron_bert.py:MegatronBertSelfOutput", "4705": "megatron_bert/modeling_megatron_bert.py:MegatronBertAttention", "4706": "megatron_bert/modeling_megatron_bert.py:MegatronBertIntermediate", "4707": "megatron_bert/modeling_megatron_bert.py:MegatronBertOutput", "4708": "megatron_bert/modeling_megatron_bert.py:MegatronBertLayer", "4709": "megatron_bert/modeling_megatron_bert.py:MegatronBertEncoder", "4710": "megatron_bert/modeling_megatron_bert.py:MegatronBertPooler", "4711": "megatron_bert/modeling_megatron_bert.py:MegatronBertPredictionHeadTransform", "4712": "megatron_bert/modeling_megatron_bert.py:MegatronBertLMPredictionHead", "4713": "megatron_bert/modeling_megatron_bert.py:MegatronBertOnlyMLMHead", "4714": "megatron_bert/modeling_megatron_bert.py:MegatronBertOnlyNSPHead", "4715": "megatron_bert/modeling_megatron_bert.py:MegatronBertPreTrainingHeads", "4716": "megatron_bert/modeling_megatron_bert.py:MegatronBertPreTrainedModel", "4717": "megatron_bert/modeling_megatron_bert.py:MegatronBertForPreTrainingOutput", "4718": "megatron_bert/modeling_megatron_bert.py:MegatronBertModel", "4719": "megatron_bert/modeling_megatron_bert.py:MegatronBertForPreTraining", "4720": "megatron_bert/modeling_megatron_bert.py:MegatronBertForCausalLM", "4721": "megatron_bert/modeling_megatron_bert.py:MegatronBertForMaskedLM", "4722": "megatron_bert/modeling_megatron_bert.py:MegatronBertForNextSentencePrediction", "4723": "megatron_bert/modeling_megatron_bert.py:MegatronBertForSequenceClassification", "4724": "megatron_bert/modeling_megatron_bert.py:MegatronBertForMultipleChoice", "4725": "megatron_bert/modeling_megatron_bert.py:MegatronBertForTokenClassification", "4726": "megatron_bert/modeling_megatron_bert.py:MegatronBertForQuestionAnswering", "4727": "timm_backbone/modeling_timm_backbone.py:TimmBackbone", "4728": "granitemoeshared/modeling_granitemoeshared.py:GraniteFlashAttentionKwargs", "4729": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedMLP", "4730": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedRMSNorm", "4731": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedParallelExperts", "4732": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedTopKGating", "4733": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedMoE", "4734": "granitemoeshared/modeling_granitemoeshared.py:rotate_half", "4735": "granitemoeshared/modeling_granitemoeshared.py:apply_rotary_pos_emb", "4736": "granitemoeshared/modeling_granitemoeshared.py:repeat_kv", "4737": "granitemoeshared/modeling_granitemoeshared.py:eager_attention_forward", "4738": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedAttention", "4739": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedDecoderLayer", "4740": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedPreTrainedModel", "4741": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedRotaryEmbedding", "4742": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedModel", "4743": "granitemoeshared/modeling_granitemoeshared.py:load_balancing_loss_func", "4744": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedForCausalLM", "4745": "dbrx/modeling_dbrx.py:DbrxRotaryEmbedding", "4746": "dbrx/modeling_dbrx.py:rotate_half", "4747": "dbrx/modeling_dbrx.py:apply_rotary_pos_emb", "4748": "dbrx/modeling_dbrx.py:repeat_kv", "4749": "dbrx/modeling_dbrx.py:eager_attention_forward", "4750": "dbrx/modeling_dbrx.py:DbrxAttention", "4751": "dbrx/modeling_dbrx.py:DbrxExpertGLU", "4752": "dbrx/modeling_dbrx.py:DbrxExperts", "4753": "dbrx/modeling_dbrx.py:DbrxRouter", "4754": "dbrx/modeling_dbrx.py:DbrxFFN", "4755": "dbrx/modeling_dbrx.py:DbrxNormAttentionNorm", "4756": "dbrx/modeling_dbrx.py:DbrxBlock", "4757": "dbrx/modeling_dbrx.py:DbrxPreTrainedModel", "4758": "dbrx/modeling_dbrx.py:DbrxModel", "4759": "dbrx/modeling_dbrx.py:load_balancing_loss_func", "4760": "dbrx/modeling_dbrx.py:DbrxForCausalLM", "4761": "mask2former/modeling_mask2former.py:Mask2FormerPixelDecoderOutput", "4762": "mask2former/modeling_mask2former.py:Mask2FormerMaskedAttentionDecoderOutput", "4763": "mask2former/modeling_mask2former.py:Mask2FormerPixelLevelModuleOutput", "4764": "mask2former/modeling_mask2former.py:Mask2FormerModelOutput", "4765": "mask2former/modeling_mask2former.py:Mask2FormerForUniversalSegmentationOutput", "4766": "mask2former/modeling_mask2former.py:sample_point", "4767": "mask2former/modeling_mask2former.py:dice_loss", "4768": "mask2former/modeling_mask2former.py:sigmoid_cross_entropy_loss", "4769": "mask2former/modeling_mask2former.py:pair_wise_dice_loss", "4770": "mask2former/modeling_mask2former.py:pair_wise_sigmoid_cross_entropy_loss", "4771": "mask2former/modeling_mask2former.py:Mask2FormerHungarianMatcher", "4772": "mask2former/modeling_mask2former.py:Mask2FormerLoss", "4773": "mask2former/modeling_mask2former.py:multi_scale_deformable_attention", "4774": "mask2former/modeling_mask2former.py:Mask2FormerSinePositionEmbedding", "4775": "mask2former/modeling_mask2former.py:Mask2FormerPixelDecoderEncoderMultiscaleDeformableAttention", "4776": "mask2former/modeling_mask2former.py:Mask2FormerPixelDecoderEncoderLayer", "4777": "mask2former/modeling_mask2former.py:Mask2FormerPixelDecoderEncoderOnly", "4778": "mask2former/modeling_mask2former.py:Mask2FormerPixelDecoder", "4779": "mask2former/modeling_mask2former.py:Mask2FormerPixelLevelModule", "4780": "mask2former/modeling_mask2former.py:Mask2FormerAttention", "4781": "mask2former/modeling_mask2former.py:Mask2FormerMaskedAttentionDecoderLayer", "4782": "mask2former/modeling_mask2former.py:Mask2FormerMaskedAttentionDecoder", "4783": "mask2former/modeling_mask2former.py:Mask2FormerPredictionBlock", "4784": "mask2former/modeling_mask2former.py:Mask2FormerMLPPredictionHead", "4785": "mask2former/modeling_mask2former.py:Mask2FormerMaskPredictor", "4786": "mask2former/modeling_mask2former.py:Mask2FormerTransformerModule", "4787": "mask2former/modeling_mask2former.py:Mask2FormerPreTrainedModel", "4788": "mask2former/modeling_mask2former.py:Mask2FormerModel", "4789": "mask2former/modeling_mask2former.py:Mask2FormerForUniversalSegmentation", "4790": "aimv2/modeling_aimv2.py:Aimv2Output", "4791": "aimv2/modeling_aimv2.py:Aimv2RMSNorm", "4792": "aimv2/modeling_aimv2.py:Aimv2MLP", "4793": "aimv2/modeling_aimv2.py:Aimv2VisionEmbeddings", "4794": "aimv2/modeling_aimv2.py:Aimv2TextEmbeddings", "4795": "aimv2/modeling_aimv2.py:eager_attention_forward", "4796": "aimv2/modeling_aimv2.py:Aimv2Attention", "4797": "aimv2/modeling_aimv2.py:Aimv2EncoderLayer", "4798": "aimv2/modeling_aimv2.py:Aimv2Encoder", "4799": "aimv2/modeling_aimv2.py:Aimv2AttentionPoolingHead", "4800": "aimv2/modeling_aimv2.py:Aimv2PreTrainedModel", "4801": "aimv2/modeling_aimv2.py:Aimv2VisionModel", "4802": "aimv2/modeling_aimv2.py:Aimv2TextModel", "4803": "aimv2/modeling_aimv2.py:_get_vector_norm", "4804": "aimv2/modeling_aimv2.py:Aimv2Model", "4805": "mistral/modeling_mistral.py:MistralMLP", "4806": "mistral/modeling_mistral.py:rotate_half", "4807": "mistral/modeling_mistral.py:apply_rotary_pos_emb", "4808": "mistral/modeling_mistral.py:repeat_kv", "4809": "mistral/modeling_mistral.py:eager_attention_forward", "4810": "mistral/modeling_mistral.py:MistralAttention", "4811": "mistral/modeling_mistral.py:MistralRMSNorm", "4812": "mistral/modeling_mistral.py:MistralDecoderLayer", "4813": "mistral/modeling_mistral.py:MistralPreTrainedModel", "4814": "mistral/modeling_mistral.py:MistralRotaryEmbedding", "4815": "mistral/modeling_mistral.py:MistralModel", "4816": "mistral/modeling_mistral.py:MistralForCausalLM", "4817": "mistral/modeling_mistral.py:MistralForTokenClassification", "4818": "mistral/modeling_mistral.py:MistralForSequenceClassification", "4819": "mistral/modeling_mistral.py:MistralForQuestionAnswering", "4820": "mvp/modeling_mvp.py:shift_tokens_right", "4821": "mvp/modeling_mvp.py:MvpLearnedPositionalEmbedding", "4822": "mvp/modeling_mvp.py:MvpAttention", "4823": "mvp/modeling_mvp.py:MvpEncoderLayer", "4824": "mvp/modeling_mvp.py:MvpDecoderLayer", "4825": "mvp/modeling_mvp.py:MvpClassificationHead", "4826": "mvp/modeling_mvp.py:MvpPrompt", "4827": "mvp/modeling_mvp.py:MvpPreTrainedModel", "4828": "mvp/modeling_mvp.py:MvpEncoder", "4829": "mvp/modeling_mvp.py:MvpDecoder", "4830": "mvp/modeling_mvp.py:MvpModel", "4831": "mvp/modeling_mvp.py:MvpForConditionalGeneration", "4832": "mvp/modeling_mvp.py:MvpForSequenceClassification", "4833": "mvp/modeling_mvp.py:MvpForQuestionAnswering", "4834": "mvp/modeling_mvp.py:MvpDecoderWrapper", "4835": "mvp/modeling_mvp.py:MvpForCausalLM", "4836": "metaclip_2/modeling_metaclip_2.py:MetaClip2TextEmbeddings", "4837": "metaclip_2/modeling_metaclip_2.py:MetaClip2VisionEmbeddings", "4838": "metaclip_2/modeling_metaclip_2.py:eager_attention_forward", "4839": "metaclip_2/modeling_metaclip_2.py:MetaClip2Attention", "4840": "metaclip_2/modeling_metaclip_2.py:MetaClip2MLP", "4841": "metaclip_2/modeling_metaclip_2.py:MetaClip2EncoderLayer", "4842": "metaclip_2/modeling_metaclip_2.py:MetaClip2PreTrainedModel", "4843": "metaclip_2/modeling_metaclip_2.py:MetaClip2Encoder", "4844": "metaclip_2/modeling_metaclip_2.py:MetaClip2TextTransformer", "4845": "metaclip_2/modeling_metaclip_2.py:MetaClip2TextModel", "4846": "metaclip_2/modeling_metaclip_2.py:MetaClip2TextModelOutput", "4847": "metaclip_2/modeling_metaclip_2.py:MetaClip2TextModelWithProjection", "4848": "metaclip_2/modeling_metaclip_2.py:MetaClip2Output", "4849": "metaclip_2/modeling_metaclip_2.py:contrastive_loss", "4850": "metaclip_2/modeling_metaclip_2.py:metaclip_2_loss", "4851": "metaclip_2/modeling_metaclip_2.py:_get_vector_norm", "4852": "metaclip_2/modeling_metaclip_2.py:MetaClip2Model", "4853": "metaclip_2/modeling_metaclip_2.py:MetaClip2VisionTransformer", "4854": "metaclip_2/modeling_metaclip_2.py:MetaClip2VisionModel", "4855": "metaclip_2/modeling_metaclip_2.py:MetaClip2VisionModelOutput", "4856": "metaclip_2/modeling_metaclip_2.py:MetaClip2VisionModelWithProjection", "4857": "metaclip_2/modeling_metaclip_2.py:MetaClip2ForImageClassification", "4858": "glm/modeling_glm.py:GlmMLP", "4859": "glm/modeling_glm.py:GlmRotaryEmbedding", "4860": "glm/modeling_glm.py:repeat_kv", "4861": "glm/modeling_glm.py:eager_attention_forward", "4862": "glm/modeling_glm.py:rotate_half", "4863": "glm/modeling_glm.py:apply_rotary_pos_emb", "4864": "glm/modeling_glm.py:GlmAttention", "4865": "glm/modeling_glm.py:GlmRMSNorm", "4866": "glm/modeling_glm.py:GlmDecoderLayer", "4867": "glm/modeling_glm.py:GlmPreTrainedModel", "4868": "glm/modeling_glm.py:GlmModel", "4869": "glm/modeling_glm.py:GlmForCausalLM", "4870": "glm/modeling_glm.py:GlmForSequenceClassification", "4871": "glm/modeling_glm.py:GlmForTokenClassification", "4872": "blip_2/modeling_blip_2.py:Blip2ForConditionalGenerationModelOutput", "4873": "blip_2/modeling_blip_2.py:Blip2ImageTextMatchingModelOutput", "4874": "blip_2/modeling_blip_2.py:Blip2TextModelOutput", "4875": "blip_2/modeling_blip_2.py:Blip2VisionModelOutput", "4876": "blip_2/modeling_blip_2.py:Blip2VisionEmbeddings", "4877": "blip_2/modeling_blip_2.py:eager_attention_forward", "4878": "blip_2/modeling_blip_2.py:Blip2Attention", "4879": "blip_2/modeling_blip_2.py:Blip2MLP", "4880": "blip_2/modeling_blip_2.py:Blip2EncoderLayer", "4881": "blip_2/modeling_blip_2.py:Blip2PreTrainedModel", "4882": "blip_2/modeling_blip_2.py:Blip2Encoder", "4883": "blip_2/modeling_blip_2.py:Blip2VisionModel", "4884": "blip_2/modeling_blip_2.py:Blip2QFormerMultiHeadAttention", "4885": "blip_2/modeling_blip_2.py:Blip2QFormerSelfOutput", "4886": "blip_2/modeling_blip_2.py:Blip2QFormerAttention", "4887": "blip_2/modeling_blip_2.py:Blip2QFormerIntermediate", "4888": "blip_2/modeling_blip_2.py:Blip2QFormerOutput", "4889": "blip_2/modeling_blip_2.py:Blip2QFormerLayer", "4890": "blip_2/modeling_blip_2.py:Blip2QFormerEncoder", "4891": "blip_2/modeling_blip_2.py:Blip2TextEmbeddings", "4892": "blip_2/modeling_blip_2.py:Blip2QFormerModel", "4893": "blip_2/modeling_blip_2.py:Blip2Model", "4894": "blip_2/modeling_blip_2.py:Blip2TextModelWithProjection", "4895": "blip_2/modeling_blip_2.py:Blip2VisionModelWithProjection", "4896": "blip_2/modeling_blip_2.py:Blip2ForConditionalGeneration", "4897": "blip_2/modeling_blip_2.py:Blip2ForImageTextRetrieval", "4898": "roc_bert/modeling_roc_bert.py:RoCBertEmbeddings", "4899": "roc_bert/modeling_roc_bert.py:eager_attention_forward", "4900": "roc_bert/modeling_roc_bert.py:RoCBertSelfAttention", "4901": "roc_bert/modeling_roc_bert.py:RoCBertCrossAttention", "4902": "roc_bert/modeling_roc_bert.py:RoCBertSelfOutput", "4903": "roc_bert/modeling_roc_bert.py:RoCBertAttention", "4904": "roc_bert/modeling_roc_bert.py:RoCBertIntermediate", "4905": "roc_bert/modeling_roc_bert.py:RoCBertOutput", "4906": "roc_bert/modeling_roc_bert.py:RoCBertLayer", "4907": "roc_bert/modeling_roc_bert.py:RoCBertEncoder", "4908": "roc_bert/modeling_roc_bert.py:RoCBertPooler", "4909": "roc_bert/modeling_roc_bert.py:RoCBertPredictionHeadTransform", "4910": "roc_bert/modeling_roc_bert.py:RoCBertLMPredictionHead", "4911": "roc_bert/modeling_roc_bert.py:RoCBertOnlyMLMHead", "4912": "roc_bert/modeling_roc_bert.py:RoCBertPreTrainedModel", "4913": "roc_bert/modeling_roc_bert.py:RoCBertModel", "4914": "roc_bert/modeling_roc_bert.py:RoCBertForPreTraining", "4915": "roc_bert/modeling_roc_bert.py:RoCBertForMaskedLM", "4916": "roc_bert/modeling_roc_bert.py:RoCBertForCausalLM", "4917": "roc_bert/modeling_roc_bert.py:RoCBertForSequenceClassification", "4918": "roc_bert/modeling_roc_bert.py:RoCBertForMultipleChoice", "4919": "roc_bert/modeling_roc_bert.py:RoCBertForTokenClassification", "4920": "roc_bert/modeling_roc_bert.py:RoCBertForQuestionAnswering", "4921": "longcat_flash/modeling_longcat_flash.py:LongcatFlashRMSNorm", "4922": "longcat_flash/modeling_longcat_flash.py:LongcatFlashRotaryEmbedding", "4923": "longcat_flash/modeling_longcat_flash.py:LongcatFlashMLP", "4924": "longcat_flash/modeling_longcat_flash.py:LongcatFlashTopkRouter", "4925": "longcat_flash/modeling_longcat_flash.py:LongcatFlashExperts", "4926": "longcat_flash/modeling_longcat_flash.py:LongcatFlashMoE", "4927": "longcat_flash/modeling_longcat_flash.py:rotate_half", "4928": "longcat_flash/modeling_longcat_flash.py:repeat_kv", "4929": "longcat_flash/modeling_longcat_flash.py:eager_attention_forward", "4930": "longcat_flash/modeling_longcat_flash.py:apply_rotary_pos_emb_interleave", "4931": "longcat_flash/modeling_longcat_flash.py:yarn_get_mscale", "4932": "longcat_flash/modeling_longcat_flash.py:LongcatFlashMLA", "4933": "longcat_flash/modeling_longcat_flash.py:LongcatFlashDecoderLayer", "4934": "longcat_flash/modeling_longcat_flash.py:LongcatFlashPreTrainedModel", "4935": "longcat_flash/modeling_longcat_flash.py:LongcatFlashModel", "4936": "longcat_flash/modeling_longcat_flash.py:LongcatFlashForCausalLM", "4937": "dpr/modeling_dpr.py:DPRContextEncoderOutput", "4938": "dpr/modeling_dpr.py:DPRQuestionEncoderOutput", "4939": "dpr/modeling_dpr.py:DPRReaderOutput", "4940": "dpr/modeling_dpr.py:DPRPreTrainedModel", "4941": "dpr/modeling_dpr.py:DPREncoder", "4942": "dpr/modeling_dpr.py:DPRSpanPredictor", "4943": "dpr/modeling_dpr.py:DPRPretrainedContextEncoder", "4944": "dpr/modeling_dpr.py:DPRPretrainedQuestionEncoder", "4945": "dpr/modeling_dpr.py:DPRPretrainedReader", "4946": "dpr/modeling_dpr.py:DPRContextEncoder", "4947": "dpr/modeling_dpr.py:DPRQuestionEncoder", "4948": "dpr/modeling_dpr.py:DPRReader", "4949": "t5gemma/modeling_t5gemma.py:T5GemmaRMSNorm", "4950": "t5gemma/modeling_t5gemma.py:T5GemmaMLP", "4951": "t5gemma/modeling_t5gemma.py:T5GemmaRotaryEmbedding", "4952": "t5gemma/modeling_t5gemma.py:rotate_half", "4953": "t5gemma/modeling_t5gemma.py:apply_rotary_pos_emb", "4954": "t5gemma/modeling_t5gemma.py:repeat_kv", "4955": "t5gemma/modeling_t5gemma.py:eager_attention_forward", "4956": "t5gemma/modeling_t5gemma.py:T5GemmaSelfAttention", "4957": "t5gemma/modeling_t5gemma.py:T5GemmaCrossAttention", "4958": "t5gemma/modeling_t5gemma.py:T5GemmaEncoderLayer", "4959": "t5gemma/modeling_t5gemma.py:T5GemmaDecoderLayer", "4960": "t5gemma/modeling_t5gemma.py:T5GemmaClassificationHead", "4961": "t5gemma/modeling_t5gemma.py:T5GemmaLMHead", "4962": "t5gemma/modeling_t5gemma.py:T5GemmaPreTrainedModel", "4963": "t5gemma/modeling_t5gemma.py:bidirectional_mask_function", "4964": "t5gemma/modeling_t5gemma.py:sliding_window_bidirectional_mask_function", "4965": "t5gemma/modeling_t5gemma.py:make_default_2d_attention_mask", "4966": "t5gemma/modeling_t5gemma.py:T5GemmaEncoder", "4967": "t5gemma/modeling_t5gemma.py:T5GemmaDecoder", "4968": "t5gemma/modeling_t5gemma.py:T5GemmaModel", "4969": "t5gemma/modeling_t5gemma.py:T5GemmaEncoderModel", "4970": "t5gemma/modeling_t5gemma.py:T5GemmaForConditionalGeneration", "4971": "t5gemma/modeling_t5gemma.py:T5GemmaForSequenceClassification", "4972": "t5gemma/modeling_t5gemma.py:T5GemmaForTokenClassification", "4973": "zamba2/modeling_zamba2.py:Zamba2RMSNormGated", "4974": "zamba2/modeling_zamba2.py:Zamba2RMSNorm", "4975": "zamba2/modeling_zamba2.py:Zamba2HybridDynamicCache", "4976": "zamba2/modeling_zamba2.py:Zamba2RotaryEmbedding", "4977": "zamba2/modeling_zamba2.py:repeat_kv", "4978": "zamba2/modeling_zamba2.py:eager_attention_forward", "4979": "zamba2/modeling_zamba2.py:rotate_half", "4980": "zamba2/modeling_zamba2.py:apply_rotary_pos_emb", "4981": "zamba2/modeling_zamba2.py:Zamba2Attention", "4982": "zamba2/modeling_zamba2.py:pad_tensor_by_size", "4983": "zamba2/modeling_zamba2.py:reshape_into_chunks", "4984": "zamba2/modeling_zamba2.py:segment_sum", "4985": "zamba2/modeling_zamba2.py:Zamba2MambaMixer", "4986": "zamba2/modeling_zamba2.py:Zamba2MLP", "4987": "zamba2/modeling_zamba2.py:Zamba2AttentionDecoderLayer", "4988": "zamba2/modeling_zamba2.py:Zamba2MambaDecoderLayer", "4989": "zamba2/modeling_zamba2.py:Zamba2HybridLayer", "4990": "zamba2/modeling_zamba2.py:Zamba2PreTrainedModel", "4991": "zamba2/modeling_zamba2.py:Zamba2Model", "4992": "zamba2/modeling_zamba2.py:Zamba2ForCausalLM", "4993": "zamba2/modeling_zamba2.py:Zamba2ForSequenceClassification", "4994": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeRMSNorm", "4995": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeModelOutputWithPast", "4996": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextRotaryEmbedding", "4997": "glm4v_moe/modeling_glm4v_moe.py:repeat_kv", "4998": "glm4v_moe/modeling_glm4v_moe.py:eager_attention_forward", "4999": "glm4v_moe/modeling_glm4v_moe.py:rotate_half", "5000": "glm4v_moe/modeling_glm4v_moe.py:apply_rotary_pos_emb", "5001": "glm4v_moe/modeling_glm4v_moe.py:apply_multimodal_rotary_pos_emb", "5002": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextAttention", "5003": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextTopkRouter", "5004": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextNaiveMoe", "5005": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextMoE", "5006": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextMLP", "5007": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextRMSNorm", "5008": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextDecoderLayer", "5009": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoePreTrainedModel", "5010": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeCausalLMOutputWithPast", "5011": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeVisionRotaryEmbedding", "5012": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeisionMlp", "5013": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeVisionPatchEmbed", "5014": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeVisionPatchMerger", "5015": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeVisionEmbeddings", "5016": "glm4v_moe/modeling_glm4v_moe.py:apply_rotary_pos_emb_vision", "5017": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeVisionAttention", "5018": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeVisionBlock", "5019": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeVisionModel", "5020": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextModel", "5021": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeModel", "5022": "glm4v_moe/modeling_glm4v_moe.py:load_balancing_loss_func", "5023": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeForConditionalGeneration", "5024": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextRMSNorm", "5025": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextExperts", "5026": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextTopKRouter", "5027": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextSparseMoeBlock", "5028": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:rotate_half", "5029": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:repeat_kv", "5030": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:eager_attention_forward", "5031": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:apply_rotary_pos_emb", "5032": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextAttention", "5033": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextMLP", "5034": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextDecoderLayer", "5035": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoePreTrainedModel", "5036": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeVisionRotaryEmbedding", "5037": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeVisionMLP", "5038": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeVisionPatchEmbed", "5039": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeVisionPatchMerger", "5040": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:apply_rotary_pos_emb_vision", "5041": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeVisionAttention", "5042": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeVisionBlock", "5043": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeVisionModel", "5044": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextRotaryEmbedding", "5045": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextModel", "5046": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeCausalLMOutputWithPast", "5047": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeModelOutputWithPast", "5048": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeModel", "5049": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:load_balancing_loss_func", "5050": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeForConditionalGeneration", "5051": "mobilevit/modeling_mobilevit.py:make_divisible", "5052": "mobilevit/modeling_mobilevit.py:MobileViTConvLayer", "5053": "mobilevit/modeling_mobilevit.py:MobileViTInvertedResidual", "5054": "mobilevit/modeling_mobilevit.py:MobileViTMobileNetLayer", "5055": "mobilevit/modeling_mobilevit.py:MobileViTSelfAttention", "5056": "mobilevit/modeling_mobilevit.py:MobileViTSelfOutput", "5057": "mobilevit/modeling_mobilevit.py:MobileViTAttention", "5058": "mobilevit/modeling_mobilevit.py:MobileViTIntermediate", "5059": "mobilevit/modeling_mobilevit.py:MobileViTOutput", "5060": "mobilevit/modeling_mobilevit.py:MobileViTTransformerLayer", "5061": "mobilevit/modeling_mobilevit.py:MobileViTTransformer", "5062": "mobilevit/modeling_mobilevit.py:MobileViTLayer", "5063": "mobilevit/modeling_mobilevit.py:MobileViTEncoder", "5064": "mobilevit/modeling_mobilevit.py:MobileViTPreTrainedModel", "5065": "mobilevit/modeling_mobilevit.py:MobileViTModel", "5066": "mobilevit/modeling_mobilevit.py:MobileViTForImageClassification", "5067": "mobilevit/modeling_mobilevit.py:MobileViTASPPPooling", "5068": "mobilevit/modeling_mobilevit.py:MobileViTASPP", "5069": "mobilevit/modeling_mobilevit.py:MobileViTDeepLabV3", "5070": "mobilevit/modeling_mobilevit.py:MobileViTForSemanticSegmentation", "5071": "exaone4/modeling_exaone4.py:Exaone4RMSNorm", "5072": "exaone4/modeling_exaone4.py:Exaone4RotaryEmbedding", "5073": "exaone4/modeling_exaone4.py:rotate_half", "5074": "exaone4/modeling_exaone4.py:apply_rotary_pos_emb", "5075": "exaone4/modeling_exaone4.py:repeat_kv", "5076": "exaone4/modeling_exaone4.py:eager_attention_forward", "5077": "exaone4/modeling_exaone4.py:Exaone4Attention", "5078": "exaone4/modeling_exaone4.py:Exaone4MLP", "5079": "exaone4/modeling_exaone4.py:Exaone4DecoderLayer", "5080": "exaone4/modeling_exaone4.py:Exaone4PreTrainedModel", "5081": "exaone4/modeling_exaone4.py:Exaone4Model", "5082": "exaone4/modeling_exaone4.py:Exaone4ForCausalLM", "5083": "exaone4/modeling_exaone4.py:Exaone4ForSequenceClassification", "5084": "exaone4/modeling_exaone4.py:Exaone4ForTokenClassification", "5085": "exaone4/modeling_exaone4.py:Exaone4ForQuestionAnswering", "5086": "dinov3_convnext/modeling_dinov3_convnext.py:drop_path", "5087": "dinov3_convnext/modeling_dinov3_convnext.py:DINOv3ConvNextDropPath", "5088": "dinov3_convnext/modeling_dinov3_convnext.py:DINOv3ConvNextLayerNorm", "5089": "dinov3_convnext/modeling_dinov3_convnext.py:DINOv3ConvNextLayer", "5090": "dinov3_convnext/modeling_dinov3_convnext.py:DINOv3ConvNextStage", "5091": "dinov3_convnext/modeling_dinov3_convnext.py:DINOv3ConvNextPreTrainedModel", "5092": "dinov3_convnext/modeling_dinov3_convnext.py:DINOv3ConvNextModel", "5093": "dinov3_convnext/modeling_dinov3_convnext.py:DINOv3ConvNextBackbone", "5094": "mobilebert/modeling_mobilebert.py:NoNorm", "5095": "mobilebert/modeling_mobilebert.py:MobileBertEmbeddings", "5096": "mobilebert/modeling_mobilebert.py:eager_attention_forward", "5097": "mobilebert/modeling_mobilebert.py:MobileBertSelfAttention", "5098": "mobilebert/modeling_mobilebert.py:MobileBertSelfOutput", "5099": "mobilebert/modeling_mobilebert.py:MobileBertAttention", "5100": "mobilebert/modeling_mobilebert.py:MobileBertIntermediate", "5101": "mobilebert/modeling_mobilebert.py:OutputBottleneck", "5102": "mobilebert/modeling_mobilebert.py:MobileBertOutput", "5103": "mobilebert/modeling_mobilebert.py:BottleneckLayer", "5104": "mobilebert/modeling_mobilebert.py:Bottleneck", "5105": "mobilebert/modeling_mobilebert.py:FFNOutput", "5106": "mobilebert/modeling_mobilebert.py:FFNLayer", "5107": "mobilebert/modeling_mobilebert.py:MobileBertLayer", "5108": "mobilebert/modeling_mobilebert.py:MobileBertEncoder", "5109": "mobilebert/modeling_mobilebert.py:MobileBertPooler", "5110": "mobilebert/modeling_mobilebert.py:MobileBertPredictionHeadTransform", "5111": "mobilebert/modeling_mobilebert.py:MobileBertLMPredictionHead", "5112": "mobilebert/modeling_mobilebert.py:MobileBertOnlyMLMHead", "5113": "mobilebert/modeling_mobilebert.py:MobileBertPreTrainingHeads", "5114": "mobilebert/modeling_mobilebert.py:MobileBertPreTrainedModel", "5115": "mobilebert/modeling_mobilebert.py:MobileBertForPreTrainingOutput", "5116": "mobilebert/modeling_mobilebert.py:MobileBertModel", "5117": "mobilebert/modeling_mobilebert.py:MobileBertForPreTraining", "5118": "mobilebert/modeling_mobilebert.py:MobileBertForMaskedLM", "5119": "mobilebert/modeling_mobilebert.py:MobileBertOnlyNSPHead", "5120": "mobilebert/modeling_mobilebert.py:MobileBertForNextSentencePrediction", "5121": "mobilebert/modeling_mobilebert.py:MobileBertForSequenceClassification", "5122": "mobilebert/modeling_mobilebert.py:MobileBertForQuestionAnswering", "5123": "mobilebert/modeling_mobilebert.py:MobileBertForMultipleChoice", "5124": "mobilebert/modeling_mobilebert.py:MobileBertForTokenClassification", "5125": "siglip/modeling_siglip.py:variance_scaling_", "5126": "siglip/modeling_siglip.py:lecun_normal_", "5127": "siglip/modeling_siglip.py:default_flax_embed_init", "5128": "siglip/modeling_siglip.py:SiglipVisionModelOutput", "5129": "siglip/modeling_siglip.py:SiglipTextModelOutput", "5130": "siglip/modeling_siglip.py:SiglipOutput", "5131": "siglip/modeling_siglip.py:SiglipVisionEmbeddings", "5132": "siglip/modeling_siglip.py:SiglipTextEmbeddings", "5133": "siglip/modeling_siglip.py:eager_attention_forward", "5134": "siglip/modeling_siglip.py:SiglipAttention", "5135": "siglip/modeling_siglip.py:SiglipMLP", "5136": "siglip/modeling_siglip.py:SiglipEncoderLayer", "5137": "siglip/modeling_siglip.py:SiglipPreTrainedModel", "5138": "siglip/modeling_siglip.py:SiglipEncoder", "5139": "siglip/modeling_siglip.py:SiglipTextTransformer", "5140": "siglip/modeling_siglip.py:SiglipTextModel", "5141": "siglip/modeling_siglip.py:SiglipVisionTransformer", "5142": "siglip/modeling_siglip.py:SiglipMultiheadAttentionPoolingHead", "5143": "siglip/modeling_siglip.py:SiglipVisionModel", "5144": "siglip/modeling_siglip.py:SiglipModel", "5145": "siglip/modeling_siglip.py:SiglipForImageClassification", "5146": "mobilenet_v2/modeling_mobilenet_v2.py:make_divisible", "5147": "mobilenet_v2/modeling_mobilenet_v2.py:apply_depth_multiplier", "5148": "mobilenet_v2/modeling_mobilenet_v2.py:apply_tf_padding", "5149": "mobilenet_v2/modeling_mobilenet_v2.py:MobileNetV2ConvLayer", "5150": "mobilenet_v2/modeling_mobilenet_v2.py:MobileNetV2InvertedResidual", "5151": "mobilenet_v2/modeling_mobilenet_v2.py:MobileNetV2Stem", "5152": "mobilenet_v2/modeling_mobilenet_v2.py:MobileNetV2PreTrainedModel", "5153": "mobilenet_v2/modeling_mobilenet_v2.py:MobileNetV2Model", "5154": "mobilenet_v2/modeling_mobilenet_v2.py:MobileNetV2ForImageClassification", "5155": "mobilenet_v2/modeling_mobilenet_v2.py:MobileNetV2DeepLabV3Plus", "5156": "mobilenet_v2/modeling_mobilenet_v2.py:MobileNetV2ForSemanticSegmentation", "5157": "deberta_v2/modeling_deberta_v2.py:DebertaV2SelfOutput", "5158": "deberta_v2/modeling_deberta_v2.py:make_log_bucket_position", "5159": "deberta_v2/modeling_deberta_v2.py:build_relative_position", "5160": "deberta_v2/modeling_deberta_v2.py:c2p_dynamic_expand", "5161": "deberta_v2/modeling_deberta_v2.py:p2c_dynamic_expand", "5162": "deberta_v2/modeling_deberta_v2.py:pos_dynamic_expand", "5163": "deberta_v2/modeling_deberta_v2.py:scaled_size_sqrt", "5164": "deberta_v2/modeling_deberta_v2.py:build_rpos", "5165": "deberta_v2/modeling_deberta_v2.py:DisentangledSelfAttention", "5166": "deberta_v2/modeling_deberta_v2.py:DebertaV2Attention", "5167": "deberta_v2/modeling_deberta_v2.py:DebertaV2Intermediate", "5168": "deberta_v2/modeling_deberta_v2.py:DebertaV2Output", "5169": "deberta_v2/modeling_deberta_v2.py:DebertaV2Layer", "5170": "deberta_v2/modeling_deberta_v2.py:ConvLayer", "5171": "deberta_v2/modeling_deberta_v2.py:DebertaV2Embeddings", "5172": "deberta_v2/modeling_deberta_v2.py:DebertaV2Encoder", "5173": "deberta_v2/modeling_deberta_v2.py:DebertaV2PreTrainedModel", "5174": "deberta_v2/modeling_deberta_v2.py:DebertaV2Model", "5175": "deberta_v2/modeling_deberta_v2.py:LegacyDebertaV2PredictionHeadTransform", "5176": "deberta_v2/modeling_deberta_v2.py:LegacyDebertaV2LMPredictionHead", "5177": "deberta_v2/modeling_deberta_v2.py:LegacyDebertaV2OnlyMLMHead", "5178": "deberta_v2/modeling_deberta_v2.py:DebertaV2LMPredictionHead", "5179": "deberta_v2/modeling_deberta_v2.py:DebertaV2OnlyMLMHead", "5180": "deberta_v2/modeling_deberta_v2.py:DebertaV2ForMaskedLM", "5181": "deberta_v2/modeling_deberta_v2.py:ContextPooler", "5182": "deberta_v2/modeling_deberta_v2.py:DebertaV2ForSequenceClassification", "5183": "deberta_v2/modeling_deberta_v2.py:DebertaV2ForTokenClassification", "5184": "deberta_v2/modeling_deberta_v2.py:DebertaV2ForQuestionAnswering", "5185": "deberta_v2/modeling_deberta_v2.py:DebertaV2ForMultipleChoice", "5186": "camembert/modeling_camembert.py:CamembertEmbeddings", "5187": "camembert/modeling_camembert.py:eager_attention_forward", "5188": "camembert/modeling_camembert.py:CamembertSelfAttention", "5189": "camembert/modeling_camembert.py:CamembertCrossAttention", "5190": "camembert/modeling_camembert.py:CamembertSelfOutput", "5191": "camembert/modeling_camembert.py:CamembertAttention", "5192": "camembert/modeling_camembert.py:CamembertIntermediate", "5193": "camembert/modeling_camembert.py:CamembertOutput", "5194": "camembert/modeling_camembert.py:CamembertLayer", "5195": "camembert/modeling_camembert.py:CamembertLMHead", "5196": "camembert/modeling_camembert.py:CamembertPreTrainedModel", "5197": "camembert/modeling_camembert.py:CamembertEncoder", "5198": "camembert/modeling_camembert.py:CamembertPooler", "5199": "camembert/modeling_camembert.py:CamembertModel", "5200": "camembert/modeling_camembert.py:CamembertForMaskedLM", "5201": "camembert/modeling_camembert.py:CamembertClassificationHead", "5202": "camembert/modeling_camembert.py:CamembertForSequenceClassification", "5203": "camembert/modeling_camembert.py:CamembertForMultipleChoice", "5204": "camembert/modeling_camembert.py:CamembertForTokenClassification", "5205": "camembert/modeling_camembert.py:CamembertForQuestionAnswering", "5206": "camembert/modeling_camembert.py:CamembertForCausalLM", "5207": "unispeech/modeling_unispeech.py:UniSpeechForPreTrainingOutput", "5208": "unispeech/modeling_unispeech.py:UniSpeechSamePadLayer", "5209": "unispeech/modeling_unispeech.py:UniSpeechPositionalConvEmbedding", "5210": "unispeech/modeling_unispeech.py:UniSpeechNoLayerNormConvLayer", "5211": "unispeech/modeling_unispeech.py:UniSpeechLayerNormConvLayer", "5212": "unispeech/modeling_unispeech.py:UniSpeechGroupNormConvLayer", "5213": "unispeech/modeling_unispeech.py:UniSpeechFeatureEncoder", "5214": "unispeech/modeling_unispeech.py:UniSpeechFeatureProjection", "5215": "unispeech/modeling_unispeech.py:eager_attention_forward", "5216": "unispeech/modeling_unispeech.py:UniSpeechAttention", "5217": "unispeech/modeling_unispeech.py:UniSpeechFeedForward", "5218": "unispeech/modeling_unispeech.py:UniSpeechEncoderLayer", "5219": "unispeech/modeling_unispeech.py:UniSpeechEncoder", "5220": "unispeech/modeling_unispeech.py:UniSpeechAttnAdapterLayer", "5221": "unispeech/modeling_unispeech.py:UniSpeechEncoderLayerStableLayerNorm", "5222": "unispeech/modeling_unispeech.py:UniSpeechEncoderStableLayerNorm", "5223": "unispeech/modeling_unispeech.py:UniSpeechGumbelVectorQuantizer", "5224": "unispeech/modeling_unispeech.py:UniSpeechPreTrainedModel", "5225": "unispeech/modeling_unispeech.py:_compute_mask_indices", "5226": "unispeech/modeling_unispeech.py:UniSpeechModel", "5227": "unispeech/modeling_unispeech.py:UniSpeechForPreTraining", "5228": "unispeech/modeling_unispeech.py:UniSpeechForCTC", "5229": "unispeech/modeling_unispeech.py:UniSpeechForSequenceClassification", "5230": "dinat/modeling_dinat.py:DinatEncoderOutput", "5231": "dinat/modeling_dinat.py:DinatModelOutput", "5232": "dinat/modeling_dinat.py:DinatImageClassifierOutput", "5233": "dinat/modeling_dinat.py:DinatEmbeddings", "5234": "dinat/modeling_dinat.py:DinatPatchEmbeddings", "5235": "dinat/modeling_dinat.py:DinatDownsampler", "5236": "dinat/modeling_dinat.py:drop_path", "5237": "dinat/modeling_dinat.py:DinatDropPath", "5238": "dinat/modeling_dinat.py:NeighborhoodAttention", "5239": "dinat/modeling_dinat.py:NeighborhoodAttentionOutput", "5240": "dinat/modeling_dinat.py:NeighborhoodAttentionModule", "5241": "dinat/modeling_dinat.py:DinatIntermediate", "5242": "dinat/modeling_dinat.py:DinatOutput", "5243": "dinat/modeling_dinat.py:DinatLayer", "5244": "dinat/modeling_dinat.py:DinatStage", "5245": "dinat/modeling_dinat.py:DinatEncoder", "5246": "dinat/modeling_dinat.py:DinatPreTrainedModel", "5247": "dinat/modeling_dinat.py:DinatModel", "5248": "dinat/modeling_dinat.py:DinatForImageClassification", "5249": "dinat/modeling_dinat.py:DinatBackbone", "5250": "jetmoe/modeling_jetmoe.py:JetMoeRMSNorm", "5251": "jetmoe/modeling_jetmoe.py:JetMoeRotaryEmbedding", "5252": "jetmoe/modeling_jetmoe.py:JetMoeParallelExperts", "5253": "jetmoe/modeling_jetmoe.py:JetMoeTopKGating", "5254": "jetmoe/modeling_jetmoe.py:JetMoeMoE", "5255": "jetmoe/modeling_jetmoe.py:JetMoeMoA", "5256": "jetmoe/modeling_jetmoe.py:rotate_half", "5257": "jetmoe/modeling_jetmoe.py:apply_rotary_pos_emb", "5258": "jetmoe/modeling_jetmoe.py:repeat_kv", "5259": "jetmoe/modeling_jetmoe.py:eager_attention_forward", "5260": "jetmoe/modeling_jetmoe.py:JetMoeAttention", "5261": "jetmoe/modeling_jetmoe.py:JetMoeDecoderLayer", "5262": "jetmoe/modeling_jetmoe.py:JetMoePreTrainedModel", "5263": "jetmoe/modeling_jetmoe.py:JetMoeModel", "5264": "jetmoe/modeling_jetmoe.py:load_balancing_loss_func", "5265": "jetmoe/modeling_jetmoe.py:JetMoeForCausalLM", "5266": "jetmoe/modeling_jetmoe.py:JetMoeForSequenceClassification", "5267": "vitdet/modeling_vitdet.py:VitDetEmbeddings", "5268": "vitdet/modeling_vitdet.py:get_rel_pos", "5269": "vitdet/modeling_vitdet.py:add_decomposed_relative_positions", "5270": "vitdet/modeling_vitdet.py:VitDetAttention", "5271": "vitdet/modeling_vitdet.py:drop_path", "5272": "vitdet/modeling_vitdet.py:VitDetDropPath", "5273": "vitdet/modeling_vitdet.py:VitDetLayerNorm", "5274": "vitdet/modeling_vitdet.py:VitDetResBottleneckBlock", "5275": "vitdet/modeling_vitdet.py:VitDetMlp", "5276": "vitdet/modeling_vitdet.py:window_partition", "5277": "vitdet/modeling_vitdet.py:window_unpartition", "5278": "vitdet/modeling_vitdet.py:VitDetLayer", "5279": "vitdet/modeling_vitdet.py:VitDetEncoder", "5280": "vitdet/modeling_vitdet.py:VitDetPreTrainedModel", "5281": "vitdet/modeling_vitdet.py:VitDetModel", "5282": "vitdet/modeling_vitdet.py:VitDetBackbone", "5283": "dac/modeling_dac.py:DacOutput", "5284": "dac/modeling_dac.py:DacEncoderOutput", "5285": "dac/modeling_dac.py:DacDecoderOutput", "5286": "dac/modeling_dac.py:Snake1d", "5287": "dac/modeling_dac.py:DacVectorQuantize", "5288": "dac/modeling_dac.py:DacResidualUnit", "5289": "dac/modeling_dac.py:DacEncoderBlock", "5290": "dac/modeling_dac.py:DacDecoderBlock", "5291": "dac/modeling_dac.py:DacResidualVectorQuantizer", "5292": "dac/modeling_dac.py:DacDecoder", "5293": "dac/modeling_dac.py:DacEncoder", "5294": "dac/modeling_dac.py:DacPreTrainedModel", "5295": "dac/modeling_dac.py:DacModel", "5296": "prophetnet/modeling_prophetnet.py:softmax", "5297": "prophetnet/modeling_prophetnet.py:ngram_attention_bias", "5298": "prophetnet/modeling_prophetnet.py:compute_relative_buckets", "5299": "prophetnet/modeling_prophetnet.py:compute_all_stream_relative_buckets", "5300": "prophetnet/modeling_prophetnet.py:ProphetNetSeq2SeqLMOutput", "5301": "prophetnet/modeling_prophetnet.py:ProphetNetSeq2SeqModelOutput", "5302": "prophetnet/modeling_prophetnet.py:ProphetNetDecoderModelOutput", "5303": "prophetnet/modeling_prophetnet.py:ProphetNetDecoderLMOutput", "5304": "prophetnet/modeling_prophetnet.py:ProphetNetPreTrainedModel", "5305": "prophetnet/modeling_prophetnet.py:ProphetNetPositionalEmbeddings", "5306": "prophetnet/modeling_prophetnet.py:ProphetNetAttention", "5307": "prophetnet/modeling_prophetnet.py:ProphetNetFeedForward", "5308": "prophetnet/modeling_prophetnet.py:ProphetNetNgramSelfAttention", "5309": "prophetnet/modeling_prophetnet.py:ProphetNetEncoderLayer", "5310": "prophetnet/modeling_prophetnet.py:ProphetNetDecoderLayer", "5311": "prophetnet/modeling_prophetnet.py:ProphetNetEncoder", "5312": "prophetnet/modeling_prophetnet.py:ProphetNetDecoder", "5313": "prophetnet/modeling_prophetnet.py:ProphetNetModel", "5314": "prophetnet/modeling_prophetnet.py:ProphetNetForConditionalGeneration", "5315": "prophetnet/modeling_prophetnet.py:ProphetNetForCausalLM", "5316": "prophetnet/modeling_prophetnet.py:ProphetNetDecoderWrapper", "5317": "mpnet/modeling_mpnet.py:MPNetPreTrainedModel", "5318": "mpnet/modeling_mpnet.py:MPNetEmbeddings", "5319": "mpnet/modeling_mpnet.py:MPNetSelfAttention", "5320": "mpnet/modeling_mpnet.py:MPNetAttention", "5321": "mpnet/modeling_mpnet.py:MPNetIntermediate", "5322": "mpnet/modeling_mpnet.py:MPNetOutput", "5323": "mpnet/modeling_mpnet.py:MPNetLayer", "5324": "mpnet/modeling_mpnet.py:MPNetEncoder", "5325": "mpnet/modeling_mpnet.py:MPNetPooler", "5326": "mpnet/modeling_mpnet.py:MPNetModel", "5327": "mpnet/modeling_mpnet.py:MPNetForMaskedLM", "5328": "mpnet/modeling_mpnet.py:MPNetLMHead", "5329": "mpnet/modeling_mpnet.py:MPNetForSequenceClassification", "5330": "mpnet/modeling_mpnet.py:MPNetForMultipleChoice", "5331": "mpnet/modeling_mpnet.py:MPNetForTokenClassification", "5332": "mpnet/modeling_mpnet.py:MPNetClassificationHead", "5333": "mpnet/modeling_mpnet.py:MPNetForQuestionAnswering", "5334": "mpnet/modeling_mpnet.py:create_position_ids_from_input_ids", "5335": "smollm3/modeling_smollm3.py:SmolLM3RotaryEmbedding", "5336": "smollm3/modeling_smollm3.py:rotate_half", "5337": "smollm3/modeling_smollm3.py:apply_rotary_pos_emb", "5338": "smollm3/modeling_smollm3.py:repeat_kv", "5339": "smollm3/modeling_smollm3.py:eager_attention_forward", "5340": "smollm3/modeling_smollm3.py:SmolLM3Attention", "5341": "smollm3/modeling_smollm3.py:SmolLM3RMSNorm", "5342": "smollm3/modeling_smollm3.py:SmolLM3MLP", "5343": "smollm3/modeling_smollm3.py:SmolLM3DecoderLayer", "5344": "smollm3/modeling_smollm3.py:SmolLM3PreTrainedModel", "5345": "smollm3/modeling_smollm3.py:SmolLM3Model", "5346": "smollm3/modeling_smollm3.py:SmolLM3ForCausalLM", "5347": "smollm3/modeling_smollm3.py:SmolLM3ForSequenceClassification", "5348": "smollm3/modeling_smollm3.py:SmolLM3ForTokenClassification", "5349": "smollm3/modeling_smollm3.py:SmolLM3ForQuestionAnswering", "5350": "pixio/modeling_pixio.py:PixioPatchEmbeddings", "5351": "pixio/modeling_pixio.py:PixioEmbeddings", "5352": "pixio/modeling_pixio.py:eager_attention_forward", "5353": "pixio/modeling_pixio.py:PixioSelfAttention", "5354": "pixio/modeling_pixio.py:PixioSelfOutput", "5355": "pixio/modeling_pixio.py:PixioAttention", "5356": "pixio/modeling_pixio.py:drop_path", "5357": "pixio/modeling_pixio.py:PixioDropPath", "5358": "pixio/modeling_pixio.py:PixioMLP", "5359": "pixio/modeling_pixio.py:PixioLayer", "5360": "pixio/modeling_pixio.py:PixioEncoder", "5361": "pixio/modeling_pixio.py:PixioPreTrainedModel", "5362": "pixio/modeling_pixio.py:PixioModel", "5363": "pixio/modeling_pixio.py:PixioBackbone", "5364": "bark/modeling_bark.py:BarkSelfAttention", "5365": "bark/modeling_bark.py:BarkSelfFlashAttention2", "5366": "bark/modeling_bark.py:BarkMLP", "5367": "bark/modeling_bark.py:BarkBlock", "5368": "bark/modeling_bark.py:BarkPreTrainedModel", "5369": "bark/modeling_bark.py:BarkCausalModel", "5370": "bark/modeling_bark.py:BarkSemanticModel", "5371": "bark/modeling_bark.py:BarkCoarseModel", "5372": "bark/modeling_bark.py:BarkFineModel", "5373": "bark/modeling_bark.py:BarkModel", "5374": "zoedepth/modeling_zoedepth.py:ZoeDepthDepthEstimatorOutput", "5375": "zoedepth/modeling_zoedepth.py:ZoeDepthReassembleStage", "5376": "zoedepth/modeling_zoedepth.py:ZoeDepthReassembleLayer", "5377": "zoedepth/modeling_zoedepth.py:ZoeDepthFeatureFusionStage", "5378": "zoedepth/modeling_zoedepth.py:ZoeDepthPreActResidualLayer", "5379": "zoedepth/modeling_zoedepth.py:ZoeDepthFeatureFusionLayer", "5380": "zoedepth/modeling_zoedepth.py:ZoeDepthNeck", "5381": "zoedepth/modeling_zoedepth.py:ZoeDepthRelativeDepthEstimationHead", "5382": "zoedepth/modeling_zoedepth.py:log_binom", "5383": "zoedepth/modeling_zoedepth.py:LogBinomialSoftmax", "5384": "zoedepth/modeling_zoedepth.py:ZoeDepthConditionalLogBinomialSoftmax", "5385": "zoedepth/modeling_zoedepth.py:ZoeDepthSeedBinRegressor", "5386": "zoedepth/modeling_zoedepth.py:inv_attractor", "5387": "zoedepth/modeling_zoedepth.py:ZoeDepthAttractorLayer", "5388": "zoedepth/modeling_zoedepth.py:ZoeDepthAttractorLayerUnnormed", "5389": "zoedepth/modeling_zoedepth.py:ZoeDepthProjector", "5390": "zoedepth/modeling_zoedepth.py:ZoeDepthMultiheadAttention", "5391": "zoedepth/modeling_zoedepth.py:ZoeDepthTransformerEncoderLayer", "5392": "zoedepth/modeling_zoedepth.py:ZoeDepthPatchTransformerEncoder", "5393": "zoedepth/modeling_zoedepth.py:ZoeDepthMLPClassifier", "5394": "zoedepth/modeling_zoedepth.py:ZoeDepthMultipleMetricDepthEstimationHeads", "5395": "zoedepth/modeling_zoedepth.py:ZoeDepthMetricDepthEstimationHead", "5396": "zoedepth/modeling_zoedepth.py:ZoeDepthPreTrainedModel", "5397": "zoedepth/modeling_zoedepth.py:ZoeDepthForDepthEstimation", "5398": "levit/modeling_levit.py:LevitForImageClassificationWithTeacherOutput", "5399": "levit/modeling_levit.py:LevitConvEmbeddings", "5400": "levit/modeling_levit.py:LevitPatchEmbeddings", "5401": "levit/modeling_levit.py:MLPLayerWithBN", "5402": "levit/modeling_levit.py:LevitSubsample", "5403": "levit/modeling_levit.py:LevitAttention", "5404": "levit/modeling_levit.py:LevitAttentionSubsample", "5405": "levit/modeling_levit.py:LevitMLPLayer", "5406": "levit/modeling_levit.py:LevitResidualLayer", "5407": "levit/modeling_levit.py:LevitStage", "5408": "levit/modeling_levit.py:LevitEncoder", "5409": "levit/modeling_levit.py:LevitClassificationLayer", "5410": "levit/modeling_levit.py:LevitPreTrainedModel", "5411": "levit/modeling_levit.py:LevitModel", "5412": "levit/modeling_levit.py:LevitForImageClassification", "5413": "levit/modeling_levit.py:LevitForImageClassificationWithTeacher", "5414": "llama/modeling_llama.py:LlamaRMSNorm", "5415": "llama/modeling_llama.py:LlamaRotaryEmbedding", "5416": "llama/modeling_llama.py:rotate_half", "5417": "llama/modeling_llama.py:apply_rotary_pos_emb", "5418": "llama/modeling_llama.py:LlamaMLP", "5419": "llama/modeling_llama.py:repeat_kv", "5420": "llama/modeling_llama.py:eager_attention_forward", "5421": "llama/modeling_llama.py:LlamaAttention", "5422": "llama/modeling_llama.py:LlamaDecoderLayer", "5423": "llama/modeling_llama.py:LlamaPreTrainedModel", "5424": "llama/modeling_llama.py:LlamaModel", "5425": "llama/modeling_llama.py:LlamaForCausalLM", "5426": "llama/modeling_llama.py:LlamaForSequenceClassification", "5427": "llama/modeling_llama.py:LlamaForQuestionAnswering", "5428": "llama/modeling_llama.py:LlamaForTokenClassification", "5429": "parakeet/modeling_parakeet.py:ParakeetEncoderModelOutput", "5430": "parakeet/modeling_parakeet.py:ParakeetEncoderRelPositionalEncoding", "5431": "parakeet/modeling_parakeet.py:ParakeetEncoderFeedForward", "5432": "parakeet/modeling_parakeet.py:ParakeetEncoderConvolutionModule", "5433": "parakeet/modeling_parakeet.py:rotate_half", "5434": "parakeet/modeling_parakeet.py:apply_rotary_pos_emb", "5435": "parakeet/modeling_parakeet.py:repeat_kv", "5436": "parakeet/modeling_parakeet.py:eager_attention_forward", "5437": "parakeet/modeling_parakeet.py:ParakeetEncoderAttention", "5438": "parakeet/modeling_parakeet.py:ParakeetEncoderSubsamplingConv2D", "5439": "parakeet/modeling_parakeet.py:ParakeetEncoderBlock", "5440": "parakeet/modeling_parakeet.py:ParakeetPreTrainedModel", "5441": "parakeet/modeling_parakeet.py:ParakeetEncoder", "5442": "parakeet/modeling_parakeet.py:ParakeetGenerateOutput", "5443": "parakeet/modeling_parakeet.py:ParakeetForCTC", "5444": "mbart/modeling_mbart.py:shift_tokens_right", "5445": "mbart/modeling_mbart.py:MBartLearnedPositionalEmbedding", "5446": "mbart/modeling_mbart.py:MBartScaledWordEmbedding", "5447": "mbart/modeling_mbart.py:eager_attention_forward", "5448": "mbart/modeling_mbart.py:MBartAttention", "5449": "mbart/modeling_mbart.py:MBartEncoderLayer", "5450": "mbart/modeling_mbart.py:MBartDecoderLayer", "5451": "mbart/modeling_mbart.py:MBartClassificationHead", "5452": "mbart/modeling_mbart.py:MBartPreTrainedModel", "5453": "mbart/modeling_mbart.py:MBartEncoder", "5454": "mbart/modeling_mbart.py:MBartDecoder", "5455": "mbart/modeling_mbart.py:MBartModel", "5456": "mbart/modeling_mbart.py:MBartForConditionalGeneration", "5457": "mbart/modeling_mbart.py:MBartForSequenceClassification", "5458": "mbart/modeling_mbart.py:MBartForQuestionAnswering", "5459": "mbart/modeling_mbart.py:MBartDecoderWrapper", "5460": "mbart/modeling_mbart.py:MBartForCausalLM", "5461": "pe_video/modeling_pe_video.py:PeVideoOutput", "5462": "pe_video/modeling_pe_video.py:PeVideoContrastiveHead", "5463": "pe_video/modeling_pe_video.py:PeVideoMaskedGroupNorm", "5464": "pe_video/modeling_pe_video.py:PeVideoConvBlock1d", "5465": "pe_video/modeling_pe_video.py:PeVideoResnetBlock1d", "5466": "pe_video/modeling_pe_video.py:PeVideoEncoderPatchEmbedder", "5467": "pe_video/modeling_pe_video.py:PeVideoEncoderEmbedder", "5468": "pe_video/modeling_pe_video.py:repeat_kv", "5469": "pe_video/modeling_pe_video.py:eager_attention_forward", "5470": "pe_video/modeling_pe_video.py:stack_freqs", "5471": "pe_video/modeling_pe_video.py:apply_rotary_pos_emb", "5472": "pe_video/modeling_pe_video.py:PeVideoEncoderRMSNorm", "5473": "pe_video/modeling_pe_video.py:PeVideoEncoderAttention", "5474": "pe_video/modeling_pe_video.py:PeVideoEncoderMLP", "5475": "pe_video/modeling_pe_video.py:PeVideoEncoderLayer", "5476": "pe_video/modeling_pe_video.py:PeVideoPreTrainedModel", "5477": "pe_video/modeling_pe_video.py:PeVideoEncoderRotaryEmbedding", "5478": "pe_video/modeling_pe_video.py:PeVideoEncoder", "5479": "pe_video/modeling_pe_video.py:PeVideoModel", "5480": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2GenerationOutput", "5481": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2TextToUnitDecoderOutput", "5482": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2TextToUnitOutput", "5483": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:shift_tokens_right", "5484": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:_compute_new_attention_mask", "5485": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:format_speech_generation_kwargs", "5486": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ConformerFeatureProjection", "5487": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ConformerFeedForward", "5488": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ConformerConvolutionModule", "5489": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ConformerSelfAttention", "5490": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ConformerEncoderLayer", "5491": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ConformerEncoder", "5492": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ConformerAdapterLayer", "5493": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ConformerAdapter", "5494": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ScaledWordEmbedding", "5495": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2SinusoidalPositionalEmbedding", "5496": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2Attention", "5497": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2FeedForwardNetwork", "5498": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2EncoderLayer", "5499": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2DecoderLayer", "5500": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2TextToUnitDecoderLayer", "5501": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2PreTrainedModel", "5502": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2SpeechEncoder", "5503": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2Encoder", "5504": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2Decoder", "5505": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2TextToUnitDecoder", "5506": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2TextToUnitModel", "5507": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2TextToUnitForConditionalGeneration", "5508": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:HifiGanResidualBlock", "5509": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2VariancePredictor", "5510": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2HifiGan", "5511": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2CodeHifiGan", "5512": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ForTextToText", "5513": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ForSpeechToText", "5514": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ForTextToSpeech", "5515": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ForSpeechToSpeech", "5516": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2Model", "5517": "lfm2/modeling_lfm2.py:Lfm2RMSNorm", "5518": "lfm2/modeling_lfm2.py:Lfm2RotaryEmbedding", "5519": "lfm2/modeling_lfm2.py:Lfm2MLP", "5520": "lfm2/modeling_lfm2.py:Lfm2HybridConvCache", "5521": "lfm2/modeling_lfm2.py:rotate_half", "5522": "lfm2/modeling_lfm2.py:apply_rotary_pos_emb", "5523": "lfm2/modeling_lfm2.py:repeat_kv", "5524": "lfm2/modeling_lfm2.py:eager_attention_forward", "5525": "lfm2/modeling_lfm2.py:Lfm2Attention", "5526": "lfm2/modeling_lfm2.py:apply_mask_to_padding_states", "5527": "lfm2/modeling_lfm2.py:Lfm2ShortConv", "5528": "lfm2/modeling_lfm2.py:Lfm2DecoderLayer", "5529": "lfm2/modeling_lfm2.py:Lfm2PreTrainedModel", "5530": "lfm2/modeling_lfm2.py:Lfm2Model", "5531": "lfm2/modeling_lfm2.py:Lfm2ForCausalLM", "5532": "phi3/modeling_phi3.py:Phi3MLP", "5533": "phi3/modeling_phi3.py:Phi3RotaryEmbedding", "5534": "phi3/modeling_phi3.py:rotate_half", "5535": "phi3/modeling_phi3.py:repeat_kv", "5536": "phi3/modeling_phi3.py:eager_attention_forward", "5537": "phi3/modeling_phi3.py:apply_rotary_pos_emb", "5538": "phi3/modeling_phi3.py:Phi3Attention", "5539": "phi3/modeling_phi3.py:Phi3RMSNorm", "5540": "phi3/modeling_phi3.py:Phi3DecoderLayer", "5541": "phi3/modeling_phi3.py:Phi3PreTrainedModel", "5542": "phi3/modeling_phi3.py:Phi3Model", "5543": "phi3/modeling_phi3.py:Phi3ForCausalLM", "5544": "phi3/modeling_phi3.py:Phi3ForSequenceClassification", "5545": "phi3/modeling_phi3.py:Phi3ForTokenClassification", "5546": "vision_encoder_decoder/modeling_vision_encoder_decoder.py:shift_tokens_right", "5547": "vision_encoder_decoder/modeling_vision_encoder_decoder.py:VisionEncoderDecoderModel", "5548": "dab_detr/modeling_dab_detr.py:DabDetrDecoderOutput", "5549": "dab_detr/modeling_dab_detr.py:DabDetrModelOutput", "5550": "dab_detr/modeling_dab_detr.py:DabDetrObjectDetectionOutput", "5551": "dab_detr/modeling_dab_detr.py:DabDetrFrozenBatchNorm2d", "5552": "dab_detr/modeling_dab_detr.py:replace_batch_norm", "5553": "dab_detr/modeling_dab_detr.py:DabDetrConvEncoder", "5554": "dab_detr/modeling_dab_detr.py:DabDetrConvModel", "5555": "dab_detr/modeling_dab_detr.py:DabDetrSinePositionEmbedding", "5556": "dab_detr/modeling_dab_detr.py:gen_sine_position_embeddings", "5557": "dab_detr/modeling_dab_detr.py:inverse_sigmoid", "5558": "dab_detr/modeling_dab_detr.py:DetrAttention", "5559": "dab_detr/modeling_dab_detr.py:DabDetrAttention", "5560": "dab_detr/modeling_dab_detr.py:DabDetrDecoderLayerSelfAttention", "5561": "dab_detr/modeling_dab_detr.py:DabDetrDecoderLayerCrossAttention", "5562": "dab_detr/modeling_dab_detr.py:DabDetrDecoderLayerFFN", "5563": "dab_detr/modeling_dab_detr.py:DabDetrEncoderLayer", "5564": "dab_detr/modeling_dab_detr.py:DabDetrDecoderLayer", "5565": "dab_detr/modeling_dab_detr.py:DabDetrMLP", "5566": "dab_detr/modeling_dab_detr.py:DabDetrPreTrainedModel", "5567": "dab_detr/modeling_dab_detr.py:DabDetrEncoder", "5568": "dab_detr/modeling_dab_detr.py:DabDetrDecoder", "5569": "dab_detr/modeling_dab_detr.py:DabDetrModel", "5570": "dab_detr/modeling_dab_detr.py:DabDetrMHAttentionMap", "5571": "dab_detr/modeling_dab_detr.py:DabDetrForObjectDetection", "5572": "xglm/modeling_xglm.py:XGLMScaledWordEmbedding", "5573": "xglm/modeling_xglm.py:XGLMSinusoidalPositionalEmbedding", "5574": "xglm/modeling_xglm.py:XGLMAttention", "5575": "xglm/modeling_xglm.py:XGLMDecoderLayer", "5576": "xglm/modeling_xglm.py:XGLMPreTrainedModel", "5577": "xglm/modeling_xglm.py:XGLMModel", "5578": "xglm/modeling_xglm.py:XGLMForCausalLM", "5579": "mt5/modeling_mt5.py:MT5LayerNorm", "5580": "mt5/modeling_mt5.py:MT5DenseActDense", "5581": "mt5/modeling_mt5.py:MT5DenseGatedActDense", "5582": "mt5/modeling_mt5.py:MT5LayerFF", "5583": "mt5/modeling_mt5.py:MT5Attention", "5584": "mt5/modeling_mt5.py:MT5LayerSelfAttention", "5585": "mt5/modeling_mt5.py:MT5LayerCrossAttention", "5586": "mt5/modeling_mt5.py:MT5Block", "5587": "mt5/modeling_mt5.py:MT5ClassificationHead", "5588": "mt5/modeling_mt5.py:MT5PreTrainedModel", "5589": "mt5/modeling_mt5.py:MT5Stack", "5590": "mt5/modeling_mt5.py:MT5Model", "5591": "mt5/modeling_mt5.py:MT5ForConditionalGeneration", "5592": "mt5/modeling_mt5.py:MT5EncoderModel", "5593": "mt5/modeling_mt5.py:MT5ForSequenceClassification", "5594": "mt5/modeling_mt5.py:MT5ForTokenClassification", "5595": "mt5/modeling_mt5.py:MT5ForQuestionAnswering", "5596": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TGenerationOutput", "5597": "seamless_m4t/modeling_seamless_m4t.py:shift_tokens_right", "5598": "seamless_m4t/modeling_seamless_m4t.py:_compute_new_attention_mask", "5599": "seamless_m4t/modeling_seamless_m4t.py:format_speech_generation_kwargs", "5600": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerPositionalConvEmbedding", "5601": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerRotaryPositionalEmbedding", "5602": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerRelPositionalEmbedding", "5603": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerSamePadLayer", "5604": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerFeatureProjection", "5605": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerFeedForward", "5606": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerConvolutionModule", "5607": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerSelfAttention", "5608": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerEncoderLayer", "5609": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerEncoder", "5610": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerAdapterLayer", "5611": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerAdapter", "5612": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TScaledWordEmbedding", "5613": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TSinusoidalPositionalEmbedding", "5614": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TAttention", "5615": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TFeedForwardNetwork", "5616": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TEncoderLayer", "5617": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TDecoderLayer", "5618": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TPreTrainedModel", "5619": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TSpeechEncoder", "5620": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TEncoder", "5621": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TDecoder", "5622": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TTextToUnitModel", "5623": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TTextToUnitForConditionalGeneration", "5624": "seamless_m4t/modeling_seamless_m4t.py:HifiGanResidualBlock", "5625": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TVariancePredictor", "5626": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4THifiGan", "5627": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TCodeHifiGan", "5628": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TForTextToText", "5629": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TForSpeechToText", "5630": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TForTextToSpeech", "5631": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TForSpeechToSpeech", "5632": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TModel", "5633": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodyOutputWithPast", "5634": "musicgen_melody/modeling_musicgen_melody.py:shift_tokens_right", "5635": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodySinusoidalPositionalEmbedding", "5636": "musicgen_melody/modeling_musicgen_melody.py:eager_attention_forward", "5637": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodyAttention", "5638": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodyDecoderLayer", "5639": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodyPreTrainedModel", "5640": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodyDecoder", "5641": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodyModel", "5642": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodyForCausalLM", "5643": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodyForConditionalGeneration", "5644": "upernet/modeling_upernet.py:UperNetConvModule", "5645": "upernet/modeling_upernet.py:UperNetPyramidPoolingBlock", "5646": "upernet/modeling_upernet.py:UperNetPyramidPoolingModule", "5647": "upernet/modeling_upernet.py:UperNetHead", "5648": "upernet/modeling_upernet.py:UperNetFCNHead", "5649": "upernet/modeling_upernet.py:UperNetPreTrainedModel", "5650": "upernet/modeling_upernet.py:UperNetForSemanticSegmentation", "5651": "deformable_detr/modeling_deformable_detr.py:MultiScaleDeformableAttention", "5652": "deformable_detr/modeling_deformable_detr.py:DeformableDetrDecoderOutput", "5653": "deformable_detr/modeling_deformable_detr.py:DeformableDetrModelOutput", "5654": "deformable_detr/modeling_deformable_detr.py:DeformableDetrObjectDetectionOutput", "5655": "deformable_detr/modeling_deformable_detr.py:inverse_sigmoid", "5656": "deformable_detr/modeling_deformable_detr.py:DeformableDetrFrozenBatchNorm2d", "5657": "deformable_detr/modeling_deformable_detr.py:replace_batch_norm", "5658": "deformable_detr/modeling_deformable_detr.py:DeformableDetrConvEncoder", "5659": "deformable_detr/modeling_deformable_detr.py:DeformableDetrConvModel", "5660": "deformable_detr/modeling_deformable_detr.py:DeformableDetrSinePositionEmbedding", "5661": "deformable_detr/modeling_deformable_detr.py:DeformableDetrLearnedPositionEmbedding", "5662": "deformable_detr/modeling_deformable_detr.py:build_position_encoding", "5663": "deformable_detr/modeling_deformable_detr.py:DeformableDetrMultiscaleDeformableAttention", "5664": "deformable_detr/modeling_deformable_detr.py:DeformableDetrMultiheadAttention", "5665": "deformable_detr/modeling_deformable_detr.py:DeformableDetrEncoderLayer", "5666": "deformable_detr/modeling_deformable_detr.py:DeformableDetrDecoderLayer", "5667": "deformable_detr/modeling_deformable_detr.py:DeformableDetrPreTrainedModel", "5668": "deformable_detr/modeling_deformable_detr.py:DeformableDetrEncoder", "5669": "deformable_detr/modeling_deformable_detr.py:DeformableDetrDecoder", "5670": "deformable_detr/modeling_deformable_detr.py:DeformableDetrModel", "5671": "deformable_detr/modeling_deformable_detr.py:DeformableDetrMLPPredictionHead", "5672": "deformable_detr/modeling_deformable_detr.py:DeformableDetrForObjectDetection", "5673": "kosmos2/modeling_kosmos2.py:_expand_mask", "5674": "kosmos2/modeling_kosmos2.py:_make_causal_mask", "5675": "kosmos2/modeling_kosmos2.py:Kosmos2ModelOutput", "5676": "kosmos2/modeling_kosmos2.py:Kosmos2ForConditionalGenerationModelOutput", "5677": "kosmos2/modeling_kosmos2.py:Kosmos2VisionEmbeddings", "5678": "kosmos2/modeling_kosmos2.py:eager_attention_forward", "5679": "kosmos2/modeling_kosmos2.py:Kosmos2VisionAttention", "5680": "kosmos2/modeling_kosmos2.py:Kosmos2VisionMLP", "5681": "kosmos2/modeling_kosmos2.py:Kosmos2VisionEncoderLayer", "5682": "kosmos2/modeling_kosmos2.py:Kosmos2VisionEncoder", "5683": "kosmos2/modeling_kosmos2.py:Kosmos2VisionTransformer", "5684": "kosmos2/modeling_kosmos2.py:Kosmos2TextSinusoidalPositionalEmbedding", "5685": "kosmos2/modeling_kosmos2.py:KosmosTextAttention", "5686": "kosmos2/modeling_kosmos2.py:Kosmos2TextFFN", "5687": "kosmos2/modeling_kosmos2.py:Kosmos2TextBlock", "5688": "kosmos2/modeling_kosmos2.py:Kosmos2TextTransformer", "5689": "kosmos2/modeling_kosmos2.py:Kosmos2PreTrainedModel", "5690": "kosmos2/modeling_kosmos2.py:Kosmos2VisionModel", "5691": "kosmos2/modeling_kosmos2.py:Kosmos2TextModel", "5692": "kosmos2/modeling_kosmos2.py:Kosmos2TextForCausalLM", "5693": "kosmos2/modeling_kosmos2.py:Kosmos2ImageToTextProjection", "5694": "kosmos2/modeling_kosmos2.py:Kosmos2Model", "5695": "kosmos2/modeling_kosmos2.py:Kosmos2ForConditionalGeneration", "5696": "lightglue/modeling_lightglue.py:LightGlueKeypointMatchingOutput", "5697": "lightglue/modeling_lightglue.py:LightGluePositionalEncoder", "5698": "lightglue/modeling_lightglue.py:rotate_half", "5699": "lightglue/modeling_lightglue.py:apply_rotary_pos_emb", "5700": "lightglue/modeling_lightglue.py:repeat_kv", "5701": "lightglue/modeling_lightglue.py:eager_attention_forward", "5702": "lightglue/modeling_lightglue.py:LightGlueAttention", "5703": "lightglue/modeling_lightglue.py:LightGlueMLP", "5704": "lightglue/modeling_lightglue.py:LightGlueTransformerLayer", "5705": "lightglue/modeling_lightglue.py:sigmoid_log_double_softmax", "5706": "lightglue/modeling_lightglue.py:LightGlueMatchAssignmentLayer", "5707": "lightglue/modeling_lightglue.py:LightGlueTokenConfidenceLayer", "5708": "lightglue/modeling_lightglue.py:LightGluePreTrainedModel", "5709": "lightglue/modeling_lightglue.py:get_matches_from_scores", "5710": "lightglue/modeling_lightglue.py:normalize_keypoints", "5711": "lightglue/modeling_lightglue.py:LightGlueForKeypointMatching", "5712": "big_bird/modeling_big_bird.py:BigBirdEmbeddings", "5713": "big_bird/modeling_big_bird.py:BigBirdSelfAttention", "5714": "big_bird/modeling_big_bird.py:BigBirdBlockSparseAttention", "5715": "big_bird/modeling_big_bird.py:BigBirdSelfOutput", "5716": "big_bird/modeling_big_bird.py:BigBirdAttention", "5717": "big_bird/modeling_big_bird.py:BigBirdIntermediate", "5718": "big_bird/modeling_big_bird.py:BigBirdOutput", "5719": "big_bird/modeling_big_bird.py:BigBirdLayer", "5720": "big_bird/modeling_big_bird.py:BigBirdEncoder", "5721": "big_bird/modeling_big_bird.py:BigBirdPredictionHeadTransform", "5722": "big_bird/modeling_big_bird.py:BigBirdLMPredictionHead", "5723": "big_bird/modeling_big_bird.py:BigBirdOnlyMLMHead", "5724": "big_bird/modeling_big_bird.py:BigBirdOnlyNSPHead", "5725": "big_bird/modeling_big_bird.py:BigBirdPreTrainingHeads", "5726": "big_bird/modeling_big_bird.py:BigBirdPreTrainedModel", "5727": "big_bird/modeling_big_bird.py:BigBirdForPreTrainingOutput", "5728": "big_bird/modeling_big_bird.py:BigBirdForQuestionAnsweringModelOutput", "5729": "big_bird/modeling_big_bird.py:BigBirdModel", "5730": "big_bird/modeling_big_bird.py:BigBirdForPreTraining", "5731": "big_bird/modeling_big_bird.py:BigBirdForMaskedLM", "5732": "big_bird/modeling_big_bird.py:BigBirdForCausalLM", "5733": "big_bird/modeling_big_bird.py:BigBirdClassificationHead", "5734": "big_bird/modeling_big_bird.py:BigBirdForSequenceClassification", "5735": "big_bird/modeling_big_bird.py:BigBirdForMultipleChoice", "5736": "big_bird/modeling_big_bird.py:BigBirdForTokenClassification", "5737": "big_bird/modeling_big_bird.py:BigBirdForQuestionAnsweringHead", "5738": "big_bird/modeling_big_bird.py:BigBirdForQuestionAnswering", "5739": "helium/modeling_helium.py:HeliumRMSNorm", "5740": "helium/modeling_helium.py:HeliumRotaryEmbedding", "5741": "helium/modeling_helium.py:HeliumMLP", "5742": "helium/modeling_helium.py:repeat_kv", "5743": "helium/modeling_helium.py:eager_attention_forward", "5744": "helium/modeling_helium.py:rotate_half", "5745": "helium/modeling_helium.py:apply_rotary_pos_emb", "5746": "helium/modeling_helium.py:HeliumAttention", "5747": "helium/modeling_helium.py:HeliumDecoderLayer", "5748": "helium/modeling_helium.py:HeliumPreTrainedModel", "5749": "helium/modeling_helium.py:HeliumModel", "5750": "helium/modeling_helium.py:HeliumForCausalLM", "5751": "helium/modeling_helium.py:HeliumForSequenceClassification", "5752": "helium/modeling_helium.py:HeliumForTokenClassification", "5753": "doge/modeling_doge.py:DogeRMSNorm", "5754": "doge/modeling_doge.py:DogeRotaryEmbedding", "5755": "doge/modeling_doge.py:rotate_half", "5756": "doge/modeling_doge.py:apply_rotary_pos_emb", "5757": "doge/modeling_doge.py:repeat_kv", "5758": "doge/modeling_doge.py:eager_attention_forward", "5759": "doge/modeling_doge.py:flex_attention_forward", "5760": "doge/modeling_doge.py:DogeAttention", "5761": "doge/modeling_doge.py:DogeMLP", "5762": "doge/modeling_doge.py:DogeCDMoE", "5763": "doge/modeling_doge.py:DogeDecoderLayer", "5764": "doge/modeling_doge.py:DogePreTrainedModel", "5765": "doge/modeling_doge.py:DogeModel", "5766": "doge/modeling_doge.py:load_balancing_loss_func", "5767": "doge/modeling_doge.py:DogeForCausalLM", "5768": "doge/modeling_doge.py:DogeForSequenceClassification", "5769": "pop2piano/modeling_pop2piano.py:Pop2PianoLayerNorm", "5770": "pop2piano/modeling_pop2piano.py:Pop2PianoDenseActDense", "5771": "pop2piano/modeling_pop2piano.py:Pop2PianoDenseGatedActDense", "5772": "pop2piano/modeling_pop2piano.py:Pop2PianoLayerFF", "5773": "pop2piano/modeling_pop2piano.py:Pop2PianoAttention", "5774": "pop2piano/modeling_pop2piano.py:Pop2PianoLayerSelfAttention", "5775": "pop2piano/modeling_pop2piano.py:Pop2PianoLayerCrossAttention", "5776": "pop2piano/modeling_pop2piano.py:Pop2PianoBlock", "5777": "pop2piano/modeling_pop2piano.py:Pop2PianoPreTrainedModel", "5778": "pop2piano/modeling_pop2piano.py:Pop2PianoStack", "5779": "pop2piano/modeling_pop2piano.py:Pop2PianoConcatEmbeddingToMel", "5780": "pop2piano/modeling_pop2piano.py:Pop2PianoForConditionalGeneration", "5781": "emu3/modeling_emu3.py:rotate_half", "5782": "emu3/modeling_emu3.py:apply_rotary_pos_emb", "5783": "emu3/modeling_emu3.py:repeat_kv", "5784": "emu3/modeling_emu3.py:eager_attention_forward", "5785": "emu3/modeling_emu3.py:Emu3Attention", "5786": "emu3/modeling_emu3.py:Emu3RMSNorm", "5787": "emu3/modeling_emu3.py:Emu3MLP", "5788": "emu3/modeling_emu3.py:Emu3DecoderLayer", "5789": "emu3/modeling_emu3.py:Emu3VQVAEVectorQuantizer", "5790": "emu3/modeling_emu3.py:Emu3VQVAEEncoderConvDownsample", "5791": "emu3/modeling_emu3.py:Emu3VQVAEEncoderConvUpsample", "5792": "emu3/modeling_emu3.py:Emu3VQVAEConv3d", "5793": "emu3/modeling_emu3.py:Emu3VQVAESpatialNorm", "5794": "emu3/modeling_emu3.py:Emu3VQVAETemporalUpsample", "5795": "emu3/modeling_emu3.py:Emu3VQVAETemporalDownsample", "5796": "emu3/modeling_emu3.py:Emu3VQVAETemporalResnetBlock", "5797": "emu3/modeling_emu3.py:Emu3VQVAEResnetBlock", "5798": "emu3/modeling_emu3.py:Emu3VQVAEAttentionBlock", "5799": "emu3/modeling_emu3.py:Emu3VQVAEGroupNorm", "5800": "emu3/modeling_emu3.py:Emu3VQVAEMiddleBlock", "5801": "emu3/modeling_emu3.py:Emu3VQVAEDownBlock", "5802": "emu3/modeling_emu3.py:Emu3VQVAEUpBlock", "5803": "emu3/modeling_emu3.py:Emu3VQVAEEncoder", "5804": "emu3/modeling_emu3.py:Emu3VQVAEDecoder", "5805": "emu3/modeling_emu3.py:Emu3VQVAE", "5806": "emu3/modeling_emu3.py:Emu3ImageVocabularyMapping", "5807": "emu3/modeling_emu3.py:Emu3PreTrainedModel", "5808": "emu3/modeling_emu3.py:Emu3RotaryEmbedding", "5809": "emu3/modeling_emu3.py:Emu3TextModel", "5810": "emu3/modeling_emu3.py:Emu3ForCausalLM", "5811": "emu3/modeling_emu3.py:Emu3Model", "5812": "emu3/modeling_emu3.py:Emu3ForConditionalGeneration", "5813": "deepseek_vl/modeling_deepseek_vl.py:DeepseekVLBaseModelOutputWithPast", "5814": "deepseek_vl/modeling_deepseek_vl.py:DeepseekVLCausalLMOutputWithPast", "5815": "deepseek_vl/modeling_deepseek_vl.py:DeepseekVLAligner", "5816": "deepseek_vl/modeling_deepseek_vl.py:DeepseekVLPreTrainedModel", "5817": "deepseek_vl/modeling_deepseek_vl.py:DeepseekVLModel", "5818": "deepseek_vl/modeling_deepseek_vl.py:DeepseekVLForConditionalGeneration", "5819": "instructblip/modeling_instructblip.py:InstructBlipForConditionalGenerationModelOutput", "5820": "instructblip/modeling_instructblip.py:InstructBlipVisionEmbeddings", "5821": "instructblip/modeling_instructblip.py:eager_attention_forward", "5822": "instructblip/modeling_instructblip.py:InstructBlipAttention", "5823": "instructblip/modeling_instructblip.py:InstructBlipMLP", "5824": "instructblip/modeling_instructblip.py:InstructBlipEncoderLayer", "5825": "instructblip/modeling_instructblip.py:InstructBlipPreTrainedModel", "5826": "instructblip/modeling_instructblip.py:InstructBlipEncoder", "5827": "instructblip/modeling_instructblip.py:InstructBlipVisionModel", "5828": "instructblip/modeling_instructblip.py:InstructBlipQFormerMultiHeadAttention", "5829": "instructblip/modeling_instructblip.py:InstructBlipQFormerSelfOutput", "5830": "instructblip/modeling_instructblip.py:InstructBlipQFormerAttention", "5831": "instructblip/modeling_instructblip.py:InstructBlipQFormerIntermediate", "5832": "instructblip/modeling_instructblip.py:InstructBlipQFormerOutput", "5833": "instructblip/modeling_instructblip.py:InstructBlipQFormerLayer", "5834": "instructblip/modeling_instructblip.py:InstructBlipQFormerEncoder", "5835": "instructblip/modeling_instructblip.py:InstructBlipQFormerEmbeddings", "5836": "instructblip/modeling_instructblip.py:InstructBlipQFormerModel", "5837": "instructblip/modeling_instructblip.py:InstructBlipModel", "5838": "instructblip/modeling_instructblip.py:InstructBlipForConditionalGeneration", "5839": "phi/modeling_phi.py:PhiRotaryEmbedding", "5840": "phi/modeling_phi.py:rotate_half", "5841": "phi/modeling_phi.py:apply_rotary_pos_emb", "5842": "phi/modeling_phi.py:repeat_kv", "5843": "phi/modeling_phi.py:eager_attention_forward", "5844": "phi/modeling_phi.py:PhiAttention", "5845": "phi/modeling_phi.py:PhiMLP", "5846": "phi/modeling_phi.py:PhiDecoderLayer", "5847": "phi/modeling_phi.py:PhiPreTrainedModel", "5848": "phi/modeling_phi.py:PhiModel", "5849": "phi/modeling_phi.py:PhiForCausalLM", "5850": "phi/modeling_phi.py:PhiForSequenceClassification", "5851": "phi/modeling_phi.py:PhiForTokenClassification", "5852": "evolla/modeling_evolla.py:create_position_ids_from_input_ids", "5853": "evolla/modeling_evolla.py:EvollaSaProtEmbeddings", "5854": "evolla/modeling_evolla.py:rotate_half_esm", "5855": "evolla/modeling_evolla.py:apply_rotary_pos_emb_esm", "5856": "evolla/modeling_evolla.py:EvollaSaProtRotaryEmbedding", "5857": "evolla/modeling_evolla.py:eager_attention_forward", "5858": "evolla/modeling_evolla.py:EvollaSaProtSelfAttention", "5859": "evolla/modeling_evolla.py:EvollaSaProtSelfOutput", "5860": "evolla/modeling_evolla.py:EvollaSaProtAttention", "5861": "evolla/modeling_evolla.py:gelu", "5862": "evolla/modeling_evolla.py:EvollaSaProtIntermediate", "5863": "evolla/modeling_evolla.py:EvollaSaProtOutput", "5864": "evolla/modeling_evolla.py:EvollaSaProtLayer", "5865": "evolla/modeling_evolla.py:EvollaSaProtEncoder", "5866": "evolla/modeling_evolla.py:EvollaSaProtPooler", "5867": "evolla/modeling_evolla.py:EvollaSaProtPreTrainedModel", "5868": "evolla/modeling_evolla.py:EvollaSaProtProteinEncoder", "5869": "evolla/modeling_evolla.py:EvollaSequenceCompressorAttention", "5870": "evolla/modeling_evolla.py:EvollaFeedForward", "5871": "evolla/modeling_evolla.py:EvollaSequenceCompressorResampler", "5872": "evolla/modeling_evolla.py:EvollaProteinEncoderModelOutput", "5873": "evolla/modeling_evolla.py:EvollaProteinEncoder", "5874": "evolla/modeling_evolla.py:EvollaSequenceAlignerCrossAttention", "5875": "evolla/modeling_evolla.py:EvollaRMSNorm", "5876": "evolla/modeling_evolla.py:EvollaRotaryEmbedding", "5877": "evolla/modeling_evolla.py:EvollaMLP", "5878": "evolla/modeling_evolla.py:rotate_half", "5879": "evolla/modeling_evolla.py:apply_rotary_pos_emb", "5880": "evolla/modeling_evolla.py:repeat_kv", "5881": "evolla/modeling_evolla.py:EvollaAttention", "5882": "evolla/modeling_evolla.py:EvollaDecoderLayer", "5883": "evolla/modeling_evolla.py:EvollaPreTrainedModel", "5884": "evolla/modeling_evolla.py:EvollaModel", "5885": "evolla/modeling_evolla.py:EvollaForProteinText2Text", "5886": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingLayer", "5887": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingPreActResidualLayer", "5888": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingFeatureFusionLayer", "5889": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingFeatureFusionStage", "5890": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingDepthEstimationHead", "5891": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingPreTrainedModel", "5892": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingReassembleLayer", "5893": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingReassembleStage", "5894": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingNeck", "5895": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingForDepthEstimation", "5896": "eomt/modeling_eomt.py:EomtForUniversalSegmentationOutput", "5897": "eomt/modeling_eomt.py:sample_point", "5898": "eomt/modeling_eomt.py:pair_wise_dice_loss", "5899": "eomt/modeling_eomt.py:pair_wise_sigmoid_cross_entropy_loss", "5900": "eomt/modeling_eomt.py:EomtHungarianMatcher", "5901": "eomt/modeling_eomt.py:dice_loss", "5902": "eomt/modeling_eomt.py:sigmoid_cross_entropy_loss", "5903": "eomt/modeling_eomt.py:EomtLoss", "5904": "eomt/modeling_eomt.py:EomtPatchEmbeddings", "5905": "eomt/modeling_eomt.py:EomtEmbeddings", "5906": "eomt/modeling_eomt.py:eager_attention_forward", "5907": "eomt/modeling_eomt.py:EomtAttention", "5908": "eomt/modeling_eomt.py:EomtLayerScale", "5909": "eomt/modeling_eomt.py:drop_path", "5910": "eomt/modeling_eomt.py:EomtDropPath", "5911": "eomt/modeling_eomt.py:EomtMLP", "5912": "eomt/modeling_eomt.py:EomtSwiGLUFFN", "5913": "eomt/modeling_eomt.py:EomtLayer", "5914": "eomt/modeling_eomt.py:EomtLayerNorm2d", "5915": "eomt/modeling_eomt.py:EomtScaleLayer", "5916": "eomt/modeling_eomt.py:EomtScaleBlock", "5917": "eomt/modeling_eomt.py:EomtMaskHead", "5918": "eomt/modeling_eomt.py:EomtPreTrainedModel", "5919": "eomt/modeling_eomt.py:EomtForUniversalSegmentation", "5920": "chameleon/modeling_chameleon.py:ChameleonRMSNorm", "5921": "chameleon/modeling_chameleon.py:ChameleonRotaryEmbedding", "5922": "chameleon/modeling_chameleon.py:rotate_half", "5923": "chameleon/modeling_chameleon.py:apply_rotary_pos_emb", "5924": "chameleon/modeling_chameleon.py:ChameleonMLP", "5925": "chameleon/modeling_chameleon.py:ChameleonLayerNorm", "5926": "chameleon/modeling_chameleon.py:repeat_kv", "5927": "chameleon/modeling_chameleon.py:eager_attention_forward", "5928": "chameleon/modeling_chameleon.py:ChameleonAttention", "5929": "chameleon/modeling_chameleon.py:ChameleonDecoderLayer", "5930": "chameleon/modeling_chameleon.py:ChameleonSwinDecoderLayer", "5931": "chameleon/modeling_chameleon.py:ChameleonVQVAEVectorQuantizer", "5932": "chameleon/modeling_chameleon.py:ChameleonVQVAEEncoderConvDownsample", "5933": "chameleon/modeling_chameleon.py:ChameleonVQVAEEncoderResnetBlock", "5934": "chameleon/modeling_chameleon.py:ChameleonVQVAEEncoderAttnBlock", "5935": "chameleon/modeling_chameleon.py:ChameleonVQVAEEncoder", "5936": "chameleon/modeling_chameleon.py:ChameleonImageVocabularyMapping", "5937": "chameleon/modeling_chameleon.py:ChameleonPreTrainedModel", "5938": "chameleon/modeling_chameleon.py:ChameleonVQVAE", "5939": "chameleon/modeling_chameleon.py:ChameleonModel", "5940": "chameleon/modeling_chameleon.py:ChameleonForConditionalGeneration", "5941": "fuyu/modeling_fuyu.py:FuyuPreTrainedModel", "5942": "fuyu/modeling_fuyu.py:FuyuModel", "5943": "fuyu/modeling_fuyu.py:FuyuForCausalLM", "5944": "aria/modeling_aria.py:AriaTextRMSNorm", "5945": "aria/modeling_aria.py:AriaProjectorMLP", "5946": "aria/modeling_aria.py:AriaCrossAttention", "5947": "aria/modeling_aria.py:AriaProjector", "5948": "aria/modeling_aria.py:AriaSharedExpertsMLP", "5949": "aria/modeling_aria.py:sequential_experts_gemm", "5950": "aria/modeling_aria.py:AriaGroupedExpertsGemm", "5951": "aria/modeling_aria.py:AriaExperts", "5952": "aria/modeling_aria.py:AriaTextMoELayer", "5953": "aria/modeling_aria.py:rotate_half", "5954": "aria/modeling_aria.py:apply_rotary_pos_emb", "5955": "aria/modeling_aria.py:repeat_kv", "5956": "aria/modeling_aria.py:eager_attention_forward", "5957": "aria/modeling_aria.py:AriaTextAttention", "5958": "aria/modeling_aria.py:AriaTextDecoderLayer", "5959": "aria/modeling_aria.py:AriaTextPreTrainedModel", "5960": "aria/modeling_aria.py:AriaPreTrainedModel", "5961": "aria/modeling_aria.py:AriaTextRotaryEmbedding", "5962": "aria/modeling_aria.py:AriaTextModel", "5963": "aria/modeling_aria.py:AriaTextForCausalLM", "5964": "aria/modeling_aria.py:AriaCausalLMOutputWithPast", "5965": "aria/modeling_aria.py:AriaModelOutputWithPast", "5966": "aria/modeling_aria.py:AriaModel", "5967": "aria/modeling_aria.py:AriaForConditionalGeneration", "5968": "depth_pro/modeling_depth_pro.py:DepthProOutput", "5969": "depth_pro/modeling_depth_pro.py:DepthProDepthEstimatorOutput", "5970": "depth_pro/modeling_depth_pro.py:split_to_patches", "5971": "depth_pro/modeling_depth_pro.py:reshape_features", "5972": "depth_pro/modeling_depth_pro.py:merge_patches", "5973": "depth_pro/modeling_depth_pro.py:reconstruct_feature_maps", "5974": "depth_pro/modeling_depth_pro.py:DepthProPatchEncoder", "5975": "depth_pro/modeling_depth_pro.py:DepthProImageEncoder", "5976": "depth_pro/modeling_depth_pro.py:DepthProEncoder", "5977": "depth_pro/modeling_depth_pro.py:DepthProFeatureUpsampleBlock", "5978": "depth_pro/modeling_depth_pro.py:DepthProFeatureUpsample", "5979": "depth_pro/modeling_depth_pro.py:DepthProFeatureProjection", "5980": "depth_pro/modeling_depth_pro.py:DepthProNeck", "5981": "depth_pro/modeling_depth_pro.py:DepthProPreTrainedModel", "5982": "depth_pro/modeling_depth_pro.py:DepthProModel", "5983": "depth_pro/modeling_depth_pro.py:DepthProPreActResidualLayer", "5984": "depth_pro/modeling_depth_pro.py:DepthProFeatureFusionLayer", "5985": "depth_pro/modeling_depth_pro.py:DepthProFeatureFusionStage", "5986": "depth_pro/modeling_depth_pro.py:DepthProFovEncoder", "5987": "depth_pro/modeling_depth_pro.py:DepthProFovHead", "5988": "depth_pro/modeling_depth_pro.py:DepthProFovModel", "5989": "depth_pro/modeling_depth_pro.py:DepthProDepthEstimationHead", "5990": "depth_pro/modeling_depth_pro.py:DepthProForDepthEstimation", "5991": "mllama/modeling_mllama.py:_prepare_cross_attention_mask", "5992": "mllama/modeling_mllama.py:_prepare_aspect_ratio_attention_mask", "5993": "mllama/modeling_mllama.py:MllamaPrecomputedAspectRatioEmbedding", "5994": "mllama/modeling_mllama.py:MllamaPrecomputedPositionEmbedding", "5995": "mllama/modeling_mllama.py:MllamaVisionMLP", "5996": "mllama/modeling_mllama.py:repeat_kv", "5997": "mllama/modeling_mllama.py:eager_attention_forward", "5998": "mllama/modeling_mllama.py:MllamaVisionAttention", "5999": "mllama/modeling_mllama.py:MllamaVisionEncoderLayer", "6000": "mllama/modeling_mllama.py:MllamaVisionEncoder", "6001": "mllama/modeling_mllama.py:MllamaTextRMSNorm", "6002": "mllama/modeling_mllama.py:MllamaTextCrossAttention", "6003": "mllama/modeling_mllama.py:rotate_half", "6004": "mllama/modeling_mllama.py:apply_rotary_pos_emb", "6005": "mllama/modeling_mllama.py:MllamaTextSelfAttention", "6006": "mllama/modeling_mllama.py:MllamaTextMLP", "6007": "mllama/modeling_mllama.py:MllamaSelfAttentionDecoderLayer", "6008": "mllama/modeling_mllama.py:MllamaCrossAttentionDecoderLayer", "6009": "mllama/modeling_mllama.py:MllamaRotaryEmbedding", "6010": "mllama/modeling_mllama.py:MllamaPreTrainedModel", "6011": "mllama/modeling_mllama.py:MllamaVisionModel", "6012": "mllama/modeling_mllama.py:MllamaTextModel", "6013": "mllama/modeling_mllama.py:MllamaForCausalLM", "6014": "mllama/modeling_mllama.py:MllamaModel", "6015": "mllama/modeling_mllama.py:MllamaForConditionalGeneration", "6016": "cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionMultiModalProjector", "6017": "cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionModelOutputWithPast", "6018": "cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionCausalLMOutputWithPast", "6019": "cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionPreTrainedModel", "6020": "cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionModel", "6021": "cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionForConditionalGeneration", "6022": "bigbird_pegasus/modeling_bigbird_pegasus.py:shift_tokens_right", "6023": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusLearnedPositionalEmbedding", "6024": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusScaledWordEmbedding", "6025": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusSelfAttention", "6026": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusBlockSparseAttention", "6027": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusEncoderAttention", "6028": "bigbird_pegasus/modeling_bigbird_pegasus.py:eager_attention_forward", "6029": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusDecoderAttention", "6030": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusEncoderLayer", "6031": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusDecoderLayer", "6032": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusClassificationHead", "6033": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusPreTrainedModel", "6034": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusEncoder", "6035": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusDecoder", "6036": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusModel", "6037": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusForConditionalGeneration", "6038": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusForSequenceClassification", "6039": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusForQuestionAnswering", "6040": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusDecoderWrapper", "6041": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusForCausalLM", "6042": "perception_lm/modeling_perception_lm.py:PerceptionLMAdaptiveAvgPooling", "6043": "perception_lm/modeling_perception_lm.py:PerceptionLMMultiModalProjector", "6044": "perception_lm/modeling_perception_lm.py:PerceptionLMPreTrainedModel", "6045": "perception_lm/modeling_perception_lm.py:PerceptionLMModelOutputWithPast", "6046": "perception_lm/modeling_perception_lm.py:PerceptionLMCausalLMOutputWithPast", "6047": "perception_lm/modeling_perception_lm.py:PerceptionLMModel", "6048": "perception_lm/modeling_perception_lm.py:PerceptionLMForConditionalGeneration", "6049": "pvt_v2/modeling_pvt_v2.py:drop_path", "6050": "pvt_v2/modeling_pvt_v2.py:PvtV2DropPath", "6051": "pvt_v2/modeling_pvt_v2.py:PvtV2OverlapPatchEmbeddings", "6052": "pvt_v2/modeling_pvt_v2.py:PvtV2DepthWiseConv", "6053": "pvt_v2/modeling_pvt_v2.py:PvtV2SelfAttention", "6054": "pvt_v2/modeling_pvt_v2.py:PvtV2ConvFeedForwardNetwork", "6055": "pvt_v2/modeling_pvt_v2.py:PvtV2BlockLayer", "6056": "pvt_v2/modeling_pvt_v2.py:PvtV2EncoderLayer", "6057": "pvt_v2/modeling_pvt_v2.py:PvtV2Encoder", "6058": "pvt_v2/modeling_pvt_v2.py:PvtV2PreTrainedModel", "6059": "pvt_v2/modeling_pvt_v2.py:PvtV2Model", "6060": "pvt_v2/modeling_pvt_v2.py:PvtV2ForImageClassification", "6061": "pvt_v2/modeling_pvt_v2.py:PvtV2Backbone", "6062": "x_clip/modeling_x_clip.py:contrastive_loss", "6063": "x_clip/modeling_x_clip.py:x_clip_loss", "6064": "x_clip/modeling_x_clip.py:XCLIPOutput", "6065": "x_clip/modeling_x_clip.py:XCLIPVisionEmbeddings", "6066": "x_clip/modeling_x_clip.py:XCLIPTextEmbeddings", "6067": "x_clip/modeling_x_clip.py:eager_attention_forward", "6068": "x_clip/modeling_x_clip.py:XCLIPAttention", "6069": "x_clip/modeling_x_clip.py:XCLIPMLP", "6070": "x_clip/modeling_x_clip.py:XCLIPEncoderLayer", "6071": "x_clip/modeling_x_clip.py:drop_path", "6072": "x_clip/modeling_x_clip.py:XCLIPDropPath", "6073": "x_clip/modeling_x_clip.py:XCLIPVisionEncoderLayer", "6074": "x_clip/modeling_x_clip.py:XCLIPPreTrainedModel", "6075": "x_clip/modeling_x_clip.py:XCLIPEncoder", "6076": "x_clip/modeling_x_clip.py:XCLIPTextTransformer", "6077": "x_clip/modeling_x_clip.py:XCLIPTextModel", "6078": "x_clip/modeling_x_clip.py:XCLIPVisionEncoder", "6079": "x_clip/modeling_x_clip.py:XCLIPVisionTransformer", "6080": "x_clip/modeling_x_clip.py:XCLIPVisionModel", "6081": "x_clip/modeling_x_clip.py:XCLIPMultiframeIntegrationTransformer", "6082": "x_clip/modeling_x_clip.py:XCLIPCrossAttention", "6083": "x_clip/modeling_x_clip.py:PromptGeneratorLayer", "6084": "x_clip/modeling_x_clip.py:XCLIPPromptGenerator", "6085": "x_clip/modeling_x_clip.py:XCLIPModel", "6086": "data2vec/modeling_data2vec_vision.py:Data2VecVisionModelOutputWithPooling", "6087": "data2vec/modeling_data2vec_vision.py:drop_path", "6088": "data2vec/modeling_data2vec_vision.py:Data2VecVisionDropPath", "6089": "data2vec/modeling_data2vec_vision.py:Data2VecVisionEmbeddings", "6090": "data2vec/modeling_data2vec_vision.py:Data2VecVisionPatchEmbeddings", "6091": "data2vec/modeling_data2vec_vision.py:Data2VecVisionSelfAttention", "6092": "data2vec/modeling_data2vec_vision.py:Data2VecVisionSdpaSelfAttention", "6093": "data2vec/modeling_data2vec_vision.py:Data2VecVisionSelfOutput", "6094": "data2vec/modeling_data2vec_vision.py:Data2VecVisionAttention", "6095": "data2vec/modeling_data2vec_vision.py:Data2VecVisionIntermediate", "6096": "data2vec/modeling_data2vec_vision.py:Data2VecVisionOutput", "6097": "data2vec/modeling_data2vec_vision.py:Data2VecVisionLayer", "6098": "data2vec/modeling_data2vec_vision.py:Data2VecVisionRelativePositionBias", "6099": "data2vec/modeling_data2vec_vision.py:Data2VecVisionEncoder", "6100": "data2vec/modeling_data2vec_vision.py:Data2VecVisionPreTrainedModel", "6101": "data2vec/modeling_data2vec_vision.py:Data2VecVisionModel", "6102": "data2vec/modeling_data2vec_vision.py:Data2VecVisionPooler", "6103": "data2vec/modeling_data2vec_vision.py:Data2VecVisionForImageClassification", "6104": "data2vec/modeling_data2vec_vision.py:Data2VecVisionConvModule", "6105": "data2vec/modeling_data2vec_vision.py:Data2VecVisionPyramidPoolingBlock", "6106": "data2vec/modeling_data2vec_vision.py:Data2VecVisionPyramidPoolingModule", "6107": "data2vec/modeling_data2vec_vision.py:Data2VecVisionUperHead", "6108": "data2vec/modeling_data2vec_vision.py:Data2VecVisionFCNHead", "6109": "data2vec/modeling_data2vec_vision.py:Data2VecVisionForSemanticSegmentation", "6110": "data2vec/modeling_data2vec_audio.py:Data2VecAudioConvLayer", "6111": "data2vec/modeling_data2vec_audio.py:Data2VecAudioPadLayer", "6112": "data2vec/modeling_data2vec_audio.py:Data2VecAudioPositionalConvLayer", "6113": "data2vec/modeling_data2vec_audio.py:Data2VecAudioPositionalConvEmbedding", "6114": "data2vec/modeling_data2vec_audio.py:Data2VecAudioFeatureEncoder", "6115": "data2vec/modeling_data2vec_audio.py:Data2VecAudioFeatureProjection", "6116": "data2vec/modeling_data2vec_audio.py:eager_attention_forward", "6117": "data2vec/modeling_data2vec_audio.py:Data2VecAudioAttention", "6118": "data2vec/modeling_data2vec_audio.py:Data2VecAudioFeedForward", "6119": "data2vec/modeling_data2vec_audio.py:Data2VecAudioEncoderLayer", "6120": "data2vec/modeling_data2vec_audio.py:Data2VecAudioEncoder", "6121": "data2vec/modeling_data2vec_audio.py:Data2VecAudioAdapterLayer", "6122": "data2vec/modeling_data2vec_audio.py:Data2VecAudioAdapter", "6123": "data2vec/modeling_data2vec_audio.py:Data2VecAudioPreTrainedModel", "6124": "data2vec/modeling_data2vec_audio.py:_compute_mask_indices", "6125": "data2vec/modeling_data2vec_audio.py:Data2VecAudioModel", "6126": "data2vec/modeling_data2vec_audio.py:Data2VecAudioForCTC", "6127": "data2vec/modeling_data2vec_audio.py:Data2VecAudioForSequenceClassification", "6128": "data2vec/modeling_data2vec_audio.py:Data2VecAudioForAudioFrameClassification", "6129": "data2vec/modeling_data2vec_audio.py:AMSoftmaxLoss", "6130": "data2vec/modeling_data2vec_audio.py:TDNNLayer", "6131": "data2vec/modeling_data2vec_audio.py:Data2VecAudioForXVector", "6132": "data2vec/modeling_data2vec_text.py:Data2VecTextEmbeddings", "6133": "data2vec/modeling_data2vec_text.py:eager_attention_forward", "6134": "data2vec/modeling_data2vec_text.py:Data2VecTextSelfAttention", "6135": "data2vec/modeling_data2vec_text.py:Data2VecTextCrossAttention", "6136": "data2vec/modeling_data2vec_text.py:Data2VecTextSelfOutput", "6137": "data2vec/modeling_data2vec_text.py:Data2VecTextAttention", "6138": "data2vec/modeling_data2vec_text.py:Data2VecTextIntermediate", "6139": "data2vec/modeling_data2vec_text.py:Data2VecTextOutput", "6140": "data2vec/modeling_data2vec_text.py:Data2VecTextLayer", "6141": "data2vec/modeling_data2vec_text.py:Data2VecTextPreTrainedModel", "6142": "data2vec/modeling_data2vec_text.py:Data2VecTextEncoder", "6143": "data2vec/modeling_data2vec_text.py:Data2VecTextPooler", "6144": "data2vec/modeling_data2vec_text.py:Data2VecTextModel", "6145": "data2vec/modeling_data2vec_text.py:Data2VecTextLMHead", "6146": "data2vec/modeling_data2vec_text.py:Data2VecTextClassificationHead", "6147": "data2vec/modeling_data2vec_text.py:Data2VecTextForCausalLM", "6148": "data2vec/modeling_data2vec_text.py:Data2VecTextForMaskedLM", "6149": "data2vec/modeling_data2vec_text.py:Data2VecTextForSequenceClassification", "6150": "data2vec/modeling_data2vec_text.py:Data2VecTextForMultipleChoice", "6151": "data2vec/modeling_data2vec_text.py:Data2VecTextForTokenClassification", "6152": "data2vec/modeling_data2vec_text.py:Data2VecTextForQuestionAnswering", "6153": "glpn/modeling_glpn.py:drop_path", "6154": "glpn/modeling_glpn.py:GLPNDropPath", "6155": "glpn/modeling_glpn.py:GLPNOverlapPatchEmbeddings", "6156": "glpn/modeling_glpn.py:GLPNEfficientSelfAttention", "6157": "glpn/modeling_glpn.py:GLPNSelfOutput", "6158": "glpn/modeling_glpn.py:GLPNAttention", "6159": "glpn/modeling_glpn.py:GLPNDWConv", "6160": "glpn/modeling_glpn.py:GLPNMixFFN", "6161": "glpn/modeling_glpn.py:GLPNLayer", "6162": "glpn/modeling_glpn.py:GLPNEncoder", "6163": "glpn/modeling_glpn.py:GLPNPreTrainedModel", "6164": "glpn/modeling_glpn.py:GLPNModel", "6165": "glpn/modeling_glpn.py:GLPNSelectiveFeatureFusion", "6166": "glpn/modeling_glpn.py:GLPNDecoderStage", "6167": "glpn/modeling_glpn.py:GLPNDecoder", "6168": "glpn/modeling_glpn.py:SiLogLoss", "6169": "glpn/modeling_glpn.py:GLPNDepthEstimationHead", "6170": "glpn/modeling_glpn.py:GLPNForDepthEstimation", "6171": "blenderbot/modeling_blenderbot.py:shift_tokens_right", "6172": "blenderbot/modeling_blenderbot.py:BlenderbotLearnedPositionalEmbedding", "6173": "blenderbot/modeling_blenderbot.py:BlenderbotScaledWordEmbedding", "6174": "blenderbot/modeling_blenderbot.py:eager_attention_forward", "6175": "blenderbot/modeling_blenderbot.py:BlenderbotAttention", "6176": "blenderbot/modeling_blenderbot.py:BlenderbotEncoderLayer", "6177": "blenderbot/modeling_blenderbot.py:BlenderbotDecoderLayer", "6178": "blenderbot/modeling_blenderbot.py:BlenderbotPreTrainedModel", "6179": "blenderbot/modeling_blenderbot.py:BlenderbotEncoder", "6180": "blenderbot/modeling_blenderbot.py:BlenderbotDecoder", "6181": "blenderbot/modeling_blenderbot.py:BlenderbotModel", "6182": "blenderbot/modeling_blenderbot.py:BlenderbotForConditionalGeneration", "6183": "blenderbot/modeling_blenderbot.py:BlenderbotDecoderWrapper", "6184": "blenderbot/modeling_blenderbot.py:BlenderbotForCausalLM", "6185": "vaultgemma/modeling_vaultgemma.py:VaultGemmaRMSNorm", "6186": "vaultgemma/modeling_vaultgemma.py:VaultGemmaMLP", "6187": "vaultgemma/modeling_vaultgemma.py:rotate_half", "6188": "vaultgemma/modeling_vaultgemma.py:apply_rotary_pos_emb", "6189": "vaultgemma/modeling_vaultgemma.py:repeat_kv", "6190": "vaultgemma/modeling_vaultgemma.py:eager_attention_forward", "6191": "vaultgemma/modeling_vaultgemma.py:VaultGemmaAttention", "6192": "vaultgemma/modeling_vaultgemma.py:VaultGemmaDecoderLayer", "6193": "vaultgemma/modeling_vaultgemma.py:VaultGemmaRotaryEmbedding", "6194": "vaultgemma/modeling_vaultgemma.py:VaultGemmaPreTrainedModel", "6195": "vaultgemma/modeling_vaultgemma.py:VaultGemmaModel", "6196": "vaultgemma/modeling_vaultgemma.py:VaultGemmaForCausalLM", "6197": "qwen3_moe/modeling_qwen3_moe.py:rotate_half", "6198": "qwen3_moe/modeling_qwen3_moe.py:apply_rotary_pos_emb", "6199": "qwen3_moe/modeling_qwen3_moe.py:repeat_kv", "6200": "qwen3_moe/modeling_qwen3_moe.py:eager_attention_forward", "6201": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeAttention", "6202": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeMLP", "6203": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeExperts", "6204": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeTopKRouter", "6205": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeSparseMoeBlock", "6206": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeRMSNorm", "6207": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeDecoderLayer", "6208": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoePreTrainedModel", "6209": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeRotaryEmbedding", "6210": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeModel", "6211": "qwen3_moe/modeling_qwen3_moe.py:load_balancing_loss_func", "6212": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeForCausalLM", "6213": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeForSequenceClassification", "6214": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeForTokenClassification", "6215": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeForQuestionAnswering", "6216": "yoso/modeling_yoso.py:load_cuda_kernels", "6217": "yoso/modeling_yoso.py:to_contiguous", "6218": "yoso/modeling_yoso.py:normalize", "6219": "yoso/modeling_yoso.py:hashing", "6220": "yoso/modeling_yoso.py:YosoCumulation", "6221": "yoso/modeling_yoso.py:YosoLSHCumulation", "6222": "yoso/modeling_yoso.py:YosoEmbeddings", "6223": "yoso/modeling_yoso.py:YosoSelfAttention", "6224": "yoso/modeling_yoso.py:YosoSelfOutput", "6225": "yoso/modeling_yoso.py:YosoAttention", "6226": "yoso/modeling_yoso.py:YosoIntermediate", "6227": "yoso/modeling_yoso.py:YosoOutput", "6228": "yoso/modeling_yoso.py:YosoLayer", "6229": "yoso/modeling_yoso.py:YosoEncoder", "6230": "yoso/modeling_yoso.py:YosoPredictionHeadTransform", "6231": "yoso/modeling_yoso.py:YosoLMPredictionHead", "6232": "yoso/modeling_yoso.py:YosoOnlyMLMHead", "6233": "yoso/modeling_yoso.py:YosoPreTrainedModel", "6234": "yoso/modeling_yoso.py:YosoModel", "6235": "yoso/modeling_yoso.py:YosoForMaskedLM", "6236": "yoso/modeling_yoso.py:YosoClassificationHead", "6237": "yoso/modeling_yoso.py:YosoForSequenceClassification", "6238": "yoso/modeling_yoso.py:YosoForMultipleChoice", "6239": "yoso/modeling_yoso.py:YosoForTokenClassification", "6240": "yoso/modeling_yoso.py:YosoForQuestionAnswering", "6241": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1RMSNorm", "6242": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1MLP", "6243": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:rotate_half", "6244": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:apply_rotary_pos_emb", "6245": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:repeat_kv", "6246": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:eager_attention_forward", "6247": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1Attention", "6248": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1Gate", "6249": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1Experts", "6250": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1Moe", "6251": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1DecoderLayer", "6252": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1PreTrainedModel", "6253": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1RotaryEmbedding", "6254": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1Model", "6255": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1ForCausalLM", "6256": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1ForSequenceClassification", "6257": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTEmbeddings", "6258": "dinov3_vit/modeling_dinov3_vit.py:get_patches_center_coordinates", "6259": "dinov3_vit/modeling_dinov3_vit.py:augment_patches_center_coordinates", "6260": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTRopePositionEmbedding", "6261": "dinov3_vit/modeling_dinov3_vit.py:rotate_half", "6262": "dinov3_vit/modeling_dinov3_vit.py:eager_attention_forward", "6263": "dinov3_vit/modeling_dinov3_vit.py:apply_rotary_pos_emb", "6264": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTAttention", "6265": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTLayerScale", "6266": "dinov3_vit/modeling_dinov3_vit.py:drop_path", "6267": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTDropPath", "6268": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTMLP", "6269": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTGatedMLP", "6270": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTLayer", "6271": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTPreTrainedModel", "6272": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTModel", "6273": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTBackbone", "6274": "owlv2/modeling_owlv2.py:contrastive_loss", "6275": "owlv2/modeling_owlv2.py:owlv2_loss", "6276": "owlv2/modeling_owlv2.py:Owlv2Output", "6277": "owlv2/modeling_owlv2.py:_upcast", "6278": "owlv2/modeling_owlv2.py:box_area", "6279": "owlv2/modeling_owlv2.py:box_iou", "6280": "owlv2/modeling_owlv2.py:generalized_box_iou", "6281": "owlv2/modeling_owlv2.py:Owlv2ObjectDetectionOutput", "6282": "owlv2/modeling_owlv2.py:Owlv2ImageGuidedObjectDetectionOutput", "6283": "owlv2/modeling_owlv2.py:Owlv2VisionEmbeddings", "6284": "owlv2/modeling_owlv2.py:Owlv2TextEmbeddings", "6285": "owlv2/modeling_owlv2.py:Owlv2Attention", "6286": "owlv2/modeling_owlv2.py:Owlv2MLP", "6287": "owlv2/modeling_owlv2.py:Owlv2EncoderLayer", "6288": "owlv2/modeling_owlv2.py:Owlv2PreTrainedModel", "6289": "owlv2/modeling_owlv2.py:Owlv2Encoder", "6290": "owlv2/modeling_owlv2.py:Owlv2TextTransformer", "6291": "owlv2/modeling_owlv2.py:Owlv2TextModel", "6292": "owlv2/modeling_owlv2.py:Owlv2VisionTransformer", "6293": "owlv2/modeling_owlv2.py:Owlv2VisionModel", "6294": "owlv2/modeling_owlv2.py:Owlv2Model", "6295": "owlv2/modeling_owlv2.py:Owlv2BoxPredictionHead", "6296": "owlv2/modeling_owlv2.py:Owlv2ClassPredictionHead", "6297": "owlv2/modeling_owlv2.py:Owlv2ForObjectDetection", "6298": "hubert/modeling_hubert.py:HubertPositionalConvEmbedding", "6299": "hubert/modeling_hubert.py:HubertSamePadLayer", "6300": "hubert/modeling_hubert.py:HubertNoLayerNormConvLayer", "6301": "hubert/modeling_hubert.py:HubertLayerNormConvLayer", "6302": "hubert/modeling_hubert.py:HubertGroupNormConvLayer", "6303": "hubert/modeling_hubert.py:HubertFeatureEncoder", "6304": "hubert/modeling_hubert.py:HubertFeatureProjection", "6305": "hubert/modeling_hubert.py:eager_attention_forward", "6306": "hubert/modeling_hubert.py:HubertAttention", "6307": "hubert/modeling_hubert.py:HubertFeedForward", "6308": "hubert/modeling_hubert.py:HubertEncoderLayer", "6309": "hubert/modeling_hubert.py:HubertEncoder", "6310": "hubert/modeling_hubert.py:HubertAttnAdapterLayer", "6311": "hubert/modeling_hubert.py:HubertEncoderLayerStableLayerNorm", "6312": "hubert/modeling_hubert.py:HubertEncoderStableLayerNorm", "6313": "hubert/modeling_hubert.py:HubertPreTrainedModel", "6314": "hubert/modeling_hubert.py:_compute_mask_indices", "6315": "hubert/modeling_hubert.py:HubertModel", "6316": "hubert/modeling_hubert.py:HubertForCTC", "6317": "hubert/modeling_hubert.py:HubertForSequenceClassification", "6318": "convbert/modeling_convbert.py:ConvBertEmbeddings", "6319": "convbert/modeling_convbert.py:ConvBertPreTrainedModel", "6320": "convbert/modeling_convbert.py:SeparableConv1D", "6321": "convbert/modeling_convbert.py:ConvBertSelfAttention", "6322": "convbert/modeling_convbert.py:ConvBertSelfOutput", "6323": "convbert/modeling_convbert.py:ConvBertAttention", "6324": "convbert/modeling_convbert.py:GroupedLinearLayer", "6325": "convbert/modeling_convbert.py:ConvBertIntermediate", "6326": "convbert/modeling_convbert.py:ConvBertOutput", "6327": "convbert/modeling_convbert.py:ConvBertLayer", "6328": "convbert/modeling_convbert.py:ConvBertEncoder", "6329": "convbert/modeling_convbert.py:ConvBertPredictionHeadTransform", "6330": "convbert/modeling_convbert.py:ConvBertSequenceSummary", "6331": "convbert/modeling_convbert.py:ConvBertModel", "6332": "convbert/modeling_convbert.py:ConvBertGeneratorPredictions", "6333": "convbert/modeling_convbert.py:ConvBertForMaskedLM", "6334": "convbert/modeling_convbert.py:ConvBertClassificationHead", "6335": "convbert/modeling_convbert.py:ConvBertForSequenceClassification", "6336": "convbert/modeling_convbert.py:ConvBertForMultipleChoice", "6337": "convbert/modeling_convbert.py:ConvBertForTokenClassification", "6338": "convbert/modeling_convbert.py:ConvBertForQuestionAnswering", "6339": "granitemoehybrid/modeling_granitemoehybrid.py:rotate_half", "6340": "granitemoehybrid/modeling_granitemoehybrid.py:apply_rotary_pos_emb", "6341": "granitemoehybrid/modeling_granitemoehybrid.py:repeat_kv", "6342": "granitemoehybrid/modeling_granitemoehybrid.py:eager_attention_forward", "6343": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridAttention", "6344": "granitemoehybrid/modeling_granitemoehybrid.py:HybridMambaAttentionDynamicCache", "6345": "granitemoehybrid/modeling_granitemoehybrid.py:pad_tensor_by_size", "6346": "granitemoehybrid/modeling_granitemoehybrid.py:reshape_into_chunks", "6347": "granitemoehybrid/modeling_granitemoehybrid.py:segment_sum", "6348": "granitemoehybrid/modeling_granitemoehybrid.py:apply_mask_to_padding_states", "6349": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridMambaLayer", "6350": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridRMSNormGated", "6351": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridMLP", "6352": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridRotaryEmbedding", "6353": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridParallelExperts", "6354": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridTopKGating", "6355": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridMoE", "6356": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteFlashAttentionKwargs", "6357": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridRMSNorm", "6358": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridDecoderLayer", "6359": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridPreTrainedModel", "6360": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridModel", "6361": "granitemoehybrid/modeling_granitemoehybrid.py:load_balancing_loss_func", "6362": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridForCausalLM", "6363": "gpt2/modeling_gpt2.py:eager_attention_forward", "6364": "gpt2/modeling_gpt2.py:GPT2Attention", "6365": "gpt2/modeling_gpt2.py:GPT2MLP", "6366": "gpt2/modeling_gpt2.py:GPT2Block", "6367": "gpt2/modeling_gpt2.py:GPT2SequenceSummary", "6368": "gpt2/modeling_gpt2.py:GPT2PreTrainedModel", "6369": "gpt2/modeling_gpt2.py:GPT2DoubleHeadsModelOutput", "6370": "gpt2/modeling_gpt2.py:GPT2Model", "6371": "gpt2/modeling_gpt2.py:GPT2LMHeadModel", "6372": "gpt2/modeling_gpt2.py:GPT2DoubleHeadsModel", "6373": "gpt2/modeling_gpt2.py:GPT2ForSequenceClassification", "6374": "gpt2/modeling_gpt2.py:GPT2ForTokenClassification", "6375": "gpt2/modeling_gpt2.py:GPT2ForQuestionAnswering", "6376": "arcee/modeling_arcee.py:ArceeMLP", "6377": "arcee/modeling_arcee.py:ArceeRMSNorm", "6378": "arcee/modeling_arcee.py:ArceeRotaryEmbedding", "6379": "arcee/modeling_arcee.py:rotate_half", "6380": "arcee/modeling_arcee.py:apply_rotary_pos_emb", "6381": "arcee/modeling_arcee.py:repeat_kv", "6382": "arcee/modeling_arcee.py:eager_attention_forward", "6383": "arcee/modeling_arcee.py:ArceeAttention", "6384": "arcee/modeling_arcee.py:ArceeDecoderLayer", "6385": "arcee/modeling_arcee.py:ArceePreTrainedModel", "6386": "arcee/modeling_arcee.py:ArceeModel", "6387": "arcee/modeling_arcee.py:ArceeForCausalLM", "6388": "arcee/modeling_arcee.py:ArceeForSequenceClassification", "6389": "arcee/modeling_arcee.py:ArceeForQuestionAnswering", "6390": "arcee/modeling_arcee.py:ArceeForTokenClassification", "6391": "ernie4_5/modeling_ernie4_5.py:Ernie4_5RotaryEmbedding", "6392": "ernie4_5/modeling_ernie4_5.py:Ernie4_5MLP", "6393": "ernie4_5/modeling_ernie4_5.py:rotate_half", "6394": "ernie4_5/modeling_ernie4_5.py:repeat_kv", "6395": "ernie4_5/modeling_ernie4_5.py:eager_attention_forward", "6396": "ernie4_5/modeling_ernie4_5.py:apply_rotary_pos_emb", "6397": "ernie4_5/modeling_ernie4_5.py:Ernie4_5Attention", "6398": "ernie4_5/modeling_ernie4_5.py:Ernie4_5RMSNorm", "6399": "ernie4_5/modeling_ernie4_5.py:Ernie4_5DecoderLayer", "6400": "ernie4_5/modeling_ernie4_5.py:Ernie4_5PreTrainedModel", "6401": "ernie4_5/modeling_ernie4_5.py:Ernie4_5Model", "6402": "ernie4_5/modeling_ernie4_5.py:Ernie4_5ForCausalLM", "6403": "vjepa2/modeling_vjepa2.py:VJEPA2WithMaskedInputPredictorOutput", "6404": "vjepa2/modeling_vjepa2.py:VJEPA2WithMaskedInputModelOutput", "6405": "vjepa2/modeling_vjepa2.py:VJEPA2PatchEmbeddings3D", "6406": "vjepa2/modeling_vjepa2.py:VJEPA2Embeddings", "6407": "vjepa2/modeling_vjepa2.py:eager_attention_forward", "6408": "vjepa2/modeling_vjepa2.py:rotate_queries_or_keys", "6409": "vjepa2/modeling_vjepa2.py:VJEPA2RopeAttention", "6410": "vjepa2/modeling_vjepa2.py:drop_path", "6411": "vjepa2/modeling_vjepa2.py:VJEPA2DropPath", "6412": "vjepa2/modeling_vjepa2.py:VJEPA2MLP", "6413": "vjepa2/modeling_vjepa2.py:VJEPA2Layer", "6414": "vjepa2/modeling_vjepa2.py:VJEPA2Encoder", "6415": "vjepa2/modeling_vjepa2.py:apply_masks", "6416": "vjepa2/modeling_vjepa2.py:VJEPA2PredictorEmbeddings", "6417": "vjepa2/modeling_vjepa2.py:VJEPA2Predictor", "6418": "vjepa2/modeling_vjepa2.py:VJEPA2PoolerSelfAttention", "6419": "vjepa2/modeling_vjepa2.py:VJEPA2PoolerCrossAttention", "6420": "vjepa2/modeling_vjepa2.py:VJEPA2PoolerSelfAttentionLayer", "6421": "vjepa2/modeling_vjepa2.py:VJEPA2PoolerCrossAttentionLayer", "6422": "vjepa2/modeling_vjepa2.py:VJEPA2AttentivePooler", "6423": "vjepa2/modeling_vjepa2.py:VJEPA2PreTrainedModel", "6424": "vjepa2/modeling_vjepa2.py:VJEPA2Model", "6425": "vjepa2/modeling_vjepa2.py:VJEPA2ForVideoClassification", "6426": "whisper/modeling_whisper.py:sinusoids", "6427": "whisper/modeling_whisper.py:shift_tokens_right", "6428": "whisper/modeling_whisper.py:_compute_mask_indices", "6429": "whisper/modeling_whisper.py:WhisperPositionalEmbedding", "6430": "whisper/modeling_whisper.py:eager_attention_forward", "6431": "whisper/modeling_whisper.py:WhisperAttention", "6432": "whisper/modeling_whisper.py:WhisperEncoderLayer", "6433": "whisper/modeling_whisper.py:WhisperDecoderLayer", "6434": "whisper/modeling_whisper.py:WhisperPreTrainedModel", "6435": "whisper/modeling_whisper.py:WhisperEncoder", "6436": "whisper/modeling_whisper.py:WhisperDecoder", "6437": "whisper/modeling_whisper.py:WhisperModel", "6438": "whisper/modeling_whisper.py:WhisperForConditionalGeneration", "6439": "whisper/modeling_whisper.py:WhisperDecoderWrapper", "6440": "whisper/modeling_whisper.py:WhisperForCausalLM", "6441": "whisper/modeling_whisper.py:WhisperForAudioClassification", "6442": "ibert/modeling_ibert.py:IBertEmbeddings", "6443": "ibert/modeling_ibert.py:IBertSelfAttention", "6444": "ibert/modeling_ibert.py:IBertSelfOutput", "6445": "ibert/modeling_ibert.py:IBertAttention", "6446": "ibert/modeling_ibert.py:IBertIntermediate", "6447": "ibert/modeling_ibert.py:IBertOutput", "6448": "ibert/modeling_ibert.py:IBertLayer", "6449": "ibert/modeling_ibert.py:IBertEncoder", "6450": "ibert/modeling_ibert.py:IBertPooler", "6451": "ibert/modeling_ibert.py:IBertPreTrainedModel", "6452": "ibert/modeling_ibert.py:IBertModel", "6453": "ibert/modeling_ibert.py:IBertForMaskedLM", "6454": "ibert/modeling_ibert.py:IBertLMHead", "6455": "ibert/modeling_ibert.py:IBertForSequenceClassification", "6456": "ibert/modeling_ibert.py:IBertForMultipleChoice", "6457": "ibert/modeling_ibert.py:IBertForTokenClassification", "6458": "ibert/modeling_ibert.py:IBertClassificationHead", "6459": "ibert/modeling_ibert.py:IBertForQuestionAnswering", "6460": "ibert/modeling_ibert.py:create_position_ids_from_input_ids", "6461": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLHybridBaseModelOutputWithPast", "6462": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLHybridCausalLMOutputWithPast", "6463": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLHybridLayerNorm", "6464": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLSamVisionNeck", "6465": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLSamVisionProj", "6466": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLHybridAligner", "6467": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLHybridPreTrainedModel", "6468": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLHybridModel", "6469": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLHybridForConditionalGeneration", "6470": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerModelOutput", "6471": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerWithHifiGanOutput", "6472": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:length_regulator", "6473": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerDurationPredictor", "6474": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerBatchNormConvLayer", "6475": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerSpeechDecoderPostnet", "6476": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerPredictorLayer", "6477": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerVariancePredictor", "6478": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerVarianceEmbedding", "6479": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerAttention", "6480": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerConvolutionModule", "6481": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerEncoderLayer", "6482": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerMultiLayeredConv1d", "6483": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerRelPositionalEncoding", "6484": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerEncoder", "6485": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerLoss", "6486": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerPreTrainedModel", "6487": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerModel", "6488": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:HifiGanResidualBlock", "6489": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerHifiGan", "6490": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerWithHifiGan", "6491": "fast_vlm/modeling_fast_vlm.py:FastVlmMultiModalProjector", "6492": "fast_vlm/modeling_fast_vlm.py:FastVlmPreTrainedModel", "6493": "fast_vlm/modeling_fast_vlm.py:FastVlmModelOutputWithPast", "6494": "fast_vlm/modeling_fast_vlm.py:FastVlmModel", "6495": "fast_vlm/modeling_fast_vlm.py:FastVlmCausalLMOutputWithPast", "6496": "fast_vlm/modeling_fast_vlm.py:FastVlmForConditionalGeneration", "6497": "grounding_dino/modeling_grounding_dino.py:MultiScaleDeformableAttention", "6498": "grounding_dino/modeling_grounding_dino.py:GroundingDinoDecoderOutput", "6499": "grounding_dino/modeling_grounding_dino.py:GroundingDinoEncoderOutput", "6500": "grounding_dino/modeling_grounding_dino.py:GroundingDinoModelOutput", "6501": "grounding_dino/modeling_grounding_dino.py:GroundingDinoObjectDetectionOutput", "6502": "grounding_dino/modeling_grounding_dino.py:GroundingDinoFrozenBatchNorm2d", "6503": "grounding_dino/modeling_grounding_dino.py:replace_batch_norm", "6504": "grounding_dino/modeling_grounding_dino.py:GroundingDinoConvEncoder", "6505": "grounding_dino/modeling_grounding_dino.py:GroundingDinoConvModel", "6506": "grounding_dino/modeling_grounding_dino.py:GroundingDinoSinePositionEmbedding", "6507": "grounding_dino/modeling_grounding_dino.py:GroundingDinoLearnedPositionEmbedding", "6508": "grounding_dino/modeling_grounding_dino.py:build_position_encoding", "6509": "grounding_dino/modeling_grounding_dino.py:GroundingDinoMultiscaleDeformableAttention", "6510": "grounding_dino/modeling_grounding_dino.py:GroundingDinoTextEnhancerLayer", "6511": "grounding_dino/modeling_grounding_dino.py:GroundingDinoBiMultiHeadAttention", "6512": "grounding_dino/modeling_grounding_dino.py:drop_path", "6513": "grounding_dino/modeling_grounding_dino.py:GroundingDinoDropPath", "6514": "grounding_dino/modeling_grounding_dino.py:GroundingDinoFusionLayer", "6515": "grounding_dino/modeling_grounding_dino.py:GroundingDinoDeformableLayer", "6516": "grounding_dino/modeling_grounding_dino.py:get_sine_pos_embed", "6517": "grounding_dino/modeling_grounding_dino.py:GroundingDinoEncoderLayer", "6518": "grounding_dino/modeling_grounding_dino.py:GroundingDinoMultiheadAttention", "6519": "grounding_dino/modeling_grounding_dino.py:GroundingDinoDecoderLayer", "6520": "grounding_dino/modeling_grounding_dino.py:GroundingDinoContrastiveEmbedding", "6521": "grounding_dino/modeling_grounding_dino.py:GroundingDinoPreTrainedModel", "6522": "grounding_dino/modeling_grounding_dino.py:GroundingDinoEncoder", "6523": "grounding_dino/modeling_grounding_dino.py:GroundingDinoDecoder", "6524": "grounding_dino/modeling_grounding_dino.py:generate_masks_with_special_tokens_and_transfer_map", "6525": "grounding_dino/modeling_grounding_dino.py:GroundingDinoModel", "6526": "grounding_dino/modeling_grounding_dino.py:GroundingDinoMLPPredictionHead", "6527": "grounding_dino/modeling_grounding_dino.py:build_label_maps", "6528": "grounding_dino/modeling_grounding_dino.py:build_text_mask", "6529": "grounding_dino/modeling_grounding_dino.py:GroundingDinoForObjectDetection", "6530": "speecht5/modeling_speecht5.py:shift_tokens_right", "6531": "speecht5/modeling_speecht5.py:shift_spectrograms_right", "6532": "speecht5/modeling_speecht5.py:_compute_mask_indices", "6533": "speecht5/modeling_speecht5.py:SpeechT5NoLayerNormConvLayer", "6534": "speecht5/modeling_speecht5.py:SpeechT5LayerNormConvLayer", "6535": "speecht5/modeling_speecht5.py:SpeechT5GroupNormConvLayer", "6536": "speecht5/modeling_speecht5.py:SpeechT5SinusoidalPositionalEmbedding", "6537": "speecht5/modeling_speecht5.py:SpeechT5PositionalConvEmbedding", "6538": "speecht5/modeling_speecht5.py:SpeechT5ScaledPositionalEncoding", "6539": "speecht5/modeling_speecht5.py:SpeechT5RelativePositionalEncoding", "6540": "speecht5/modeling_speecht5.py:SpeechT5SamePadLayer", "6541": "speecht5/modeling_speecht5.py:SpeechT5FeatureEncoder", "6542": "speecht5/modeling_speecht5.py:SpeechT5FeatureProjection", "6543": "speecht5/modeling_speecht5.py:SpeechT5SpeechEncoderPrenet", "6544": "speecht5/modeling_speecht5.py:SpeechT5SpeechDecoderPrenet", "6545": "speecht5/modeling_speecht5.py:SpeechT5BatchNormConvLayer", "6546": "speecht5/modeling_speecht5.py:SpeechT5SpeechDecoderPostnet", "6547": "speecht5/modeling_speecht5.py:SpeechT5TextEncoderPrenet", "6548": "speecht5/modeling_speecht5.py:SpeechT5TextDecoderPrenet", "6549": "speecht5/modeling_speecht5.py:SpeechT5TextDecoderPostnet", "6550": "speecht5/modeling_speecht5.py:SpeechT5Attention", "6551": "speecht5/modeling_speecht5.py:SpeechT5FeedForward", "6552": "speecht5/modeling_speecht5.py:SpeechT5EncoderLayer", "6553": "speecht5/modeling_speecht5.py:SpeechT5DecoderLayer", "6554": "speecht5/modeling_speecht5.py:SpeechT5PreTrainedModel", "6555": "speecht5/modeling_speecht5.py:SpeechT5Encoder", "6556": "speecht5/modeling_speecht5.py:SpeechT5EncoderWithSpeechPrenet", "6557": "speecht5/modeling_speecht5.py:SpeechT5EncoderWithTextPrenet", "6558": "speecht5/modeling_speecht5.py:SpeechT5EncoderWithoutPrenet", "6559": "speecht5/modeling_speecht5.py:SpeechT5Decoder", "6560": "speecht5/modeling_speecht5.py:SpeechT5DecoderWithSpeechPrenet", "6561": "speecht5/modeling_speecht5.py:SpeechT5DecoderWithTextPrenet", "6562": "speecht5/modeling_speecht5.py:SpeechT5DecoderWithoutPrenet", "6563": "speecht5/modeling_speecht5.py:SpeechT5GuidedMultiheadAttentionLoss", "6564": "speecht5/modeling_speecht5.py:SpeechT5SpectrogramLoss", "6565": "speecht5/modeling_speecht5.py:SpeechT5Model", "6566": "speecht5/modeling_speecht5.py:SpeechT5ForSpeechToText", "6567": "speecht5/modeling_speecht5.py:_generate_speech", "6568": "speecht5/modeling_speecht5.py:SpeechT5ForTextToSpeech", "6569": "speecht5/modeling_speecht5.py:SpeechT5ForSpeechToSpeech", "6570": "speecht5/modeling_speecht5.py:HifiGanResidualBlock", "6571": "speecht5/modeling_speecht5.py:SpeechT5HifiGan", "6572": "gpt_oss/modeling_gpt_oss.py:GptOssRMSNorm", "6573": "gpt_oss/modeling_gpt_oss.py:GptOssExperts", "6574": "gpt_oss/modeling_gpt_oss.py:GptOssTopKRouter", "6575": "gpt_oss/modeling_gpt_oss.py:GptOssMLP", "6576": "gpt_oss/modeling_gpt_oss.py:GptOssRotaryEmbedding", "6577": "gpt_oss/modeling_gpt_oss.py:repeat_kv", "6578": "gpt_oss/modeling_gpt_oss.py:_apply_rotary_emb", "6579": "gpt_oss/modeling_gpt_oss.py:apply_rotary_pos_emb", "6580": "gpt_oss/modeling_gpt_oss.py:eager_attention_forward", "6581": "gpt_oss/modeling_gpt_oss.py:GptOssAttention", "6582": "gpt_oss/modeling_gpt_oss.py:GptOssDecoderLayer", "6583": "gpt_oss/modeling_gpt_oss.py:GptOssPreTrainedModel", "6584": "gpt_oss/modeling_gpt_oss.py:GptOssModel", "6585": "gpt_oss/modeling_gpt_oss.py:load_balancing_loss_func", "6586": "gpt_oss/modeling_gpt_oss.py:GptOssForCausalLM", "6587": "gpt_oss/modeling_gpt_oss.py:GptOssForSequenceClassification", "6588": "gpt_oss/modeling_gpt_oss.py:GptOssForTokenClassification", "6589": "clipseg/modeling_clipseg.py:contrastive_loss", "6590": "clipseg/modeling_clipseg.py:clipseg_loss", "6591": "clipseg/modeling_clipseg.py:CLIPSegOutput", "6592": "clipseg/modeling_clipseg.py:CLIPSegDecoderOutput", "6593": "clipseg/modeling_clipseg.py:CLIPSegImageSegmentationOutput", "6594": "clipseg/modeling_clipseg.py:CLIPSegVisionEmbeddings", "6595": "clipseg/modeling_clipseg.py:CLIPSegTextEmbeddings", "6596": "clipseg/modeling_clipseg.py:eager_attention_forward", "6597": "clipseg/modeling_clipseg.py:CLIPSegAttention", "6598": "clipseg/modeling_clipseg.py:CLIPSegMLP", "6599": "clipseg/modeling_clipseg.py:CLIPSegEncoderLayer", "6600": "clipseg/modeling_clipseg.py:CLIPSegPreTrainedModel", "6601": "clipseg/modeling_clipseg.py:CLIPSegEncoder", "6602": "clipseg/modeling_clipseg.py:CLIPSegTextTransformer", "6603": "clipseg/modeling_clipseg.py:CLIPSegTextModel", "6604": "clipseg/modeling_clipseg.py:CLIPSegVisionTransformer", "6605": "clipseg/modeling_clipseg.py:CLIPSegVisionModel", "6606": "clipseg/modeling_clipseg.py:CLIPSegModel", "6607": "clipseg/modeling_clipseg.py:CLIPSegDecoderLayer", "6608": "clipseg/modeling_clipseg.py:CLIPSegDecoder", "6609": "clipseg/modeling_clipseg.py:CLIPSegForImageSegmentation", "6610": "gemma/modeling_gemma.py:GemmaRMSNorm", "6611": "gemma/modeling_gemma.py:GemmaMLP", "6612": "gemma/modeling_gemma.py:GemmaRotaryEmbedding", "6613": "gemma/modeling_gemma.py:rotate_half", "6614": "gemma/modeling_gemma.py:apply_rotary_pos_emb", "6615": "gemma/modeling_gemma.py:repeat_kv", "6616": "gemma/modeling_gemma.py:eager_attention_forward", "6617": "gemma/modeling_gemma.py:GemmaAttention", "6618": "gemma/modeling_gemma.py:GemmaDecoderLayer", "6619": "gemma/modeling_gemma.py:GemmaPreTrainedModel", "6620": "gemma/modeling_gemma.py:GemmaModel", "6621": "gemma/modeling_gemma.py:GemmaForCausalLM", "6622": "gemma/modeling_gemma.py:GemmaForSequenceClassification", "6623": "gemma/modeling_gemma.py:GemmaForTokenClassification", "6624": "patchtst/modeling_patchtst.py:eager_attention_forward", "6625": "patchtst/modeling_patchtst.py:PatchTSTAttention", "6626": "patchtst/modeling_patchtst.py:PatchTSTBatchNorm", "6627": "patchtst/modeling_patchtst.py:random_masking", "6628": "patchtst/modeling_patchtst.py:forecast_masking", "6629": "patchtst/modeling_patchtst.py:PatchTSTPatchify", "6630": "patchtst/modeling_patchtst.py:PatchTSTMasking", "6631": "patchtst/modeling_patchtst.py:PatchTSTEncoderLayer", "6632": "patchtst/modeling_patchtst.py:PatchTSTPreTrainedModel", "6633": "patchtst/modeling_patchtst.py:PatchTSTEmbedding", "6634": "patchtst/modeling_patchtst.py:PatchTSTPositionalEncoding", "6635": "patchtst/modeling_patchtst.py:PatchTSTEncoder", "6636": "patchtst/modeling_patchtst.py:PatchTSTModelOutput", "6637": "patchtst/modeling_patchtst.py:PatchTSTForPretrainingOutput", "6638": "patchtst/modeling_patchtst.py:PatchTSTForRegressionOutput", "6639": "patchtst/modeling_patchtst.py:PatchTSTForPredictionOutput", "6640": "patchtst/modeling_patchtst.py:PatchTSTForClassificationOutput", "6641": "patchtst/modeling_patchtst.py:SamplePatchTSTOutput", "6642": "patchtst/modeling_patchtst.py:nll", "6643": "patchtst/modeling_patchtst.py:weighted_average", "6644": "patchtst/modeling_patchtst.py:PatchTSTStdScaler", "6645": "patchtst/modeling_patchtst.py:PatchTSTMeanScaler", "6646": "patchtst/modeling_patchtst.py:PatchTSTNOPScaler", "6647": "patchtst/modeling_patchtst.py:PatchTSTScaler", "6648": "patchtst/modeling_patchtst.py:PatchTSTModel", "6649": "patchtst/modeling_patchtst.py:PatchTSTMaskPretrainHead", "6650": "patchtst/modeling_patchtst.py:PatchTSTForPretraining", "6651": "patchtst/modeling_patchtst.py:PatchTSTClassificationHead", "6652": "patchtst/modeling_patchtst.py:PatchTSTForClassification", "6653": "patchtst/modeling_patchtst.py:PatchTSTPredictionHead", "6654": "patchtst/modeling_patchtst.py:PatchTSTForPrediction", "6655": "patchtst/modeling_patchtst.py:PatchTSTRegressionHead", "6656": "patchtst/modeling_patchtst.py:PatchTSTForRegression", "6657": "tapas/modeling_tapas.py:TableQuestionAnsweringOutput", "6658": "tapas/modeling_tapas.py:TapasEmbeddings", "6659": "tapas/modeling_tapas.py:TapasSelfAttention", "6660": "tapas/modeling_tapas.py:TapasSelfOutput", "6661": "tapas/modeling_tapas.py:TapasAttention", "6662": "tapas/modeling_tapas.py:TapasIntermediate", "6663": "tapas/modeling_tapas.py:TapasOutput", "6664": "tapas/modeling_tapas.py:TapasLayer", "6665": "tapas/modeling_tapas.py:TapasEncoder", "6666": "tapas/modeling_tapas.py:TapasPooler", "6667": "tapas/modeling_tapas.py:TapasPredictionHeadTransform", "6668": "tapas/modeling_tapas.py:TapasLMPredictionHead", "6669": "tapas/modeling_tapas.py:TapasOnlyMLMHead", "6670": "tapas/modeling_tapas.py:TapasPreTrainedModel", "6671": "tapas/modeling_tapas.py:TapasModel", "6672": "tapas/modeling_tapas.py:TapasForMaskedLM", "6673": "tapas/modeling_tapas.py:TapasForQuestionAnswering", "6674": "tapas/modeling_tapas.py:TapasForSequenceClassification", "6675": "tapas/modeling_tapas.py:AverageApproximationFunction", "6676": "tapas/modeling_tapas.py:IndexMap", "6677": "tapas/modeling_tapas.py:ProductIndexMap", "6678": "tapas/modeling_tapas.py:gather", "6679": "tapas/modeling_tapas.py:flatten", "6680": "tapas/modeling_tapas.py:range_index_map", "6681": "tapas/modeling_tapas.py:_segment_reduce", "6682": "tapas/modeling_tapas.py:reduce_sum", "6683": "tapas/modeling_tapas.py:reduce_mean", "6684": "tapas/modeling_tapas.py:reduce_max", "6685": "tapas/modeling_tapas.py:reduce_min", "6686": "tapas/modeling_tapas.py:compute_column_logits", "6687": "tapas/modeling_tapas.py:_single_column_cell_selection_loss", "6688": "tapas/modeling_tapas.py:compute_token_logits", "6689": "tapas/modeling_tapas.py:_calculate_aggregate_mask", "6690": "tapas/modeling_tapas.py:_calculate_aggregation_loss_known", "6691": "tapas/modeling_tapas.py:_calculate_aggregation_loss_unknown", "6692": "tapas/modeling_tapas.py:_calculate_aggregation_loss", "6693": "tapas/modeling_tapas.py:_calculate_expected_result", "6694": "tapas/modeling_tapas.py:huber_loss", "6695": "tapas/modeling_tapas.py:_calculate_regression_loss", "6696": "smolvlm/modeling_smolvlm.py:SmolVLMPreTrainedModel", "6697": "smolvlm/modeling_smolvlm.py:SmolVLMVisionEmbeddings", "6698": "smolvlm/modeling_smolvlm.py:eager_attention_forward", "6699": "smolvlm/modeling_smolvlm.py:SmolVLMVisionAttention", "6700": "smolvlm/modeling_smolvlm.py:SmolVLMVisionMLP", "6701": "smolvlm/modeling_smolvlm.py:SmolVLMEncoderLayer", "6702": "smolvlm/modeling_smolvlm.py:SmolVLMEncoder", "6703": "smolvlm/modeling_smolvlm.py:SmolVLMVisionTransformer", "6704": "smolvlm/modeling_smolvlm.py:SmolVLMBaseModelOutputWithPast", "6705": "smolvlm/modeling_smolvlm.py:SmolVLMSimpleMLP", "6706": "smolvlm/modeling_smolvlm.py:SmolVLMConnector", "6707": "smolvlm/modeling_smolvlm.py:SmolVLMModel", "6708": "smolvlm/modeling_smolvlm.py:SmolVLMCausalLMOutputWithPast", "6709": "smolvlm/modeling_smolvlm.py:SmolVLMForConditionalGeneration", "6710": "bart/modeling_bart.py:shift_tokens_right", "6711": "bart/modeling_bart.py:BartLearnedPositionalEmbedding", "6712": "bart/modeling_bart.py:BartScaledWordEmbedding", "6713": "bart/modeling_bart.py:eager_attention_forward", "6714": "bart/modeling_bart.py:BartAttention", "6715": "bart/modeling_bart.py:BartEncoderLayer", "6716": "bart/modeling_bart.py:BartDecoderLayer", "6717": "bart/modeling_bart.py:BartClassificationHead", "6718": "bart/modeling_bart.py:BartPreTrainedModel", "6719": "bart/modeling_bart.py:PretrainedBartModel", "6720": "bart/modeling_bart.py:BartPretrainedModel", "6721": "bart/modeling_bart.py:BartEncoder", "6722": "bart/modeling_bart.py:BartDecoder", "6723": "bart/modeling_bart.py:BartModel", "6724": "bart/modeling_bart.py:BartForConditionalGeneration", "6725": "bart/modeling_bart.py:BartForSequenceClassification", "6726": "bart/modeling_bart.py:BartForQuestionAnswering", "6727": "bart/modeling_bart.py:BartDecoderWrapper", "6728": "bart/modeling_bart.py:BartForCausalLM", "6729": "swin2sr/modeling_swin2sr.py:Swin2SREncoderOutput", "6730": "swin2sr/modeling_swin2sr.py:window_partition", "6731": "swin2sr/modeling_swin2sr.py:window_reverse", "6732": "swin2sr/modeling_swin2sr.py:drop_path", "6733": "swin2sr/modeling_swin2sr.py:Swin2SRDropPath", "6734": "swin2sr/modeling_swin2sr.py:Swin2SREmbeddings", "6735": "swin2sr/modeling_swin2sr.py:Swin2SRPatchEmbeddings", "6736": "swin2sr/modeling_swin2sr.py:Swin2SRPatchUnEmbeddings", "6737": "swin2sr/modeling_swin2sr.py:Swin2SRPatchMerging", "6738": "swin2sr/modeling_swin2sr.py:Swin2SRSelfAttention", "6739": "swin2sr/modeling_swin2sr.py:Swin2SRSelfOutput", "6740": "swin2sr/modeling_swin2sr.py:Swin2SRAttention", "6741": "swin2sr/modeling_swin2sr.py:Swin2SRIntermediate", "6742": "swin2sr/modeling_swin2sr.py:Swin2SROutput", "6743": "swin2sr/modeling_swin2sr.py:Swin2SRLayer", "6744": "swin2sr/modeling_swin2sr.py:Swin2SRStage", "6745": "swin2sr/modeling_swin2sr.py:Swin2SREncoder", "6746": "swin2sr/modeling_swin2sr.py:Swin2SRPreTrainedModel", "6747": "swin2sr/modeling_swin2sr.py:Swin2SRModel", "6748": "swin2sr/modeling_swin2sr.py:Upsample", "6749": "swin2sr/modeling_swin2sr.py:UpsampleOneStep", "6750": "swin2sr/modeling_swin2sr.py:PixelShuffleUpsampler", "6751": "swin2sr/modeling_swin2sr.py:NearestConvUpsampler", "6752": "swin2sr/modeling_swin2sr.py:PixelShuffleAuxUpsampler", "6753": "swin2sr/modeling_swin2sr.py:Swin2SRForImageSuperResolution", "6754": "cohere/modeling_cohere.py:CohereLayerNorm", "6755": "cohere/modeling_cohere.py:CohereRotaryEmbedding", "6756": "cohere/modeling_cohere.py:CohereMLP", "6757": "cohere/modeling_cohere.py:repeat_kv", "6758": "cohere/modeling_cohere.py:eager_attention_forward", "6759": "cohere/modeling_cohere.py:rotate_half", "6760": "cohere/modeling_cohere.py:apply_rotary_pos_emb", "6761": "cohere/modeling_cohere.py:CohereAttention", "6762": "cohere/modeling_cohere.py:CohereDecoderLayer", "6763": "cohere/modeling_cohere.py:CoherePreTrainedModel", "6764": "cohere/modeling_cohere.py:CohereModel", "6765": "cohere/modeling_cohere.py:CohereForCausalLM", "6766": "pegasus_x/modeling_pegasus_x.py:DimensionInfo", "6767": "pegasus_x/modeling_pegasus_x.py:shift_tokens_right", "6768": "pegasus_x/modeling_pegasus_x.py:PegasusXScaledWordEmbedding", "6769": "pegasus_x/modeling_pegasus_x.py:PegasusXSinusoidalPositionalEmbedding", "6770": "pegasus_x/modeling_pegasus_x.py:eager_attention_forward", "6771": "pegasus_x/modeling_pegasus_x.py:PegasusXAttention", "6772": "pegasus_x/modeling_pegasus_x.py:PegasusXGlobalLocalAttention", "6773": "pegasus_x/modeling_pegasus_x.py:PegasusXEncoderLayer", "6774": "pegasus_x/modeling_pegasus_x.py:PegasusXDecoderLayer", "6775": "pegasus_x/modeling_pegasus_x.py:PegasusXPreTrainedModel", "6776": "pegasus_x/modeling_pegasus_x.py:PegasusXEncoder", "6777": "pegasus_x/modeling_pegasus_x.py:PegasusXDecoder", "6778": "pegasus_x/modeling_pegasus_x.py:PegasusXModel", "6779": "pegasus_x/modeling_pegasus_x.py:PegasusXForConditionalGeneration", "6780": "pegasus_x/modeling_pegasus_x.py:PegasusXDecoderWrapper", "6781": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaEmbeddings", "6782": "xlm_roberta/modeling_xlm_roberta.py:eager_attention_forward", "6783": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaSelfAttention", "6784": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaCrossAttention", "6785": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaSelfOutput", "6786": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaAttention", "6787": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaIntermediate", "6788": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaOutput", "6789": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaLayer", "6790": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaLMHead", "6791": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaPreTrainedModel", "6792": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaEncoder", "6793": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaPooler", "6794": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaModel", "6795": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaForCausalLM", "6796": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaForMaskedLM", "6797": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaClassificationHead", "6798": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaForSequenceClassification", "6799": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaForMultipleChoice", "6800": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaForTokenClassification", "6801": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaForQuestionAnswering", "6802": "lasr/modeling_lasr.py:LasrEncoderSubsampling", "6803": "lasr/modeling_lasr.py:LasrEncoderRotaryEmbedding", "6804": "lasr/modeling_lasr.py:rotate_half", "6805": "lasr/modeling_lasr.py:apply_rotary_pos_emb", "6806": "lasr/modeling_lasr.py:repeat_kv", "6807": "lasr/modeling_lasr.py:eager_attention_forward", "6808": "lasr/modeling_lasr.py:LasrEncoderAttention", "6809": "lasr/modeling_lasr.py:LasrEncoderConvolutionModule", "6810": "lasr/modeling_lasr.py:LasrEncoderFeedForward", "6811": "lasr/modeling_lasr.py:LasrEncoderBlock", "6812": "lasr/modeling_lasr.py:LasrPreTrainedModel", "6813": "lasr/modeling_lasr.py:LasrEncoder", "6814": "lasr/modeling_lasr.py:LasrGenerateOutput", "6815": "lasr/modeling_lasr.py:LasrForCTC", "6816": "llama4/modeling_llama4.py:Llama4TextExperts", "6817": "llama4/modeling_llama4.py:Llama4TextMLP", "6818": "llama4/modeling_llama4.py:Llama4TextL2Norm", "6819": "llama4/modeling_llama4.py:Llama4TextRMSNorm", "6820": "llama4/modeling_llama4.py:Llama4Router", "6821": "llama4/modeling_llama4.py:Llama4TextMoe", "6822": "llama4/modeling_llama4.py:Llama4TextRotaryEmbedding", "6823": "llama4/modeling_llama4.py:apply_rotary_emb", "6824": "llama4/modeling_llama4.py:repeat_kv", "6825": "llama4/modeling_llama4.py:eager_attention_forward", "6826": "llama4/modeling_llama4.py:vision_eager_attention_forward", "6827": "llama4/modeling_llama4.py:Llama4TextAttention", "6828": "llama4/modeling_llama4.py:Llama4TextDecoderLayer", "6829": "llama4/modeling_llama4.py:Llama4PreTrainedModel", "6830": "llama4/modeling_llama4.py:Llama4TextModel", "6831": "llama4/modeling_llama4.py:Llama4ForCausalLM", "6832": "llama4/modeling_llama4.py:Llama4CausalLMOutputWithPast", "6833": "llama4/modeling_llama4.py:Llama4VisionMLP2", "6834": "llama4/modeling_llama4.py:Llama4MultiModalProjector", "6835": "llama4/modeling_llama4.py:pixel_shuffle", "6836": "llama4/modeling_llama4.py:Llama4VisionPixelShuffleMLP", "6837": "llama4/modeling_llama4.py:reshape_for_broadcast", "6838": "llama4/modeling_llama4.py:vision_apply_rotary_emb", "6839": "llama4/modeling_llama4.py:Llama4VisionAttention", "6840": "llama4/modeling_llama4.py:Llama4VisionMLP", "6841": "llama4/modeling_llama4.py:Llama4VisionEncoderLayer", "6842": "llama4/modeling_llama4.py:Llama4VisionEncoder", "6843": "llama4/modeling_llama4.py:Llama4UnfoldConvolution", "6844": "llama4/modeling_llama4.py:Llama4VisionRotaryEmbedding", "6845": "llama4/modeling_llama4.py:Llama4VisionModel", "6846": "llama4/modeling_llama4.py:Llama4ForConditionalGeneration", "6847": "git/modeling_git.py:GitVisionModelOutput", "6848": "git/modeling_git.py:token_type_ids_mask_function", "6849": "git/modeling_git.py:create_causal_mask_mapping", "6850": "git/modeling_git.py:GitEmbeddings", "6851": "git/modeling_git.py:GitSelfAttention", "6852": "git/modeling_git.py:GitSelfOutput", "6853": "git/modeling_git.py:GitAttention", "6854": "git/modeling_git.py:GitIntermediate", "6855": "git/modeling_git.py:GitOutput", "6856": "git/modeling_git.py:GitLayer", "6857": "git/modeling_git.py:GitEncoder", "6858": "git/modeling_git.py:GitPreTrainedModel", "6859": "git/modeling_git.py:GitVisionEmbeddings", "6860": "git/modeling_git.py:GitVisionMLP", "6861": "git/modeling_git.py:eager_attention_forward", "6862": "git/modeling_git.py:GitVisionAttention", "6863": "git/modeling_git.py:GitVisionEncoderLayer", "6864": "git/modeling_git.py:GitVisionEncoder", "6865": "git/modeling_git.py:GitVisionTransformer", "6866": "git/modeling_git.py:GitVisionModel", "6867": "git/modeling_git.py:GitProjection", "6868": "git/modeling_git.py:GitModel", "6869": "git/modeling_git.py:GitForCausalLM", "6870": "lxmert/modeling_lxmert.py:GeLU", "6871": "lxmert/modeling_lxmert.py:LxmertModelOutput", "6872": "lxmert/modeling_lxmert.py:LxmertForQuestionAnsweringOutput", "6873": "lxmert/modeling_lxmert.py:LxmertForPreTrainingOutput", "6874": "lxmert/modeling_lxmert.py:LxmertEmbeddings", "6875": "lxmert/modeling_lxmert.py:LxmertAttention", "6876": "lxmert/modeling_lxmert.py:LxmertAttentionOutput", "6877": "lxmert/modeling_lxmert.py:LxmertCrossAttentionLayer", "6878": "lxmert/modeling_lxmert.py:LxmertSelfAttentionLayer", "6879": "lxmert/modeling_lxmert.py:LxmertIntermediate", "6880": "lxmert/modeling_lxmert.py:LxmertOutput", "6881": "lxmert/modeling_lxmert.py:LxmertLayer", "6882": "lxmert/modeling_lxmert.py:LxmertXLayer", "6883": "lxmert/modeling_lxmert.py:LxmertVisualFeatureEncoder", "6884": "lxmert/modeling_lxmert.py:LxmertEncoder", "6885": "lxmert/modeling_lxmert.py:LxmertPooler", "6886": "lxmert/modeling_lxmert.py:LxmertPredictionHeadTransform", "6887": "lxmert/modeling_lxmert.py:LxmertLMPredictionHead", "6888": "lxmert/modeling_lxmert.py:LxmertVisualAnswerHead", "6889": "lxmert/modeling_lxmert.py:LxmertVisualObjHead", "6890": "lxmert/modeling_lxmert.py:LxmertPreTrainingHeads", "6891": "lxmert/modeling_lxmert.py:LxmertPreTrainedModel", "6892": "lxmert/modeling_lxmert.py:LxmertModel", "6893": "lxmert/modeling_lxmert.py:LxmertForPreTraining", "6894": "lxmert/modeling_lxmert.py:LxmertForQuestionAnswering", "6895": "dpt/modeling_dpt.py:BaseModelOutputWithIntermediateActivations", "6896": "dpt/modeling_dpt.py:BaseModelOutputWithPoolingAndIntermediateActivations", "6897": "dpt/modeling_dpt.py:DPTViTHybridEmbeddings", "6898": "dpt/modeling_dpt.py:DPTViTEmbeddings", "6899": "dpt/modeling_dpt.py:DPTViTPatchEmbeddings", "6900": "dpt/modeling_dpt.py:eager_attention_forward", "6901": "dpt/modeling_dpt.py:DPTSelfAttention", "6902": "dpt/modeling_dpt.py:DPTViTSelfOutput", "6903": "dpt/modeling_dpt.py:DPTViTAttention", "6904": "dpt/modeling_dpt.py:DPTViTIntermediate", "6905": "dpt/modeling_dpt.py:DPTViTOutput", "6906": "dpt/modeling_dpt.py:DPTViTLayer", "6907": "dpt/modeling_dpt.py:DPTViTEncoder", "6908": "dpt/modeling_dpt.py:DPTReassembleStage", "6909": "dpt/modeling_dpt.py:_get_backbone_hidden_size", "6910": "dpt/modeling_dpt.py:DPTReassembleLayer", "6911": "dpt/modeling_dpt.py:DPTFeatureFusionStage", "6912": "dpt/modeling_dpt.py:DPTPreActResidualLayer", "6913": "dpt/modeling_dpt.py:DPTFeatureFusionLayer", "6914": "dpt/modeling_dpt.py:DPTPreTrainedModel", "6915": "dpt/modeling_dpt.py:DPTModel", "6916": "dpt/modeling_dpt.py:DPTViTPooler", "6917": "dpt/modeling_dpt.py:DPTNeck", "6918": "dpt/modeling_dpt.py:DPTDepthEstimationHead", "6919": "dpt/modeling_dpt.py:DPTForDepthEstimation", "6920": "dpt/modeling_dpt.py:DPTSemanticSegmentationHead", "6921": "dpt/modeling_dpt.py:DPTAuxiliaryHead", "6922": "dpt/modeling_dpt.py:DPTForSemanticSegmentation", "6923": "rwkv/modeling_rwkv.py:load_wkv_cuda_kernel", "6924": "rwkv/modeling_rwkv.py:RwkvLinearAttention", "6925": "rwkv/modeling_rwkv.py:rwkv_linear_attention_cpu", "6926": "rwkv/modeling_rwkv.py:rwkv_linear_attention", "6927": "rwkv/modeling_rwkv.py:RwkvSelfAttention", "6928": "rwkv/modeling_rwkv.py:RwkvFeedForward", "6929": "rwkv/modeling_rwkv.py:RwkvBlock", "6930": "rwkv/modeling_rwkv.py:RwkvPreTrainedModel", "6931": "rwkv/modeling_rwkv.py:RwkvOutput", "6932": "rwkv/modeling_rwkv.py:RwkvCausalLMOutput", "6933": "rwkv/modeling_rwkv.py:RwkvModel", "6934": "rwkv/modeling_rwkv.py:RwkvForCausalLM", "6935": "encodec/modeling_encodec.py:EncodecOutput", "6936": "encodec/modeling_encodec.py:EncodecEncoderOutput", "6937": "encodec/modeling_encodec.py:EncodecDecoderOutput", "6938": "encodec/modeling_encodec.py:EncodecConv1d", "6939": "encodec/modeling_encodec.py:EncodecConvTranspose1d", "6940": "encodec/modeling_encodec.py:EncodecLSTM", "6941": "encodec/modeling_encodec.py:EncodecResnetBlock", "6942": "encodec/modeling_encodec.py:EncodecEncoder", "6943": "encodec/modeling_encodec.py:EncodecDecoder", "6944": "encodec/modeling_encodec.py:EncodecEuclideanCodebook", "6945": "encodec/modeling_encodec.py:EncodecVectorQuantization", "6946": "encodec/modeling_encodec.py:EncodecResidualVectorQuantizer", "6947": "encodec/modeling_encodec.py:EncodecPreTrainedModel", "6948": "encodec/modeling_encodec.py:EncodecModel", "6949": "falcon/modeling_falcon.py:FalconLinear", "6950": "falcon/modeling_falcon.py:rotate_half", "6951": "falcon/modeling_falcon.py:apply_rotary_pos_emb", "6952": "falcon/modeling_falcon.py:FalconRotaryEmbedding", "6953": "falcon/modeling_falcon.py:build_alibi_tensor", "6954": "falcon/modeling_falcon.py:dropout_add", "6955": "falcon/modeling_falcon.py:FalconAttention", "6956": "falcon/modeling_falcon.py:FalconFlashAttention2", "6957": "falcon/modeling_falcon.py:FalconMLP", "6958": "falcon/modeling_falcon.py:FalconDecoderLayer", "6959": "falcon/modeling_falcon.py:FalconPreTrainedModel", "6960": "falcon/modeling_falcon.py:FalconModel", "6961": "falcon/modeling_falcon.py:FalconForCausalLM", "6962": "falcon/modeling_falcon.py:FalconForSequenceClassification", "6963": "falcon/modeling_falcon.py:FalconForTokenClassification", "6964": "falcon/modeling_falcon.py:FalconForQuestionAnswering", "6965": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrDecoderOutput", "6966": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrModelOutput", "6967": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrObjectDetectionOutput", "6968": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrSegmentationOutput", "6969": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrFrozenBatchNorm2d", "6970": "conditional_detr/modeling_conditional_detr.py:replace_batch_norm", "6971": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrConvEncoder", "6972": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrConvModel", "6973": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrSinePositionEmbedding", "6974": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrLearnedPositionEmbedding", "6975": "conditional_detr/modeling_conditional_detr.py:build_position_encoding", "6976": "conditional_detr/modeling_conditional_detr.py:gen_sine_position_embeddings", "6977": "conditional_detr/modeling_conditional_detr.py:inverse_sigmoid", "6978": "conditional_detr/modeling_conditional_detr.py:DetrAttention", "6979": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrAttention", "6980": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrEncoderLayer", "6981": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrDecoderLayer", "6982": "conditional_detr/modeling_conditional_detr.py:MLP", "6983": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrPreTrainedModel", "6984": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrEncoder", "6985": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrDecoder", "6986": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrModel", "6987": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrMLPPredictionHead", "6988": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrForObjectDetection", "6989": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrForSegmentation", "6990": "conditional_detr/modeling_conditional_detr.py:_expand", "6991": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrMaskHeadSmallConv", "6992": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrMHAttentionMap", "6993": "flaubert/modeling_flaubert.py:create_sinusoidal_embeddings", "6994": "flaubert/modeling_flaubert.py:get_masks", "6995": "flaubert/modeling_flaubert.py:MultiHeadAttention", "6996": "flaubert/modeling_flaubert.py:TransformerFFN", "6997": "flaubert/modeling_flaubert.py:FlaubertPredLayer", "6998": "flaubert/modeling_flaubert.py:FlaubertSquadHeadOutput", "6999": "flaubert/modeling_flaubert.py:FlaubertPoolerStartLogits", "7000": "flaubert/modeling_flaubert.py:FlaubertPoolerEndLogits", "7001": "flaubert/modeling_flaubert.py:FlaubertPoolerAnswerClass", "7002": "flaubert/modeling_flaubert.py:FlaubertSQuADHead", "7003": "flaubert/modeling_flaubert.py:FlaubertSequenceSummary", "7004": "flaubert/modeling_flaubert.py:FlaubertPreTrainedModel", "7005": "flaubert/modeling_flaubert.py:FlaubertModel", "7006": "flaubert/modeling_flaubert.py:FlaubertWithLMHeadModel", "7007": "flaubert/modeling_flaubert.py:FlaubertForSequenceClassification", "7008": "flaubert/modeling_flaubert.py:FlaubertForTokenClassification", "7009": "flaubert/modeling_flaubert.py:FlaubertForQuestionAnsweringSimple", "7010": "flaubert/modeling_flaubert.py:FlaubertForQuestionAnsweringOutput", "7011": "flaubert/modeling_flaubert.py:FlaubertForQuestionAnswering", "7012": "flaubert/modeling_flaubert.py:FlaubertForMultipleChoice", "7013": "regnet/modeling_regnet.py:RegNetConvLayer", "7014": "regnet/modeling_regnet.py:RegNetEmbeddings", "7015": "regnet/modeling_regnet.py:RegNetShortCut", "7016": "regnet/modeling_regnet.py:RegNetSELayer", "7017": "regnet/modeling_regnet.py:RegNetXLayer", "7018": "regnet/modeling_regnet.py:RegNetYLayer", "7019": "regnet/modeling_regnet.py:RegNetStage", "7020": "regnet/modeling_regnet.py:RegNetEncoder", "7021": "regnet/modeling_regnet.py:RegNetPreTrainedModel", "7022": "regnet/modeling_regnet.py:RegNetModel", "7023": "regnet/modeling_regnet.py:RegNetForImageClassification", "7024": "glm4_moe/modeling_glm4_moe.py:Glm4MoeRotaryEmbedding", "7025": "glm4_moe/modeling_glm4_moe.py:repeat_kv", "7026": "glm4_moe/modeling_glm4_moe.py:eager_attention_forward", "7027": "glm4_moe/modeling_glm4_moe.py:rotate_half", "7028": "glm4_moe/modeling_glm4_moe.py:apply_rotary_pos_emb", "7029": "glm4_moe/modeling_glm4_moe.py:Glm4MoeAttention", "7030": "glm4_moe/modeling_glm4_moe.py:Glm4MoeMLP", "7031": "glm4_moe/modeling_glm4_moe.py:Glm4MoeTopkRouter", "7032": "glm4_moe/modeling_glm4_moe.py:Glm4MoeRMSNorm", "7033": "glm4_moe/modeling_glm4_moe.py:Glm4MoeNaiveMoe", "7034": "glm4_moe/modeling_glm4_moe.py:Glm4MoeMoE", "7035": "glm4_moe/modeling_glm4_moe.py:Glm4MoeDecoderLayer", "7036": "glm4_moe/modeling_glm4_moe.py:Glm4MoePreTrainedModel", "7037": "glm4_moe/modeling_glm4_moe.py:Glm4MoeModel", "7038": "glm4_moe/modeling_glm4_moe.py:Glm4MoeForCausalLM", "7039": "swin/modeling_swin.py:SwinEncoderOutput", "7040": "swin/modeling_swin.py:SwinModelOutput", "7041": "swin/modeling_swin.py:SwinMaskedImageModelingOutput", "7042": "swin/modeling_swin.py:SwinImageClassifierOutput", "7043": "swin/modeling_swin.py:window_partition", "7044": "swin/modeling_swin.py:window_reverse", "7045": "swin/modeling_swin.py:SwinEmbeddings", "7046": "swin/modeling_swin.py:SwinPatchEmbeddings", "7047": "swin/modeling_swin.py:SwinPatchMerging", "7048": "swin/modeling_swin.py:drop_path", "7049": "swin/modeling_swin.py:SwinDropPath", "7050": "swin/modeling_swin.py:SwinSelfAttention", "7051": "swin/modeling_swin.py:SwinSelfOutput", "7052": "swin/modeling_swin.py:SwinAttention", "7053": "swin/modeling_swin.py:SwinIntermediate", "7054": "swin/modeling_swin.py:SwinOutput", "7055": "swin/modeling_swin.py:SwinLayer", "7056": "swin/modeling_swin.py:SwinStage", "7057": "swin/modeling_swin.py:SwinEncoder", "7058": "swin/modeling_swin.py:SwinPreTrainedModel", "7059": "swin/modeling_swin.py:SwinModel", "7060": "swin/modeling_swin.py:SwinForMaskedImageModeling", "7061": "swin/modeling_swin.py:SwinForImageClassification", "7062": "swin/modeling_swin.py:SwinBackbone", "7063": "jamba/modeling_jamba.py:JambaRMSNorm", "7064": "jamba/modeling_jamba.py:HybridMambaAttentionDynamicCache", "7065": "jamba/modeling_jamba.py:rotate_half", "7066": "jamba/modeling_jamba.py:apply_rotary_pos_emb", "7067": "jamba/modeling_jamba.py:repeat_kv", "7068": "jamba/modeling_jamba.py:eager_attention_forward", "7069": "jamba/modeling_jamba.py:JambaAttention", "7070": "jamba/modeling_jamba.py:JambaMambaMixer", "7071": "jamba/modeling_jamba.py:JambaMLP", "7072": "jamba/modeling_jamba.py:JambaExperts", "7073": "jamba/modeling_jamba.py:JambaSparseMoeBlock", "7074": "jamba/modeling_jamba.py:JambaAttentionDecoderLayer", "7075": "jamba/modeling_jamba.py:JambaMambaDecoderLayer", "7076": "jamba/modeling_jamba.py:JambaPreTrainedModel", "7077": "jamba/modeling_jamba.py:JambaModel", "7078": "jamba/modeling_jamba.py:load_balancing_loss_func", "7079": "jamba/modeling_jamba.py:JambaForCausalLM", "7080": "jamba/modeling_jamba.py:JambaForSequenceClassification", "7081": "m2m_100/modeling_m2m_100.py:shift_tokens_right", "7082": "m2m_100/modeling_m2m_100.py:M2M100ScaledWordEmbedding", "7083": "m2m_100/modeling_m2m_100.py:M2M100SinusoidalPositionalEmbedding", "7084": "m2m_100/modeling_m2m_100.py:eager_attention_forward", "7085": "m2m_100/modeling_m2m_100.py:M2M100Attention", "7086": "m2m_100/modeling_m2m_100.py:M2M100EncoderLayer", "7087": "m2m_100/modeling_m2m_100.py:M2M100DecoderLayer", "7088": "m2m_100/modeling_m2m_100.py:M2M100PreTrainedModel", "7089": "m2m_100/modeling_m2m_100.py:M2M100Encoder", "7090": "m2m_100/modeling_m2m_100.py:M2M100Decoder", "7091": "m2m_100/modeling_m2m_100.py:M2M100Model", "7092": "m2m_100/modeling_m2m_100.py:M2M100ForConditionalGeneration", "7093": "xcodec/modeling_xcodec.py:XcodecOutput", "7094": "xcodec/modeling_xcodec.py:XcodecEncoderOutput", "7095": "xcodec/modeling_xcodec.py:XcodecDecoderOutput", "7096": "xcodec/modeling_xcodec.py:ResidualUnit", "7097": "xcodec/modeling_xcodec.py:SemanticEncoderBlock", "7098": "xcodec/modeling_xcodec.py:SemanticEncoder", "7099": "xcodec/modeling_xcodec.py:SemanticDecoderBlock", "7100": "xcodec/modeling_xcodec.py:SemanticDecoder", "7101": "xcodec/modeling_xcodec.py:XcodecEuclideanCodebook", "7102": "xcodec/modeling_xcodec.py:XcodecVectorQuantization", "7103": "xcodec/modeling_xcodec.py:XcodecResidualVectorQuantization", "7104": "xcodec/modeling_xcodec.py:XcodecPreTrainedModel", "7105": "xcodec/modeling_xcodec.py:XcodecModel", "7106": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionMLP", "7107": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionPatchEmbed", "7108": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionRotaryEmbedding", "7109": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionPatchMerger", "7110": "qwen3_vl/modeling_qwen3_vl.py:rotate_half", "7111": "qwen3_vl/modeling_qwen3_vl.py:apply_rotary_pos_emb_vision", "7112": "qwen3_vl/modeling_qwen3_vl.py:repeat_kv", "7113": "qwen3_vl/modeling_qwen3_vl.py:eager_attention_forward", "7114": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionAttention", "7115": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionBlock", "7116": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextRotaryEmbedding", "7117": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextRMSNorm", "7118": "qwen3_vl/modeling_qwen3_vl.py:apply_rotary_pos_emb", "7119": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextAttention", "7120": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextMLP", "7121": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextDecoderLayer", "7122": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLModelOutputWithPast", "7123": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLPreTrainedModel", "7124": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionModel", "7125": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextModel", "7126": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLModel", "7127": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLCausalLMOutputWithPast", "7128": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLForConditionalGeneration", "7129": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaCache", "7130": "falcon_mamba/modeling_falcon_mamba.py:rms_forward", "7131": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaMixer", "7132": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaRMSNorm", "7133": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaBlock", "7134": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaPreTrainedModel", "7135": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaOutput", "7136": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaCausalLMOutput", "7137": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaModel", "7138": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaForCausalLM"}