identifier
stringlengths
24
117
embedding
listlengths
2.56k
2.56k
tokens
listlengths
4
448
depth_pro/modeling_depth_pro.py:DepthProFovModel
[ -0.00012688309652730823, 0.038842201232910156, -0.006388519890606403, 0.01084628701210022, -0.00029635633109137416, 0.04429373890161514, 0.05860402062535286, -0.053606778383255005, 0.01232274528592825, 0.01203881110996008, 0.02816627360880375, 0.012152384035289288, 0.0056218975223600864, -...
[ "Conv2d", "ModelFovEncoder", "ModelFovHead", "ModelFovModel", "Module", "ReLU", "True", "__init__", "activation", "class", "config", "conv", "def", "flatten", "forward", "fov_encoder", "fov_features", "fov_output", "fusion_hidden_size", "global_features", "head", "inplace",...
depth_pro/modeling_depth_pro.py:DepthProDepthEstimationHead
[ -0.00022063942742533982, 0.02551973983645439, 0.027107635512948036, 0.015538685955107212, -0.0008896465296857059, 0.03924369066953659, 0.03992421552538872, -0.04332684725522995, 0.00771263288334012, 0.01264644879847765, 0.009697501547634602, -0.029262635856866837, 0.0019139805808663368, 0....
[ "Conv2d", "ConvTranspose2d", "ModelModelEstimationHead", "Module", "ModuleList", "ReLU", "True", "__init__", "class", "config", "def", "dim", "features", "for", "forward", "fusion_hidden_size", "hidden_states", "in", "in_channels", "kernel_size", "layer", "layers", "nn", ...
depth_pro/modeling_depth_pro.py:DepthProForDepthEstimation
[ -0.00019587513816077262, 0.04398999363183975, -0.024274270981550217, 0.0382918082177639, -0.0006909050280228257, 0.05926113203167915, 0.04353414103388786, -0.016638701781630516, 0.006752350367605686, 0.012934882193803787, 0.024388235062360764, 0.015043210238218307, 0.0060400767251849174, -...
[ "Model", "ModelFeatureFusionStage", "ModelForModelEstimation", "ModelFovModel", "ModelModel", "ModelModelEstimationHead", "ModelModelEstimatorOutput", "ModelPreTrainedModel", "Model_outputs", "None", "True", "__init__", "attentions", "auto_docstring", "class", "config", "def", "det...
mllama/modeling_mllama.py:_prepare_cross_attention_mask
[ -0.0000703664481989108, 0.033550720661878586, 0.02972278743982315, -0.01733829267323017, -0.0007036644383333623, 0.030398305505514145, 0.012271908111870289, -0.056743502616882324, 0.01812639646232128, 0.033550720661878586, 0.019027087837457657, 0.021503986790776253, 0.0032650032080709934, ...
[ "None", "_", "_prepare_cross_attention_mask", "any", "batch_size", "bool", "cross_attention_mask", "def", "dim", "dtype", "finfo", "full_text_row_masked_out_mask", "inverted_cross_attn_mask", "masked_fill", "min", "negative_inf_value", "num_vision_tokens", "repeat_interleave", "r...
mllama/modeling_mllama.py:_prepare_aspect_ratio_attention_mask
[ -0.00009095249697566032, 0.02384086884558201, 0.05102391913533211, -0.024397898465394974, -0.0006788799073547125, 0.030970849096775055, -0.01125199906527996, -0.04679049178957939, 0.009135286323726177, 0.028074294328689575, 0.009079583920538425, 0.009915128350257874, -0.0013438340974971652, ...
[ "_prepare_aspect_ratio_attention_mask", "aspect_ratio_mask", "attention_mask", "batch_size", "def", "dtype", "finfo", "max_num_tiles", "min", "num_patches", "pad_patches", "repeat", "reshape", "return", "shape", "target_length", "torch", "transpose", "unsqueeze", "view" ]
mllama/modeling_mllama.py:MllamaPrecomputedAspectRatioEmbedding
[ -0.0003803886938840151, -0.0005496525554917753, 0.04170079156756401, 0.0018346018623560667, -0.0016817183932289481, 0.030285492539405823, 0.01129881851375103, -0.03121735341846943, 0.002970307832583785, 0.006319184321910143, 0.022131705656647682, -0.0026791011914610863, -0.001215787720866501...
[ "Embedding", "ModelPrecomputedAspectRatioEmbedding", "Module", "Parameter", "True", "__init__", "aspect_ratio_ids", "class", "config", "def", "embedding", "embeddings", "forward", "gate", "hidden_size", "hidden_state", "if", "is_gated", "max_aspect_ratio_id", "max_num_tiles", ...
mllama/modeling_mllama.py:MllamaPrecomputedPositionEmbedding
[ -0.00032226037001237273, 0.012328709475696087, 0.01371136773377657, 0.003802312072366476, -0.0016419074963778257, 0.01947244629263878, 0.023620424792170525, -0.012155876494944096, -0.00024844653671607375, -0.006221965420991182, 0.0022468208335340023, 0.007835067808628082, -0.0035142581909894...
[ "Embedding", "ModelPrecomputedPositionEmbedding", "Module", "Parameter", "__init__", "aspect_ratio_ids", "batch_size", "class", "config", "def", "embedding", "forward", "gate", "gated_position_embedding", "gated_tile_position_embedding", "hidden_size", "hidden_state", "image_size",...
mllama/modeling_mllama.py:MllamaVisionMLP
[ -0.0001662031572777778, 0.03708730265498161, 0.032536715269088745, 0.034129418432712555, -0.00047283468302339315, 0.04709859937429428, 0.03162659704685211, -0.03481200709939003, 0.003341838950291276, 0, 0.049373894929885864, -0.020136358216404915, 0.0019482210045680404, -0.0018913387320935...
[ "ACT2FN", "Linear", "ModelVisionMLP", "Module", "__init__", "activation_fn", "class", "config", "def", "fc1", "fc2", "forward", "hidden_act", "hidden_size", "hidden_states", "intermediate_size", "nn", "return", "self", "super" ]
mllama/modeling_mllama.py:repeat_kv
[ -0.00025096136960200965, -0.0023088448215276003, -0.004072744864970446, -0.009292741306126118, -0.000559285341296345, 0.03143470734357834, 0.009694280102849007, -0.058739304542541504, 0.011185707524418831, 0.05162634328007698, 0.005736260209232569, -0.02317449077963829, 0.0006775957299396396...
[ "Model_kv", "None", "batch", "def", "expand", "head_dim", "hidden_states", "if", "n_rep", "num_key_value_heads", "reshape", "return", "shape", "slen" ]
mllama/modeling_mllama.py:eager_attention_forward
[ 0, 0.020139193162322044, 0.013972979038953781, -0.018215786665678024, 0, 0.038468122482299805, 0.0527239553630352, -0.03326360881328583, 0.020704900845885277, 0.007297629024833441, 0.0280591007322073, 0.020704900845885277, 0.002489113714545965, -0.017876362428069115, -0.03484759107232094...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "key_states", "kwargs", "matmul", "module", "nn", "not", "num_key_value_groups", "p", "query",...
mllama/modeling_mllama.py:MllamaVisionAttention
[ -0.00008038675878196955, 0.03103719651699066, 0.03441080451011658, 0.01743030920624733, -0.0002934336371254176, 0.022715630009770393, 0.04588107392191887, -0.018779752776026726, 0.004132670350372791, 0.035985156893730164, 0.027888495475053787, 0.018554845824837685, -0.0011456211796030402, ...
[ "ALL_ATTENTION_FUNCTIONS", "Linear", "ModelVisionAttention", "Module", "None", "_", "__init__", "_attn_implementation", "attention_heads", "attention_interface", "attention_mask", "attn_output", "attn_weights", "batch_size", "class", "config", "contiguous", "def", "dropout", "e...
mllama/modeling_mllama.py:MllamaVisionEncoderLayer
[ -0.00004559535227599554, 0.019641075283288956, 0.010830650106072426, 0.028283147141337395, 0, 0.03411935269832611, 0.024803871288895607, -0.03748639300465584, 0.004826092626899481, 0.021885769441723824, 0.03524170070886612, 0.002483193064108491, 0.0010802590986713767, 0.004573564510792494,...
[ "False", "LayerNorm", "ModelVisionAttention", "ModelVisionEncoderLayer", "ModelVisionMLP", "Module", "None", "Parameter", "__init__", "attention_heads", "attention_mask", "attn_weights", "class", "config", "def", "eps", "forward", "gate_attn", "gate_ffn", "hidden_size", "hidd...
mllama/modeling_mllama.py:MllamaVisionEncoder
[ -0.00006536524597322568, 0.010232310742139816, 0.007038246840238571, 0.04432116821408272, 0, 0.04567793756723404, 0.020238492637872696, -0.04771309345960617, 0.012437062337994576, 0.014472218230366707, 0.029396692290902138, -0.01639430969953537, 0.00032859283965080976, 0.011136824265122414...
[ "BaseModelOutput", "False", "ModelVisionEncoder", "ModelVisionEncoderLayer", "Module", "ModuleList", "None", "_", "__init__", "attention_mask", "class", "config", "def", "encoder_layer", "encoder_states", "for", "forward", "gradient_checkpointing", "hidden_state", "hidden_state...
mllama/modeling_mllama.py:MllamaTextRMSNorm
[ -0.0000997777096927166, 0.04746946692466736, 0.03865370899438858, 0.048825737088918686, -0.0005686446093022823, 0.03435885161161423, 0.027125408872961998, -0.03481094166636467, 0.008702735416591167, 0.035715121775865555, 0.021474281325936317, 0.011415275745093822, 0.002090916968882084, 0.0...
[ "ModelTextRMSNorm", "Module", "Parameter", "True", "__init__", "class", "def", "dtype", "eps", "extra_repr", "f", "float32", "forward", "hidden_size", "hidden_states", "input_dtype", "keepdim", "mean", "nn", "ones", "pow", "return", "rsqrt", "self", "shape", "super"...
mllama/modeling_mllama.py:MllamaTextCrossAttention
[ -0.00011495164653751999, 0.04790543019771576, 0.0469929464161396, 0.005218270234763622, -0.000691492052283138, 0.018819989636540413, 0.022698048502206802, -0.034902527928352356, 0.0027802258264273405, 0.03353380039334297, 0.023838654160499573, 0.026005804538726807, 0.00042238045716658235, ...
[ "ALL_ATTENTION_FUNCTIONS", "Cross", "Linear", "ModelTextCrossAttention", "ModelTextRMSNorm", "Module", "None", "ValueError", "_", "__init__", "_attn_implementation", "attention", "attention_interface", "attention_mask", "attn_output", "attn_weights", "bsz", "cache_position", "cac...
mllama/modeling_mllama.py:rotate_half
[ 0.00002049485374300275, 0.013860220089554787, 0.03434192016720772, 0.002974055241793394, 0.0003507140791043639, 0.028506038710474968, 0.01975221559405327, -0.02008890174329281, 0.014477476477622986, 0.018742159008979797, -0.001487027620896697, -0.01217679213732481, 0.00022445700597018003, ...
[ "Model_half", "cat", "def", "dim", "return", "shape", "torch", "x", "x1", "x2" ]
mllama/modeling_mllama.py:apply_rotary_pos_emb
[ -0.0001444592053303495, 0.027112245559692383, 0.028019767254590988, 0.0028360087890177965, -0.0005707467789761722, 0.021667107939720154, 0.046510547399520874, -0.002183726755902171, 0.01293220091611147, 0.03652779385447502, 0.007884104736149311, 0.0017654155381023884, -0.0007834474672563374,...
[ "Model_rotary_pos_emb", "cos", "def", "k", "k_embed", "q", "q_embed", "return", "rotate_half", "sin", "unsqueeze", "unsqueeze_dim" ]
mllama/modeling_mllama.py:MllamaTextSelfAttention
[ -0.000044882690417580307, 0.044608116149902344, 0.028386982157826424, -0.009068063460290432, -0.0005737944156862795, 0.018586713820695877, 0.03266755864024162, -0.013123346492648125, 0.0022529351990669966, 0.011884232982993126, 0.013968197628855705, 0.02478228509426117, -0.001626337529160082...
[ "ALL_ATTENTION_FUNCTIONS", "Linear", "ModelTextSelfAttention", "Module", "None", "True", "_", "__init__", "_attn_implementation", "apply_rotary_pos_emb", "attention_interface", "attention_mask", "attn_output", "attn_weights", "bsz", "cache_kwargs", "cache_position", "class", "con...
mllama/modeling_mllama.py:MllamaTextMLP
[ -0.00021370963077060878, 0.03204750269651413, 0.03410769999027252, 0.01705384999513626, -0.0006760020041838288, 0.056312039494514465, 0.02415008284151554, -0.014020781964063644, -0.0031904433853924274, -0.009557022713124752, 0.033878788352012634, -0.025981368497014046, -0.0016596028581261635...
[ "ACT2FN", "Linear", "ModelTextMLP", "Module", "__init__", "act_fn", "class", "config", "def", "down_proj", "forward", "gate_proj", "hidden_act", "hidden_size", "intermediate_size", "nn", "return", "self", "super", "up_proj", "x" ]
mllama/modeling_mllama.py:MllamaSelfAttentionDecoderLayer
[ -0.00006500193558167666, 0.04902200028300285, 0.028333816677331924, 0.007814287208020687, -0.00027933265664614737, 0.03957739472389221, 0.024960743263363838, -0.029907917603850365, 0.006296403706073761, 0.002318988088518381, 0.014223126694560051, 0.02968304604291916, -0.000822186644654721, ...
[ "False", "GradientCheckpointingLayer", "ModelSelfAttentionDecoderLayer", "ModelTextMLP", "ModelTextRMSNorm", "ModelTextSelfAttention", "None", "Tensor", "__init__", "attention_mask", "cache_position", "class", "config", "cross_attention_mask", "cross_attention_states", "def", "eps", ...
mllama/modeling_mllama.py:MllamaCrossAttentionDecoderLayer
[ -0.00022159525542519987, 0.03820016235113144, 0.03339654952287674, 0.001279533957131207, -0.001136569306254387, 0.04025885462760925, 0.022416863590478897, -0.03179534524679184, 0.0013295715907588601, -0.0003913658147212118, 0.01109405979514122, 0.030422884970903397, -0.0007648610626347363, ...
[ "False", "GradientCheckpointingLayer", "ModelCrossAttentionDecoderLayer", "ModelTextCrossAttention", "ModelTextMLP", "ModelTextRMSNorm", "None", "Parameter", "Tensor", "__init__", "attention_mask", "attn_weights", "cache_position", "class", "config", "cross_attention_mask", "cross_at...
mllama/modeling_mllama.py:MllamaRotaryEmbedding
[ -0.0003249177534598857, 0.04832743480801582, 0.002047526417300105, -0.005750500131398439, -0.0015901003498584032, 0.04112478718161583, 0.04089244455099106, 0.007057431619614363, -0.00312211480922997, 0.021143252030014992, 0.0019603976979851723, -0.004269310738891363, -0.0014811893925070763, ...
[ "False", "ModelRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "arange", "attention_factor", "attention_scaling", "base", "cat", "class", "clone", "compute_default_rope_parameters", "config", "cos", "cpu", "def", "default", "device", ...
mllama/modeling_mllama.py:MllamaPreTrainedModel
[ -0.00022467288363259286, 0.04933212697505951, 0.000628728826995939, 0.016254594549536705, -0.0011082678101956844, 0.0336458720266819, 0.01670926809310913, -0.028644459322094917, -0.0009519736049696803, 0.010855341330170631, 0.007729457225650549, 0.004916162230074406, -0.0036373918410390615, ...
[ "Conv2d", "Embedding", "False", "LayerNorm", "Linear", "ModelConfig", "ModelCrossAttentionDecoderLayer", "ModelPreTrainedModel", "ModelPrecomputedAspectRatioEmbedding", "ModelPrecomputedPositionEmbedding", "ModelRotaryEmbedding", "ModelSelfAttentionDecoderLayer", "ModelTextCrossAttention", ...
mllama/modeling_mllama.py:MllamaVisionModel
[ 0.0000462499592686072, 0.04043468087911606, 0.0030716953333467245, 0.031275443732738495, 0.0003944335912819952, 0.02926487848162651, 0.03194563090801239, -0.015190929174423218, 0.00720452144742012, 0.024126769974827766, 0.029488274827599525, 0.015637721866369247, -0.000617829617112875, -0....
[ "BaseModelOutput", "Conv2d", "F", "False", "LayerNorm", "ModelPreTrainedModel", "ModelPrecomputedAspectRatioEmbedding", "ModelPrecomputedPositionEmbedding", "ModelVisionConfig", "ModelVisionEncoder", "ModelVisionModel", "None", "Parameter", "True", "_", "__init__", "_prepare_aspect_r...
mllama/modeling_mllama.py:MllamaTextModel
[ -0.00010132812894880772, 0.04669199883937836, 0.012688043527305126, 0.001860913005657494, -0.0007718559936620295, 0.04172956570982933, 0.026616696268320084, -0.019173042848706245, 0.005892891436815262, 0.018157999962568283, 0.025150522589683533, 0.016015131026506424, -0.0006661223014816642, ...
[ "BaseModelOutputWithPast", "DynamicCache", "Embedding", "False", "ModelPreTrainedModel", "ModelRotaryEmbedding", "ModelTextConfig", "ModelTextModel", "ModelTextRMSNorm", "ModuleList", "None", "Tensor", "ValueError", "You", "__init__", "and", "arange", "attention_mask", "auto_docs...
mllama/modeling_mllama.py:MllamaForCausalLM
[ -0.0002982082078233361, 0.04782828316092491, 0.018395494669675827, 0.008277972228825092, -0.0013724763412028551, 0.03196217119693756, 0.03725087642669678, -0.019200297072529793, -0.0006431236979551613, 0.030582508072257042, 0.02483391761779785, -0.0016886488301679492, -0.0019545212853699923,...
[ "CausalLMOutputWithPast", "GenerationMixin", "Linear", "ModelForCausalLM", "ModelPreTrainedModel", "ModelTextConfig", "ModelTextModel", "None", "Tensor", "True", "__init__", "_can_compile_fullgraph", "_from_config", "attention_mask", "attentions", "auto_docstring", "base_model_prefix...
mllama/modeling_mllama.py:MllamaModel
[ -0.00002278784813825041, 0.044002681970596313, 0.008500518277287483, 0.017334390431642532, 0, 0.042447034269571304, 0.018001098185777664, -0.029779594391584396, 0.003208528971299529, 0.03822455182671547, 0.04622504115104675, 0.01722327247262001, 0.00009636003960622475, 0.004083582200109959...
[ "BaseModelOutputWithPast", "Linear", "ModelModel", "ModelPreTrainedModel", "ModelTextModel", "ModelVisionModel", "None", "ValueError", "You", "__init__", "_checkpoint_conversion_mapping", "_from_config", "_prepare_cross_attention_mask", "and", "aspect_ratio_ids", "aspect_ratio_mask", ...
mllama/modeling_mllama.py:MllamaForConditionalGeneration
[ -0.0003075887216255069, 0.03982916474342346, 0.027353936806321144, 0.01888451725244522, -0.001258967793546617, 0.03502219542860985, 0.007611032575368881, -0.03708232566714287, -0.004435000009834766, 0.02746838890016079, 0.04120258241891861, 0.008068839088082314, 0.0014234919799491763, -0.0...
[ "CausalLMOutputWithPast", "False", "GenerationMixin", "Linear", "ModelForConditionalGeneration", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "_checkpoint_conversion_mapping", "_update_model_kwargs_for_generation", "and", "aspect_ratio_ids", "aspect_ratio_mask", "attention_m...
cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionMultiModalProjector
[ -0.00037384219467639923, 0.04144103080034256, 0.031138012185692787, 0.028962930664420128, -0.001130470191128552, 0.02117842808365822, 0.019690213724970818, -0.06685514748096466, -0.00024147701333276927, 0.025414112955331802, 0.0295353215187788, -0.011218843050301075, -0.002289559692144394, ...
[ "Linear", "ModelMultiModalProjector", "Module", "SiLU", "__init__", "act", "alignment_intermediate_size", "batch_size", "channels", "chunk", "class", "config", "def", "dim", "downsample_factor", "feature_dim", "forward", "gate", "height", "hidden_size", "hidden_states", "im...
cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionModelOutputWithPast
[ -0.000123380494187586, 0.02538112923502922, 0.014777458272874355, 0.01703355833888054, -0.0007931602885946631, 0.02910369634628296, 0.04331712797284126, -0.038353707641363144, 0.016920752823352814, -0.007952754385769367, 0.02763723023235798, 0.026170765981078148, -0.002354804892092943, 0.0...
[ "BaseModelOutputWithPast", "ModelModelOutputWithPast", "None", "class", "image_hidden_states", "r" ]
cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionCausalLMOutputWithPast
[ -0.00020158913685008883, 0.02637176588177681, 0.02978190779685974, 0.009434727020561695, -0.0011793408775702119, 0.03387407958507538, 0.03773890808224678, -0.030463937669992447, 0.01784641109406948, -0.004205842036753893, 0.020006166771054268, 0.018642110750079155, -0.0008596400148235261, ...
[ "ModelCausalLMOutputWithPast", "ModelOutput", "None", "attentions", "class", "hidden_states", "image_hidden_states", "logits", "loss", "past_key_values", "r" ]
cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionPreTrainedModel
[ -0.000341487058904022, 0.03185274824500084, 0.015693873167037964, 0.01767013780772686, -0.0018454832024872303, 0.03022523783147335, 0.028248971328139305, -0.032085251063108444, -0.0034439333248883486, 0.01057883258908987, 0.012903851456940174, 0.009474449791014194, -0.004679099190980196, 0...
[ "Attention", "DecoderLayer", "False", "ModelConfig", "ModelPreTrainedModel", "PreTrainedModel", "True", "_can_compile_fullgraph", "_can_record_outputs", "_skip_keys_device_placement", "_supports_attention_backend", "_supports_flash_attn", "_supports_flex_attn", "_supports_sdpa", "attenti...
cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionModel
[ -0.00013851301628164947, 0.03858153894543648, -0.0024531614035367966, 0.013715403154492378, -0.0004425447550602257, 0.02598121017217636, 0.017172129824757576, -0.034567274153232574, -0.00301069812849164, 0.031891100108623505, 0.027319299057126045, 0.012544576078653336, -0.0011987038888037205...
[ "AutoModel", "Image", "ModelModel", "ModelModelOutputWithPast", "ModelMultiModalProjector", "ModelPreTrainedModel", "None", "Obtains", "True", "ValueError", "You", "__init__", "_checkpoint_conversion_mapping", "all", "and", "apply", "attention_mask", "attentions", "auto_docstring...
cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionForConditionalGeneration
[ -0.00022356834961101413, 0.03650325536727905, 0.00805550254881382, 0.021180901676416397, -0.0008062543929554522, 0.019490934908390045, 0.021969551220536232, -0.040784500539302826, -0.004731903318315744, 0.01735031232237816, 0.03605259582400322, 0.0030278547201305628, 0.0021969550289213657, ...
[ "False", "GenerationMixin", "Linear", "ModelCausalLMOutputWithPast", "ModelForConditionalGeneration", "ModelModel", "ModelPreTrainedModel", "None", "True", "__init__", "_checkpoint_conversion_mapping", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "cache_posit...
bigbird_pegasus/modeling_bigbird_pegasus.py:shift_tokens_right
[ -0.0001357423752779141, 0.026739485561847687, 0.00459761219099164, -0.04174518957734108, -0.0005782272783108056, 0.05189942196011543, 0.04039129242300987, -0.057089366018772125, 0.006572046782821417, 0.007784913759678602, 0.020759768784046173, -0.009872172959148884, -0.0006698974757455289, ...
[ "Model_tokens_right", "Modeled_input_ids", "None", "clone", "decoder_start_token_id", "def", "if", "input_ids", "is", "masked_fill_", "new_zeros", "pad_token_id", "return", "shape" ]
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusLearnedPositionalEmbedding
[ -0.00023594559752382338, 0.01361334603279829, -0.007550259120762348, 0.0048333099111914635, -0.0015300714876502752, 0.026082713156938553, 0.026883497834205627, -0.005862890742719173, 0.004518715664744377, -0.0017946164589375257, 0.01023860927671194, 0.005805691704154015, 0.001758867176249623...
[ "Embedding", "ModelLearnedPositionalEmbedding", "None", "__init__", "arange", "class", "def", "device", "dtype", "embedding_dim", "forward", "if", "input_ids_shape", "is", "long", "nn", "num_embeddings", "past_key_values_length", "position_ids", "return", "self", "seq_len",...
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusScaledWordEmbedding
[ -0.00018138372979592532, 0.03397626802325249, 0.021971320733428, 0.020159251987934113, -0.0008741811034269631, 0.019253218546509743, 0.04462216794490814, -0.020725524052977562, 0.00569102494046092, -0.00651211803779006, 0.005974160507321358, 0.018913455307483673, 0.002505749696865678, -0.0...
[ "Embedding", "ModelScaledWordEmbedding", "__init__", "class", "def", "embed_scale", "embedding_dim", "forward", "input_ids", "nn", "num_embeddings", "padding_idx", "return", "self", "super" ]
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusSelfAttention
[ 0.00003463905522949062, 0.04127058386802673, 0.03145487606525421, -0.0030534653924405575, -0.00009847077308222651, 0.011990777216851711, 0.021416086703538895, -0.018850617110729218, 0.0006936525460332632, 0.019966037943959236, 0.01115421112626791, 0.01806982234120369, -0.00009759935346664861...
[ "Dropout", "False", "Linear", "ModelSelfAttention", "Module", "None", "The", "ValueError", "_", "__init__", "a", "all_head_size", "and", "attention", "attention_head_size", "attention_mask", "attention_probs", "attention_probs_dropout_prob", "attention_scores", "batch_size", ...
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusBlockSparseAttention
[ -0.000024794891942292452, -0.005554055795073509, 0.02709018997848034, -0.031510766595602036, -0.00021164210920687765, 0.03491120785474777, 0.02164948172867298, -0.045339230448007584, 0.011448155157268047, 0.008784475736320019, 0.013375072740018368, 0.01025800034403801, -0.0012539130402728915...
[ "Key", "Linear", "Make", "ModelBlockSparseAttention", "Modelbird_block_sparse_attention", "Module", "None", "Query", "The", "Value", "ValueError", "_", "_Modelbird_block_rand_mask", "_Modelbird_block_rand_mask_with_head", "__init__", "_create_rand_mask_from_inputs", "_get_rand_attn_p...
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusEncoderAttention
[ -0.000011840564184240066, 0.03592514991760254, 0.03704781085252762, -0.010103948414325714, 0, 0.03749687224626541, 0.018748436123132706, -0.042661115527153015, 0, 0.017738042399287224, 0.01897296868264675, 0.005164240021258593, 0.001431392622180283, -0.013471931219100952, -0.020881492644...
[ "False", "Linear", "ModelBlockSparseAttention", "ModelEncoderAttention", "ModelSelfAttention", "Module", "None", "ValueError", "__init__", "attention_mask", "attention_output", "attention_type", "attn_weights", "band_mask", "be", "block_sparse", "but", "can", "class", "config",...
bigbird_pegasus/modeling_bigbird_pegasus.py:eager_attention_forward
[ 0.00002116747054969892, 0.02788461372256279, 0.02686857432126999, -0.011007084511220455, 0.00009084372140932828, 0.03590003028512001, 0.06141388416290283, -0.02427203208208084, 0.02099812962114811, 0.0128698218613863, 0.022578634321689606, 0.028110399842262268, 0.0027376594953238964, -0.00...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "dtype", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "size", "softmax", "to"...
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusDecoderAttention
[ -0.0002549925120547414, 0.055920034646987915, 0.039486389607191086, -0.012553476728498936, -0.001291010994464159, 0.00935804657638073, 0.027960017323493958, -0.0424535758793354, -0.0023252463433891535, 0.012838782742619514, -0.0013694701483473182, 0.013409395702183247, -0.003095573280006647,...
[ "ALL_ATTENTION_FUNCTIONS", "EncoderDecoderCache", "False", "Instantiating", "Linear", "ModelDecoderAttention", "Module", "None", "Please", "True", "ValueError", "__class__", "__init__", "__name__", "_attn_implementation", "a", "and", "attention_interface", "attention_mask", "at...
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusEncoderLayer
[ -0.0000737809605197981, 0.038675274699926376, 0.02945617027580738, -0.0031339339911937714, -0.00020992440113332123, 0.030130738392472267, 0.007364042568951845, -0.02405961975455284, -0.0064084031619131565, 0.015739938244223595, 0.013716232031583786, 0.015290225856006145, 0.000346067827194929...
[ "ACT2FN", "False", "GradientCheckpointingLayer", "LayerNorm", "Linear", "ModelEncoderAttention", "ModelEncoderLayer", "None", "ValueError", "__init__", "activation_dropout", "activation_fn", "activation_function", "and", "any", "attention_mask", "attention_type", "band_mask", "be...
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusDecoderLayer
[ -0.00012360607797745615, 0.05175916105508804, 0.020229017361998558, -0.006582906469702721, -0.0006745360442437232, 0.02983497641980648, 0.044074397534132004, -0.028817875310778618, 0.007034951820969582, -0.0063568842597305775, -0.005311529617756605, 0.02169816382229328, -0.000427323888288810...
[ "ACT2FN", "False", "GradientCheckpointingLayer", "LayerNorm", "Linear", "ModelDecoderAttention", "ModelDecoderLayer", "None", "True", "__init__", "activation_dropout", "activation_fn", "activation_function", "attention_dropout", "attention_mask", "cache_position", "class", "config"...
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusClassificationHead
[ -0.0002954535011667758, 0.02414887398481369, 0.032806020230054855, 0.012131392024457455, -0.0010394267737865448, 0.019706392660737038, 0.0478421114385128, 0.004157707095146179, -0.003844455350190401, -0.0045848689042031765, 0.01788383722305298, -0.013384399935603142, -0.0007190555334091187, ...
[ "Dropout", "Linear", "ModelClassificationHead", "Module", "__init__", "class", "def", "dense", "dropout", "forward", "hidden_states", "inner_dim", "input_dim", "nn", "num_classes", "out_proj", "p", "pooler_dropout", "return", "self", "super", "tanh", "torch" ]
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusPreTrainedModel
[ -0.0002963812730740756, 0.03633705899119377, -0.0009641318465583026, 0.017939994111657143, -0.0012640840141102672, 0.02331056445837021, 0.03610852360725403, -0.026852857321500778, -0.0011426748242229223, 0.0004820659232791513, 0.010684008710086346, -0.02479604259133339, -0.00275670294649899,...
[ "ModelConfig", "ModelDecoderLayer", "ModelEncoderLayer", "ModelForConditionalGeneration", "ModelPreTrainedModel", "PreTrainedModel", "True", "_can_compile_fullgraph", "_init_weights", "_no_split_modules", "_skip_keys_device_placement", "attention_mask", "base_model_prefix", "class", "con...
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusEncoder
[ -0.00008247972436947748, 0.02639351226389408, 0.013533694669604301, -0.013028287328779697, -0.0005054076900705695, 0.05233777314424515, 0.001221401966176927, -0.03481697291135788, 0.00023866473929956555, -0.0010037957690656185, 0.010725874453783035, 0.015499169006943703, 0.001249480061233043...
[ "Attention", "BaseModelOutput", "Changing", "False", "Input", "LayerNorm", "ModelEncoder", "ModelEncoderLayer", "ModelLearnedPositionalEmbedding", "ModelPreTrainedModel", "ModelScaledWordEmbedding", "ModuleList", "None", "Sequence", "True", "ValueError", "__init__", "_pad_to_block_...
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusDecoder
[ -0.00022999616339802742, 0.07257181406021118, 0.002439029049128294, -0.020310979336500168, -0.0013692794600501657, 0.025103457272052765, 0.049978699535131454, -0.036057692021131516, 0.005305957980453968, -0.014605646952986717, -0.0010412229457870126, 0.017230099067091942, -0.0025103455409407...
[ "BaseModelOutputWithPastAndCrossAttentions", "DynamicCache", "EncoderDecoderCache", "False", "LayerNorm", "ModelDecoder", "ModelDecoderLayer", "ModelLearnedPositionalEmbedding", "ModelPreTrainedModel", "ModelScaledWordEmbedding", "ModuleList", "None", "Setting", "True", "ValueError", "...
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusModel
[ -0.0000630183276371099, 0.05646442249417305, 0.0053495559841394424, -0.02296667918562889, -0.00017855192709248513, 0.04279644787311554, 0.02890440635383129, -0.040555793792009354, 0.008850574493408203, -0.006385857239365578, 0.02834424376487732, -0.002072602743282914, -0.0017295030411332846,...
[ "BaseModelOutput", "If", "ModelDecoder", "ModelEncoder", "ModelModel", "ModelPreTrainedModel", "ModelScaledWordEmbedding", "None", "Please", "Seq2SeqModelOutput", "ValueError", "__init__", "_tied_weights_keys", "and", "are", "attention_mask", "attentions", "auto_docstring", "be",...
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusForConditionalGeneration
[ -0.00028457859298214316, 0.03847502917051315, 0.009334178641438484, -0.0037564376834779978, -0.0013161760289222002, 0.03824736550450325, 0.011952301487326622, -0.02014816552400589, -0.003870269050821662, -0.009846420027315617, 0.024701423943042755, 0.0031445936765521765, -0.00003201509389327...
[ "CrossEntropyLoss", "False", "GenerationMixin", "Linear", "ModelForConditionalGeneration", "ModelModel", "ModelPreTrainedModel", "None", "Seq2SeqLMOutput", "True", "__init__", "_keys_to_ignore_on_load_missing", "_resize_final_logits_bias", "_tied_weights_keys", "and", "attention_mask",...
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusForSequenceClassification
[ -0.00025072647258639336, 0.034579064697027206, 0.0134474141523242, 0.031640972942113876, -0.0007945557008497417, 0.019888611510396004, 0.016272500157356262, 0.007853741757571697, 0.0030793447513133287, 0.0017021149396896362, 0.04565340653061867, 0.001730365795083344, 0.0017515539657324553, ...
[ "All", "BCEWithLogitsLoss", "CrossEntropyLoss", "False", "MSELoss", "ModelClassificationHead", "ModelForSequenceClassification", "ModelModel", "ModelPreTrainedModel", "None", "NotImplementedError", "Passing", "Seq2SeqSequenceClassifierOutput", "__class__", "__init__", "__name__", "an...
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusForQuestionAnswering
[ -0.00021988233493175358, 0.03354876488447189, 0.011314265429973602, 0.018125340342521667, -0.0006578879547305405, 0.04615769907832146, 0.033999085426330566, 0.012214903719723225, 0.006417045835405588, 0.029270736500620842, 0.02589334361255169, 0.015085686929523945, 0.0024345372803509235, -...
[ "CrossEntropyLoss", "False", "Linear", "ModelForQuestionAnswering", "ModelModel", "ModelPreTrainedModel", "None", "Seq2SeqQuestionAnsweringModelOutput", "__init__", "and", "attention_mask", "auto_docstring", "cache_position", "clamp", "class", "config", "contiguous", "cross_attenti...
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusDecoderWrapper
[ -0.00007960084622027352, 0.058670200407505035, -0.005794242024421692, -0.009013265371322632, -0.0005808238056488335, 0.02396073006093502, 0.03000689670443535, -0.02955903299152851, -0.0018194479634985328, -0.014107719995081425, -0.008117537014186382, 0.00895728264003992, -0.00405876850709319...
[ "ModelDecoder", "ModelDecoderWrapper", "ModelPreTrainedModel", "__init__", "args", "class", "config", "decoder", "def", "forward", "kwargs", "post_init", "return", "self", "super" ]
bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusForCausalLM
[ -0.0001852803980000317, 0.05443361774086952, 0.016262324526906013, -0.0016234092181548476, -0.0009811038617044687, 0.015132997184991837, 0.029814263805747032, -0.0156976617872715, 0.0035150337498635054, 0.014116602018475533, 0.01953737623989582, -0.010615685023367405, 0.00021527818171307445,...
[ "CausalLMOutputWithCrossAttentions", "CrossEntropyLoss", "False", "GenerationMixin", "Linear", "ModelDecoderWrapper", "ModelForCausalLM", "ModelPreTrainedModel", "None", "True", "__init__", "attention_mask", "attentions", "auto_docstring", "cache_position", "class", "config", "cros...
perception_lm/modeling_perception_lm.py:PerceptionLMAdaptiveAvgPooling
[ -0.00025583416572771966, 0.008130233734846115, 0.03410181403160095, 0.018067186698317528, -0.0009104167693294585, 0.005702455528080463, 0.01067093200981617, -0.026197420433163643, 0.0038110469467937946, 0.023374421522021294, 0.012816410511732101, -0.014792508445680141, -0.001002164208330214,...
[ "F", "ModelAdaptiveAvgPooling", "Module", "__init__", "adaptive_avg_pool2d", "b", "c", "class", "def", "flatten", "forward", "h", "hidden_states", "if", "int", "math", "nn", "num_tokens", "permute", "pooling_ratio", "reshape", "return", "self", "shape", "sqrt", "sup...
perception_lm/modeling_perception_lm.py:PerceptionLMMultiModalProjector
[ -0.00031123144435696304, 0.03159535676240921, 0.030450597405433655, 0.05082731321454048, -0.0007870219997130334, 0.034342776983976364, 0.029191361740231514, -0.02724527008831501, -0.00023074053751770407, 0.02701631933450699, 0.019689859822392464, -0.010302833281457424, -0.0018316148780286312...
[ "GELU", "Identity", "Linear", "ModelAdaptiveAvgPooling", "ModelMultiModalProjector", "Module", "__init__", "class", "config", "def", "else", "embed_dim", "features", "forward", "gelu", "hidden_size", "if", "in_features", "input_size", "linear_1", "linear_2", "model_args", ...
perception_lm/modeling_perception_lm.py:PerceptionLMPreTrainedModel
[ -0.00040332425851374865, 0.03276687115430832, 0.01826929859817028, 0.019447963684797287, -0.0021068628411740065, 0.03158820793032646, 0.028523679822683334, -0.03323833644390106, -0.003683326533064246, 0.01744423434138298, 0.015440504997968674, 0.008898916654288769, -0.004272658843547106, 0...
[ "ModelConfig", "ModelPreTrainedModel", "PreTrainedModel", "True", "_can_compile_fullgraph", "_skip_keys_device_placement", "_supports_attention_backend", "_supports_flash_attn", "_supports_flex_attn", "_supports_sdpa", "base_model_prefix", "class", "config", "image", "input_modalities", ...
perception_lm/modeling_perception_lm.py:PerceptionLMModelOutputWithPast
[ -0.00013376322749536484, 0.032666388899087906, 0.004590190481394529, 0.015319408848881721, -0.0008800211944617331, 0.031089389696717262, 0.045057084411382675, -0.05857421085238457, 0.016558479517698288, -0.0008765011443756521, 0.028385963290929794, 0.032666388899087906, -0.00266118417493999,...
[ "BaseModelOutputWithPast", "ModelModelOutputWithPast", "None", "class", "image_hidden_states", "r", "video_hidden_states" ]
perception_lm/modeling_perception_lm.py:PerceptionLMCausalLMOutputWithPast
[ -0.0002032790653174743, 0.031223664060235023, 0.021155163645744324, 0.007070576306432486, -0.0010040218476206064, 0.032807473093271255, 0.04185781255364418, -0.05022937431931496, 0.017761288210749626, -0.0009121043258346617, 0.02601972036063671, 0.02092890627682209, -0.0012373508652672172, ...
[ "ModelCausalLMOutputWithPast", "ModelOutput", "None", "attentions", "class", "hidden_states", "image_hidden_states", "logits", "loss", "past_key_values", "r", "video_hidden_states" ]
perception_lm/modeling_perception_lm.py:PerceptionLMModel
[ -0.00008251098915934563, 0.037798721343278885, -0.013729829341173172, 0.009505266323685646, -0.0002015005302382633, 0.036242302507162094, 0.023568613454699516, -0.04624784365296364, 0.000712200126145035, 0.02623675763607025, 0.030906012281775475, 0.012117824517190456, -0.0005489152390509844,...
[ "AutoModel", "Image", "ModelModel", "ModelModelOutputWithPast", "ModelMultiModalProjector", "ModelPreTrainedModel", "None", "Obtains", "True", "ValueError", "Video", "You", "_", "__init__", "_checkpoint_conversion_mapping", "all", "and", "apply", "at", "attention_mask", "atte...
perception_lm/modeling_perception_lm.py:PerceptionLMForConditionalGeneration
[ -0.0002118964766850695, 0.037139032036066055, 0.0005029243766330183, 0.01868205890059471, -0.000752628140617162, 0.027460375800728798, 0.022283419966697693, -0.04771802946925163, -0.00472678616642952, 0.01496815588325262, 0.038939714431762695, 0.0035028860438615084, 0.0003640047216322273, ...
[ "False", "GenerationMixin", "Linear", "ModelCausalLMOutputWithPast", "ModelForConditionalGeneration", "ModelModel", "ModelPreTrainedModel", "None", "True", "__init__", "_checkpoint_conversion_mapping", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "cache_posit...
pvt_v2/modeling_pvt_v2.py:drop_path
[ 0, 0.017753854393959045, 0.03775503486394882, -0.02123720571398735, 0, 0.043598074465990067, 0.0339345820248127, -0.027304979041218758, 0.014832334592938423, 0.0017486985307186842, 0.020113544538617134, -0.04944111406803131, 0.0016574009787291288, -0.018315685912966728, -0.05034004524350...
[ "False", "Model_path", "Model_prob", "def", "device", "div", "dtype", "floor_", "if", "input", "keep_prob", "ndim", "not", "or", "output", "rand", "random_tensor", "return", "shape", "torch", "training" ]
pvt_v2/modeling_pvt_v2.py:PvtV2DropPath
[ -0.00011107319733127952, 0.018094833940267563, 0.03596488758921623, -0.025063030421733856, -0.00047414645086973906, 0.031918834894895554, 0.033717080950737, -0.026861274614930153, 0.0066872211173176765, -0.002430439693853259, 0.009721758775413036, -0.033717080950737, -0.000818341679405421, ...
[ "ModelDropPath", "Module", "None", "__init__", "class", "def", "drop_path", "drop_prob", "extra_repr", "f", "forward", "hidden_states", "nn", "p", "return", "self", "super", "training" ]
pvt_v2/modeling_pvt_v2.py:PvtV2OverlapPatchEmbeddings
[ -0.00012412518844939768, 0.008620098233222961, 0.02693076618015766, 0.012845637276768684, -0.0004824156640097499, 0.021522076800465584, 0.018479688093066216, -0.01701483502984047, 0.00498613528907299, 0.0031268985476344824, 0.023099610581994057, 0.020733309909701347, -0.001542321522720158, ...
[ "Conv2d", "LayerNorm", "ModelOverlapPatchEmbeddings", "Module", "_", "__init__", "class", "config", "def", "else", "embeddings", "eps", "flatten", "forward", "height", "hidden_size", "hidden_sizes", "if", "int", "isinstance", "kernel_size", "layer_idx", "layer_norm", "l...
pvt_v2/modeling_pvt_v2.py:PvtV2DepthWiseConv
[ -0.00006560274050571024, -0.0015427651815116405, 0.026600372046232224, -0.03967513144016266, -0.00018756221106741577, 0.02231726050376892, 0.03449031338095665, -0.043732814490795135, 0.01408917922526598, 0.009636999107897282, 0.032236043363809586, -0.010820490308105946, 0.0015004976885393262...
[ "Conv2d", "ModelDepthWiseConv", "Module", "__init__", "batch_size", "class", "config", "def", "dim", "dwconv", "flatten", "forward", "groups", "height", "hidden_states", "nn", "num_channels", "return", "self", "seq_len", "shape", "super", "transpose", "view", "width" ...
pvt_v2/modeling_pvt_v2.py:PvtV2SelfAttention
[ -0.0000353922659996897, 0.03736025094985962, 0.029082830995321274, -0.008333349600434303, 0, 0.009955276735126972, 0.005704709328711033, -0.0236018355935812, 0.007326636463403702, 0.03154368698596954, 0.01409398764371872, 0.013310988433659077, 0.0014121951535344124, -0.026845689862966537, ...
[ "AdaptiveAvgPool2d", "Conv2d", "Dropout", "False", "GELU", "LayerNorm", "Linear", "ModelSelfAttention", "Module", "The", "ValueError", "__init__", "a", "act", "all_head_size", "attention", "attention_head_size", "attention_probs", "attention_probs_dropout_prob", "attention_scor...
pvt_v2/modeling_pvt_v2.py:PvtV2ConvFeedForwardNetwork
[ -0.00014040115638636053, 0.032551877200603485, 0.027917755767703056, -0.016615020111203194, -0.0005015588831156492, 0.02430088073015213, 0.037525080144405365, -0.050636254251003265, 0.006583843380212784, -0.008307510055601597, 0.02893500216305256, -0.013450254686176777, 0.005397056229412556,...
[ "ACT2FN", "Dropout", "Identity", "Linear", "ModelConvFeedForwardNetwork", "ModelDepthWiseConv", "Module", "None", "ReLU", "__init__", "class", "config", "def", "dense1", "dense2", "dropout", "dwconv", "else", "forward", "height", "hidden_act", "hidden_dropout_prob", "hidd...
pvt_v2/modeling_pvt_v2.py:PvtV2BlockLayer
[ -0.00011197711137356237, 0.00014371865836437792, 0.026634681969881058, -0.014276640489697456, -0.00019133096793666482, 0.05823514983057976, 0.027311833575367928, -0.029343292117118835, 0.006517596542835236, -0.007674399763345718, 0.0020455659832805395, -0.01422021072357893, 0.000518445216584...
[ "False", "Identity", "LayerNorm", "ModelBlockLayer", "ModelConvFeedForwardNetwork", "ModelDropPath", "ModelSelfAttention", "Module", "__init__", "attention", "attention_output", "class", "config", "def", "drop_path", "else", "eps", "forward", "height", "hidden_features", "hid...
pvt_v2/modeling_pvt_v2.py:PvtV2EncoderLayer
[ -0.00007066135731292889, 0.008539052680134773, 0.028987836092710495, 0.013819782063364983, -0.00010050590935861692, 0.04179641231894493, 0.015842188149690628, -0.020898206159472466, 0.0007092469604685903, -0.004831305705010891, -0.004718950018286705, 0.008426696062088013, -0.0023875639308243...
[ "GradientCheckpointingLayer", "LayerNorm", "ModelBlockLayer", "ModelEncoderLayer", "ModelOverlapPatchEmbeddings", "ModuleList", "None", "__init__", "all_self_attentions", "append", "block", "block_idx", "block_layers", "blocks", "class", "config", "cpu", "def", "depths", "devic...
pvt_v2/modeling_pvt_v2.py:PvtV2Encoder
[ -0.00005757914550486021, 0.010858284309506416, 0.016765641048550606, 0.030155649408698082, -0.00009889547800412402, 0.03465649113059044, 0.014965303242206573, -0.04838406294584274, 0.009057946503162384, 0.020253794267773628, 0.01462774071842432, -0.0027427012100815773, -0.00118850392755121, ...
[ "BaseModelOutput", "False", "ModelEncoder", "ModelEncoderLayer", "Module", "ModuleList", "None", "True", "__init__", "all_hidden_states", "all_self_attentions", "attentions", "batch_size", "class", "config", "contiguous", "def", "else", "enumerate", "for", "forward", "gradi...
pvt_v2/modeling_pvt_v2.py:PvtV2PreTrainedModel
[ -0.00022046589583624154, 0.04628020152449608, 0.0004938435740768909, 0.027880998328328133, -0.0006949084927327931, 0.02381737157702446, 0.007280665449798107, -0.0022716806270182133, -0.0016296838875859976, 0.02562342770397663, 0.006208319682627916, -0.011513610370457172, -0.00170023296959698...
[ "Conv2d", "LayerNorm", "Linear", "Model", "ModelConfig", "ModelPreTrainedModel", "None", "PreTrainedModel", "True", "_init_weights", "base_model_prefix", "bias", "class", "config", "def", "elif", "fan_out", "groups", "if", "image", "init", "initializer_range", "input_moda...
pvt_v2/modeling_pvt_v2.py:PvtV2Model
[ 0.000051370163419051096, 0.031337983906269073, 0.0062396167777478695, 0.03044261410832405, 0.0004389416426420212, 0.02518230862915516, 0.029323400929570198, -0.02507038787007332, 0.008897749707102776, 0.01634052023291588, 0.0212650615721941, 0.016116678714752197, 0.00013028347166255116, 0....
[ "BaseModelOutput", "ModelEncoder", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "attentions", "auto_docstring", "class", "config", "def", "else", "encoder", "encoder_outputs", "forward", "hidden_states", "if", "is", "kwargs", "last_hidden_state", "not", "output...
pvt_v2/modeling_pvt_v2.py:PvtV2ForImageClassification
[ -0.00010905389353865758, 0.0381915457546711, 0.006979449186474085, 0.029927877709269524, 0, -0.00036816595820710063, 0.021887551993131638, -0.004997285548597574, 0.005723148584365845, 0.014852267690002918, 0.04198836535215378, 0.0030290810391306877, 0.0004920511855743825, -0.01278635114431...
[ "Identity", "ImageClassifierOutput", "Linear", "Model", "ModelForImageClassification", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "attentions", "auto_docstring", "batch_size", "class", "classifier", "config", "def", "dim", "else", "forward", "hidden_sizes", "hi...
pvt_v2/modeling_pvt_v2.py:PvtV2Backbone
[ -0.00025237759109586477, -0.011690553277730942, 0.026882626116275787, 0.03795194253325462, -0.0005894693313166499, 0.026882626116275787, 0.01361074112355709, -0.021009111776947975, -0.0006918322760611773, 0.006579466164112091, 0.0022872821427881718, -0.007285417523235083, -0.0011012840550392...
[ "BackboneMixin", "BackboneOutput", "ModelBackbone", "ModelModel", "None", "True", "__init__", "attentions", "auto_docstring", "class", "config", "def", "else", "encoder", "enumerate", "feature_maps", "for", "forward", "hidden_sizes", "hidden_states", "idx", "if", "in", ...
x_clip/modeling_x_clip.py:contrastive_loss
[ 0.00008122165309032425, 0.037291333079338074, 0.037291333079338074, -0.0017656880663707852, 0.00034784057061187923, 0.061926212161779404, 0.017854638397693634, 0.025877924636006355, 0.023052824661135674, -0.006780242547392845, 0.030059074983000755, -0.031415123492479324, 0.002161202253773808...
[ "Model_loss", "arange", "cross_entropy", "def", "device", "functional", "len", "logits", "nn", "return", "torch" ]
x_clip/modeling_x_clip.py:x_clip_loss
[ 0.00016533616872038692, 0.03712017834186554, 0.07967258244752884, -0.018786432221531868, 0.0008841506205499172, 0.04391045495867729, -0.004498558584600687, -0.012165912427008152, 0.026595251634716988, 0.01935228891670704, 0.033272355794906616, -0.005517099983990192, 0.0037346521858125925, ...
[ "Model_loss", "caption_loss", "contrastive_loss", "def", "image_loss", "return", "similarity", "t" ]
x_clip/modeling_x_clip.py:XCLIPOutput
[ -0.00009337403025710955, 0.01939360983669758, 0.009245791472494602, -0.002466483972966671, -0.00039816097705625, 0.04848402738571167, 0.05592576041817665, -0.02142317406833172, 0.01928085647523403, 0.014996222220361233, 0.01758955419063568, 0.016800278797745705, -0.002720179269090295, -0.0...
[ "Model", "ModelOutput", "None", "class", "def", "else", "for", "getattr", "if", "in", "k", "keys", "logits_per_text", "logits_per_video", "loss", "mit_output", "not", "r", "return", "self", "text_embeds", "text_model_output", "to_tuple", "tuple", "video_embeds", "vi...
x_clip/modeling_x_clip.py:XCLIPVisionEmbeddings
[ -0.00016189027519430965, 0.030407218262553215, 0.020834574475884438, 0.01723075658082962, -0.0006933127297088504, 0.010360977612435818, 0.02477625198662281, -0.02038409747183323, 0.0004645547305699438, 0.016780279576778412, 0.019145285710692406, 0.016442421823740005, -0.0020412253215909004, ...
[ "Conv2d", "Embedding", "False", "Input", "Model", "Module", "Parameter", "ValueError", "_", "__init__", "align_corners", "and", "arange", "batch_size", "bicubic", "cat", "class", "class_embedding", "class_embeds", "class_pos_embed", "config", "def", "dim", "doesn", "d...
x_clip/modeling_x_clip.py:XCLIPTextEmbeddings
[ -0.0002546473406255245, 0.017330365255475044, 0.004734288435429335, -0.006685389205813408, -0.0014202866004779935, 0.04039926454424858, 0.0438423827290535, -0.0007244896260090172, 0.00889472384005785, 0.0007854615105316043, 0.020543944090604782, 0.011649219319224358, -0.0009898967109620571, ...
[ "Embedding", "False", "Model", "Module", "None", "Sequence", "ValueError", "__init__", "and", "arange", "be", "class", "config", "def", "else", "embed_dim", "embeddings", "expand", "f", "forward", "got", "hidden_size", "if", "input_ids", "inputs_embeds", "is", "le...
x_clip/modeling_x_clip.py:eager_attention_forward
[ 0.000051929029723396525, 0.031000085175037384, 0.02477744035422802, -0.010239078663289547, 0.00008883178816176951, 0.035978201776742935, 0.05928483232855797, -0.020365018397569656, 0.022175243124365807, 0.01170988567173481, 0.02477744035422802, 0.031452640891075134, 0.0025173425674438477, ...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "softmax", "...
x_clip/modeling_x_clip.py:XCLIPAttention
[ -0.00017653060785960406, 0.03996865823864937, 0.037243522703647614, 0.003477386897429824, -0.0006599938496947289, 0.021119803190231323, 0.03860609233379364, -0.008686370216310024, -0.0008870884776115417, 0.03519967198371887, 0.009367654100060463, -0.0035057738423347473, -0.002725135767832398...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "Model", "Module", "None", "ValueError", "__init__", "_attn_implementation", "and", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "batch_size", "be", "by", "class", "config", "contigu...
x_clip/modeling_x_clip.py:XCLIPMLP
[ -0.0001473212760174647, 0.03635108098387718, 0.042485326528549194, 0.011530108749866486, -0.0003603158984333277, 0.02885367162525654, 0.0422581322491169, -0.022037843242287636, 0.006560234352946281, -0.008349388837814331, 0.03021683730185032, -0.019879497587680817, 0.002456537913531065, -0...
[ "ACT2FN", "Linear", "Model", "Module", "__init__", "activation_fn", "class", "config", "def", "fc1", "fc2", "forward", "hidden_act", "hidden_size", "hidden_states", "intermediate_size", "nn", "return", "self", "super" ]
x_clip/modeling_x_clip.py:XCLIPEncoderLayer
[ -0.00006329204916255549, 0.03353072330355644, 0.023178955540060997, 0.009789170697331429, 0, 0.04163210466504097, 0.036906298249959946, 0.0013853926211595535, 0.008326420560479164, 0.010633064433932304, 0.013727342709898949, 0.002939564175903797, 0.0010548675199970603, -0.00666676275432109...
[ "False", "GradientCheckpointingLayer", "LayerNorm", "Model", "ModelAttention", "ModelMLP", "__init__", "attention_mask", "attn_weights", "class", "config", "def", "embed_dim", "eps", "forward", "hidden_size", "hidden_states", "if", "kwargs", "layer_norm1", "layer_norm2", "l...
x_clip/modeling_x_clip.py:drop_path
[ 0, 0.017753854393959045, 0.03775503486394882, -0.02123720571398735, 0, 0.043598074465990067, 0.0339345820248127, -0.027304979041218758, 0.014832334592938423, 0.0017486985307186842, 0.020113544538617134, -0.04944111406803131, 0.0016574009787291288, -0.018315685912966728, -0.05034004524350...
[ "False", "Model_path", "Model_prob", "def", "device", "div", "dtype", "floor_", "if", "input", "keep_prob", "ndim", "not", "or", "output", "rand", "random_tensor", "return", "shape", "torch", "training" ]
x_clip/modeling_x_clip.py:XCLIPDropPath
[ -0.00015407476166728884, 0.015301218256354332, 0.040123194456100464, -0.0190415158867836, -0.0008465257124044001, 0.03150917589664459, 0.042390041053295135, -0.025162002071738243, 0.0022101758513599634, -0.0017993098590523005, 0.013204384595155716, -0.04375014826655388, -0.001430947217158973...
[ "Model", "Module", "None", "__init__", "class", "def", "drop_path", "drop_prob", "extra_repr", "f", "forward", "hidden_states", "nn", "p", "return", "self", "super", "training" ]
x_clip/modeling_x_clip.py:XCLIPVisionEncoderLayer
[ -0.00011392598389647901, 0.025660723447799683, 0.012943404726684093, -0.0032923726830631495, -0.0003885847399942577, 0.042730193585157394, 0.02859983779489994, -0.014921654015779495, 0.013282532803714275, 0.008704298175871372, 0.023173781111836433, 0.012604276649653912, 0.0014271658146753907...
[ "False", "GradientCheckpointingLayer", "Identity", "LayerNorm", "Linear", "Model", "ModelAttention", "ModelDropPath", "ModelMLP", "__init__", "attention_mask", "attn_weights", "batch_size", "batch_time", "cat", "class", "config", "def", "dim", "drop_path", "drop_path_rate", ...
x_clip/modeling_x_clip.py:XCLIPPreTrainedModel
[ -0.00013923647929914296, 0.04872924089431763, 0.0063167535699903965, 0.025379812344908714, -0.0006944198976270854, 0.04241248965263367, 0.01680707558989525, -0.00834713876247406, 0.005611758679151535, 0.010264724493026733, 0.016468679532408714, 0.0016073881415650249, -0.0025520811323076487, ...
[ "LayerNorm", "Linear", "Model", "ModelAttention", "ModelConfig", "ModelMLP", "ModelModel", "ModelMultiframeIntegrationTransformer", "ModelTextEmbeddings", "ModelVisionEmbeddings", "None", "PreTrainedModel", "True", "_init_weights", "arange", "base_model_prefix", "bias", "class", ...
x_clip/modeling_x_clip.py:XCLIPEncoder
[ -0.00008836816414259374, 0.020821474492549896, 0.019245795905590057, 0.030613195151090622, -0.00018816704687196761, 0.03151358291506767, 0.023297540843486786, -0.010073091834783554, 0.006668499205261469, -0.008441138081252575, 0.018232857808470726, -0.013505821116268635, -0.00088631949620321...
[ "BaseModelOutput", "False", "Model", "ModelLayer", "Module", "ModuleList", "None", "_", "__init__", "all_attentions", "attention_mask", "attentions", "can_return_tuple", "class", "config", "def", "else", "encoder_layer", "encoder_states", "enumerate", "for", "forward", "g...
x_clip/modeling_x_clip.py:XCLIPTextTransformer
[ -0.00011487881420180202, 0.03838881477713585, 0.018857663497328758, 0.02200060710310936, -0.0006699802470393479, 0.02660277672111988, 0.02738851122558117, -0.011617667973041534, 0.013469760306179523, 0.016724951565265656, 0.03816432133316994, 0.01100030355155468, -0.00028062000637874007, 0...
[ "BaseModelOutputWithPooling", "LayerNorm", "Model", "ModelEncoder", "ModelPreTrainedModel", "ModelTextEmbeddings", "None", "True", "__init__", "arange", "argmax", "attention_mask", "attentions", "auto_docstring", "cache_position", "class", "config", "create_causal_mask", "def", ...
x_clip/modeling_x_clip.py:XCLIPTextModel
[ 0.000032221862056758255, 0.05216806381940842, 0.016497593373060226, -0.0000892632597242482, -0.00018462254956830293, 0.019730228930711746, 0.03254930302500725, 0.003525245701894164, -0.0025916825979948044, 0.000919629295822233, 0.021067870780825615, 0.010255260393023491, -0.00423586834222078...
[ "Model", "ModelPreTrainedModel", "ModelTextConfig", "ModelTextTransformer", "None", "__init__", "attention_mask", "auto_docstring", "class", "config", "def", "embeddings", "forward", "get_input_embeddings", "input_ids", "input_modalities", "kwargs", "output_attentions", "output_h...
x_clip/modeling_x_clip.py:XCLIPVisionEncoder
[ -0.00005532349678105675, 0.024616321548819542, 0.018771350383758545, 0.03214734420180321, 0.00009396213135914877, 0.031248116865754128, 0.015399252064526081, -0.00651939120143652, 0.004889543633908033, -0.009104667231440544, 0.021693836897611618, -0.010060095228254795, -0.0002028528251685202...
[ "BaseModelOutput", "False", "Model", "ModelLayer", "Module", "ModuleList", "None", "_", "__init__", "all_attentions", "attention_mask", "attentions", "class", "config", "def", "else", "encoder_layer", "encoder_states", "enumerate", "for", "forward", "gradient_checkpointing"...
x_clip/modeling_x_clip.py:XCLIPVisionTransformer
[ -0.00004367959627415985, 0.03975813835859299, 0.006268683820962906, 0.06370338052511215, -0.00009353610221296549, 0.030722197145223618, 0.02767256833612919, -0.024284090846776962, 0.011859672144055367, 0.038176849484443665, 0.04133942723274231, 0.020443815737962723, 0.0023436970077455044, ...
[ "BaseModelOutputWithPooling", "False", "LayerNorm", "Model", "ModelVisionEmbeddings", "ModelVisionEncoder", "Module", "None", "__init__", "attentions", "auto_docstring", "class", "config", "def", "else", "embed_dim", "embeddings", "encoder", "encoder_outputs", "eps", "forward...
x_clip/modeling_x_clip.py:XCLIPVisionModel
[ 0.00004254153463989496, 0.04646670073270798, 0.00804231408983469, 0.034403230994939804, 0.00018238319898955524, 0.001801143167540431, 0.042222145944833755, -0.002792470157146454, 0.0053894673474133015, 0.02278655581176281, 0.01921219378709793, 0.0029600183479487896, -0.0003316058136988431, ...
[ "Model", "ModelPreTrainedModel", "ModelVisionConfig", "ModelVisionTransformer", "None", "__init__", "auto_docstring", "class", "config", "def", "embeddings", "forward", "get_input_embeddings", "image", "input_modalities", "kwargs", "main_input_name", "output_attentions", "output_...
x_clip/modeling_x_clip.py:XCLIPMultiframeIntegrationTransformer
[ -0.00005461640830617398, 0.01522212103009224, 0.005919713992625475, 0.01849205791950226, -0.0004651203635148704, 0.028301870450377464, 0.018604815006256104, -0.02221302129328251, 0.009866190142929554, 0.017364494502544403, 0.00535593181848526, 0.006709008943289518, 0.001064139069058001, -0...
[ "BaseModelOutputWithPooling", "False", "Model", "ModelEncoder", "Module", "None", "Parameter", "__init__", "attentions", "class", "config", "def", "dim", "dtype", "empty", "encoder", "encoder_outputs", "forward", "hidden_size", "hidden_states", "if", "inputs_embeds", "kee...
x_clip/modeling_x_clip.py:XCLIPCrossAttention
[ -0.00013503212539944798, 0.041120368987321854, 0.030049500986933708, 0.016380367800593376, -0.00035479027428664267, 0.014572878368198872, 0.04902813211083412, 0.010901416651904583, 0.006693356670439243, 0.020108312368392944, 0.0050835623405873775, 0.00020563714497257024, -0.00247117597609758...
[ "Dropout", "Linear", "Model", "Module", "__init__", "_shape", "attn", "attn_drop", "batch_size", "class", "config", "contiguous", "def", "dim", "forward", "head_dim", "hidden_size", "k_proj", "key_seq_len", "keys", "nn", "num_heads", "permute", "proj", "proj_drop", ...
x_clip/modeling_x_clip.py:PromptGeneratorLayer
[ -0.00023683997278567404, 0.045131370425224304, 0.018462833017110825, 0.024959014728665352, -0.0007372311665676534, 0.06245452165603638, 0.02404727041721344, -0.009117448702454567, 0.0030201547779142857, 0.008490623906254768, 0.027808217331767082, 0.023705365136265755, 0.001061327988281846, ...
[ "ACT2FN", "Dropout", "LayerNorm", "Linear", "ModelCrossAttention", "ModelGeneratorLayer", "Model_attention_dropout", "Model_hidden_act", "Module", "Sequential", "__init__", "class", "config", "cross_attn", "def", "embed_dim", "eps", "forward", "layer_norm_eps", "mlp", "nn", ...
x_clip/modeling_x_clip.py:XCLIPPromptGenerator
[ -0.00016144043183885515, 0.05904107540845871, 0.010786349885165691, 0.05722442641854286, -0.0005535100935958326, 0.041782915592193604, 0.045416209846735, -0.01759878173470497, 0.009083242155611515, 0.008799390867352486, 0.019188350066542625, -0.0014902193797752261, -0.001504412037320435, -...
[ "LayerNorm", "Model", "Module", "ModuleList", "Parameter", "PromptGeneratorLayer", "_", "__init__", "alpha", "class", "config", "decoder", "def", "embed_dim", "eps", "for", "forward", "in", "layer", "layer_norm_eps", "layernorm", "nn", "ones", "projection_dim", "promp...
x_clip/modeling_x_clip.py:XCLIPModel
[ -0.00008882679685484618, 0.05974734202027321, 0.017723558470606804, 0.016162993386387825, -0.0002821556990966201, 0.037453558295965195, 0.03455536440014839, -0.0325489267706871, 0.003567005507647991, 0.020733218640089035, 0.03344067558646202, 0.01309759821742773, -0.002452316228300333, -0....
[ "False", "LayerNorm", "Linear", "Model", "ModelConfig", "ModelMultiframeIntegrationTransformer", "ModelOutput", "ModelPreTrainedModel", "ModelPromptGenerator", "ModelTextConfig", "ModelTextTransformer", "ModelVisionConfig", "ModelVisionTransformer", "Model_loss", "None", "Parameter", ...
data2vec/modeling_data2vec_vision.py:Data2VecVisionModelOutputWithPooling
[ -0.00007458389882231131, 0.017858019098639488, 0.011736873537302017, 0.04245491325855255, -0.00007984865078469738, 0.040657877922058105, 0.04200565442442894, -0.004296032711863518, 0.01561172679066658, 0.009041323326528072, 0.03391900286078453, 0.017633389681577682, 0.0006317695369943976, ...
[ "BaseModelOutputWithPooling", "ModelVisionModelOutputWithPooling", "class", "r" ]
data2vec/modeling_data2vec_vision.py:drop_path
[ 0, 0.017753854393959045, 0.03775503486394882, -0.02123720571398735, 0, 0.043598074465990067, 0.0339345820248127, -0.027304979041218758, 0.014832334592938423, 0.0017486985307186842, 0.020113544538617134, -0.04944111406803131, 0.0016574009787291288, -0.018315685912966728, -0.05034004524350...
[ "False", "Model_path", "Model_prob", "def", "device", "div", "dtype", "floor_", "if", "input", "keep_prob", "ndim", "not", "or", "output", "rand", "random_tensor", "return", "shape", "torch", "training" ]