identifier
stringlengths
24
117
embedding
listlengths
2.56k
2.56k
tokens
listlengths
4
448
lxmert/modeling_lxmert.py:LxmertLMPredictionHead
[ -0.00033065120805986226, 0.03450273349881172, 0.03381267935037613, 0.019896578043699265, -0.0016029395628720522, 0.0272571612149477, 0.05704452097415924, -0.02748717926442623, 0.00044386330409906805, 0.002113292459398508, 0.021161677315831184, 0.008568178862333298, -0.00024978542933240533, ...
[ "Linear", "ModelLMPredictionHead", "ModelPredictionHeadTransform", "Module", "Parameter", "__init__", "bias", "class", "config", "decoder", "def", "forward", "hidden_size", "hidden_states", "nn", "return", "self", "super", "torch", "transform", "vocab_size", "zeros" ]
lxmert/modeling_lxmert.py:LxmertVisualAnswerHead
[ -0.0002647011133376509, 0.032050296664237976, 0.015910683199763298, 0.04235217720270157, -0.0010802666656672955, 0.03846035525202751, 0.05837732553482056, 0.0007547558634541929, -0.000461438437923789, 0.030905643478035927, 0.03662890940904617, 0.010702509433031082, -0.0009443390881642699, ...
[ "GeLU", "LayerNorm", "Linear", "ModelVisualAnswerHead", "Module", "Sequential", "__init__", "class", "config", "def", "eps", "forward", "hid_dim", "hidden_size", "hidden_states", "logit_fc", "nn", "num_labels", "return", "self", "super" ]
lxmert/modeling_lxmert.py:LxmertVisualObjHead
[ -0.0003618190821725875, 0.029346752911806107, 0.014788012020289898, 0.01811244897544384, -0.0013971232110634446, 0.032327283173799515, 0.050669003278017044, -0.04837628826498985, 0, 0.0008311091805808246, 0.009686720557510853, 0.016736820340156555, -0.0008239444578066468, -0.00166938314214...
[ "ModelPredictionHeadTransform", "ModelVisualObjHead", "Module", "ModuleDict", "__init__", "class", "config", "decoder_dict", "def", "for", "forward", "hidden_states", "if", "in", "key", "nn", "num", "num_attr_labels", "num_object_labels", "output", "return", "self", "shap...
lxmert/modeling_lxmert.py:LxmertPreTrainingHeads
[ -0.0003449733485467732, 0.013582197017967701, 0.02519931085407734, 0.022887447848916054, -0.0015316094504669309, 0.047855570912361145, 0.02820473350584507, -0.008033725433051586, 0.0029187274631112814, 0.015027112327516079, 0.019419651478528976, 0.011096944101154804, -0.00037748392787761986,...
[ "Linear", "ModelLMPredictionHead", "ModelPreTrainingHeads", "Module", "__init__", "class", "config", "def", "forward", "hidden_size", "nn", "pooled_output", "prediction_scores", "predictions", "return", "self", "seq_relationship", "seq_relationship_score", "sequence_output", "s...
lxmert/modeling_lxmert.py:LxmertPreTrainedModel
[ -0.00023054261691868305, 0.0383622907102108, 0.009931067004799843, 0.01225777342915535, -0.001276851398870349, 0.028260977938771248, 0.012428020127117634, -0.025537028908729553, -0.0004610852338373661, 0.005476274061948061, 0.011860530823469162, -0.002284145215526223, -0.002553702797740698, ...
[ "Model", "ModelConfig", "ModelLMPredictionHead", "ModelPreTrainedModel", "PreTrainedModel", "_init_weights", "base_model_prefix", "bias", "class", "config", "def", "if", "image", "init", "input_modalities", "isinstance", "module", "no_grad", "self", "super", "text", "torch"...
lxmert/modeling_lxmert.py:LxmertModel
[ 0.000022717236788594164, 0.04205135256052017, 0.0056478544138371944, 0.023374289274215698, 0.00009698512440081686, 0.04652490094304085, 0.02494003064930439, -0.030420128256082535, 0.008891177363693714, 0.018788902088999748, 0.03265690058469772, 0.02214406430721283, 0.004138031974434853, -0...
[ "ModelEmbeddings", "ModelEncoder", "ModelModel", "ModelModelOutput", "ModelPooler", "ModelPreTrainedModel", "None", "__init__", "all_attentions", "and", "attention_mask", "auto_docstring", "class", "config", "cross_encoder_attentions", "def", "device", "dtype", "elif", "else", ...
lxmert/modeling_lxmert.py:LxmertForPreTraining
[ -0.00024346869031433016, 0.0377129465341568, 0.006153759080916643, 0.012533344328403473, -0.0008397905621677637, 0.052165813744068146, 0.019872691482305527, -0.0075651719234883785, -0.001397298532538116, 0.011968779377639294, 0.018517734482884407, 0.04990755394101143, 0.000522222719155252, ...
[ "CrossEntropyLoss", "FloatTensor", "Linear", "Model", "ModelForPreTraining", "ModelForPreTrainingOutput", "ModelModel", "ModelPreTrainedModel", "ModelPreTrainingHeads", "ModelVisualAnswerHead", "ModelVisualObjHead", "Model_output", "None", "Parameter", "SmoothL1Loss", "True", "__init...
lxmert/modeling_lxmert.py:LxmertForQuestionAnswering
[ -0.0002347014524275437, 0.04901554435491562, 0.0018776116194203496, 0.02394307777285576, -0.0008046907023526728, 0.03975454345345497, 0.03659224882721901, 0.014569136314094067, -0.0021317244973033667, 0.03297819942235947, 0.019651392474770546, 0.0230395644903183, 0.002625832799822092, 0.00...
[ "CrossEntropyLoss", "Linear", "Model", "ModelForQuestionAnswering", "ModelForQuestionAnsweringOutput", "ModelModel", "ModelPreTrainedModel", "ModelVisualAnswerHead", "Model_output", "None", "__init__", "_get_resized_qa_labels", "_init_weights", "_resize_qa_labels", "_set_qa_logit_layer",...
dpt/modeling_dpt.py:BaseModelOutputWithIntermediateActivations
[ -0.00026755224098451436, 0.020362894982099533, 0.02915596216917038, 0.021519877016544342, -0.0017427051207050681, 0.057617735117673874, 0.03725484013557434, -0.021519877016544342, 0.010759938508272171, -0.03494087606668472, 0.032626911997795105, 0.020941385999321938, -0.003962665796279907, ...
[ "ModelModelOutputWithIntermediateActivations", "ModelOutput", "None", "class", "intermediate_activations", "last_hidden_states", "r" ]
dpt/modeling_dpt.py:BaseModelOutputWithPoolingAndIntermediateActivations
[ -0.0002918439859058708, 0.011817879043519497, 0.026402872055768967, 0.024558130651712418, -0.0014628228964284062, 0.04796329513192177, 0.035511285066604614, -0.009223710745573044, 0.008243691176176071, -0.02813231758773327, 0.035050101578235626, 0.014700288884341717, -0.0033868313767015934, ...
[ "ModelModelOutputWithPoolingAndIntermediateActivations", "ModelOutput", "None", "attentions", "class", "hidden_states", "intermediate_activations", "last_hidden_state", "pooler_output", "r" ]
dpt/modeling_dpt.py:DPTViTHybridEmbeddings
[ -0.0002669061068445444, 0.02607850730419159, 0.02118166908621788, 0.01844855025410652, -0.0011886218562722206, 0.009736735373735428, 0.008882635273039341, -0.015829311683773994, -0.006832796614617109, 0.0013736768160015345, -0.002747353632003069, -0.002163718920201063, -0.0011459168745204806...
[ "BaseModelOutputWithIntermediateActivations", "Conv2d", "Expected", "False", "Input", "Iterable", "Make", "Model", "Module", "None", "Parameter", "ValueError", "__init__", "_resize_pos_embed", "abc", "backbone", "backbone_featmap_shape", "backbone_output", "batch_size", "biline...
dpt/modeling_dpt.py:DPTViTEmbeddings
[ -0.00029726314824074507, 0.033492833375930786, 0.018341314047574997, 0.014695835299789906, -0.0010751313529908657, 0.011221238411962986, 0.0382775254547596, -0.01868307776749134, -0.008031444624066353, 0.017771707847714424, 0.009455460123717785, -0.0076896813698112965, -0.0013172138715162873...
[ "BaseModelOutputWithIntermediateActivations", "Dropout", "Model", "ModelViTPatchEmbeddings", "Module", "Parameter", "_", "__init__", "_resize_pos_embed", "batch_size", "bilinear", "cat", "class", "cls_token", "cls_tokens", "config", "def", "dim", "dropout", "embeddings", "exp...
dpt/modeling_dpt.py:DPTViTPatchEmbeddings
[ -0.000143062075949274, 0.020911788567900658, 0.018877128139138222, 0.022155191749334335, -0.00009405006130691618, 0.008138641715049744, 0.018311945721507072, -0.017972834408283234, 0.008816862478852272, 0.010964559391140938, 0.017520688474178314, -0.01220796350389719, -0.001978142186999321, ...
[ "Conv2d", "Iterable", "Make", "Model", "Module", "ValueError", "__init__", "abc", "batch_size", "channel", "class", "collections", "config", "configuration", "def", "dimension", "else", "embeddings", "flatten", "forward", "height", "hidden_size", "if", "image_size", "...
dpt/modeling_dpt.py:eager_attention_forward
[ 0.000042374336771899834, 0.02700657583773136, 0.026102591305971146, -0.010847830213606358, 0.0003637130430433899, 0.03683742135763168, 0.06101904436945915, -0.026328587904572487, 0.02000068686902523, 0.012542803771793842, 0.02395562455058098, 0.026893578469753265, 0.002500085858628154, -0....
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "size", "softmax", "torch", "tra...
dpt/modeling_dpt.py:DPTSelfAttention
[ -0.00024049222702160478, 0.027020638808608055, 0.040815986692905426, 0.001895435038022697, -0.0009334661299362779, 0.02405634894967079, 0.04150005429983139, -0.017215680330991745, -0.0004596073704306036, 0.022688215598464012, 0.013339302502572536, -0.0045319427736103535, -0.00200944626703858...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "Model", "Module", "None", "The", "ValueError", "__init__", "_attn_implementation", "a", "all_head_size", "and", "attention", "attention_head_size", "attention_interface", "attention_probs", "attention_probs_dropout_prob", "batch_size...
dpt/modeling_dpt.py:DPTViTSelfOutput
[ -0.00016798802244011313, 0.020370759069919586, 0.061564959585666656, -0.008770743384957314, -0.0005729275871999562, 0.025802960619330406, 0.03259321302175522, -0.02851906232535839, 0.0024614667054265738, 0.013637091033160686, 0.02172880806028843, -0.022181492298841476, 0.0021643931977450848,...
[ "Dropout", "Linear", "Model", "Module", "__init__", "class", "config", "def", "dense", "dropout", "forward", "hidden_dropout_prob", "hidden_size", "hidden_states", "input_tensor", "nn", "return", "self", "super" ]
dpt/modeling_dpt.py:DPTViTAttention
[ 0.000031113224395085126, 0.05654017999768257, 0.03185994178056717, 0.013405856676399708, 0.00008895753126125783, 0.03275740519165993, 0.05205286666750908, -0.025241151452064514, 0.010208643972873688, 0.010601283982396126, 0.022436579689383507, 0.016042154282331467, 0.0033374412450939417, -...
[ "Model", "ModelSelfAttention", "ModelViTSelfOutput", "Module", "_", "__init__", "attention", "class", "config", "def", "forward", "hidden_states", "nn", "output", "return", "self", "self_attn_output", "super" ]
dpt/modeling_dpt.py:DPTViTIntermediate
[ -0.00024672862491570413, 0.024029221385717392, 0.0379890538752079, 0.019795501604676247, -0.0009654597961343825, 0.03089471347630024, 0.03570055589079857, -0.01762142963707447, -0.0008653380209580064, -0.0033612302504479885, 0.027004268020391464, -0.03089471347630024, -0.00003039410512428730...
[ "ACT2FN", "Linear", "Model", "Module", "__init__", "class", "config", "def", "dense", "else", "forward", "hidden_act", "hidden_size", "hidden_states", "if", "intermediate_act_fn", "intermediate_size", "isinstance", "nn", "return", "self", "str", "super" ]
dpt/modeling_dpt.py:DPTViTOutput
[ -0.00025999039644375443, 0.028918102383613586, 0.060131292790174484, 0.005995916202664375, -0.0011618880089372396, 0.0390164889395237, 0.03190171718597412, -0.028114821761846542, -0.002036890247836709, 0.0031987782567739487, 0.014918068423867226, -0.01950824446976185, 0.0008893464109860361, ...
[ "Dropout", "Linear", "Model", "Module", "__init__", "class", "config", "def", "dense", "dropout", "forward", "hidden_dropout_prob", "hidden_size", "hidden_states", "input_tensor", "intermediate_size", "nn", "return", "self", "super" ]
dpt/modeling_dpt.py:DPTViTLayer
[ -0.00008778493065619841, 0.046063944697380066, 0.030187536031007767, 0.025268085300922394, -0.00018692512821871787, 0.030858369544148445, 0.03220003843307495, -0.0014325103256851435, 0.0002201174502260983, 0.021354885771870613, 0.017329880967736244, 0.01475835032761097, 0.0021382838021963835...
[ "GradientCheckpointingLayer", "LayerNorm", "Model", "ModelViTAttention", "ModelViTIntermediate", "ModelViTOutput", "__init__", "attention", "attention_output", "chunk_size_feed_forward", "class", "config", "def", "eps", "forward", "hidden_size", "hidden_states", "hidden_states_norm...
dpt/modeling_dpt.py:DPTViTEncoder
[ -0.00014313950669020414, 0.03079870343208313, 0.009273332543671131, 0.03686852008104324, -0.00037585097015835345, 0.026976965367794037, 0.030124278739094734, -0.01719781570136547, -0.0011029645102098584, 0.014612522907555103, 0.029000239446759224, -0.034845247864723206, -0.001770363422110676...
[ "BaseModelOutput", "False", "Model", "ModelViTLayer", "Module", "ModuleList", "None", "_", "__init__", "all_hidden_states", "class", "config", "def", "else", "enumerate", "for", "forward", "gradient_checkpointing", "hidden_states", "i", "if", "in", "last_hidden_state", ...
dpt/modeling_dpt.py:DPTReassembleStage
[ -0.0003655755426734686, 0.0076445103622972965, -0.011930068954825401, 0.034516122192144394, -0.0011437808861956, 0.029303956776857376, -0.010076854377985, -0.04864688590168953, -0.005125296767801046, -0.009903116151690483, -0.004401384852826595, -0.01969040557742119, -0.0004108200373593718, ...
[ "ACT2FN", "Hybrid", "Identity", "Linear", "Model", "ModelReassembleLayer", "Module", "ModuleList", "None", "Readout", "Sequential", "ValueError", "_", "__init__", "_get_backbone_hidden_size", "_init_reassemble_Model", "_init_reassemble_Model_hybrid", "add", "and", "append", "...
dpt/modeling_dpt.py:_get_backbone_hidden_size
[ -0.0002768740523606539, 0.0040870183147490025, 0.05601787567138672, 0.006287720520049334, -0.0012289636069908738, -0.005373143125325441, 0.02435062639415264, -0.02892351523041725, 0.004658629186451435, 0.02686571516096592, 0.003415375482290983, -0.006773590110242367, -0.0026151202619075775, ...
[ "None", "_get_backbone_hidden_size", "and", "backbone_config", "config", "def", "else", "hasattr", "hidden_size", "if", "is", "not", "return" ]
dpt/modeling_dpt.py:DPTReassembleLayer
[ -0.0002784844837151468, 0.013565919362008572, 0.05108505114912987, 0.005676116794347763, -0.000666943727992475, 0.026564227417111397, 0.025315482169389725, -0.04154917597770691, 0.0026819652412086725, 0.010160249657928944, -0.005250408314168453, -0.02440730296075344, -0.0008301321067847311, ...
[ "Conv2d", "ConvTranspose2d", "Identity", "Model", "Module", "__init__", "_get_backbone_hidden_size", "channels", "class", "config", "def", "elif", "factor", "forward", "hidden_size", "hidden_state", "if", "in_channels", "int", "kernel_size", "nn", "out_channels", "padding...
dpt/modeling_dpt.py:DPTFeatureFusionStage
[ -0.00014170487702358514, 0.016223007813096046, 0.020166100934147835, 0.03402325510978699, -0.00020155539095867425, 0.02129269763827324, 0.02174333669245243, -0.05700584873557091, 0.007829854264855385, 0.0025066800881177187, 0.009632411412894726, -0.02129269763827324, 0.003055896610021591, ...
[ "Model", "ModelFeatureFusionLayer", "Module", "ModuleList", "None", "_", "__init__", "append", "class", "config", "def", "else", "for", "forward", "fused_hidden_state", "fused_hidden_states", "hidden_state", "hidden_states", "if", "in", "is", "layer", "layers", "len", ...
dpt/modeling_dpt.py:DPTPreActResidualLayer
[ -0.00008072388300206512, 0.029294785112142563, 0.017486033961176872, -0.01498802937567234, 0.00029450911097228527, 0.02384459227323532, -0.0025121981743723154, -0.040649354457855225, 0.015896394848823547, -0.013341616839170456, 0.034744977951049805, -0.02089240401983261, 0.004087644629180431...
[ "BatchNorm2d", "Conv2d", "Model", "Module", "None", "ReLU", "__init__", "activation1", "activation2", "batch_norm1", "batch_norm2", "class", "config", "convolution1", "convolution2", "def", "else", "forward", "fusion_hidden_size", "hidden_state", "if", "is", "kernel_size"...
dpt/modeling_dpt.py:DPTFeatureFusionLayer
[ -0.00023980655532795936, 0.027625715360045433, 0.03547005355358124, -0.0018900309223681688, -0.000682116427924484, 0.04797552153468132, 0.02182772569358349, -0.0536598265171051, 0.003893747925758362, -0.01261915359646082, 0.019667690619826317, -0.017280282452702522, -0.00043342815479263663, ...
[ "Conv2d", "False", "Model", "ModelPreActResidualLayer", "Module", "None", "True", "__init__", "align_corners", "bilinear", "class", "config", "def", "forward", "functional", "fusion_hidden_size", "hidden_state", "if", "interpolate", "is", "kernel_size", "mode", "nn", "n...
dpt/modeling_dpt.py:DPTPreTrainedModel
[ -0.00017762993229553103, 0.03962026908993721, -0.004811836406588554, 0.01677107997238636, -0.0010130181908607483, 0.016996193677186966, 0.02431243658065796, -0.014013418927788734, -0.002377778757363558, 0.005036951508373022, 0.007653915323317051, 0.011537151411175728, -0.001435109181329608, ...
[ "Model", "ModelConfig", "ModelSelfAttention", "ModelViTEmbeddings", "ModelViTHybridEmbeddings", "PreTrainedModel", "True", "_can_record_outputs", "_init_weights", "_supports_attention_backend", "_supports_flash_attn", "_supports_flex_attn", "_supports_sdpa", "attentions", "base_model_pre...
dpt/modeling_dpt.py:DPTModel
[ -0.00008786725084064528, 0.02999202162027359, -0.005539570935070515, 0.04028778895735741, -0.00009355020301882178, 0.016003204509615898, 0.03581136837601662, -0.012198247015476227, 0, 0.023277388885617256, 0.024508405476808548, -0.0050079962238669395, -0.0018884901655837893, -0.01141487341...
[ "BaseModelOutputWithPoolingAndIntermediateActivations", "False", "LayerNorm", "Model", "ModelPreTrainedModel", "ModelViTEmbeddings", "ModelViTEncoder", "ModelViTHybridEmbeddings", "ModelViTPooler", "None", "True", "__init__", "add_pooling_layer", "auto_docstring", "capture_outputs", "c...
dpt/modeling_dpt.py:DPTViTPooler
[ -0.00023585045710206032, 0.018982388079166412, 0.03773607313632965, 0.028244879096746445, -0.0010077246697619557, 0.03430552035570145, 0.038650885224342346, -0.021726829931139946, 0.001729569979943335, -0.007775918114930391, 0.01692405715584755, -0.01440831832587719, -0.0003037467831745744, ...
[ "ACT2FN", "Linear", "Model", "Module", "__init__", "activation", "class", "config", "def", "dense", "first_token_tensor", "forward", "hidden_size", "hidden_states", "nn", "pooled_output", "pooler_act", "pooler_output_size", "return", "self", "super" ]
dpt/modeling_dpt.py:DPTNeck
[ -0.00008347707625944167, 0.014545115642249584, -0.009734039194881916, 0.016670940443873405, 0, 0.005202676169574261, 0.005230647511780262, -0.04788699746131897, 0.0038320785388350487, 0.0027132234536111355, 0.020027505233883858, -0.02360784262418747, 0.001818139455281198, -0.00068180233938...
[ "Model", "ModelFeatureFusionStage", "ModelReassembleStage", "Module", "ModuleList", "None", "The", "TypeError", "ValueError", "__init__", "a", "and", "backbone_config", "be", "channel", "class", "config", "convs", "def", "else", "enumerate", "equal", "feature", "feature...
dpt/modeling_dpt.py:DPTDepthEstimationHead
[ -0.00023595368838869035, 0.018734367564320564, 0.029974989593029022, 0.012943744659423828, -0.0005038409726694226, 0.031110404059290886, 0.04359998181462288, -0.03951248526573181, 0.0009509110823273659, -0.00016765130567364395, 0.019642701372504234, -0.0191885344684124, -0.000567708106245845...
[ "Conv2d", "Model", "Module", "None", "ReLU", "Sequential", "True", "Upsample", "__init__", "add_projection", "align_corners", "bilinear", "class", "config", "def", "dim", "features", "forward", "fusion_hidden_size", "head", "head_in_index", "hidden_states", "if", "is", ...
dpt/modeling_dpt.py:DPTForDepthEstimation
[ -0.00017126543389167637, -0.0027826218865811825, 0.023729974403977394, 0.01932297833263874, -0.0005473431665450335, 0.03706395998597145, 0.0517539419233799, -0.014689983800053596, 0.004293995443731546, 0.006808242294937372, 0.0015537482686340809, 0.006977742072194815, 0.00018009354243986309,...
[ "DepthEstimatorOutput", "False", "Model", "ModelDepthEstimationHead", "ModelModel", "ModelNeck", "ModelPreTrainedModel", "None", "True", "_", "__init__", "add_pooling_layer", "and", "attentions", "auto_docstring", "backbone", "backbone_config", "backbone_hidden_states", "backbone...
dpt/modeling_dpt.py:DPTSemanticSegmentationHead
[ -0.0003572010318748653, 0.026786470785737038, 0.020782606676220894, 0.005426569376140833, -0.0010824274504557252, 0.035099513828754425, 0.03417584300041199, -0.027941059321165085, 0.003001932054758072, -0.002900905441492796, 0.03671593964099884, -0.01454782485961914, 0.0004203426360618323, ...
[ "BatchNorm2d", "Conv2d", "Dropout", "Model", "Module", "ReLU", "Sequential", "True", "Upsample", "__init__", "align_corners", "bilinear", "class", "config", "def", "features", "forward", "fusion_hidden_size", "head", "head_in_index", "hidden_states", "kernel_size", "logit...
dpt/modeling_dpt.py:DPTAuxiliaryHead
[ -0.00018998893210664392, 0.023864030838012695, 0.024091307073831558, -0.001484399545006454, -0.0006711758906021714, 0.03295509144663811, 0.03636423870921135, -0.02431858330965042, 0.01215929165482521, -0.0012287135468795896, 0.03477330133318901, -0.017954843118786812, 0.0021307170391082764, ...
[ "BatchNorm2d", "Conv2d", "Dropout", "False", "Model", "Module", "ReLU", "Sequential", "__init__", "class", "config", "def", "features", "forward", "fusion_hidden_size", "head", "hidden_states", "kernel_size", "logits", "nn", "num_labels", "padding", "return", "self", ...
dpt/modeling_dpt.py:DPTForSemanticSegmentation
[ -0.0002723265206441283, 0.031236205250024796, -0.01652350090444088, 0.006960241589695215, -0.0007285618339665234, 0.03395239636301994, 0.027614615857601166, -0.007412940263748169, 0.000682584650348872, 0.00950667168945074, 0.020031914114952087, 0.015278578735888004, -0.003041568910703063, ...
[ "CrossEntropyLoss", "False", "Model", "ModelAuxiliaryHead", "ModelModel", "ModelNeck", "ModelPreTrainedModel", "ModelSemanticSegmentationHead", "None", "SemanticSegmenterOutput", "The", "True", "ValueError", "__init__", "add_pooling_layer", "align_corners", "and", "attentions", "...
rwkv/modeling_rwkv.py:load_wkv_cuda_kernel
[ -0.00017561235290486366, 0.02433226816356182, -0.005851331166923046, -0.02989392913877964, -0.0008472843328490853, 0.03406517580151558, 0.03591906279325485, -0.03151608258485794, 0.017727795988321304, 0.026533760130405426, 0.023868797346949577, -0.01830713450908661, 0.0014845581026747823, ...
[ "ImportError", "Model", "Model_cuda_kernel", "Model_wkv_cuda_kernel", "community", "context_length", "def", "get_kernel", "global", "if", "install", "installed", "is", "is_kernels_available", "it", "kernels", "max_seq_length", "not", "pip", "please", "raise", "with" ]
rwkv/modeling_rwkv.py:RwkvLinearAttention
[ -0.00036578462459146976, 0.008327514864504337, 0.020149091258645058, -0.041462868452072144, -0.001666958793066442, 0.02620546519756317, 0.020964372903108597, -0.029699528589844704, 0.005881671328097582, 0.03657118231058121, 0.018052654340863228, -0.003144655842334032, 0.0024312848690897226, ...
[ "CUDA", "Calling", "Cannot", "False", "Function", "ModelLinearAttention", "Model_cuda_kernel", "None", "The", "ValueError", "a", "all", "and", "at", "attention", "autograd", "backward", "backward_bf16", "backward_func", "batch", "batch_size", "be", "bfloat16", "cat", ...
rwkv/modeling_rwkv.py:rwkv_linear_attention_cpu
[ -0.00021797692170366645, 0.005812717601656914, 0.042191825807094574, -0.0016162190586328506, -0.0009782378328964114, 0.03039626032114029, 0.015878643840551376, -0.045594390481710434, 0.00918692909181118, 0.030849935486912727, 0.03606720641255379, -0.008392997086048126, 0.0042532081715762615,...
[ "False", "Model_linear_attention_cpu", "None", "_", "current_index", "current_key", "current_value", "def", "den_state", "denominator", "dtype", "e1", "e2", "else", "exp", "float", "float32", "for", "if", "in", "is", "key", "max_for_output", "max_for_state", "max_stat...
rwkv/modeling_rwkv.py:rwkv_linear_attention
[ -0.00020711187971755862, 0.008723481558263302, 0.04282436519861221, -0.02175205759704113, -0.0007293170783668756, 0.029455911368131638, 0.015407707542181015, -0.05573964864015579, 0.009516525082290173, 0.03398758918046951, 0.027643240988254547, 0.0014444725820794702, 0.004191802814602852, ...
[ "False", "Model_cuda_kernel", "Model_linear_attention", "None", "any", "cuda", "def", "device", "else", "for", "if", "in", "is", "key", "no_cuda", "one_token", "or", "return_state", "size", "state", "t", "time_decay", "time_first", "type", "value" ]
rwkv/modeling_rwkv.py:RwkvSelfAttention
[ -0.0001313970424234867, 0.03183159604668617, 0.017157455906271935, -0.028896767646074295, -0.0005114784580655396, 0.029799792915582657, 0.015577164478600025, -0.040636081248521805, 0.006236509419977665, 0.03499218076467514, 0.020318040624260902, -0.003315791254863143, 0.0024409866891801357, ...
[ "Exception", "False", "Linear", "Model", "ModelSelfAttention", "Model_cuda_kernel", "Model_linear_attention", "Module", "None", "Parameter", "ZeroPad2d", "__init__", "and", "attention_hidden_size", "class", "config", "context_length", "def", "else", "empty", "except", "extr...
rwkv/modeling_rwkv.py:RwkvFeedForward
[ -0.0002633359981700778, 0.03554140403866768, 0.02109554223716259, -0.018458599224686623, -0.0009243630920536816, 0.03622930124402046, 0.029006371274590492, -0.04861146956682205, 0.0028949046973139048, 0.02625477872788906, 0.037834398448467255, -0.02166879177093506, 0.0030095542315393686, -...
[ "Linear", "ModelFeedForward", "Module", "None", "Parameter", "ZeroPad2d", "__init__", "and", "class", "config", "def", "else", "empty", "forward", "hidden", "hidden_size", "if", "intermediate_size", "is", "key", "layer_id", "nn", "not", "receptance", "relu", "return...
rwkv/modeling_rwkv.py:RwkvBlock
[ 0, -0.00932429451495409, 0.018874632194638252, -0.005735853686928749, 0, 0.052216049283742905, 0.02283039316534996, -0.023847589269280434, 0.010624044574797153, -0.0007487690891139209, 0.00751594640314579, 0.002557117026299238, 0.0025995003525167704, 0.008363609202206135, -0.019778806716...
[ "False", "GradientCheckpointingLayer", "LayerNorm", "ModelBlock", "ModelFeedForward", "ModelSelfAttention", "None", "__init__", "attention", "class", "config", "def", "else", "eps", "feed_forward", "forward", "hidden", "hidden_size", "if", "layer_id", "layer_norm_epsilon", ...
rwkv/modeling_rwkv.py:RwkvPreTrainedModel
[ -0.00022670072212349623, 0.035818714648485184, 0.0036838867235928774, -0.000591547170188278, -0.0009138872846961021, 0.03377840667963028, 0.02969779446721077, -0.009351404383778572, -0.0004817390290554613, 0.00844460166990757, 0.011448386125266552, -0.017796006053686142, -0.00111225037835538...
[ "Embedding", "LayerNorm", "Linear", "Model", "ModelBlock", "ModelConfig", "ModelFeedForward", "ModelPreTrainedModel", "ModelSelfAttention", "None", "PreTrainedModel", "True", "_init_weights", "_is_stateful", "_keep_in_fp32_modules", "_no_split_modules", "and", "attention_hidden_siz...
rwkv/modeling_rwkv.py:RwkvOutput
[ -0.0001520880323369056, 0.024931762367486954, 0.0217441413551569, 0.012238194234669209, -0.001024593017064035, 0.04212215542793274, 0.04940815269947052, -0.01775961183011532, 0.02071954682469368, -0.0029314744751900434, 0.02402101457118988, 0.033242352306842804, -0.0011668975930660963, -0....
[ "ModelOutput", "None", "attentions", "class", "hidden_states", "last_hidden_state", "r", "state" ]
rwkv/modeling_rwkv.py:RwkvCausalLMOutput
[ -0.00016286378377117217, 0.03126984462141991, 0.034895334392786026, -0.006401254795491695, -0.0009559395839460194, 0.045771803706884384, 0.03965378925204277, -0.0019968515262007713, 0.021073156967759132, -0.00844059232622385, 0.025265129283070564, 0.01387882698327303, -0.0015224223025143147,...
[ "ModelCausalLMOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "logits", "loss", "r", "state" ]
rwkv/modeling_rwkv.py:RwkvModel
[ -0.000011896131582034286, 0.026384195312857628, 0.004687404725700617, -0.006118887569755316, 0, 0.036151960492134094, 0.018412798643112183, -0.007017072755843401, 0.007634575478732586, 0.004406721796840429, 0.009094126522541046, 0.0016630463069304824, 0.0016840974567458034, 0.0249246433377...
[ "Embedding", "False", "ImportError", "LayerNorm", "ModelBlock", "ModelModel", "ModelOutput", "ModelPreTrainedModel", "ModuleList", "None", "Params4bit", "Please", "SCB", "Setting", "True", "__init__", "_bnb_4bit_dequantize_and_rescale", "_rescale_layers", "all_hidden_states", "...
rwkv/modeling_rwkv.py:RwkvForCausalLM
[ -0.00023808353580534458, 0.04605064541101456, 0.01862342283129692, 0.0020598634146153927, -0.0009241168154403567, 0.026411399245262146, 0.03205486014485359, -0.017156122252345085, -0.001142800902016461, 0.03318355232477188, 0.029797477647662163, -0.004063292406499386, 0.0001604859426151961, ...
[ "GenerationMixin", "Linear", "Model", "ModelCausalLMOutput", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "Model_outputs", "None", "__init__", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "class", "config", "def", "else", "embeddings", "f...
encodec/modeling_encodec.py:EncodecOutput
[ -0.00009126806980930269, 0.026099996641278267, -0.010656548663973808, 0.01629825122654438, -0.0003900040464941412, 0.0335082933306694, 0.07248733192682266, -0.04558951407670975, 0.018121832981705666, -0.013391920365393162, 0.020173359662294388, 0.024390390142798424, -0.004900872707366943, ...
[ "ModelOutput", "None", "audio_codes", "audio_values", "class", "r" ]
encodec/modeling_encodec.py:EncodecEncoderOutput
[ -0.0001739181170705706, 0.015868468210101128, 0.006935319397598505, 0.01895083300769329, -0.0008383745444007218, 0.03721669316291809, 0.0648438110947609, -0.06895363330841064, 0.01849418506026268, -0.012158215045928955, 0.00850504171103239, 0.04155483841896057, -0.003995657432824373, 0, ...
[ "ModelEncoderOutput", "ModelOutput", "None", "audio_codes", "audio_scales", "class", "last_frame_pad_length", "r" ]
encodec/modeling_encodec.py:EncodecDecoderOutput
[ -0.0001343319599982351, 0.04190445318818092, 0.0007116924971342087, 0.008198697119951248, -0.0008611478842794895, 0.028012216091156006, 0.07834310829639435, -0.06149023026227951, 0.015030945651233196, -0.02004126086831093, 0.011159338057041168, 0.0202690027654171, -0.004811041057109833, -0...
[ "ModelDecoderOutput", "ModelOutput", "None", "audio_values", "class", "r" ]
encodec/modeling_encodec.py:EncodecConv1d
[ -0.00016131515440065414, 0.023581983521580696, -0.0035260142758488655, -0.008744515478610992, -0.00019657528901007026, 0.020761171355843544, 0.0010578042129054666, -0.053256917744874954, 0.010154920630156994, 0.0212125014513731, 0.03881436586380005, -0.007503358181566, 0.00406196853145957, ...
[ "Conv1d", "False", "GroupNorm", "ModelConv1d", "Module", "ValueError", "__init__", "_get_extra_padding_for_conv1d", "_pad1d", "and", "be", "been", "causal", "ceil", "class", "config", "conv", "def", "dilation", "dtype", "elif", "else", "end", "extra_pad", "extra_paddi...
encodec/modeling_encodec.py:EncodecConvTranspose1d
[ -0.0003673782921396196, 0.04932954162359238, 0.019939277321100235, -0.005272958893328905, -0.0013254432706162333, 0.02040030062198639, 0.005762796849012375, -0.06684844195842743, -0.004235655535012484, 0.015559551306068897, 0.022244395688176155, -0.005042447242885828, -0.002031385898590088, ...
[ "ConvTranspose1d", "GroupNorm", "ModelConvTranspose1d", "Module", "ValueError", "__init__", "be", "causal", "ceil", "class", "config", "conv", "convolutions", "def", "elif", "else", "end", "f", "for", "forward", "got", "hasattr", "hidden_states", "if", "in", "in_cha...
encodec/modeling_encodec.py:EncodecLSTM
[ -0.00002087353823299054, 0.008999790996313095, 0.06524848192930222, 0.013105944730341434, 0, 0.05107381194829941, -0.005877988412976265, -0.05669867992401123, 0.01709960214793682, 0.013162193819880486, 0.055123716592788696, -0.030149297788739204, 0.00424677599221468, -0.014399665407836437,...
[ "LSTM", "ModelLSTM", "Module", "__init__", "class", "config", "def", "dimension", "forward", "hidden_states", "lstm", "nn", "num_lstm_layers", "permute", "return", "self", "super" ]
encodec/modeling_encodec.py:EncodecResnetBlock
[ -0.00004540539885056205, -0.01128357369452715, 0.02075270377099514, -0.016443399712443352, 0.00048550553037784994, 0.051031239330768585, -0.005103123839944601, -0.051938463002443314, 0.009582532569766045, -0.01031965110450983, 0.008051595650613308, -0.02993832714855671, 0.003798992373049259,...
[ "ELU", "Identity", "ModelConv1d", "ModelResnetBlock", "Module", "ModuleList", "Number", "ValueError", "__init__", "block", "class", "compress", "config", "def", "dilation", "dilations", "dim", "else", "enumerate", "for", "forward", "hidden", "hidden_states", "i", "if"...
encodec/modeling_encodec.py:EncodecEncoder
[ -0.00007812786498107016, 0.013503319583833218, 0.020113730803132057, 0.025198662653565407, -0.0002860274398699403, 0.05378727987408638, 0.0114128477871418, -0.07457900047302246, 0.011921340599656105, -0.011073851957917213, 0.028475619852542877, -0.016836775466799736, 0.0012076713610440493, ...
[ "ELU", "ModelConv1d", "ModelEncoder", "ModelLSTM", "ModelResnetBlock", "Module", "ModuleList", "__init__", "audio_channels", "class", "config", "current_scale", "def", "dilation_growth_rate", "for", "forward", "hidden_size", "hidden_states", "in", "j", "kernel_size", "last_...
encodec/modeling_encodec.py:EncodecDecoder
[ -0.0001914795720949769, 0.042241983115673065, 0.01214174646884203, 0.017958490177989006, -0.0006741493125446141, 0.03749823197722435, 0.02880135178565979, -0.08267682045698166, 0.004997881595045328, -0.030495548620820045, 0.008583932183682919, -0.016377238556742668, -0.002287166193127632, ...
[ "ELU", "ModelConv1d", "ModelConvTranspose1d", "ModelDecoder", "ModelLSTM", "ModelResnetBlock", "Module", "ModuleList", "__init__", "audio_channels", "class", "config", "current_scale", "def", "dilation_growth_rate", "for", "forward", "hidden_size", "hidden_states", "in", "int...
encodec/modeling_encodec.py:EncodecEuclideanCodebook
[ -0.0001791936083463952, 0.005734195467084646, 0.0196438767015934, 0.006869679782539606, -0.0007664518780075014, 0.009083873592317104, 0.045419368892908096, -0.0374709777534008, 0.007380647584795952, -0.0005145163158886135, 0.036562591791152954, 0.0026541943661868572, -0.0016535489121451974, ...
[ "ModelEuclideanCodebook", "Module", "Tensor", "True", "__init__", "class", "clone", "cluster_size", "codebook_dim", "codebook_size", "config", "decode", "def", "dim", "dist", "embed", "embed_avg", "embed_ind", "embedding", "encode", "functional", "hidden_states", "indices...
encodec/modeling_encodec.py:EncodecVectorQuantization
[ -0.00013040936028119177, 0.01680518500506878, 0.01714354380965233, 0.006569812074303627, -0.00037360520218499005, 0.012632085010409355, 0.046242453157901764, -0.045565735548734665, 0.004821621812880039, 0.01827140897512436, 0.042633287608623505, 0.005611126776784658, -0.0032144144643098116, ...
[ "ModelEuclideanCodebook", "ModelVectorQuantization", "Module", "__init__", "class", "codebook", "config", "decode", "def", "embed_in", "embed_ind", "encode", "hidden_states", "nn", "permute", "quantize", "return", "self", "super" ]
encodec/modeling_encodec.py:EncodecResidualVectorQuantizer
[ -0.00013507652329280972, 0.008305882103741169, -0.013334613293409348, -0.0025567426346242428, -0.00028781010769307613, 0.040907882153987885, 0.03367554768919945, -0.04407202824950218, 0.006780311930924654, -0.008588395081460476, 0.03209347650408745, 0.007966866716742516, -0.00020482191757764...
[ "ModelResidualVectorQuantizer", "ModelVectorQuantization", "Module", "ModuleList", "None", "_", "__init__", "all_indices", "and", "append", "bandwidth", "bw_per_q", "class", "codebook_size", "codes", "config", "decode", "def", "device", "embeddings", "encode", "enumerate", ...
encodec/modeling_encodec.py:EncodecPreTrainedModel
[ -0.00008015034109121189, 0.03632781282067299, -0.011997148394584656, -0.007904662750661373, -0.0002943225554190576, 0.02253669872879982, 0.021079102531075478, -0.03744904324412346, 0.009418321773409843, -0.018612397834658623, 0.041036974638700485, 0.00818497035652399, -0.0019341197330504656,...
[ "Conv1d", "ConvTranspose1d", "GroupNorm", "LSTM", "Model", "ModelConfig", "ModelConv1d", "ModelEuclideanCodebook", "ModelPreTrainedModel", "None", "PreTrainedAudioTokenizerBase", "Tensor", "True", "_init_weights", "a", "b", "base_model_prefix", "bias", "class", "cluster_size", ...
encodec/modeling_encodec.py:EncodecModel
[ 0.000023728960513835773, 0.02216109074652195, -0.01664894074201584, -0.00548402639105916, 0.00020301443873904645, 0.02216109074652195, 0.03329788148403168, -0.03959748148918152, 0.010011862963438034, -0.018673812970519066, 0.0339728407561779, 0.021823612973093987, -0.002292041666805744, -0...
[ "Expected", "False", "ModelDecoder", "ModelDecoderOutput", "ModelEncoder", "ModelEncoderOutput", "ModelModel", "ModelOutput", "ModelPreTrainedModel", "ModelResidualVectorQuantizer", "None", "Number", "Select", "This", "True", "ValueError", "_", "__init__", "_decode_frame", "_en...
falcon/modeling_falcon.py:FalconLinear
[ -0.0000680610173731111, 0.008037347346544266, 0.02551717311143875, 0.022482091560959816, -0.0005796164041385055, 0.02012147195637226, 0.02034629136323929, -0.024505479261279106, 0.014388537965714931, 0.027990203350782394, 0.011522071436047554, -0.005170880816876888, 0.001580771990120411, -...
[ "Linear", "ModelLinear", "None", "T", "bias", "class", "def", "forward", "hidden_states", "if", "input", "is", "nn", "return", "self", "weight" ]
falcon/modeling_falcon.py:rotate_half
[ 0.00002049485374300275, 0.013860220089554787, 0.03434192016720772, 0.002974055241793394, 0.0003507140791043639, 0.028506038710474968, 0.01975221559405327, -0.02008890174329281, 0.014477476477622986, 0.018742159008979797, -0.001487027620896697, -0.01217679213732481, 0.00022445700597018003, ...
[ "Model_half", "cat", "def", "dim", "return", "shape", "torch", "x", "x1", "x2" ]
falcon/modeling_falcon.py:apply_rotary_pos_emb
[ -0.0001444592053303495, 0.027112245559692383, 0.028019767254590988, 0.0028360087890177965, -0.0005707467789761722, 0.021667107939720154, 0.046510547399520874, -0.002183726755902171, 0.01293220091611147, 0.03652779385447502, 0.007884104736149311, 0.0017654155381023884, -0.0007834474672563374,...
[ "Model_rotary_pos_emb", "cos", "def", "k", "k_embed", "q", "q_embed", "return", "rotate_half", "sin", "unsqueeze", "unsqueeze_dim" ]
falcon/modeling_falcon.py:FalconRotaryEmbedding
[ -0.00029968636226840317, 0.04852752387523651, 0.0020364229567348957, -0.005228263325989246, -0.0013792794197797775, 0.04136393964290619, 0.040670689195394516, 0.0062681385315954685, -0.0026430170983076096, 0.021144136786460876, 0.0021808501332998276, -0.004043960478156805, -0.001364836702123...
[ "False", "ModelRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "arange", "attention_factor", "attention_scaling", "base", "cat", "class", "clone", "compute_default_rope_parameters", "config", "cos", "cpu", "def", "default", "device", ...
falcon/modeling_falcon.py:build_alibi_tensor
[ -0.0002907592570409179, 0.01952189765870571, 0.06804123520851135, -0.015526187606155872, -0.0011915777577087283, 0.038358815014362335, 0.030824048444628716, -0.02968241646885872, 0.0030253231525421143, -0.011644640006124973, 0.028883274644613266, -0.007363522425293922, 0.0008312503923662007,...
[ "Model_alibi_tensor", "None", "alibi", "arange", "arange_tensor", "attention_mask", "base", "batch_size", "bfloat16", "cat", "closest_power_of_2", "cumsum", "def", "device", "dim", "dtype", "extra_base", "extra_powers", "float32", "floor", "if", "int32", "log2", "math",...
falcon/modeling_falcon.py:dropout_add
[ -0.00008448258449789137, 0.01585264876484871, 0.012002719566226006, 0.009398356080055237, -0.000360930833267048, 0.07246924936771393, 0.01743791252374649, -0.012399035505950451, 0.001273873494938016, -0.024684837087988853, 0.017098212614655495, -0.0023071265313774347, -0.0006864762981422246,...
[ "F", "Model", "Model_add", "def", "out", "p", "prob", "residual", "return", "training", "x" ]
falcon/modeling_falcon.py:FalconAttention
[ -0.00030962948221713305, 0.03802894428372383, 0.023710817098617554, -0.01557812187820673, -0.001267154235392809, 0.01889992691576481, 0.03001079335808754, -0.03505077213048935, -0.00979359820485115, 0.012370861135423183, 0.010194506496191025, 0.01844174787402153, -0.0031070334371179342, -0...
[ "Dropout", "F", "False", "Instantiating", "ModelAttention", "ModelLinear", "Module", "None", "Please", "Tensor", "True", "ValueError", "_", "__class__", "__init__", "__name__", "_merge_heads", "_split_heads", "a", "alibi", "and", "apply_rotary_pos_emb", "attention_dropout...
falcon/modeling_falcon.py:FalconFlashAttention2
[ -0.00009292854520026594, 0.01941329799592495, 0.05161917582154274, -0.01189485378563404, -0.00029807270038872957, 0.027268391102552414, 0.03613342344760895, -0.058352112770080566, 0.008640601299703121, 0.008360061794519424, 0.020759886130690575, 0.018066711723804474, 0.0024126353673636913, ...
[ "False", "ModelAttention", "ModelFlashAttention2", "None", "Tensor", "The", "We", "_", "__init__", "_flash_attention_forward", "_flash_attn_uses_top_left_mask", "_is_quantized", "_split_heads", "alibi", "apply_rotary_pos_emb", "args", "attention_dropout", "attention_mask", "attn_...
falcon/modeling_falcon.py:FalconMLP
[ -0.0002459005918353796, 0.0269952192902565, 0.03836767375469208, 0.021251555532217026, -0.0008184720645658672, 0.042732857167720795, 0.04296260327100754, -0.02929268404841423, 0.000642572354990989, -0.015163271687924862, 0.03078603744506836, -0.025616740807890892, 0.0007789843948557973, 0....
[ "ModelLinear", "ModelMLP", "Module", "__init__", "act", "activation", "class", "config", "def", "dense_4h_to_h", "dense_h_to_4h", "ffn_hidden_size", "forward", "get_activation", "hidden_dropout", "hidden_size", "nn", "return", "self", "super", "x" ]
falcon/modeling_falcon.py:FalconDecoderLayer
[ -0.00026863490347750485, 0.045315682888031006, 0.020152954384684563, 0.00038783051422797143, -0.0010460750199854374, 0.03210809826850891, 0.029830927029252052, -0.03506841883063316, -0.007742378395050764, -0.02243012562394142, 0.0027610687538981438, 0.016395624727010727, -0.00296032102778553...
[ "False", "GradientCheckpointingLayer", "LayerNorm", "ModelDecoderLayer", "ModelMLP", "Model_ATTENTION_CLASSES", "None", "Tensor", "__init__", "_attn_implementation", "alibi", "and", "attention_dropout", "attention_layernorm_out", "attention_mask", "attention_output", "attn_weights", ...
falcon/modeling_falcon.py:FalconPreTrainedModel
[ -0.00028953002765774727, 0.049870654940605164, 0.0065483832731842995, 0.007434845436364412, -0.0011938646202906966, 0.020474420860409737, 0.021847007796168327, -0.018987450748682022, -0.0017872224561870098, 0.006948721129447222, -0.008006756193935871, -0.00643400102853775, -0.000409273954574...
[ "False", "ModelConfig", "ModelDecoderLayer", "ModelLinear", "ModelPreTrainedModel", "None", "PreTrainedModel", "True", "_can_compile_fullgraph", "_check_and_enable_sdpa", "_init_weights", "_is_bettertransformer", "_no_split_modules", "_supports_flash_attn", "_supports_sdpa", "base_mode...
falcon/modeling_falcon.py:FalconModel
[ -0.00013081190991215408, 0.03730861097574234, 0.011967972852289677, -0.005787329748272896, -0.0007023458601906896, 0.03416210040450096, 0.035510607063770294, -0.015058294869959354, 0.0010956594487652183, -0.016519173979759216, 0.021126562729477882, 0.00455120112746954, -0.0018822868587449193...
[ "BaseModelOutputWithPastAndCrossAttentions", "DynamicCache", "Embedding", "False", "LayerNorm", "ModelDecoderLayer", "ModelModel", "ModelPreTrainedModel", "ModelRotaryEmbedding", "ModuleList", "None", "Setting", "True", "ValueError", "You", "_", "__init__", "alibi", "all_hidden_s...
falcon/modeling_falcon.py:FalconForCausalLM
[ -0.0002791432198137045, 0.05097830668091774, 0.018775492906570435, -0.004153366666287184, -0.0012872591614723206, 0.020937519147992134, 0.03686824068427086, -0.014679021202027798, 0.0020197879057377577, 0.015020393766462803, 0.014792812056839466, 0.002887443173676729, 0.0011307967361062765, ...
[ "CausalLMOutputWithCrossAttentions", "GenerationMixin", "Linear", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "cache_position", "class", "config", "def", "else", "forward", "...
falcon/modeling_falcon.py:FalconForSequenceClassification
[ -0.00038114553899504244, 0.035502009093761444, 0.0052680401131510735, 0.01717839017510414, -0.0012955369893461466, 0.020385023206472397, 0.01242570299655199, -0.017636481672525406, -0.0036074621602892876, -0.0039796605706214905, 0.047183312475681305, 0.023248089477419853, 0.00075871229637414...
[ "BCEWithLogitsLoss", "Cannot", "CrossEntropyLoss", "Linear", "MSELoss", "ModelForSequenceClassification", "ModelModel", "ModelPreTrainedModel", "None", "Results", "SequenceClassifierOutputWithPast", "ValueError", "__class__", "__init__", "__name__", "and", "arange", "argmax", "at...
falcon/modeling_falcon.py:FalconForTokenClassification
[ -0.0002678140881471336, 0.03655041754245758, 0.006356594152748585, -0.031101906672120094, -0.0008655184064991772, 0.029966801404953003, 0.033372119069099426, 0.0013975993497297168, -0.004058004356920719, 0.00485257850959897, 0.05675530433654785, 0.019183292984962463, -0.0003901927266269922, ...
[ "CrossEntropyLoss", "Dropout", "Linear", "ModelForTokenClassification", "ModelModel", "ModelPreTrainedModel", "None", "TokenClassifierOutput", "__init__", "attention_mask", "attentions", "auto_docstring", "batch_size", "class", "classifier", "classifier_dropout", "config", "def", ...
falcon/modeling_falcon.py:FalconForQuestionAnswering
[ -0.00019118325144518167, 0.029771726578474045, 0.02199047990143299, 0.0166902095079422, -0.0006378648686222732, 0.05142389237880707, 0.04420650377869606, 0.01815624162554741, 0.007048230618238449, 0.026614118367433548, 0.02052444778382778, 0.017141297459602356, 0.0016281412681564689, 0.002...
[ "CrossEntropyLoss", "Linear", "ModelForQuestionAnswering", "ModelModel", "ModelPreTrainedModel", "None", "QuestionAnsweringModelOutput", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "clamp", "class", "config", "contiguous", "def", "dim", "else", "end_log...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrDecoderOutput
[ -0.00015174728468991816, 0.05270511656999588, 0.01681111566722393, 0.00732646556571126, -0.0012068790383636951, 0.03475811704993248, 0.07269671559333801, -0.04248214140534401, 0.012835513800382614, -0.002924906788393855, -0.008291969075798988, 0.016583938151597977, -0.004174381494522095, 0...
[ "BaseModelOutputWithCrossAttentions", "ModelDecoderOutput", "None", "class", "intermediate_hidden_states", "r", "reference_points" ]
conditional_detr/modeling_conditional_detr.py:ConditionalDetrModelOutput
[ -0.00013925493112765253, 0.029745563864707947, 0.024977190420031548, 0.014645716175436974, -0.0008231119718402624, 0.052452102303504944, 0.06267004460096359, -0.030880890786647797, 0.012942725792527199, -0.0003051190869882703, 0, -0.003193106735125184, -0.0024125694762915373, 0.00982057768...
[ "ModelModelOutput", "None", "Seq2SeqModelOutput", "class", "intermediate_hidden_states", "r", "reference_points" ]
conditional_detr/modeling_conditional_detr.py:ConditionalDetrObjectDetectionOutput
[ -0.00018221529899165034, 0.011206685565412045, -0.0008568563498556614, -0.0022612474858760834, -0.0009528526570647955, 0.05256333574652672, 0.052108243107795715, -0.021730730310082436, 0.014051022008061409, -0.014904322102665901, -0.004209617618471384, 0.03526977077126503, -0.004693155176937...
[ "ModelObjectDetectionOutput", "ModelOutput", "None", "auxiliary_outputs", "class", "cross_attentions", "decoder_attentions", "decoder_hidden_states", "encoder_attentions", "encoder_hidden_states", "encoder_last_hidden_state", "last_hidden_state", "logits", "loss", "loss_dict", "pred_bo...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrSegmentationOutput
[ -0.0002218172448920086, 0.021436987444758415, -0.011687719263136387, -0.007183671463280916, -0.0011687719961628318, 0.05039972811937332, 0.03968123346567154, -0.03375185281038284, 0.01573566161096096, -0.017218006774783134, 0.007468738127499819, 0.027594421058893204, -0.003876902163028717, ...
[ "ModelOutput", "ModelSegmentationOutput", "None", "auxiliary_outputs", "class", "cross_attentions", "decoder_attentions", "decoder_hidden_states", "encoder_attentions", "encoder_hidden_states", "encoder_last_hidden_state", "last_hidden_state", "logits", "loss", "loss_dict", "pred_boxes...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrFrozenBatchNorm2d
[ -0.00022441272449214011, 0.023253431543707848, 0.018465960398316383, 0.0027356978971511126, -0.0008050361066125333, 0.029408752918243408, -0.01937786117196083, -0.014704376459121704, 0.01687013730406761, 0.020061785355210304, 0.01607222482562065, -0.003034914843738079, 0.00237948726862669, ...
[ "ModelFrozenBatchNorm2d", "Module", "__init__", "_load_from_state_dict", "bias", "class", "def", "del", "epsilon", "error_msgs", "forward", "if", "in", "local_metadata", "missing_keys", "n", "nn", "num_batches_tracked", "num_batches_tracked_key", "ones", "prefix", "register...
conditional_detr/modeling_conditional_detr.py:replace_batch_norm
[ 0.00006147739622974768, 0.005075003486126661, 0.02018596976995468, -0.013628380373120308, 0.0004401425248943269, 0.03170451521873474, -0.014312650077044964, -0.013457313179969788, 0.028169121593236923, 0.00610140897333622, 0.011119389906525612, -0.007698039058595896, 0.0019815322011709213, ...
[ "BatchNorm2d", "ModelFrozenBatchNorm2d", "Model_batch_norm", "_modules", "bias", "children", "copy_", "def", "device", "for", "if", "in", "isinstance", "len", "list", "meta", "model", "module", "name", "named_children", "new_module", "nn", "num_features", "r", "runnin...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrConvEncoder
[ -0.00012956744467373937, 0.014343466609716415, 0.026445765048265457, 0.007900112308561802, -0.00030290766153484583, 0.030031632632017136, -0.00829231645911932, -0.037875715643167496, 0.004734464455395937, 0.010197307914495468, 0.010981716215610504, 0.014567582868039608, -0.000400958699174225...
[ "False", "ModelConvEncoder", "Module", "None", "True", "__init__", "_backbone", "and", "append", "backbone", "backbone_config", "backbone_model_type", "bool", "channels", "class", "config", "def", "dict", "else", "feature_map", "feature_maps", "features", "float", "for"...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrSinePositionEmbedding
[ -0.0002054105862043798, 0.04598351567983627, 0.009617839939892292, 0.013999932445585728, -0.0011310920817777514, 0.0038699000142514706, 0.04052012786269188, -0.013430829159915447, 0.011666609905660152, 0.019463321194052696, 0.013317008502781391, 0.019008038565516472, -0.00047306681517511606,...
[ "False", "ModelSinePositionEmbedding", "Module", "None", "__init__", "and", "arange", "bool", "cat", "class", "compile_compatible_method_lru_cache", "cos", "cumsum", "def", "device", "dim", "dim_t", "div", "dtype", "else", "eps", "flatten", "floor", "forward", "if", ...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrLearnedPositionEmbedding
[ -0.0002498500398360193, -0.006110617890954018, 0.018960047513246536, 0.02010221965610981, -0.0013277744874358177, 0.025698861107230186, 0.03266610577702522, -0.017132574692368507, 0.0038548291195183992, 0.015419316478073597, -0.006510377861559391, 0.01633305288851261, -0.0007067186525091529,...
[ "Embedding", "ModelLearnedPositionEmbedding", "Module", "None", "__init__", "arange", "cat", "class", "column_embeddings", "compile_compatible_method_lru_cache", "def", "device", "dim", "dtype", "embedding_dim", "flatten", "forward", "height", "height_values", "mask", "maxsiz...
conditional_detr/modeling_conditional_detr.py:eager_attention_forward
[ 0.000042374336771899834, 0.02700657583773136, 0.026102591305971146, -0.010847830213606358, 0.0003637130430433899, 0.03683742135763168, 0.06101904436945915, -0.026328587904572487, 0.02000068686902523, 0.012542803771793842, 0.02395562455058098, 0.026893578469753265, 0.002500085858628154, -0....
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "size", "softmax", "torch", "tra...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrSelfAttention
[ 0.000010863224815693684, 0.04719248041510582, 0.03056275099515915, -0.006123786326497793, -0.00009305065032094717, 0.020674802362918854, 0.040450699627399445, -0.004943974316120148, 0.007191235199570656, 0.014607196673750877, 0.014607196673750877, 0.015730828046798706, 0.00018258995260111988...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "ModelSelfAttention", "Module", "None", "__init__", "_attn_implementation", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "class", "config", "contiguous", "def", "dropout", "eager_attenti...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrDecoderSelfAttention
[ -0.00012499306467361748, 0.05565888434648514, 0.035603661090135574, -0.012506349012255669, -0.0005809537251479924, 0.0126190185546875, 0.03763171657919884, -0.014759744517505169, 0.00022710008488502353, 0.02805478125810623, 0.0025773218367248774, 0.026702744886279106, -0.0025914055295288563,...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "ModelDecoderSelfAttention", "Module", "None", "__init__", "_attn_implementation", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "class", "config", "contiguous", "def", "dropout", "eager_...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrDecoderCrossAttention
[ -0.0002832967438735068, 0.06793384253978729, 0.037180010229349136, 0.0029405485838651657, -0.001355521148070693, 0.009524508379399776, 0.04727828502655029, -0.010040897876024246, -0.0009467132040299475, 0.02306537702679634, -0.007631082087755203, 0.02914729155600071, -0.0020942443516105413, ...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "ModelDecoderCrossAttention", "Module", "None", "__init__", "_attn_implementation", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "cat", "class", "config", "contiguous", "def", "dim", "...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrMLP
[ -0.00025981778162531555, 0.036926377564668655, 0.040366724133491516, 0.008371507748961449, -0.0009675972396507859, 0.04610063508152962, 0.031192468479275703, -0.018348511308431625, 0.00021681345242541283, -0.019839327782392502, 0.0451832078397274, -0.03715573623776436, -0.0001675376697676256...
[ "ACT2FN", "Linear", "ModelMLP", "Module", "__init__", "activation_dropout", "activation_fn", "activation_function", "class", "config", "def", "dropout", "fc1", "fc2", "forward", "functional", "hidden_size", "hidden_states", "intermediate_size", "nn", "p", "return", "self"...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrEncoderLayer
[ -0.00014732932322658598, 0.03907138481736183, 0.026085229590535164, 0.010614769533276558, -0.000582259614020586, 0.03952307626605034, 0.018519384786486626, -0.014792922884225845, 0.00570261524990201, 0.013889538124203682, 0.011800461448729038, 0.028117846697568893, 0.0009810192277655005, -...
[ "GradientCheckpointingLayer", "LayerNorm", "ModelEncoderLayer", "ModelMLP", "ModelSelfAttention", "None", "_", "__init__", "any", "attention_dropout", "attention_mask", "clamp", "clamp_value", "class", "config", "d_model", "def", "dropout", "dtype", "encoder_attention_heads", ...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrDecoderLayer
[ -0.00017760855553206056, 0.0506705716252327, 0.01074487529695034, -0.0011946605518460274, -0.0006644857348874211, 0.03619326651096344, 0.046146415174007416, -0.026240117847919464, 0.006701409351080656, 0, 0.0032234627287834883, 0.02148975059390068, -0.0016117313643917441, 0.004439330194145...
[ "False", "GradientCheckpointingLayer", "LayerNorm", "ModelDecoderCrossAttention", "ModelDecoderLayer", "ModelDecoderSelfAttention", "ModelMLP", "None", "_", "__init__", "attention_dropout", "attention_mask", "class", "config", "d_model", "decoder_attention_heads", "decoder_ffn_dim", ...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrMLPPredictionHead
[ -0.00024052357184700668, -0.0006734660128131509, 0.02770831435918808, 0.04241766780614853, -0.0006734660128131509, 0.040137145668268204, 0.04583844542503357, -0.009635196067392826, -0.0012685391120612621, -0.013569092378020287, 0.022007014602422714, -0.0588374100625515, -0.002750876825302839...
[ "Linear", "ModelMLPPredictionHead", "Module", "ModuleList", "__init__", "class", "def", "else", "enumerate", "for", "forward", "functional", "h", "hidden_dim", "i", "if", "in", "input_dim", "k", "layer", "layers", "n", "nn", "num_layers", "output_dim", "relu", "re...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrConvBlock
[ 0.00006208741251612082, -0.009833931922912598, -0.0085761034861207, 0.0008147298358380795, 0.000707528495695442, 0.041851382702589035, 0.004774030763655901, -0.014865245670080185, 0.02275526151061058, -0.00428805174306035, 0.04230877757072449, -0.016123075038194656, 0.004888379015028477, -...
[ "ACT2FN", "Conv2d", "GroupNorm", "ModelConvBlock", "Module", "__init__", "activation", "class", "conv", "def", "forward", "in_channels", "kernel_size", "min", "nn", "norm", "out_channels", "padding", "relu", "return", "self", "super", "x" ]
conditional_detr/modeling_conditional_detr.py:ConditionalDetrFPNFusionStage
[ -0.00029496633214876056, 0.020822836086153984, -0.005920778959989548, 0.005320120137184858, -0.0008866866701282561, 0.023797526955604553, 0.03615393489599228, -0.06681612879037857, 0.0011012076865881681, -0.00903848372399807, 0.010754651390016079, -0.01613197661936283, 0.003389431396499276, ...
[ "Conv2d", "ModelConvBlock", "ModelFPNFusionStage", "Module", "__init__", "activation", "class", "current_channels", "def", "features", "forward", "fpn_adapter", "fpn_channels", "fpn_features", "functional", "interpolate", "kernel_size", "mode", "nearest", "nn", "output_channe...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrMaskHeadSmallConv
[ -0.0001942921371664852, 0.014243380166590214, 0.0050304001197218895, -0.005539092235267162, -0.00041507865535095334, 0.029391102492809296, 0.017747703939676285, -0.06285174190998077, -0.0007312449743039906, 0.009947757236659527, 0.01396077312529087, 0.002642373088747263, 0.002981501165777445...
[ "Conv2d", "ModelConvBlock", "ModelFPNFusionStage", "ModelMaskHeadSmallConv", "Module", "ModuleList", "ValueError", "__init__", "activation_function", "attention_masks", "be", "by", "cat", "class", "conv1", "conv2", "def", "dim", "divisible", "expand", "f", "features", "fl...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrMHAttentionMap
[ -0.00018104938499163836, 0.009721910580992699, 0.06014018878340721, -0.030070094391703606, -0.0005793580203317106, 0.026791775599122047, 0.04024418815970421, -0.030296185985207558, 0.007913182489573956, 0.02407868579030037, 0.028261367231607437, 0.027696140110492706, 0.00013512466102838516, ...
[ "Linear", "ModelMHAttentionMap", "Module", "None", "_", "__init__", "attention_dropout", "attention_mask", "attn_weights", "attn_weights_shape", "batch_size", "bias", "class", "contiguous", "conv2d", "def", "dim", "dropout", "flatten", "forward", "functional", "head_dim", ...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrPreTrainedModel
[ -0.0002648759400472045, 0.05961664021015167, -0.008362257853150368, 0.004095799755305052, -0.001130611402913928, 0.0235508494079113, 0.01427841279655695, -0.026850242167711258, -0.0007288532797247171, 0.010979019105434418, -0.004181128926575184, -0.0009599530603736639, -0.0009528422961011529...
[ "Conv2d", "Embedding", "False", "GroupNorm", "LayerNorm", "Linear", "ModelConfig", "ModelConvEncoder", "ModelDecoderLayer", "ModelEncoderLayer", "ModelLearnedPositionEmbedding", "ModelMHAttentionMap", "ModelMaskHeadSmallConv", "ModelPreTrainedModel", "None", "PreTrainedModel", "True"...