operator name
stringclasses
180 values
used in model
stringclasses
155 values
args
stringlengths
19
5.24k
aten.mm.default
HuggingFace/DistilBertForQuestionAnswering
((T([2, 4096], f16, stride=(1, 2)), T([4096, 768], f16)), {})
aten.mm.default
HuggingFace/GPT2ForSequenceClassification
((T([2, 4096], f16, stride=(1, 2)), T([4096, 768], f16)), {})
aten.mm.default
HuggingFace/DebertaV2ForQuestionAnswering
((T([2, 512], f16, stride=(1, 2)), T([512, 1536], f16)), {})
aten.mm.default
TIMM/cait_m36_384
((T([2, 768], f16), T([768, 3072], f16)), {})
aten.mm.default
TIMM/cait_m36_384
((T([2, 768], f16), T([768, 768], f16)), {})
aten.mm.default
HuggingFace/BertForQuestionAnswering
((T([2, 8192], f16, stride=(1, 2)), T([8192, 768], f16)), {})
aten.mm.default
HuggingFace/RobertaForQuestionAnswering
((T([2, 8192], f16, stride=(1, 2)), T([8192, 768], f16)), {})
aten.mm.default
TIMM/jx_nest_base
((T([200704, 128], f16), T([128, 128], f16)), {})
aten.mm.default
TIMM/swin_base_patch4_window7_224
((T([200704, 128], f16), T([128, 128], f16)), {})
aten.mm.default
TIMM/jx_nest_base
((T([200704, 128], f16), T([128, 128], f16, stride=(1, 128))), {})
aten.mm.default
TIMM/jx_nest_base
((T([200704, 128], f16), T([128, 384], f16, stride=(1, 128))), {})
aten.mm.default
TIMM/jx_nest_base
((T([200704, 128], f16), T([128, 512], f16)), {})
aten.mm.default
TIMM/swin_base_patch4_window7_224
((T([200704, 128], f16), T([128, 512], f16)), {})
aten.mm.default
TIMM/jx_nest_base
((T([200704, 128], f16), T([128, 512], f16, stride=(1, 128))), {})
aten.mm.default
TIMM/tnt_s_patch16_224
((T([200704, 24], f16), T([24, 24], f16)), {})
aten.mm.default
TIMM/tnt_s_patch16_224
((T([200704, 24], f16), T([24, 24], f16, stride=(1, 24))), {})
aten.mm.default
TIMM/tnt_s_patch16_224
((T([200704, 24], f16), T([24, 48], f16, stride=(1, 24))), {})
aten.mm.default
TIMM/tnt_s_patch16_224
((T([200704, 24], f16), T([24, 96], f16)), {})
aten.mm.default
TIMM/jx_nest_base
((T([200704, 384], f16), T([384, 128], f16)), {})
aten.mm.default
TIMM/swin_base_patch4_window7_224
((T([200704, 384], f16), T([384, 128], f16)), {})
aten.mm.default
TIMM/tnt_s_patch16_224
((T([200704, 48], f16), T([48, 24], f16)), {})
aten.mm.default
TIMM/jx_nest_base
((T([200704, 512], f16), T([512, 128], f16)), {})
aten.mm.default
TIMM/swin_base_patch4_window7_224
((T([200704, 512], f16), T([512, 128], f16)), {})
aten.mm.default
TIMM/jx_nest_base
((T([200704, 512], f16), T([512, 128], f16, stride=(1, 512))), {})
aten.mm.default
TIMM/tnt_s_patch16_224
((T([200704, 96], f16), T([96, 24], f16)), {})
aten.mm.default
TorchBench/speech_transformer
((T([2040, 2048], f16), T([2048, 512], f16)), {})
aten.mm.default
TorchBench/speech_transformer
((T([2040, 512], f16), T([512, 2048], f16)), {})
aten.mm.default
TorchBench/speech_transformer
((T([2040, 512], f16), T([512, 512], f16)), {})
aten.mm.default
HuggingFace/BartForConditionalGeneration
((T([2048, 1024], f16), T([1024, 1024], f16)), {})
aten.mm.default
HuggingFace/MBartForCausalLM
((T([2048, 1024], f16), T([1024, 1024], f16)), {})
aten.mm.default
TIMM/levit_128
((T([2048, 1024], f16), T([1024, 384], f16, stride=(1, 1024))), {})
aten.mm.default
HuggingFace/BartForConditionalGeneration
((T([2048, 1024], f16), T([1024, 4096], f16)), {})
aten.mm.default
HuggingFace/MBartForCausalLM
((T([2048, 1024], f16), T([1024, 4096], f16)), {})
aten.mm.default
HuggingFace/XLNetLMHeadModel
((T([2048, 1024], f16), T([1024, 4096], f16)), {})
aten.mm.default
HuggingFace/BartForConditionalGeneration
((T([2048, 1024], f16), T([1024, 50265], f16, stride=(1, 1024))), {})
aten.mm.default
HuggingFace/MBartForCausalLM
((T([2048, 1024], f16), T([1024, 50265], f16, stride=(1, 1024))), {})
aten.mm.default
TIMM/jx_nest_base
((T([2048, 12544], f16, stride=(1, 2048)), T([12544, 512], f16)), {})
aten.mm.default
TIMM/swin_base_patch4_window7_224
((T([2048, 12544], f16, stride=(1, 2048)), T([12544, 512], f16)), {})
aten.mm.default
HuggingFace/MobileBertForMaskedLM
((T([2048, 128], f16), T([128, 128], f16)), {})
aten.mm.default
HuggingFace/GPTNeoForCausalLM
((T([2048, 128], f16), T([128, 2048], f16)), {})
aten.mm.default
HuggingFace/GPTNeoForSequenceClassification
((T([2048, 128], f16), T([128, 2048], f16)), {})
aten.mm.default
HuggingFace/MobileBertForMaskedLM
((T([2048, 128], f16), T([128, 512], f16)), {})
aten.mm.default
HuggingFace/GPTNeoForCausalLM
((T([2048, 128], f16, stride=(1, 2048)), T([128, 2048], f16)), {})
aten.mm.default
HuggingFace/GPTNeoForSequenceClassification
((T([2048, 128], f16, stride=(1, 2048)), T([128, 2048], f16)), {})
aten.mm.default
HuggingFace/GPTNeoForCausalLM
((T([2048, 128], f16, stride=(1, 2048)), T([128, 8192], f16)), {})
aten.mm.default
HuggingFace/GPTNeoForSequenceClassification
((T([2048, 128], f16, stride=(1, 2048)), T([128, 8192], f16)), {})
aten.mm.default
TIMM/twins_pcpvt_base
((T([2048, 1568], f16, stride=(1, 2048)), T([1568, 512], f16)), {})
aten.mm.default
TIMM/pit_b_224
((T([2048, 16448], f16, stride=(1, 2048)), T([16448, 512], f16)), {})
aten.mm.default
TorchBench/speech_transformer
((T([2048, 2040], f16, stride=(1, 2048)), T([2040, 512], f16)), {})
aten.mm.default
TorchBench/speech_transformer
((T([2048, 220], f16, stride=(1, 2048)), T([220, 512], f16)), {})
aten.mm.default
HuggingFace/DebertaForMaskedLM
((T([2048, 2304], f16), T([2304, 768], f16)), {})
aten.mm.default
HuggingFace/DebertaForQuestionAnswering
((T([2048, 2304], f16), T([2304, 768], f16)), {})
aten.mm.default
TorchBench/hf_GPT2
((T([2048, 2304], f16), T([2304, 768], f16, stride=(1, 2304))), {})
aten.mm.default
TIMM/levit_128
((T([2048, 256], f16), T([256, 256], f16)), {})
aten.mm.default
TIMM/levit_128
((T([2048, 256], f16), T([256, 256], f16, stride=(1, 256))), {})
aten.mm.default
HuggingFace/DebertaForQuestionAnswering
((T([2048, 2], f16), T([2, 768], f16)), {})
aten.mm.default
HuggingFace/MobileBertForMaskedLM
((T([2048, 30522], f16), T([30522, 512], f16, stride=(1, 30522))), {})
aten.mm.default
HuggingFace/DistilBertForMaskedLM
((T([2048, 30522], f16), T([30522, 768], f16)), {})
aten.mm.default
TorchBench/hf_Bert
((T([2048, 30522], f16, stride=(0, 0)), T([30522, 768], f16)), {})
aten.mm.default
HuggingFace/DebertaForMaskedLM
((T([2048, 3072], f16), T([3072, 768], f16)), {})
aten.mm.default
HuggingFace/DebertaForQuestionAnswering
((T([2048, 3072], f16), T([3072, 768], f16)), {})
aten.mm.default
HuggingFace/DistilBertForMaskedLM
((T([2048, 3072], f16), T([3072, 768], f16)), {})
aten.mm.default
HuggingFace/PLBartForCausalLM
((T([2048, 3072], f16), T([3072, 768], f16)), {})
aten.mm.default
TorchBench/BERT_pytorch
((T([2048, 3072], f16), T([3072, 768], f16)), {})
aten.mm.default
TorchBench/hf_Bart
((T([2048, 3072], f16), T([3072, 768], f16)), {})
aten.mm.default
TorchBench/hf_Bert
((T([2048, 3072], f16), T([3072, 768], f16)), {})
aten.mm.default
TorchBench/hf_BigBird
((T([2048, 3072], f16), T([3072, 768], f16)), {})
aten.mm.default
TorchBench/hf_Longformer
((T([2048, 3072], f16), T([3072, 768], f16)), {})
aten.mm.default
TorchBench/hf_GPT2
((T([2048, 3072], f16), T([3072, 768], f16, stride=(1, 3072))), {})
aten.mm.default
HuggingFace/XLNetLMHeadModel
((T([2048, 32000], f16), T([32000, 1024], f16)), {})
aten.mm.default
TIMM/levit_128
((T([2048, 384], f16), T([384, 1024], f16)), {})
aten.mm.default
TIMM/levit_128
((T([2048, 384], f16), T([384, 384], f16)), {})
aten.mm.default
TIMM/levit_128
((T([2048, 384], f16), T([384, 384], f16, stride=(1, 384))), {})
aten.mm.default
TIMM/levit_128
((T([2048, 384], f16), T([384, 768], f16)), {})
aten.mm.default
TIMM/levit_128
((T([2048, 384], f16), T([384, 768], f16, stride=(1, 384))), {})
aten.mm.default
HuggingFace/BartForConditionalGeneration
((T([2048, 4096], f16), T([4096, 1024], f16)), {})
aten.mm.default
HuggingFace/MBartForCausalLM
((T([2048, 4096], f16), T([4096, 1024], f16)), {})
aten.mm.default
HuggingFace/XLNetLMHeadModel
((T([2048, 4096], f16), T([4096, 1024], f16)), {})
aten.mm.default
HuggingFace/PLBartForCausalLM
((T([2048, 50005], f16), T([50005, 768], f16)), {})
aten.mm.default
TorchBench/hf_GPT2
((T([2048, 50257], f16, stride=(0, 0)), T([50257, 768], f16)), {})
aten.mm.default
HuggingFace/BartForConditionalGeneration
((T([2048, 50265], f16), T([50265, 1024], f16)), {})
aten.mm.default
HuggingFace/MBartForCausalLM
((T([2048, 50265], f16), T([50265, 1024], f16)), {})
aten.mm.default
HuggingFace/DebertaForMaskedLM
((T([2048, 50265], f16), T([50265, 768], f16)), {})
aten.mm.default
TorchBench/hf_Bart
((T([2048, 50265], f16, stride=(0, 0)), T([50265, 768], f16)), {})
aten.mm.default
TorchBench/hf_Longformer
((T([2048, 50265], f16, stride=(0, 0)), T([50265, 768], f16)), {})
aten.mm.default
TorchBench/hf_BigBird
((T([2048, 50358], f16, stride=(0, 0)), T([50358, 768], f16)), {})
aten.mm.default
HuggingFace/MobileBertForMaskedLM
((T([2048, 512], f16), T([512, 128], f16)), {})
aten.mm.default
HuggingFace/MobileBertForMaskedLM
((T([2048, 512], f16), T([512, 30522], f16)), {})
aten.mm.default
HuggingFace/MobileBertForMaskedLM
((T([2048, 512], f16), T([512, 384], f16)), {})
aten.mm.default
HuggingFace/MobileBertForMaskedLM
((T([2048, 512], f16), T([512, 512], f16)), {})
aten.mm.default
TIMM/convnext_base
((T([2048, 6272], f16, stride=(1, 2048)), T([6272, 512], f16)), {})
aten.mm.default
TIMM/coat_lite_mini
((T([2048, 6400], f16, stride=(1, 2048)), T([6400, 512], f16)), {})
aten.mm.default
HuggingFace/DebertaForMaskedLM
((T([2048, 768], f16), T([768, 2304], f16, stride=(1, 768))), {})
aten.mm.default
HuggingFace/DebertaForQuestionAnswering
((T([2048, 768], f16), T([768, 2304], f16, stride=(1, 768))), {})
aten.mm.default
HuggingFace/DebertaForMaskedLM
((T([2048, 768], f16), T([768, 3072], f16)), {})
aten.mm.default
HuggingFace/DebertaForQuestionAnswering
((T([2048, 768], f16), T([768, 3072], f16)), {})
aten.mm.default
HuggingFace/DistilBertForMaskedLM
((T([2048, 768], f16), T([768, 3072], f16)), {})
aten.mm.default
HuggingFace/PLBartForCausalLM
((T([2048, 768], f16), T([768, 3072], f16)), {})
aten.mm.default
TorchBench/BERT_pytorch
((T([2048, 768], f16), T([768, 3072], f16)), {})
aten.mm.default
TorchBench/hf_Bart
((T([2048, 768], f16), T([768, 3072], f16)), {})