operator name
stringclasses
180 values
used in model
stringclasses
155 values
args
stringlengths
19
5.24k
aten.copy_.default
TorchBench/Background_Matting
((T([3, 1, 512, 512], f16), T([3, 1, 512, 512], f16)), {})
aten.copy_.default
TorchBench/vision_maskrcnn
((T([3, 1155, 800], f16, stride=(1439744, 1216, 1)), T([3, 1155, 800], f16)), {})
aten.copy_.default
TorchBench/Background_Matting
((T([3, 3, 512, 512], f16), T([3, 3, 512, 512], f16)), {})
aten.copy_.default
TorchBench/Background_Matting
((T([3, 4, 512, 512], f16), T([3, 4, 512, 512], f16)), {})
aten.copy_.default
TorchBench/vision_maskrcnn
((T([3, 799, 1115], f16, stride=(1439744, 1216, 1)), T([3, 799, 1115], f16)), {})
aten.copy_.default
TorchBench/vision_maskrcnn
((T([3, 799, 1199], f16, stride=(1439744, 1216, 1)), T([3, 799, 1199], f16)), {})
aten.copy_.default
TorchBench/vision_maskrcnn
((T([3, 800, 800], f16, stride=(1439744, 1216, 1)), T([3, 800, 800], f16)), {})
aten.copy_.default
HuggingFace/MobileBertForMaskedLM
((T([30522, 128], f16), T([30522, 128], f16, stride=(1, 30522))), {})
aten.copy_.default
TIMM/convnext_base
((T([32, 1024], f16), T([32, 1024], f16)), {})
aten.copy_.default
HuggingFace/DistilBertForQuestionAnswering
((T([32, 128], i64), T([32, 128], i64)), {})
aten.copy_.default
HuggingFace/MobileBertForQuestionAnswering
((T([32, 128], i64), T([32, 128], i64)), {})
aten.copy_.default
TIMM/convmixer_768_32
((T([32, 3, 224, 224], f16), T([32, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/convnext_base
((T([32, 3, 224, 224], f16), T([32, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/dpn107
((T([32, 3, 224, 224], f16), T([32, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/gluon_senet154
((T([32, 3, 224, 224], f16), T([32, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/legacy_senet154
((T([32, 3, 224, 224], f16), T([32, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/swsl_resnext101_32x16d
((T([32, 3, 224, 224], f16), T([32, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/twins_pcpvt_base
((T([32, 3, 224, 224], f16), T([32, 3, 224, 224], f16)), {})
aten.copy_.default
TorchBench/mnasnet1_0
((T([32, 3, 224, 224], f16), T([32, 3, 224, 224], f16)), {})
aten.copy_.default
TorchBench/mobilenet_v3_large
((T([32, 3, 224, 224], f16), T([32, 3, 224, 224], f16)), {})
aten.copy_.default
TorchBench/resnet50
((T([32, 3, 224, 224], f16), T([32, 3, 224, 224], f16)), {})
aten.copy_.default
TorchBench/squeezenet1_1
((T([32, 3, 224, 224], f16), T([32, 3, 224, 224], f16)), {})
aten.copy_.default
TorchBench/timm_efficientnet
((T([32, 3, 224, 224], f16), T([32, 3, 224, 224], f16)), {})
aten.copy_.default
TorchBench/timm_regnet
((T([32, 3, 224, 224], f16), T([32, 3, 224, 224], f16)), {})
aten.copy_.default
TorchBench/timm_resnest
((T([32, 3, 224, 224], f16), T([32, 3, 224, 224], f16)), {})
aten.copy_.default
TorchBench/timm_vovnet
((T([32, 3, 224, 224], f16), T([32, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/resnest101e
((T([32, 3, 256, 256], f16), T([32, 3, 256, 256], f16)), {})
aten.copy_.default
TIMM/gluon_xception65
((T([32, 3, 299, 299], f16), T([32, 3, 299, 299], f16)), {})
aten.copy_.default
TorchBench/dcgan
((T([32, 3, 64, 64], f16), T([32, 3, 64, 64], f16)), {})
aten.copy_.default
TIMM/twins_pcpvt_base
((T([320, 320, 2, 2], f16), T([320, 320, 2, 2], f16, stride=(1280, 1, 640, 320))), {})
aten.copy_.default
HuggingFace/DistilBertForQuestionAnswering
((T([32], i64), T([32], i64)), {})
aten.copy_.default
HuggingFace/MobileBertForQuestionAnswering
((T([32], i64), T([32], i64)), {})
aten.copy_.default
TIMM/gmixer_24_224
((T([384, 196], f16), T([384, 196], f16, stride=(1, 384))), {})
aten.copy_.default
TIMM/mixer_b16_224
((T([384, 196], f16), T([384, 196], f16, stride=(1, 384))), {})
aten.copy_.default
HuggingFace/BartForCausalLM
((T([4, 1024], i64), T([4, 1024], i64)), {})
aten.copy_.default
HuggingFace/GPT2ForSequenceClassification
((T([4, 1024], i64), T([4, 1024], i64)), {})
aten.copy_.default
HuggingFace/OPTForCausalLM
((T([4, 128], i64), T([4, 128], i64)), {})
aten.copy_.default
HuggingFace/PegasusForConditionalGeneration
((T([4, 128], i64), T([4, 128], i64)), {})
aten.copy_.default
HuggingFace/RobertaForCausalLM
((T([4, 128], i64), T([4, 128], i64)), {})
aten.copy_.default
TorchBench/densenet121
((T([4, 3, 224, 224], f16), T([4, 3, 224, 224], f16)), {})
aten.copy_.default
HuggingFace/DebertaForMaskedLM
((T([4, 512], i64), T([4, 512], i64)), {})
aten.copy_.default
HuggingFace/DebertaForQuestionAnswering
((T([4, 512], i64), T([4, 512], i64)), {})
aten.copy_.default
HuggingFace/XLNetLMHeadModel
((T([4, 512], i64), T([4, 512], i64)), {})
aten.copy_.default
TorchBench/hf_Bart
((T([4, 512], i64), T([4, 512], i64)), {})
aten.copy_.default
TorchBench/hf_Bert
((T([4, 512], i64), T([4, 512], i64)), {})
aten.copy_.default
TorchBench/hf_GPT2
((T([4, 512], i64), T([4, 512], i64)), {})
aten.copy_.default
TorchBench/pytorch_struct
((T([40, 29], i64, stride=(1, 40)), T([40, 29], i64, stride=(1, 40))), {})
aten.copy_.default
TorchBench/fastNLP_Bert
((T([474], i64), T([474], i64)), {})
aten.copy_.default
HuggingFace/DebertaForQuestionAnswering
((T([4], i64), T([4], i64)), {})
aten.copy_.default
HuggingFace/GPT2ForSequenceClassification
((T([4], i64), T([4], i64)), {})
aten.copy_.default
TorchBench/maml_omniglot
((T([5, 1, 28, 28], f16), T([5, 1, 28, 28], f16)), {})
aten.copy_.default
TIMM/convnext_base
((T([512, 256, 2, 2], f16), T([512, 256, 2, 2], f16, stride=(1024, 1, 512, 256))), {})
aten.copy_.default
TIMM/jx_nest_base
((T([512, 256, 3, 3], f16), T([512, 256, 3, 3], f16, stride=(2304, 1, 768, 256))), {})
aten.copy_.default
HuggingFace/YituTechConvBert
((T([54, 384], f16), T([54, 384], f16, stride=(1, 54))), {})
aten.copy_.default
TorchBench/Super_SloMo
((T([6, 3, 352, 352], f16), T([6, 3, 352, 352], f16)), {})
aten.copy_.default
TorchBench/fastNLP_Bert
((T([6, 474, 768], f16), T([6, 474, 768], f16)), {})
aten.copy_.default
TorchBench/fastNLP_Bert
((T([6, 474], i64), T([6, 474], i64)), {})
aten.copy_.default
TorchBench/fastNLP_Bert
((T([6, 474], i64, stride=(475, 1)), T([6, 474], i64)), {})
aten.copy_.default
TorchBench/pytorch_struct
((T([60, 60, 256], f16), T([60, 60, 256], f16, stride=(60, 1, 3600))), {})
aten.copy_.default
HuggingFace/BertForMaskedLM
((T([64, 128], i64), T([64, 128], i64)), {})
aten.copy_.default
HuggingFace/BertForQuestionAnswering
((T([64, 128], i64), T([64, 128], i64)), {})
aten.copy_.default
HuggingFace/BlenderbotSmallForCausalLM
((T([64, 128], i64), T([64, 128], i64)), {})
aten.copy_.default
HuggingFace/BlenderbotSmallForConditionalGeneration
((T([64, 128], i64), T([64, 128], i64)), {})
aten.copy_.default
HuggingFace/RobertaForQuestionAnswering
((T([64, 128], i64), T([64, 128], i64)), {})
aten.copy_.default
HuggingFace/Speech2Text2ForCausalLM
((T([64, 128], i64), T([64, 128], i64)), {})
aten.copy_.default
TorchBench/pytorch_CycleGAN_and_pix2pix
((T([64, 256, 256], f16), T([64, 256, 256], f16)), {})
aten.copy_.default
TIMM/beit_base_patch16_224
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/deit_base_distilled_patch16_224
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/densenet121
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/dla102
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/ecaresnet101d
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/gmixer_24_224
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/gmlp_s16_224
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/jx_nest_base
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/mixer_b16_224
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/mixnet_l
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/pit_b_224
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/poolformer_m36
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/res2net101_26w_4s
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/swin_base_patch4_window7_224
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/tf_mixnet_l
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/tnt_s_patch16_224
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/vit_base_patch16_224
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/volo_d1_224
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TorchBench/vgg16
((T([64, 3, 224, 224], f16), T([64, 3, 224, 224], f16)), {})
aten.copy_.default
TIMM/crossvit_9_240
((T([64, 3, 240, 240], f16), T([64, 3, 240, 240], f16)), {})
aten.copy_.default
TIMM/cspdarknet53
((T([64, 3, 256, 256], f16), T([64, 3, 256, 256], f16)), {})
aten.copy_.default
TIMM/mobilevit_s
((T([64, 3, 256, 256], f16), T([64, 3, 256, 256], f16)), {})
aten.copy_.default
TIMM/sebotnet33ts_256
((T([64, 3, 256, 256], f16), T([64, 3, 256, 256], f16)), {})
aten.copy_.default
TorchBench/tts_angular
((T([64, 50, 40], f16), T([64, 50, 40], f16)), {})
aten.copy_.default
TIMM/jx_nest_base
((T([64, 512], f16), T([64, 512], f16)), {})
aten.copy_.default
HuggingFace/ElectraForQuestionAnswering
((T([64, 512], i64), T([64, 512], i64)), {})
aten.copy_.default
TorchBench/maml_omniglot
((T([64, 64, 3, 3], f16), T([64, 64, 3, 3], f16, stride=(576, 1, 192, 64))), {})
aten.copy_.default
TIMM/twins_pcpvt_base
((T([64, 64, 8, 8], f16), T([64, 64, 8, 8], f16, stride=(4096, 1, 512, 64))), {})
aten.copy_.default
TIMM/levit_128
((T([640, 128], f16), T([640, 128], f16, stride=(1, 640))), {})
aten.copy_.default
TorchBench/pytorch_stargan
((T([64], f16), T([64], f16)), {})
aten.copy_.default
HuggingFace/BertForQuestionAnswering
((T([64], i64), T([64], i64)), {})
aten.copy_.default
HuggingFace/ElectraForQuestionAnswering
((T([64], i64), T([64], i64)), {})
aten.copy_.default
HuggingFace/RobertaForQuestionAnswering
((T([64], i64), T([64], i64)), {})
aten.copy_.default
TorchBench/Super_SloMo
((T([6], i64), T([6], i64)), {})