text
stringlengths
15
129
model:slone/mbart-large-51-mul-myv-v1 [43, 40]
model:oeg/esT5s-base [40]
model:Luciano/bertimbau-base-finetuned-lener-br-finetuned-brazilian_court_decisions [32]
model:YSKartal/distilbert-base-uncased-finetuned-svident [32]
model:juliensimon/autotrain-food101-1471154053 [10]
model:theodotus/stt_uk_contextnet_512 [4]
model:pszemraj/pegasus-x-large-book-summary [26, 40]
model:lambdalabs/sd-pokemon-diffusers [37]
model:pnr-svc/bert-finetuned-ner [42]
model:mediabiasgroup/da-roberta-babe-ft [32]
model:KoichiYasuoka/deberta-base-japanese-wikipedia-ud-goeswith [42]
model:KoichiYasuoka/roberta-classical-chinese-base-ud-goeswith [42]
model:morenolq/distilgpt2-fables-demo [33]
model:KoichiYasuoka/deberta-large-japanese-wikipedia-ud-goeswith [42]
model:morenolq/bart-base-xsum [26, 40]
model:KoichiYasuoka/roberta-classical-chinese-large-ud-goeswith [42]
model:morenolq/distilbert-base-cased-hate-speech [32]
model:bdotloh/just-another-emotion-classifier [32]
model:morenolq/distilbert-base-cased-emotion [32]
model:pszemraj/long-t5-tglobal-large-pubmed-3k-booksum-16384-WIP14 [26, 40]
model:ku-nlp/roberta-base-japanese-char-wwm [8]
model:theodotus/stt_uk_squeezeformer_ctc_xs [4]
model:nhero/finetuning-sentiment-model-3000-samples [32]
model:pszemraj/long-t5-tglobal-large-pubmed-3k-booksum-16384-WIP15 [26, 40]
model:bdotloh/distilbert-base-uncased-go-emotion-empathetic-dialogues-context-v2 [32]
model:bdotloh/distilbert-base-uncased-empathetic-dialogues-context [32]
model:farleyknight/arxiv-summarization-t5-base-2022-09-21 [40]
model:omarques/autotrain-dogs-and-cats-1527055142 [10]
model:farleyknight/arxiv-summarization-fb-bart-base-2022-09-21 [40]
model:uaritm/multilingual_en_ru_uk [25, 7]
model:EleutherAI/polyglot-ko-5.8b [33]
model:cartesinus/xlm-r-base-amazon-massive-intent [32]
model:opennyaiorg/en_legal_ner_trf [42]
model:cartesinus/xlm-r-base-amazon-massive-domain [32]
model:Gabriel/bart-base-cnn-xsum-swe [26, 40]
model:cartesinus/bert-base-uncased-amazon-massive-intent [32]
model:lcw99/t5-base-korean-text-summary [40]
model:theodotus/stt_uk_squeezeformer_ctc_sm [4]
model:gokuls/distilroberta-emotion-intent [32]
model:nickprock/bert-finetuned-ner-ontonotes [42]
model:nickprock/distilbert-finetuned-ner-ontonotes [42]
model:bond005/wav2vec2-large-ru-golos-with-lm [4]
model:shibing624/bart4csc-base-chinese [40]
model:facebook/esm2_t12_35M_UR50D [8]
model:jmassot/bert-base-uncased-issues-128 [8]
model:cardiffnlp/roberta-base-tweet-topic-multi-all [32]
model:cardiffnlp/twitter-roberta-base-dec2020-tweet-topic-multi-all [32]
model:cardiffnlp/twitter-roberta-base-2019-90m-tweet-topic-multi-all [32]
model:cardiffnlp/roberta-base-tweet-topic-single-all [32]
model:cardiffnlp/roberta-base-tweet-topic-multi-2020 [32]
model:cardiffnlp/twitter-roberta-base-2019-90m-tweet-topic-multi-2020 [32]
model:cardiffnlp/twitter-roberta-base-dec2020-tweet-topic-multi-2020 [32]
model:bigscience/bloomz-7b1-mt [33]
model:irenepap/t5-small-asqa-ob [40]
model:stevhliu/my_awesome_model [32]
model:postbot/gpt-neo-1.3B-emailgen [33]
model:lschlessinger/bert-finetuned-math-prob-classification [32]
model:cardiffnlp/roberta-large-tweet-topic-multi-2020 [32]
model:cardiffnlp/twitter-roberta-base-dec2021-tweet-topic-multi-all [32]
model:cardiffnlp/roberta-large-tweet-topic-multi-all [32]
model:cardiffnlp/twitter-roberta-base-dec2021-tweet-topic-multi-2020 [32]
model:cardiffnlp/twitter-roberta-base-dec2021-tweet-topic-single-all [32]
model:cardiffnlp/twitter-roberta-base-dec2020-tweet-topic-single-all [32]
model:cardiffnlp/twitter-roberta-base-2019-90m-tweet-topic-single-all [32]
model:cardiffnlp/twitter-roberta-base-dec2021-tweet-topic-single-2020 [32]
model:cardiffnlp/twitter-roberta-base-dec2020-tweet-topic-single-2020 [32]
model:nielsr/lilt-roberta-en-base-finetuned-funsd [42]
model:cardiffnlp/twitter-roberta-base-2019-90m-tweet-topic-single-2020 [32]
model:cardiffnlp/roberta-base-tweet-topic-single-2020 [32]
model:cardiffnlp/roberta-large-tweet-topic-single-2020 [32]
model:cardiffnlp/roberta-large-tweet-topic-single-all [32]
model:nvidia/stt_be_conformer_ctc_large [4]
model:irenepap/t5-base-asqa-ob [40]
model:din0s/t5-base-asqa-cb [40]
model:nvidia/stt_be_conformer_transducer_large [4]
model:carbon225/vit-base-patch16-224-hentai [10]
model:esc-bench/wav2vec2-aed-librispeech [4]
model:esc-bench/wav2vec2-aed-common_voice [4]
model:esc-bench/wav2vec2-aed-tedlium [4]
model:esc-bench/wav2vec2-aed-voxpopuli [4]
model:esc-bench/wav2vec2-aed-gigaspeech [4]
model:esc-bench/wav2vec2-aed-spgispeech [4]
model:esc-bench/wav2vec2-aed-earnings22 [4]
model:esc-bench/wav2vec2-aed-ami [4]
model:esc-bench/wav2vec2-aed-switchboard [4]
model:esc-bench/wav2vec2-aed-chime4 [4]
model:esc-bench/wav2vec2-ctc-common_voice [4]
model:esc-bench/wav2vec2-ctc-tedlium [4]
model:esc-bench/wav2vec2-ctc-voxpopuli [4]
model:esc-bench/wav2vec2-ctc-gigaspeech [4]
model:esc-bench/wav2vec2-ctc-spgispeech [4]
model:esc-bench/wav2vec2-ctc-earnings22 [4]
model:esc-bench/wav2vec2-ctc-ami [4]
model:esc-bench/wav2vec2-ctc-switchboard [4]
model:esc-bench/wav2vec2-ctc-chime4 [4]
model:esc-bench/wav2vec2-ctc-librispeech [4]
model:RamAnanth1/distilroberta-base-finetuned-topic-news [32]
model:ashwinperti/newSentiment_1Oct22 [32]
model:cartesinus/xlm-r-base-amazon-massive-slot [42]
model:din0s/t5-base-msmarco-nlgen-cb [40]