index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
3,600 | Anishsavla2/xlnet-base-cased-finetuned-wikitext2 | xlnet | ["XLNetLMHeadModel"] | 32000 | float32 | 4.24.0 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,601 | Anitha008/English-QA | bert | ["BertForQuestionAnswering"] | 30522 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,602 | Anitha008/my_Malayalam_qa_model | bert | ["BertForQuestionAnswering"] | 197285 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,603 | AnjaliSarawgi/test-ocr-v2 | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | 1004 | float32 | 4.48.0 | null | null | null | null | null | null | null | null | null | null | null | null | null | 1000 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,604 | Anjalis404/ModalVerbsToLogic | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.46.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,605 | Anjnay/vit-base-oxford-iiit-pets | vit | ["ViTForImageClassification"] | null | float32 | 4.49.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
3,606 | Anjoe/german-poetry-gpt2-large | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.19.4 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
3,607 | Anjoe/german-poetry-xlm-roberta | xlm-roberta | ["XLMRobertaForMaskedLM"] | 250002 | float32 | 4.25.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,608 | Anjoe/kant-gpt2 | gpt2 | ["GPT2LMHeadModel"] | 50265 | float32 | 4.19.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
3,609 | AnkesAccount/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.40.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
3,610 | AnkesAccount/xlm-roberta-base-finetuned-panx-de | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,611 | Ankhitan/1000-model5 | segformer | ["SegformerForSemanticSegmentation"] | null | float32 | 4.20.1 | null | null | null | [1, 2, 5, 8] | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,612 | Ankit393/skin_lesion | vit | ["ViTForImageClassification"] | null | float32 | 4.46.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
3,613 | Ankita802/test_part1 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.40.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,614 | Ankitja/fun2 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
3,615 | Ankur464221/t5-small-finetuned-transcripts | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.30.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,616 | Ankur464221/t5-small-transcripts | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.33.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,617 | Ankur464221/transcripts-t5-small-finetuned | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.33.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,618 | Ankur87/Llama2_Time_series_forecasting_6.0 | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.41.2 | 2,048 | 5,632 | 22 | 32 | 4 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
3,619 | Ankur87/Llama2_Time_series_forecasting_7.0 | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.41.2 | 2,048 | 5,632 | 22 | 32 | 4 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
3,620 | Ankur9514/Fine_Tune_T5_Model_Review_Summarization | t5 | ["T5ForConditionalGeneration"] | 32128 | null | 4.38.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,621 | Ankursingh/Qwen2.5-7B-Instruct-1M | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.47.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 10,000,000 | 0.00001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
3,622 | AnkushJindal28/bio-gpt-3.2 | gpt2 | ["GPT2ForTokenClassification"] | 42384 | float32 | 4.38.2 | null | 4,096 | null | null | null | gelu | null | True | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
3,623 | Anmol0130/bacardi_bottle_prediction | vit | ["ViTForImageClassification"] | null | float32 | 4.29.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
3,624 | Anmol0130/bacardi_bottles | vit | ["ViTForImageClassification"] | null | float32 | 4.29.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
3,625 | Anmol0130/bacardi_bottles_2 | vit | ["ViTForImageClassification"] | null | float32 | 4.29.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
3,626 | AnnOl/my_awesome_qa_model | distilbert | ["DistilBertForQuestionAnswering"] | 30522 | float32 | 4.42.4 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
3,627 | Anna-Kulinchenko/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,628 | Anna-UoC/bert_finetuned_banking77test | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.26.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,629 | Anna15/Reshard1 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.46.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
3,630 | Anna1MED/BERT | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,631 | Anna1MED/PubMedBERT | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,632 | Anna1MED/Screening_NER_simple | bert | ["BertForSequenceClassification"] | 28895 | float32 | 4.27.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,633 | Anna1MED/Screening_baseline_simple_text_stratified | bert | ["BertForSequenceClassification"] | 28895 | float32 | 4.27.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,634 | Anna1MED/Screening_sent_NER_simple_stratified | bert | ["BertForSequenceClassification"] | 28895 | float32 | 4.27.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,635 | Anna567/inner-clf-v10 | bert | ["BertForSequenceClassification"] | 55083 | float32 | 4.50.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,636 | AnnaBurikova/llm-course-hw1 | null | [] | 1024 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,637 | AnnaBurikova/trainer_output | llama | ["LlamaForSequenceClassification"] | 49152 | float32 | 4.50.0 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | medium |
3,638 | AnnaR/literature_summarizer | bart | ["BartForConditionalGeneration"] | 50264 | null | 4.17.0 | null | null | 1 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
3,639 | AnnaelleMyriam/MNLP_M2_dpo_1024_beta0.5_2e-5 | qwen3 | ["Qwen3ForCausalLM"] | 151936 | bfloat16 | 4.51.3 | 1,024 | 3,072 | 28 | 16 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 64 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.352322 | small | very_long |
3,640 | AnnaelleMyriam/MNLP_M2_dpo_1024_beta0.5_2e-5_check30000 | qwen3 | ["Qwen3ForCausalLM"] | 151936 | float32 | 4.51.3 | 1,024 | 3,072 | 28 | 16 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 64 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.352322 | small | very_long |
3,641 | AnnaelleMyriam/MNLP_M2_dpo_model | qwen3 | ["Qwen3ForCausalLM"] | 151936 | float32 | 4.51.3 | 1,024 | 3,072 | 28 | 16 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 64 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.352322 | small | very_long |
3,642 | Annamaziarz1/finetuning-distilbert-sentiment-model | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.27.4 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
3,643 | Annastya/bart | mbart | ["MBartForConditionalGeneration"] | 250054 | float32 | 4.40.2 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
3,644 | AnneMona/esm2_t12_35M_UR50D-finetuned-localization | esm | ["EsmForSequenceClassification"] | 33 | float32 | 4.35.2 | 480 | 1,920 | 12 | 20 | null | gelu | null | True | 1,026 | null | null | 0.02 | null | null | null | 24 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.033178 | small | short |
3,645 | AnnekaThiesson/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.16.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
3,646 | Anni0401/Christian_Lindner_llama_80B | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.51.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
3,647 | Annie-Li/whisper_medium_500 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.27.0.dev0 | null | null | 24 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
3,648 | Annie-Li/whisper_medium_mix | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.27.0.dev0 | null | null | 24 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
3,649 | Annieee215/Deepseek-R-Distilled-Llama-70B-awq-shortGPT10 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.47.1 | 8,192 | 28,672 | 70 | 64 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 56.371446 | large | very_long |
3,650 | Annieee215/Deepseek-R-Distilled-Llama-70B-shortGPT10 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.47.1 | 8,192 | 28,672 | 70 | 64 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 56.371446 | large | very_long |
3,651 | Annvasileva/test-bert-finetuned-squad-accelerate | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.34.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,652 | Anon1s/8bit_deepseek | llama | ["LlamaForCausalLM"] | 32256 | float16 | 4.38.2 | 2,048 | 5,504 | 24 | 16 | 16 | silu | 0 | True | 16,384 | 100,000 | 0.000001 | 0.02 | 32013 | 32021 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
3,653 | Anon25/DialoGPT-Medium-BaymaxBot | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.21.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
3,654 | Anony100/FashionVLM | null | [] | 58498 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,655 | AnonyResearcher/codellama_13b_instruct_AdaDecode_gsm8k | llama | ["MultiHeadLlamaForCausalLM"] | 32016 | bfloat16 | 4.43.2 | 5,120 | 13,824 | 40 | 40 | 40 | silu | 0 | True | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | long |
3,656 | AnonyResearcher/codellama_13b_instruct_AdaDecode_humaneval | llama | ["MultiHeadLlamaForCausalLM"] | 32016 | bfloat16 | 4.43.2 | 5,120 | 13,824 | 40 | 40 | 40 | silu | 0 | True | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | long |
3,657 | AnonyResearcher/codellama_34b_instruct_AdaDecode_gsm8k | llama | ["MultiHeadLlamaForCausalLM"] | 32000 | bfloat16 | 4.43.2 | 8,192 | 22,016 | 48 | 64 | 8 | silu | 0 | True | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 38.654706 | large | long |
3,658 | AnonyResearcher/codellama_34b_instruct_AdaDecode_humaneval | llama | ["MultiHeadLlamaForCausalLM"] | 32000 | bfloat16 | 4.43.2 | 8,192 | 22,016 | 48 | 64 | 8 | silu | 0 | True | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 38.654706 | large | long |
3,659 | AnonyResearcher/codellama_34b_instruct_AdaDecode_xsum | llama | ["MultiHeadLlamaForCausalLM"] | 32000 | bfloat16 | 4.43.2 | 8,192 | 22,016 | 48 | 64 | 8 | silu | 0 | True | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 38.654706 | large | long |
3,660 | AnonyResearcher/llama_3.1_8b_instruct_AdaDecode_humaneval | llama | ["MultiHeadLlamaForCausalLM"] | 128256 | bfloat16 | 4.43.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
3,661 | Anonymezzh/qwen2 | qwen2 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,662 | Anonymous-3982/OICT_GRPO_Qwen2.5VL_7B_Full | qwen2_5_vl | ["Qwen2_5_VLForConditionalGeneration"] | 152064 | bfloat16 | 4.49.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 128,000 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
3,663 | Anonymous-G/Genixer-shikra-7b | Genixer_shikra | ["ShikraLlamaForCausalLM"] | 32003 | float32 | 4.28.0.dev0 | 4,096 | 11,008 | 32 | 32 | null | silu | null | False | 2,048 | null | 0.000001 | 0.02 | 0 | 1 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 6.442451 | medium | medium |
3,664 | Anonymous-G/llava-Genixer-915K-FT-8K-v1.5-7b | llava | ["LlavaLlamaForCausalLM"] | 32000 | bfloat16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
3,665 | Anonymous2023/codet5-base-kg | t5 | ["T5ForConditionalGeneration"] | 32100 | float32 | 4.10.2 | null | null | null | null | null | null | null | True | null | null | null | null | 1 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,666 | Anonymous2023/codet5-large-kg | t5 | ["T5WithLMHeadModel"] | 32100 | null | 4.19.2 | null | null | null | null | null | null | null | True | null | null | null | null | 1 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,667 | Anonymous2023/codet5-small-kg | t5 | ["T5ForConditionalGeneration"] | 32100 | float32 | 4.10.2 | null | null | null | null | null | null | null | True | null | null | null | null | 1 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,668 | AnonymousCS/BERT_hateval | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,669 | AnonymousCS/BERT_offenseval | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,670 | AnonymousCS/HateBERT | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,671 | AnonymousCS/HateBERT-Twitter-toxicity | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,672 | AnonymousCS/HateBERT_abuseval | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,673 | AnonymousCS/HateBERT_hateval | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,674 | AnonymousCS/HateBERT_offenseval | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,675 | AnonymousCS/germanic_antielite | bert | ["BertForSequenceClassification"] | 119547 | float32 | 4.46.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,676 | AnonymousCS/germanic_peoplecentric | bert | ["BertForSequenceClassification"] | 119547 | float32 | 4.46.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,677 | AnonymousCS/populism_model180 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,678 | AnonymousCS/populism_model181 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,679 | AnonymousCS/populism_model182 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,680 | AnonymousCS/populism_model183 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,681 | AnonymousCS/populism_model184 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,682 | AnonymousCS/populism_model185 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,683 | AnonymousCS/populism_model186 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,684 | AnonymousCS/populism_model187 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,685 | AnonymousCS/populism_model189 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,686 | AnonymousCS/populism_model190 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,687 | AnonymousCS/populism_model191 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,688 | AnonymousCS/populism_model192 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,689 | AnonymousCS/populism_model193 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,690 | AnonymousCS/populism_model194 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,691 | AnonymousCS/populism_model195 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,692 | AnonymousCS/populism_model196 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,693 | AnonymousCS/populism_model197 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,694 | AnonymousCS/populism_model198 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,695 | AnonymousCS/populism_model199 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,696 | AnonymousCS/populism_model202 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,697 | AnonymousCS/populism_model203 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,698 | AnonymousCS/populism_model204 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,699 | AnonymousCS/populism_model207 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.