index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
3,700 | AnonymousCS/populism_model208 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,701 | AnonymousCS/populism_model209 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,702 | AnonymousCS/populism_model213 | bert | ["BertForSequenceClassification"] | 105879 | float32 | 4.49.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,703 | AnonymousCS/populism_model214 | bert | ["BertForSequenceClassification"] | 105879 | float32 | 4.49.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,704 | AnonymousCS/populism_model217 | bert | ["BertForSequenceClassification"] | 105879 | float32 | 4.49.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,705 | AnonymousCS/populism_model219 | bert | ["BertForSequenceClassification"] | 105879 | float32 | 4.49.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,706 | AnonymousCS/populism_model220 | bert | ["BertForSequenceClassification"] | 105879 | float32 | 4.49.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,707 | AnonymousCS/populism_model230 | bert | ["BertForSequenceClassification"] | 105879 | float32 | 4.49.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,708 | AnonymousCS/populism_model237 | bert | ["BertForSequenceClassification"] | 105879 | float32 | 4.49.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,709 | AnonymousCS/populism_model239 | bert | ["BertForSequenceClassification"] | 105879 | float32 | 4.49.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,710 | AnonymousCS/refined-bert-base-uncased-Twitter-toxicity | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,711 | AnonymousOrca/parser_final | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.48.1 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
3,712 | AnonymousOrca/parser_new_era_V10 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.49.0 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
3,713 | AnonymousOrca/parser_new_era_V5 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.49.0 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
3,714 | AnonymousOrca/parser_new_era_V6 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.49.0 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
3,715 | AnonymousOrca/paser_new_modified | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.49.0 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
3,716 | AnonymousPage/checkpoint-all | pixel | ["PIXELForSequenceClassification"] | null | float32 | 4.17.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
3,717 | AnonymousPaper/ITS_Bert_french_difficulty | camembert | ["CamembertForSequenceClassification"] | 32005 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 5 | 6 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,718 | AnonymousResearcher2025/Llama3_8B_Instruct_QSpec | llama_quarot | ["QuarotLlamaForCausalLM"] | 128256 | float16 | 4.37.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
3,719 | AnonymousSub/BART-LARGE_MedQuAD_question_generation | bart | ["BartModel"] | 50265 | float32 | 4.28.1 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
3,720 | AnonymousSub/CLIP_bloodmnist_224x224_img_then_txt_bert-base-uncased_1_epochs | bert | ["BertModel"] | 30522 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,721 | AnonymousSub/CLIP_bloodmnist_224x224_img_then_txt_bert-base-uncased_1_epochs_10000_samples | bert | ["BertModel"] | 30522 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,722 | AnonymousSub/ONLY_HTM_FPDM_roberta | roberta | ["RobertaModel"] | 50265 | float32 | 4.33.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,723 | AnonymousSub/Recipe_QA_CLIP_and_ranking_with_attention_ViT_Finetuned_1_epochs_bert-base-uncased | bert | ["BertModel"] | 30522 | float32 | 4.27.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,724 | AnonymousSub/Recipe_QA_CLIP_and_ranking_with_attention_ViT_Finetuned_1_epochs_roberta-base | roberta | ["RobertaModel"] | 50265 | float32 | 4.27.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,725 | AnonymousSub/SQuAD_CLIP_text_8 | roberta | ["RobertaForQuestionAnswering"] | 50265 | float32 | 4.33.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,726 | AnonymousSub/Sci_HTM_2_FPDM_bert | bert | ["BertModel"] | 30522 | float32 | 4.33.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,727 | AnonymousSub/Sci_HTM_FPDM_bert | bert | ["BertModel"] | 30522 | float32 | 4.33.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,728 | AnonymousSub/bloodmnist_224x224_img_then_txt_bert-base-uncased_3_epochs | bert | ["BertModel"] | 30522 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,729 | AnonymousSub/breastmnist_224x224_img_then_txt_bert-base-uncased_3_epochs | bert | ["BertModel"] | 30522 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,730 | AnonymousSub/end2end_breastmnist_224x224_img_then_txt_bert-base-uncased | bert | ["BertModel"] | 30522 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,731 | AnonymousSub/fpdm_models_scibert_hybrid_epochs_4 | bert | ["BertModel"] | 31090 | float32 | 4.11.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,732 | AnonymousSub/mask_1_or_2_steps_RECIPES_HALVED_pretraining_plus_contr_recipes-roberta-large | roberta | ["RobertaModel"] | 50265 | float32 | 4.27.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,733 | AnonymousSub/mask_step_pretraining_bert-base-uncased_epochs_1 | bert | ["BertModel"] | 30522 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,734 | AnonymousSub/mask_step_pretraining_recipe-bert-base-uncased_epochs_1 | bert | ["BertModel"] | 30522 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,735 | AnonymousSub/mask_step_pretraining_recipes-roberta-base_epochs_1 | roberta | ["RobertaModel"] | 50265 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,736 | AnonymousSub/mask_step_pretraining_recipes-roberta-base_squadv2_epochs_3 | roberta | ["RobertaForQuestionAnswering"] | 50265 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,737 | AnonymousSub/mask_step_pretraining_roberta-base_epochs_1 | roberta | ["RobertaModel"] | 50265 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,738 | AnonymousSub/mask_step_pretraining_roberta-base_squadv2_epochs_3 | roberta | ["RobertaForQuestionAnswering"] | 50265 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,739 | AnonymousSub/mask_step_pretraining_spanbert-base-cased_epochs_1 | bert | ["BertModel"] | 28996 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,740 | AnonymousSub/output_HTM_FPDM_roberta | roberta | ["RobertaForQuestionAnswering"] | 50265 | float32 | 4.33.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,741 | AnonymousSub/output_ONLY_HTM_FPDM_roberta | roberta | ["RobertaForQuestionAnswering"] | 50265 | float32 | 4.33.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,742 | AnonymousSub/recipe_RL_pretraining_recipes-roberta-base | roberta | ["RobertaModel"] | 50265 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,743 | AnonymousSub/recipe_RL_pretraining_roberta-base | roberta | ["RobertaModel"] | 50265 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,744 | AnonymousSub/recipe_triplet_bert-base-uncased_VERYHARD_epochs_1 | bert | ["BertModel"] | 30522 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,745 | AnonymousSub/recipe_triplet_mask_step_pretraining_bert-base-uncased_epochs_1_epochs_1 | bert | ["BertModel"] | 30522 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,746 | AnonymousSub/recipe_triplet_mask_step_pretraining_recipe-bert-base-uncased_epochs_1_epochs_1 | bert | ["BertModel"] | 30522 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,747 | AnonymousSub/recipe_triplet_mask_step_pretraining_recipes-roberta-base_epochs_1_epochs_1 | roberta | ["RobertaModel"] | 50265 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,748 | AnonymousSub/recipe_triplet_mask_step_pretraining_roberta-base_epochs_1_epochs_1 | roberta | ["RobertaModel"] | 50265 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,749 | AnonymousSub/recipe_triplet_recipe-bert-base-uncased_VERYHARD_epochs_1 | bert | ["BertModel"] | 30522 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,750 | AnonymousSub/recipe_triplet_recipes-roberta-base_VERYHARD_epochs_1 | roberta | ["RobertaModel"] | 50265 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,751 | AnonymousSub/recipe_triplet_roberta-base_VERYHARD_epochs_1 | roberta | ["RobertaModel"] | 50265 | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,752 | AnonymousSub/recipes-roberta-base-no-ingr | roberta | ["RobertaForMaskedLM"] | 50265 | float32 | 4.11.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,753 | AnonymousSub/recipes-roberta-base-tokenwise-token-and-step-losses_no_ingr | roberta | ["RobertaModel"] | 50265 | float32 | 4.11.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,754 | AnonymousSub/recipes-roberta-base-tokenwise-token-and-step-losses_with_ingr | roberta | ["RobertaModel"] | 50265 | float32 | 4.11.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,755 | AnonymousSub/recipes-roberta-base-tokenwise-token-and-step-losses_with_ingr_only_step_loss_full_lr_decay | roberta | ["RobertaModel"] | 50265 | float32 | 4.11.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,756 | AnonymousSub/recipes-roberta-base-tokenwise-token-and-step-losses_with_pos_with_ingr | roberta | ["RobertaModel"] | 50265 | float32 | 4.11.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,757 | AnonymousSubmissionOnly/RobustGen | t5 | ["T5ForConditionalGeneration"] | 32442 | float32 | 4.31.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,758 | Anopheles/summary_cz_eurlex | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.35.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,759 | Anoshor/prism-v2 | llama | ["LlamaForCausalLM"] | 128258 | float16 | 4.45.1 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
3,760 | AnotherMiner/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-sizable_eager_shrimp | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.51.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
3,761 | AnotherMiner/Qwen3-0.6B-Gensyn-Swarm-prickly_beaked_cat | qwen3 | ["Qwen3ForCausalLM"] | 151936 | float32 | 4.51.3 | 1,024 | 3,072 | 28 | 16 | 8 | silu | 0 | True | 40,960 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.352322 | small | very_long |
3,762 | Anoubhav/sd-class-butterflies-32 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,763 | AnranZZ/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,764 | Anren/PPO-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,765 | Anrom97/LLama-3.1-WIS-v6.2-GGUF | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
3,766 | Ans0nWr0ng/Qwen2.5-14b-Cantonese | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.51.3 | 5,120 | 13,824 | 48 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 15.099494 | large | very_long |
3,767 | Ansah-AI/E1 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.48.3 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
3,768 | AnsenIO/dia_ttv | dia | ["DiaForConditionalGeneration"] | null | float32 | 4.53.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | 1026 | 1024 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
3,769 | AnsenIO/my-finetuned-bert | bert | ["BertModel"] | 28996 | float32 | 4.53.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,770 | AnshDube77/distilroberta-dark-pattern | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.35.2 | 768 | 3,072 | 6 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
3,771 | Anshler/selective-blind-guessing | gpt2 | ["GPT2Model"] | 50257 | float32 | 4.35.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
3,772 | Anshler/vietnamese-poem-classifier | bert | ["BertForSequenceClassification"] | 32000 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,773 | AnshuPatel-Vikalp/Nutritions_Llama-3.2-7b-chat-finetune2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,774 | Anshul2000s/twitter_suicide_detection_model | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.32.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,775 | Anshul99/masked-stance-detection | t5 | ["T5ForConditionalGeneration"] | 32000 | float32 | 4.42.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,776 | Anshul99/masked-stance-detection_v4 | t5 | ["T5ForConditionalGeneration"] | 32000 | float32 | 4.42.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,777 | AnshulProjects/results | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.52.0.dev0 | 768 | null | null | null | null | gelu | 0.1 | True | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,778 | Ansoi/birdstruct2 | llama | ["LlamaForCausalLM"] | 32032 | float16 | 4.33.1 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
3,779 | Ansoi/chatp | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.33.1 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
3,780 | Ansoi/kundachat | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.33.1 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
3,781 | Ansoi/peter | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.36.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
3,782 | Ansu/whisper-small-dv | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.48.3 | null | null | 12 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
3,783 | Ant3wan95/t5-finetuned | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.47.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,784 | AntOG96/llama | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.51.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
3,785 | AntResearchNLP/ViLaSR | qwen2_5_vl | ["Qwen2_5_VLForConditionalGeneration"] | 152064 | float32 | 4.51.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 128,000 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
3,786 | AntResearchNLP/ViLaSR-cold-start | qwen2_5_vl | ["Qwen2_5_VLForConditionalGeneration"] | 152064 | bfloat16 | 4.51.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 128,000 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
3,787 | AntX-ai/AntX-13B | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.29.0.dev0 | 5,120 | 13,824 | 40 | 40 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 12.582912 | large | medium |
3,788 | AntX-ai/AntX-7B | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.29.0.dev0 | 4,096 | 11,008 | 32 | 32 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 6.442451 | medium | medium |
3,789 | AntaraIIITD/my_NMT_model2 | marian | ["MarianMTModel"] | 58101 | float32 | 4.28.1 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
3,790 | AntaraIIITD/wav2vec_on_grid | wav2vec2 | ["Wav2Vec2ForCTC"] | 32 | float32 | 4.27.1 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
3,791 | Antdochi/results | resnet | ["ResNetForImageClassification"] | null | float32 | 4.34.0 | null | null | null | null | null | relu | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
3,792 | Anteia/Qwen2-7B-Instruct-v1.2 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float16 | 4.45.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
3,793 | Anteia/Qwen2-7B-Instruct-v2 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float16 | 4.45.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
3,794 | Anteia/Qwen2-7B-Instruct-v2-dpo | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float16 | 4.45.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
3,795 | AnthoneoJ/ms6-train | convnextv2 | ["ConvNextV2ForImageClassification"] | null | float32 | 4.45.1 | null | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,796 | Anthony10/whisper-small-hi | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.46.0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
3,797 | Anthos23/FS-finbert-fine-tuned-f1 | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.17.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,798 | Anthuni/Final_Thesis_Model | vit | ["ViTForImageClassification"] | null | float32 | 4.41.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
3,799 | Antiraedus/sd-class-butterflies-32 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.