index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
4,100 | Ariffiq99/e_care_KUCI_albert_base_finetuned | albert | ["AlbertForMultipleChoice"] | 30000 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu_new | null | null | 512 | null | null | 0.02 | 2 | 3 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,101 | Ariffiq99/e_care_KUCI_bert_base_uncased_finetuned | bert | ["BertForMultipleChoice"] | 30522 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,102 | Ariffiq99/e_care_KUCI_xlm_roberta_base_finetuned | xlm-roberta | ["XLMRobertaForMultipleChoice"] | 250002 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,103 | Ariffiq99/e_care_KUCI_xlm_roberta_large_finetuned | xlm-roberta | ["XLMRobertaForMultipleChoice"] | 250002 | float32 | 4.41.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
4,104 | Ariffiq99/e_care_albert_base_finetuned | albert | ["AlbertForMultipleChoice"] | 30000 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu_new | null | null | 512 | null | null | 0.02 | 2 | 3 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,105 | Ariffnaz/a01-suicide-bert-huggingface-finetune | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.26.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,106 | Ariffnaz/a02-suicide-bert-huggingface-finetune | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.26.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,107 | Arihant-Bhandari/feedflash-flan-t5 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.51.3 | null | null | null | null | null | null | null | False | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,108 | Arikkod/PPO-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,109 | ArinaKomar/distilT5-finetuned-cnn | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.40.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,110 | ArinaOwl/ast-ser-ru | audio-spectrogram-transformer | ["ASTForAudioClassification"] | null | float32 | 4.29.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
4,111 | Arindam75/a2c-AntBulletEnv-v0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,112 | Arindam75/poca-SoccerTwos | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,113 | Arindam75/rl_course_vizdoom_health_gathering_supreme | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,114 | ArindamSingh/Llama-3.2-3B-Instruct-FineTome100k-16bit | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.52.4 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,115 | ArindamSingh/gemma-3-1b-it-medical-o1-reasoning-finetune-16bit | gemma3_text | ["Gemma3ForCausalLM"] | 262144 | float16 | 4.51.3 | 1,152 | 6,912 | 26 | 4 | 1 | null | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 2 | 106 | null | 288 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.414056 | small | very_long |
4,116 | Aris-AI/Aris-Qwen1.5-14B-Chat-Agent-DPO-16K-MLX-20240531 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.41.2 | 5,120 | 13,696 | 40 | 40 | 40 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 12.582912 | large | very_long |
4,117 | ArisuNguyen/bart_finetuned_5e_5_7epoch | bart | ["BartForQuestionAnswering"] | 50265 | float32 | 4.27.4 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
4,118 | ArisuNguyen/processed_bart | bart | ["BartForQuestionAnswering"] | 50265 | float32 | 4.29.2 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
4,119 | ArisuNguyen/retrain_non_seg_bart | bart | ["BartForQuestionAnswering"] | 50265 | float32 | 4.30.2 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
4,120 | ArisuNguyen/retrain_non_seg_mbart | mbart | ["MBartForQuestionAnswering"] | 250054 | float32 | 4.30.2 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
4,121 | ArisuNguyen/retrain_seg_bart | bart | ["BartForQuestionAnswering"] | 50265 | float32 | 4.30.2 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
4,122 | ArisuNguyen/retrain_seg_mbart | mbart | ["MBartForQuestionAnswering"] | 250054 | float32 | 4.30.2 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
4,123 | ArisuNguyen/uitviquad_noseg_bart | bart | ["BartForQuestionAnswering"] | 50265 | float32 | 4.28.1 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
4,124 | AritraRay2005/SuperiorLLM | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
4,125 | Arix-ALIMAGNIDOKPO/Mistral_Instruct_IDSR | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.37.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
4,126 | Arix-ALIMAGNIDOKPO/Mistral_Instruct_Isheero_v1 | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.38.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
4,127 | AriyanH22/frame_classification_longformer | longformer | ["LongformerForSequenceClassification"] | 50265 | float32 | 4.32.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 4,098 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | medium |
4,128 | AriyanH22/framing_classification_longformer_30_augmented_multi_undersampled_second | longformer | ["LongformerForSequenceClassification"] | 50265 | float32 | 4.32.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 4,098 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | medium |
4,129 | ArjTheHacker/vit_detection_of_retinology | vit | ["ViTForImageClassification"] | null | float32 | 4.46.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
4,130 | ArjTheHacker/vit_detection_of_skin_diseases | vit | ["ViTForImageClassification"] | null | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
4,131 | ArjanvD95/animals_5 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,132 | ArjanvD95/animals_dbnl_bertje | bert | ["BertForTokenClassification"] | 30073 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,133 | ArjanvD95/animals_mdeberta_s42 | deberta-v2 | ["DebertaV2ForTokenClassification"] | 251000 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,134 | ArjanvD95/animals_nf_gysbert_512 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,135 | ArjanvD95/animals_nf_mdeberta_512 | deberta-v2 | ["DebertaV2ForTokenClassification"] | 251000 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,136 | ArjanvD95/by_the_horns_D43Gysbert | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.42.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,137 | ArjanvD95/by_the_horns_D44Gysbert | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.42.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,138 | ArjanvD95/by_the_horns_D46Gysbert | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.42.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,139 | ArjanvD95/by_the_horns_T43Gysbert | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.42.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,140 | ArjanvD95/by_the_horns_T44Gysbert | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.42.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,141 | ArjanvD95/by_the_horns_T45Gysbert | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.42.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,142 | ArjanvD95/by_the_horns_T46Gysbert | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.42.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,143 | Arjav/TOS-Longformer | led | ["LEDForConditionalGeneration"] | 50265 | float32 | 4.30.0 | null | null | 12 | null | null | null | 0 | False | null | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,144 | Arjs/Llama-2-7b-chatbot-finetune | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
4,145 | Arjun-G-Ravi/GPT2-Alpaca | gpt2 | ["GPT2LMHeadModel"] | 50259 | float32 | 4.32.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
4,146 | Arjun9/t5_long_tglobal_base_samsum | longt5 | ["LongT5ForConditionalGeneration"] | 32128 | float32 | 4.40.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,147 | ArjunK47/zinemind_v6 | detr | ["DetrModel"] | null | float32 | 4.44.2 | null | null | 6 | null | null | null | 0 | null | 1,024 | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
4,148 | Ark2044/eye | resnet | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,149 | ArkadiusDS/polbert-base-polish-manipulation | bert | ["BertForSequenceClassification"] | 50000 | float32 | 4.47.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,150 | ArkadiusDS/polbert-large-polish-manipulation | bert | ["BertForSequenceClassification"] | 50000 | float32 | 4.47.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
4,151 | ArkadiusDS/polberta-base-polish-manipulation | roberta | ["RobertaForSequenceClassification"] | 50001 | float32 | 4.47.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,152 | Arkajyoti/Arkajyoti-Mistral-7B-v0.1-authors-random-standardized-many-random-names-easy | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.42.4 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
4,153 | Arkana08/B-NIMITA-L3-8B-v0.02 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
4,154 | Arkana08/Mythorica-L3-8B | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
4,155 | Arkash/my_awesome_qa_model | distilbert | ["DistilBertForQuestionAnswering"] | 30522 | float32 | 4.28.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
4,156 | Arkintea/Llama-2-7b-chat-finetune | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
4,157 | Arkong/chatglm2-6b-torchkeras | chatglm | ["ChatGLMForConditionalGeneration"] | 65024 | float16 | 4.35.0 | 4,096 | null | null | 32 | null | null | 0 | True | null | null | null | null | null | 2 | false | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
4,158 | Arkya/xlm-roberta-base-finetuned-panx-de-fr | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,159 | Arkya/xlm-roberta-base-finetuned-panx-fr | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,160 | Arlethh/09arly | vit | ["ViTForImageClassification"] | null | float32 | 4.30.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
4,161 | ArliAI/Llama-3.1-8B-ArliAI-Formax-v1.0 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
4,162 | ArliAI/Llama-3.1-8B-ArliAI-Indo-Formax-v1.0 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
4,163 | ArliAI/Mistral-Small-22B-ArliAI-RPMax-v1.1 | mistral | ["MistralForCausalLM"] | 32768 | bfloat16 | 4.45.0.dev0 | 6,144 | 16,384 | 56 | 48 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 6 | false | null | null | false | false | false | false | false | false | true | false | true | true | 25.367151 | large | very_long |
4,164 | ArliAI/Qwen3-30B-A3B-ArliAI-RpR-v4-Fast | qwen3_moe | ["Qwen3MoeForCausalLM"] | 151936 | bfloat16 | 4.51.3 | 2,048 | 6,144 | 48 | 32 | 4 | silu | 0 | True | 40,960 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | false | 64 | 8 | false | null | 8 | false | false | false | false | false | false | false | false | true | true | 2.415919 | medium | very_long |
4,165 | Armaan11/14B_GRPO_checkpoint-250-merged | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.50.1 | 5,120 | 13,824 | 48 | 40 | 8 | silu | 0 | True | 131,072 | 1,000,000 | 0.00001 | 0.02 | 151643 | 151643 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 15.099494 | large | very_long |
4,166 | Armaan11/checkpoint-110-merged | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.47.1 | 5,120 | 13,824 | 48 | 40 | 8 | silu | 0 | True | 131,072 | 1,000,000 | 0.00001 | 0.02 | 151643 | 151643 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 15.099494 | large | very_long |
4,167 | Armaan11/checkpoint-130-merged | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.50.1 | 5,120 | 13,824 | 48 | 40 | 8 | silu | 0 | True | 131,072 | 1,000,000 | 0.00001 | 0.02 | 151643 | 151643 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 15.099494 | large | very_long |
4,168 | Armaan11/checkpoint-150-merged | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.50.1 | 5,120 | 13,824 | 48 | 40 | 8 | silu | 0 | True | 131,072 | 1,000,000 | 0.00001 | 0.02 | 151643 | 151643 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 15.099494 | large | very_long |
4,169 | Armaan11/checkpoint-200-merged | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.50.1 | 5,120 | 13,824 | 48 | 40 | 8 | silu | 0 | True | 131,072 | 1,000,000 | 0.00001 | 0.02 | 151643 | 151643 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 15.099494 | large | very_long |
4,170 | Armaan11/deepseek-r1-distill-qwen-7b_s | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.49.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 131,072 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
4,171 | ArmaanSeth/Llama-2-7b-chat-hf-shards-mental-health-counselling | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.38.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
4,172 | ArmaanSeth/minilm-finetuned-emotion | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.32.1 | 384 | 1,536 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.021234 | small | short |
4,173 | Arman51/Qwen2-0.5B-SFT-full | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.51.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
4,174 | ArmanAsq/aya-23-35B-IF | cohere | ["CohereForCausalLM"] | 255029 | float32 | 4.42.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 10,000 | null | 0.02 | 5 | 255001 | null | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
4,175 | ArmandoRockYourCloud/T58000 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.52.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,176 | Armandodelca/ClasificadorV7 | bert | ["BertForSequenceClassification"] | 31002 | float32 | 4.41.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,177 | Armandodelca/ClasificadorV8 | bert | ["BertForSequenceClassification"] | 31002 | float32 | 4.41.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,178 | Armandodelca/Prototipo_5_EMI | bert | ["BertForSequenceClassification"] | 31002 | float32 | 4.40.1 | 768 | 3,072 | 12 | 12 | null | gelu | 0.2 | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,179 | ArmelR/Stack10K2048 | gpt_bigcode | ["GPTBigCodeForCausalLM"] | 49280 | float32 | 4.29.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
4,180 | ArmelR/starcoder-gradio-v2.1 | gpt_bigcode | ["GPTBigCodeForCausalLM"] | 49152 | float16 | 4.28.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 0 | 0 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
4,181 | ArmsOpen/my_awesome_model | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.39.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
4,182 | ArmurAI/Pentest_AI | mistral | ["MistralForCausalLM"] | 32002 | bfloat16 | 4.34.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
4,183 | Arnaba2002/bargain-bot-modell | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.52.4 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
4,184 | Arnaudmkonan/xlm-roberta-base-finetuned-panx-de | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.11.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,185 | Arnaudmkonan/xlm-roberta-base-finetuned-panx-de-fr | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.11.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,186 | Arnaudmkonan/xlm-roberta-base-finetuned-panx-fr | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.11.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,187 | Arnav0400/SmolVLM2-500M-Video-Instruct-video-feedback | smolvlm | ["SmolVLMForConditionalGeneration"] | 49280 | bfloat16 | 4.52.0.dev0 | null | null | null | null | null | null | null | False | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
4,188 | Arnav0805/bert-phishing-classifier_teacher | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.47.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,189 | ArnavL/roberta-10M-imdb-0 | roberta | ["RobertaForMaskedLM"] | 50265 | float32 | 4.21.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,190 | ArnavL/roberta-one_mil-imdb-0 | roberta | ["RobertaForMaskedLM"] | 50265 | float32 | 4.20.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,191 | ArnavL/twteval-pretrained | bert | ["BertForMaskedLM"] | 30522 | float32 | 4.17.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,192 | ArnavL/yelp-pretrained | bert | ["BertForMaskedLM"] | 30522 | float32 | 4.16.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,193 | Arnavaz/gpt2-arnavaz-beta | gpt2 | ["GPT2LMHeadModel"] | 25000 | float32 | 4.12.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 8 | 9 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
4,194 | Arnavbhatia/Llama-2-7b-chat-finetune | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
4,195 | ArneL2206/a2c-AntBulletEnv-v0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,196 | ArneL2206/a2c-PandaReachDense-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,197 | Arnic/Gemma-2-2b-it-chat-MedQuad | gemma2 | ["Gemma2ForCausalLM"] | 256002 | float16 | 4.42.3 | 2,304 | 9,216 | 26 | 8 | 4 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 256000 | 256001 | null | 288 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.656226 | medium | long |
4,198 | ArnoChristie/gpt2-course-generator | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.51.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
4,199 | Arodrigo/ClasificadorCorreoSoporteDistilEspanol | distilbert | ["DistilBertForSequenceClassification"] | 31002 | float32 | 4.42.4 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.