index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
2,200 | Aivesa/349aacef-3660-4206-b561-6d7b5d5a9b62 | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.47.1 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
2,201 | Aivesa/45d81bfd-b7dd-4169-b979-bdc0739428a0 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.47.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
2,202 | Aivesa/4dc6507a-6049-4f97-a49d-9c4d38d41625 | gpt_neox | ["GPTNeoXForCausalLM"] | 50304 | float16 | 4.47.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | False | 2,048 | 10,000 | null | 0.02 | 0 | 0 | false | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.30199 | small | medium |
2,203 | Aivesa/514decb9-bd0e-4782-86b5-6131d662a51b | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.47.1 | 16 | 64 | 2 | 4 | 4 | silu | 0 | False | 2,048 | 10,000 | 0.000001 | 0.02 | 0 | 2 | false | 4 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.000006 | small | medium |
2,204 | Aivesa/53638570-6e01-4257-8d08-795d436a977c | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.47.1 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
2,205 | Aivesa/5e50f4e8-6556-4450-9721-0660047938a0 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.47.1 | 16 | 64 | 2 | 4 | 4 | silu | 0 | False | 2,048 | 10,000 | 0.000001 | 0.02 | 0 | 2 | false | 4 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.000006 | small | medium |
2,206 | Aivesa/85a3b62a-09c1-44d6-af05-0f30aee76da5 | mixtral | ["MixtralForCausalLM"] | 32000 | float32 | 4.47.1 | 1,024 | 3,584 | 2 | 32 | 8 | silu | 0 | False | 131,072 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 32 | 4 | false | 8 | 2 | false | false | false | false | false | false | false | true | true | true | 0.025166 | small | very_long |
2,207 | Aivesa/9cc26248-48d3-4b39-8479-e5d69049c582 | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.47.1 | 2,048 | 5,632 | 22 | 32 | 4 | silu | 0 | False | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
2,208 | Aivesa/a05b476c-463b-4946-89c4-d93aca81d7e7 | opt | ["OPTForCausalLM"] | 50272 | float16 | 4.47.1 | 2,048 | null | 24 | 32 | null | null | 0 | False | 2,048 | null | null | null | 2 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 1.20796 | medium | medium |
2,209 | Aivesa/ad5f3025-5478-4dad-b670-bbd401699c03 | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.47.1 | 16 | 64 | 2 | 4 | 4 | silu | 0 | False | 2,048 | 10,000 | 0.000001 | 0.02 | 0 | 2 | false | 4 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.000006 | small | medium |
2,210 | Aivesa/b87a98bd-2eb0-4bfa-8f4e-6c92644bdc42 | codegen | ["CodeGenForCausalLM"] | 51200 | float32 | 4.47.1 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 50256 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,211 | Aivesa/c1093e50-e6e8-4bfe-8dc8-afe74f920cfd | gpt_neox | ["GPTNeoXForCausalLM"] | 50304 | float16 | 4.47.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | False | 2,048 | 10,000 | null | 0.02 | 0 | 0 | false | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.30199 | small | medium |
2,212 | Aivesa/d04a0ea9-5c92-49fd-8fd8-300b97476de2 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.47.1 | 2,048 | 5,504 | 24 | 16 | 16 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151643 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | very_long |
2,213 | Aivesa/d8b02868-3f22-4fc4-bf4b-cd5c2a0ad840 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.47.1 | 16 | 64 | 2 | 4 | 4 | silu | 0 | False | 2,048 | 10,000 | 0.000001 | 0.02 | 0 | 2 | false | 4 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.000006 | small | medium |
2,214 | Aivesa/d9ca8632-4ba2-424e-a050-6fb0f4780371 | mixtral | ["MixtralForCausalLM"] | 32000 | float32 | 4.47.1 | 1,024 | 3,584 | 2 | 32 | 8 | silu | 0 | False | 131,072 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 32 | 4 | false | 8 | 2 | false | false | false | false | false | false | false | true | true | true | 0.025166 | small | very_long |
2,215 | Aivesa/df5493b9-6228-4cae-a15d-b3a2fd8fb6e3 | gpt_neox | ["GPTNeoXForCausalLM"] | 50304 | float16 | 4.47.1 | 512 | 2,048 | 6 | 8 | null | gelu | 0 | False | 2,048 | 10,000 | null | 0.02 | 0 | 0 | false | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.018874 | small | medium |
2,216 | Aivesa/f2a098be-735f-44f7-99ac-6937ddb1e108 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.47.1 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
2,217 | Aivesa/f8549911-86bb-4e2b-8478-525efbbc4415 | gpt_neox | ["GPTNeoXForCausalLM"] | 50304 | float16 | 4.47.1 | 512 | 2,048 | 6 | 8 | null | gelu | 0 | False | 2,048 | 10,000 | null | 0.02 | 0 | 0 | false | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.018874 | small | medium |
2,218 | Aixile/CLIP-ViT-H-14-laion2B-s32B-b79K | clip | ["CLIPModel"] | null | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,219 | Aixos/webbot | bert | ["BertForQuestionAnswering"] | 30522 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,220 | Ajas2002/cover | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.38.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
2,221 | Ajaxpax/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,222 | Ajay2525/wav2vec2-ks | wav2vec2 | ["Wav2Vec2ForSequenceClassification"] | 32 | float32 | 4.44.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
2,223 | AjayMukundS/Llama-2-7b-LTS-finetuned | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,224 | Ajhl1992/lilt-en-funsd | lilt | ["LiltForTokenClassification"] | 50265 | float32 | 4.28.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,225 | Ajit1498/donut-base-sroie | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | 4.41.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,226 | Ajita15/distilgpt2-finetuned-wikitext2 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.40.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
2,227 | Ajohe/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,228 | Ak1104/3_test_q8 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
2,229 | Ak1104/QA_2k | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
2,230 | Ak1104/codellama-7b_rust | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
2,231 | Ak1104/llama-3-test | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
2,232 | Ak1104/prompt1 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
2,233 | Ak128umar/bert-finetuned-ner-accelerate | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.52.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,234 | Ak1R401/wiki-newmm | roberta | ["RobertaForTokenClassification"] | 97982 | float32 | 4.30.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,235 | Ak729/MT5_with_MLP | mt5 | ["MT5EncoderModel"] | 250112 | float32 | 4.44.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
2,236 | Ak729/Model | bert | ["BertForMaskedLM"] | 30522 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,237 | AkRl1/Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,238 | Akajackson/weights | bert | ["BertForSequenceClassification"] | 83828 | float32 | 4.39.3 | 312 | 600 | 3 | 12 | null | gelu | null | True | 2,048 | null | null | 0.02 | null | null | null | 26 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.003504 | small | medium |
2,239 | AkankshaLoria21/gemma-Code-Instruct-Finetune-test | gemma | ["GemmaForCausalLM"] | 256000 | float16 | 4.38.0 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
2,240 | Akash2106/TRAIL | mixformer-sequential | ["MixFormerSequentialForCausalLM"] | 51200 | float32 | 4.34.1 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,241 | Akash2106/phi1_5_sqldatabase | mixformer-sequential | ["MixFormerSequentialForCausalLM"] | 51200 | float32 | 4.34.1 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,242 | Akash24/my_awesome_model | albert | ["AlbertForSequenceClassification"] | 30000 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu_new | null | null | 512 | null | null | 0.02 | 2 | 3 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,243 | Akash82228/results | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.35.0 | null | null | 6 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
2,244 | Akash82228/tmp_trainer | roberta | ["RobertaForQuestionAnswering"] | 50265 | float32 | 4.36.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,245 | Akbar0627/dummy-model | camembert | ["CamembertForMaskedLM"] | 32005 | float32 | 4.49.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 5 | 6 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,246 | Akchunks/MLAgents-Pyramids | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,247 | Akchunks/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,248 | Akeda01/MontirOnlinePro | llama | ["LlamaForCausalLM"] | 128258 | float16 | 4.48.0 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
2,249 | Akenar/speecht5_finetuned_voxpopuli_nl | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.40.2 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
2,250 | Akerrules/swotmodelQA | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.46.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
2,251 | AkhiVasu/distilbert-base-uncased-finetuned-imdb | distilbert | ["DistilBertForMaskedLM"] | 30522 | null | 4.40.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,252 | Akhil-Theerthala/Kuvera-14B-v0.1.0 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.52.4 | 5,120 | 13,824 | 48 | 40 | 8 | silu | 0 | True | 131,072 | 1,000,000 | 0.00001 | 0.02 | 151643 | 151643 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 15.099494 | large | very_long |
2,253 | Akhil-Theerthala/Kuvera-8B-v0.1.0 | qwen3 | ["Qwen3ForCausalLM"] | 151936 | float32 | 4.51.2 | 4,096 | 12,288 | 36 | 32 | 8 | silu | 0 | False | 40,960 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 7.247757 | large | very_long |
2,254 | Akhil0-o/saved_model_body | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,255 | Akhil0-o/saved_model_links | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,256 | Akhil06042002/model_model_super_model | bert | ["BertForMultipleChoice"] | 30522 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,257 | Akhil123/emotions_classifier | vit | ["ViTForImageClassification"] | null | null | 4.31.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
2,258 | Akhil199797/emotion-distilbert-final | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.52.4 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,259 | AkhilGhosh/llama-cnn-210k | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.29.0.dev0 | 4,096 | 11,008 | 32 | 32 | null | silu | null | False | 2,048 | null | 0.000001 | 0.02 | 0 | 1 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 6.442451 | medium | medium |
2,260 | AkhilKashyap1998/gpt2-GPTQ | gpt2 | ["GPT2LMHeadModel"] | 50257 | float16 | 4.37.2 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
2,261 | Akhilamyaka/SAWIT_Translation_Model | mbart | ["MBartForConditionalGeneration"] | 250054 | float32 | 4.48.3 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
2,262 | Akhilsplendid/T5-model | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.32.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
2,263 | Akhilsplendid/bart-1 | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.29.2 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
2,264 | Akhilsplendid/bart-model | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.32.0 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
2,265 | Akhilsplendid/pegasus-1 | pegasus | ["PegasusForConditionalGeneration"] | 96103 | float32 | 4.29.1 | null | null | 16 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
2,266 | Akhilsplendid/pegasus-model | pegasus | ["PegasusForConditionalGeneration"] | 96103 | float32 | 4.32.0 | null | null | 16 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
2,267 | AkiiraBino/rubert-tiny2-russian-sentiment-reviews | bert | ["BertForSequenceClassification"] | 83828 | float32 | 4.44.2 | 312 | 600 | 3 | 12 | null | gelu | null | True | 2,048 | null | null | 0.02 | null | null | null | 26 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.003504 | small | medium |
2,268 | Akila/Mistral-of-Realms-7b-Instruct | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
2,269 | Akila/Mistral-of-Realms-7b-gguf | mistral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | false | null | null | null |
2,270 | AkilaYoga/Animal_Classy | vit | ["ViTForImageClassification"] | null | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
2,271 | AkilaYoga/identify_countries | vit | ["ViTForImageClassification"] | null | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
2,272 | Akina21336/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.46.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,273 | Akira812/nllb-200-distilled-1.3B-ct2-int8 | m2m_100 | ["M2M100ForConditionalGeneration"] | 256206 | float32 | 4.21.0.dev0 | null | null | 24 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
2,274 | Akirastroworld/td8 | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.48.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,275 | Akirpap/llama-7b-qlora-sutd-qa | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.35.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,276 | AkitoP/whisper-large-v3-turbo-japanese-phone-voicevox-gen | whisper | ["WhisperForConditionalGeneration"] | 51866 | float32 | 4.47.1 | null | null | 32 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
2,277 | AkitooKun/DialoGPT-small-crybaby | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.44.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
2,278 | AkkyMa/LunarLander | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,279 | AkmalAshirmatov/first_try | wav2vec2 | ["Wav2Vec2ForCTC"] | 34 | float32 | 4.11.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
2,280 | AksayKanthan/model | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.41.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
2,281 | Akseluhr/whisper-small-sv-SE-auhr-v2 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.26.0.dev0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
2,282 | AksharaBalan/malayalam-bert-finetuned-ner | bert | ["BertForTokenClassification"] | 197285 | float32 | 4.47.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,283 | Akshay-Vasisht/Qwen2_VL_2B_FT_OCRLP | qwen2_vl | ["Qwen2VLForConditionalGeneration"] | 151936 | bfloat16 | 4.48.2 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
2,284 | Akshay0706/All-Plants-18-Epochs-Model | vit | ["ViTForImageClassification"] | null | float32 | 4.35.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
2,285 | Akshay1791/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,286 | Akshay47/Llama-3.1-8B-Instruct_bvr_finetune_v3 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.44.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
2,287 | AkshaySalunke/model | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.38.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,288 | Akshayxx/bert-base-cased-finetuned-cora | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,289 | Akshayxx/citeseer-finetuned-embeddings | bert | ["BertModel"] | 30522 | float32 | 4.38.2 | 384 | 1,536 | 6 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.010617 | small | short |
2,290 | Akshayxx/citeseer-finetuned-embeddingsV2 | bert | ["BertModel"] | 30522 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,291 | AkshithaSai19/my-updated-model | albert | ["AlbertModel"] | 200000 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | 2 | 3 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,292 | Aktsvigun/bart-large_aeslc_12345 | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.21.3 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
2,293 | Aktsvigun/bart-large_aeslc_23419 | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.21.3 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
2,294 | Aktsvigun/bart-large_aeslc_42 | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.21.3 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
2,295 | Aktsvigun/bart-large_aeslc_4837 | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.21.3 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
2,296 | Aktsvigun/bart-large_aeslc_705525 | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.21.3 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
2,297 | Akul/MiniLlama-1.8b-Math | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.42.4 | 2,048 | 5,632 | 40 | 32 | 4 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 2.013266 | medium | medium |
2,298 | Akul/MiniLlama-1.8b-Math-safetensors | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.42.4 | 2,048 | 5,632 | 40 | 32 | 4 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 2.013266 | medium | medium |
2,299 | AkylaiBva/my_medium_wspr | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.42.3 | null | null | 24 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.