index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1,000 | A01794620/distilbert-base-uncased-finetuned-emotion-short-sets | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.41.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,001 | A0B0C/Flacon | RefinedWebModel | ["RWForCausalLM"] | 65024 | bfloat16 | 4.27.4 | 4,544 | null | null | null | null | null | 0 | True | null | null | null | 0.02 | 11 | 11 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,002 | A2Amir/SF_A68_IDEFICS_9B_IDL_SFT | idefics | ["IdeficsForVisionText2Text"] | 32000 | float32 | 4.44.2 | 4,096 | 11,008 | 32 | 32 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 6.442451 | medium | medium |
1,003 | A5medAshraf/Magicllama | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.38.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | long |
1,004 | A7it/kjnn | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.47.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
1,005 | A7m0d/distilhubert-finetuned-gtzan | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.45.1 | 768 | 3,072 | 2 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.014156 | small | null |
1,006 | AAA01101312/distilbert-base-uncased-finetuned-clinc | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.39.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,007 | AAA01101312/pegasus-samsum | pegasus | ["PegasusForConditionalGeneration"] | 96103 | float32 | 4.39.0 | null | null | 16 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
1,008 | AAA01101312/xlm-roberta-base-finetuned-panx-de-fr | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.39.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,009 | AAA01101312/xlm-roberta-base-finetuned-panx-en | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.39.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,010 | AAA01101312/xlm-roberta-base-finetuned-panx-fr | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.39.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,011 | AAA01101312/xlm-roberta-base-finetuned-panx-it | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.39.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,012 | AAAI2025/MathSpeech_Ablation_Study_1stage_fine-tuned_with_errors_T5_base | t5 | ["T5ForConditionalGeneration"] | 32106 | float32 | 4.44.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,013 | AAAI2025/MathSpeech_Ablation_Study_1stage_fine-tuned_with_errors_T5_small | t5 | ["T5ForConditionalGeneration"] | 32106 | float32 | 4.44.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,014 | AAAI2025/MathSpeech_Ablation_Study_Error_corrector_T5_base | t5 | ["T5ForConditionalGeneration"] | 32106 | float32 | 4.44.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,015 | AAAI2025/MathSpeech_Ablation_Study_Error_corrector_T5_small | t5 | ["T5ForConditionalGeneration"] | 32106 | float32 | 4.44.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,016 | AAAI2025/MathSpeech_Ablation_Study_LaTeX_translator_T5_base | t5 | ["T5ForConditionalGeneration"] | 32106 | float32 | 4.44.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,017 | AAAI2025/MathSpeech_Ablation_Study_LaTeX_translator_T5_small | t5 | ["T5ForConditionalGeneration"] | 32106 | float32 | 4.44.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,018 | AAAndyZ/JD_Qwen2.5_7B_lora | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
1,019 | AAAndyZ/JD_Qwen2.5_7B_lora_20250225 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
1,020 | AAK1423/wav2vec2-ALC-output | wav2vec2 | ["Wav2Vec2ForCTC"] | 95 | float32 | 4.51.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
1,021 | AAK1423/wave2vecWvald15Epoch | wav2vec2 | ["Wav2Vec2ForCTC"] | 95 | float32 | 4.51.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
1,022 | AAbduallah1/Finetuned-Llama-3.1-8B | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.53.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
1,023 | AAkhilesh/wav2vec2-large-xls-r-300m-hsb-colab | wav2vec2 | ["Wav2Vec2ForCTC"] | 49 | float32 | 4.18.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
1,024 | AAkhilesh/wav2vec2-large-xls-r-300m-ta-colab | wav2vec2 | ["Wav2Vec2ForCTC"] | 53 | float32 | 4.18.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
1,025 | ABHIiiii1/mt5-Finetuned-Bi-En-Mn-trans | mt5 | ["MT5ForConditionalGeneration"] | 250102 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,026 | ABHIiiii1/mt5-Finetuned-Bi-En-Mn-trans2 | mt5 | ["MT5ForConditionalGeneration"] | 250102 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,027 | ABX-AI/Infinite-Laymons-9B | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.39.1 | 4,096 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 8.053064 | large | very_long |
1,028 | ABX-AI/Infinite-Sumika-9B | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.39.1 | 4,096 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 8.053064 | large | very_long |
1,029 | ABX-AI/Kuno-Lemon-7B | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.39.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | long |
1,030 | ABX-AI/Laymonade-7B | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.39.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
1,031 | ABX-AI/Norosumika-7B | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.39.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
1,032 | ABX-AI/Quantum-Citrus-9B | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.39.1 | 4,096 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 8.053064 | large | very_long |
1,033 | ABX-AI/Silver-Sun-v2-11B | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.39.1 | 4,096 | 14,336 | 48 | 32 | 8 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 9.663676 | large | medium |
1,034 | ABX-AI/Starfinite-Laymospice-v2-7B | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.39.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
1,035 | ABrinkmann/bert-finetuned-ner | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.30.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,036 | ABrinkmann/bert-finetuned-ner-10epochs | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.30.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,037 | ABrinkmann/bert-finetuned-ner-3epochs | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.30.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,038 | ABrinkmann/deberta-v3-large-ad-opentag-finetuned-ner-5epochs | deberta-v2 | ["DebertaV2ForTokenClassification"] | 128100 | float32 | 4.30.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
1,039 | ABrinkmann/deberta-v3-large-finetuned-ner-10epochs | deberta-v2 | ["DebertaV2ForTokenClassification"] | 128100 | float32 | 4.30.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
1,040 | ABrinkmann/deberta-v3-large-finetuned-ner-10epochs-V2 | deberta-v2 | ["DebertaV2ForTokenClassification"] | 128100 | float32 | 4.30.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
1,041 | ABrinkmann/sbert_xtremedistil-l6-h256-uncased-mean-cosine-h32 | bert | ["BertModel"] | 30533 | float32 | 4.17.0 | 256 | 1,024 | 6 | 8 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.004719 | small | short |
1,042 | ACATECH/ncos | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.47.1 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
1,043 | ACCORD-NLP/ner-albert-large | albert | ["AlbertForTokenClassification"] | 30000 | float32 | 4.16.2 | 1,024 | 4,096 | 24 | 16 | null | gelu_new | null | null | 512 | null | null | 0.02 | 2 | 3 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
1,044 | ACCORD-NLP/ner-roberta-large-lm | roberta | ["RobertaForTokenClassification"] | 50265 | float32 | 4.16.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
1,045 | ACCORD-NLP/re-albert-large | albert | ["AlbertForSequenceClassification"] | 30004 | float32 | 4.16.2 | 1,024 | 4,096 | 24 | 16 | null | gelu_new | null | null | 512 | null | null | 0.02 | 2 | 3 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
1,046 | ACCORD-NLP/re-bert-large | bert | ["BertForSequenceClassification"] | 29000 | float32 | 4.16.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
1,047 | ACE-Step/ACE-Step-v1-chinese-rap-LoRA | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,048 | ACIDE/User-VLM-3B-base | paligemma | ["PaliGemmaForConditionalGeneration"] | null | bfloat16 | 4.48.0.dev0 | 2,048 | null | 26 | null | null | null | null | null | null | null | null | null | 2 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 1.308623 | medium | null |
1,049 | ADHIZ/causal-omni_thanvi | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.46.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,050 | ADHIZ/code-omni_kali | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.46.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,051 | ADHIZ/code-omni_madasu | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.46.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,052 | ADHIZ/image_omni_netz | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,053 | ADI10HERO/PPO_LunarLander_v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,054 | ADP-EPFL/GPT2-Einstein-EPFL | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.40.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,055 | ADRIANRICO/Distilbert-finetuned-IMDB | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.34.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,056 | ADSah/billsum | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.52.4 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,057 | ADT7/SN11-EMP-5000-VAL-QL | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.46.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
1,058 | ADT7/SN11-EMP-6000-Val-Data-QL-v2-rename | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.46.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
1,059 | ADT7/SN11-MK-500-Q-v2 | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.46.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
1,060 | ADT7/SN11-NM-FAT-5000-QLORA-v1 | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.48.3 | 5,120 | 32,768 | 40 | 32 | 8 | silu | 0 | True | 32,768 | 100,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
1,061 | ADT7/SN11-T-100-Q | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.46.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
1,062 | AELLM/gemma-2-aeria-9b | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.43.1 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
1,063 | AELLM/gemma-2-inko-9b | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.43.1 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 107 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
1,064 | AFSA1729/movie-classifier | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,065 | AFZAL0008/bert-finetuned-ner | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.40.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,066 | AGBonnet/medinote-13b | llama | ["LlamaForCausalLM"] | 32019 | float32 | 4.36.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
1,067 | AGBonnet/medinote-7b | llama | ["LlamaForCausalLM"] | 32019 | float32 | 4.36.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,068 | AGCobra/EXAONE-Deep-32B-mlx-4Bit | exaone | ["ExaoneForCausalLM"] | 102400 | bfloat16 | 4.43.1 | 5,120 | 27,392 | null | 40 | 8 | null | 0 | True | 32,768 | 1,000,000 | null | 0.02 | 1 | 361 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | null | null | very_long |
1,069 | AGI-CEO/ppo-PyramidsV1 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,070 | AGI-CEO/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,071 | AGnatkiv/distilbert-reviews | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.35.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,072 | AGresse/ast-finetuned-audioset-10-10-0.4593-finetuned-rfcx-reprod | audio-spectrogram-transformer | ["ASTForAudioClassification"] | null | float32 | 4.48.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
1,073 | AGudden/xlm-roberta-base-finetuned-marc | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,074 | AGundawar/moondream-ft | moondream1 | ["Moondream"] | null | float16 | 4.41.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,075 | AGuzhvenko/ppo-Pyramids | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,076 | AHDMK/Sentence-GISTEmbedLoss-BioBert-Allnli-scinli | bert | ["BertModel"] | 28996 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,077 | AHMED13245/xlm-roberta-base-finetuned-panx-de | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,078 | AHarbury/debiasNLPFinal | distilbert | ["DistilBertForSequenceClassification"] | 30522 | null | 4.29.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,079 | AHedya/Al-Fatiha-Recitation-Model | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.39.3 | null | null | 12 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
1,080 | AHedya/w2v-bert-fine-tuning | wav2vec2-bert | ["Wav2Vec2BertForCTC"] | 14 | float32 | 4.40.2 | 1,024 | 4,096 | 24 | 16 | null | swish | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | null |
1,081 | AHedya/whisper-small-dv | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.39.3 | null | null | 12 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
1,082 | AI-Enthusiast11/pii-entity-extractor | deberta-v2 | ["DebertaV2ForTokenClassification"] | 251000 | float32 | 4.51.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,083 | AI-MO/Kimina-Autoformalizer-7B | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.49.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
1,084 | AI-MO/Kimina-Prover-Preview-Distill-7B | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.49.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 16,384 | 100,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | long |
1,085 | AI-MO/NuminaMath-7B-CoT | llama | ["LlamaForCausalLM"] | 102400 | float32 | 4.42.3 | 4,096 | 11,008 | 30 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.000001 | 0.02 | 100000 | 100001 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.039798 | medium | medium |
1,086 | AI-MO/NuminaMath-7B-TIR | llama | ["LlamaForCausalLM"] | 102400 | bfloat16 | 4.42.3 | 4,096 | 11,008 | 30 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.000001 | 0.02 | 100000 | 100001 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.039798 | medium | medium |
1,087 | AI-Roadmap/SmolLM2-135M-rm-60k | llama | ["RewardModel"] | 49152 | bfloat16 | 4.51.3 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | False | 8,192 | 100,000 | 0.00001 | 0.041667 | 1 | 2 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | long |
1,088 | AI-Sweden-Models/Llama-3-8B | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
1,089 | AI-Sweden-Models/ModernBERT-large | modernbert | ["ModernBertForMaskedLM"] | 50368 | null | 4.56.2 | 1,024 | 2,624 | 28 | 16 | null | null | 0 | null | 8,192 | null | null | 0.02 | 50281 | 50282 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | true | 0.352322 | small | long |
1,090 | AI-Sweden-Models/bert-large-nordic-pile-1M-steps | bert | ["BertForMaskedLM"] | 64000 | float32 | 4.31.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
1,091 | AI-Sweden-Models/gpt-sw3-1.3b-instruct | gpt2 | ["GPT2LMHeadModel"] | 64000 | float32 | 4.22.1 | null | null | null | null | null | null | null | True | null | null | null | 0.01 | 2 | 3 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,092 | AI-Sweden-Models/gpt-sw3-126m-instruct | gpt2 | ["GPT2LMHeadModel"] | 64000 | float32 | 4.22.1 | null | null | null | null | null | null | null | True | null | null | null | 0.023 | 2 | 3 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,093 | AI-Sweden-Models/gpt-sw3-20b-instruct | gpt2 | ["GPT2LMHeadModel"] | 64000 | float32 | 4.22.1 | null | null | null | null | null | null | null | True | null | null | null | 0.008165 | 2 | 3 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,094 | AI-Sweden-Models/gpt-sw3-20b-instruct-4bit-gptq | gpt2 | ["GPT2LMHeadModel"] | 64000 | float16 | 4.36.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | 0.008165 | 2 | 3 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,095 | AI-Sweden-Models/gpt-sw3-356m-instruct | gpt2 | ["GPT2LMHeadModel"] | 64000 | float32 | 4.22.1 | null | null | null | null | null | null | null | True | null | null | null | 0.01 | 2 | 3 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,096 | AI-Sweden-Models/gpt-sw3-6.7b-v2 | gpt2 | ["GPT2LMHeadModel"] | 64000 | float32 | 4.22.1 | null | null | null | null | null | null | null | True | null | null | null | 0.01 | 1 | 1 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,097 | AI-Sweden-Models/gpt-sw3-6.7b-v2-instruct | gpt2 | ["GPT2LMHeadModel"] | 64000 | float32 | 4.22.1 | null | null | null | null | null | null | null | True | null | null | null | 0.01 | 1 | 1 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,098 | AI-newbie89/codeparrot-ds | gpt2 | ["GPT2LMHeadModel"] | 50000 | float32 | 4.41.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 0 | 0 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,099 | AI-newbie89/marian-finetuned-kde4-en-to-fr | marian | ["MarianMTModel"] | 59514 | float32 | 4.41.1 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.