index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
14,900 | Hplm/student_1750_1820 | llama | ["LlamaForCausalLM"] | 16000 | float32 | 4.46.3 | 960 | 2,560 | 32 | 15 | 5 | silu | 0 | True | 256 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.353894 | small | short |
14,901 | Hplm/student_1820_1850 | llama | ["LlamaForCausalLM"] | 16000 | float32 | 4.46.3 | 960 | 2,560 | 32 | 15 | 5 | silu | 0 | True | 256 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.353894 | small | short |
14,902 | Hplm/student_1850_1880 | llama | ["LlamaForCausalLM"] | 16000 | float32 | 4.46.3 | 960 | 2,560 | 32 | 15 | 5 | silu | 0 | True | 256 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.353894 | small | short |
14,903 | Hplm/student_1880_1910 | llama | ["LlamaForCausalLM"] | 16000 | float32 | 4.46.3 | 960 | 2,560 | 32 | 15 | 5 | silu | 0 | True | 256 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.353894 | small | short |
14,904 | Hplm/student_1910_1940 | llama | ["LlamaForCausalLM"] | 16000 | float32 | 4.46.3 | 960 | 2,560 | 32 | 15 | 5 | silu | 0 | True | 256 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.353894 | small | short |
14,905 | HrantDinkFoundation/turkish-hs-span-prediction | bert | ["BertForTokenClassification"] | 32000 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,906 | HrantDinkFoundation/turkish-news-hs-2class-prediction | bert | ["BertForSequenceClassification"] | 32000 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,907 | HrantDinkFoundation/turkish-news-hs-4class-prediction | bert | ["BertForSequenceClassification"] | 32000 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,908 | HrantDinkFoundation/turkish-news-hs-group-prediction | bert | ["BertForSequenceClassification"] | 32000 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,909 | Hrbk/whisper-small-cv13_hrbk_b16g4 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.37.2 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
14,910 | Hrimurr/bert-base-multilingual-cased-finetuned-multibert | bert | ["BertForMaskedLM"] | 119547 | null | 4.24.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,911 | HrishiPatel/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,912 | Hrishith123/Telugu-Distilbert-base-Multilingual-uncased | distilbert | ["DistilBertForSequenceClassification"] | 119547 | float32 | 4.40.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
14,913 | Hsawa/20250122HunIshHubOther15wordsplit1300Epoch3 | bert | ["BertForSequenceClassification"] | 64000 | float32 | 4.48.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,914 | Hsawa/20250201HunIshHubOther15wordsplit1300Epoch5 | bert | ["BertForSequenceClassification"] | 64000 | float32 | 4.48.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,915 | Hsawa/ARBERT20250211HunIshHubOther15wordsplit1300Epoch3 | bert | ["BertForSequenceClassification"] | 100000 | float32 | 4.48.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,916 | Hsianchengfun/modelall | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.51.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
14,917 | HsiehMengTing/esm2_t12_35M_UR50D-finetuned-localization | esm | ["EsmForSequenceClassification"] | 33 | float32 | 4.35.2 | 320 | 1,280 | 6 | 20 | null | gelu | null | True | 1,026 | null | null | 0.02 | null | null | null | 16 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.007373 | small | short |
14,918 | Hsuan0929/llama-3.2-custom-energy_saving_assistant | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.48.3 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
14,919 | HsuanLLM/BreezeDolphinDPO-7B-DARE | mistral | ["MistralForCausalLM"] | 61875 | bfloat16 | 4.35.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
14,920 | HsuuHsuu/mt5-small | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.46.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
14,921 | HsuuHsuu/my_awesome_billsum_model | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.34.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
14,922 | HuShifang83/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,923 | Huamin/santacoder-finetuned-the-stack-bash | gpt2 | ["GPT2LMHeadCustomModel"] | 49280 | float32 | 4.30.2 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 49152 | 49152 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
14,924 | HuanYangChang/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,925 | HuangAurelianus/test-trainer | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.46.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
14,926 | HuangJordan/whisper-hi-small | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.40.0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
14,927 | HuangXinBa/test-trainer | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,928 | Huanghz/align2llava-7b-lora-question | llava_llama | ["LlavaLlamaForCausalLM"] | 32000 | float16 | 4.37.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
14,929 | HuanjinYao/DenseConnector-v1.5-SigLIP-7B-AnyRes | llava_llama | ["LlavaLlamaForCausalLM"] | 32000 | bfloat16 | 4.36.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
14,930 | HuanjinYao/DenseConnector-with-mgm-7B | mgm | ["MGMLlamaForCausalLM"] | 32000 | float16 | 4.36.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
14,931 | HuaxinTang/marian-finetuned-kde4-en-to-fr | marian | ["MarianMTModel"] | 59514 | float32 | 4.46.2 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
14,932 | Hubert0314/bert_tutorial | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,933 | Hubert0314/unsloth | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
14,934 | HubertBaginski/bert-alternatives-to-suicide | bert | ["BertForSequenceClassification"] | 30522 | null | 4.17.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,935 | HubertBaginski/bert-celebrity-suicide | bert | ["BertForSequenceClassification"] | 30522 | null | 4.17.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,936 | HubertBaginski/bert-enhancing-myths | bert | ["BertForSequenceClassification"] | 30522 | null | 4.17.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,937 | HubertBaginski/bert-healing-story | bert | ["BertForSequenceClassification"] | 30522 | null | 4.17.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,938 | HubertBaginski/bert-main-focus | bert | ["BertForSequenceClassification"] | 30522 | null | 4.17.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,939 | HubertBaginski/bert-monocausality | bert | ["BertForSequenceClassification"] | 30522 | null | 4.17.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,940 | HubertBaginski/bert-positive-outcome-crisis | bert | ["BertForSequenceClassification"] | 30522 | null | 4.17.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,941 | HubertBaginski/bert-problem-or-solution | bert | ["BertForSequenceClassification"] | 30522 | null | 4.17.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,942 | HubertBaginski/bert-suicidal-ideation | bert | ["BertForSequenceClassification"] | 30522 | null | 4.17.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,943 | HubertBaginski/bert-suicide-death-completed | bert | ["BertForSequenceClassification"] | 30522 | null | 4.17.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,944 | HubertBaginski/bert-twitter-about-suicide | bert | ["BertForMaskedLM"] | 30522 | null | null | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,945 | HubertBaginski/bert-twitter-main-categories | bert | ["BertForMaskedLM"] | 30522 | null | null | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,946 | Huertas97/smollm-gec-kto | llama | ["LlamaForCausalLM"] | 49152 | float32 | 4.44.2 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 0 | 0 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | medium |
14,947 | Huertas97/smollm-gec-sftt-dpo | llama | ["LlamaForCausalLM"] | 49152 | float32 | 4.44.2 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 0 | 0 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | medium |
14,948 | Huertas97/smollm-gec-sftt-kto | llama | ["LlamaForCausalLM"] | 49152 | float32 | 4.44.2 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 0 | 0 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | medium |
14,949 | HueyWoo/qwen2.5_3B-gguf | qwen2 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,950 | HugMi/M3-Assignment2 | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.48.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
14,951 | HugeFighter/ddpm-celebahq-finetuned-butterflies-2epochs | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,952 | HugeFighter/dummy-model | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,953 | HugeFighter/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,954 | HugeFighter/ppo-LunarLander-v2-2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,955 | Hugging-Steven/chatglm2-6b-torchkeras | chatglm | ["ChatGLMForConditionalGeneration"] | 65024 | float16 | 4.30.2 | 4,096 | null | null | 32 | null | null | 0 | True | null | null | null | null | null | 2 | false | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
14,956 | HuggingAlex1247/distilbert-base-german-europeana-cased-germeval_14 | distilbert | ["DistilBertForTokenClassification"] | 32000 | null | 4.17.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | true | null | null | short |
14,957 | HuggingDLRL/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,958 | HuggingDreamer/fine-tuning_test | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
14,959 | HuggingFaceAdminstrator/saseucm.ai | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,960 | HuggingFaceFW/ablation-model-fineweb-edu | llama | ["LlamaForCausalLM"] | 50272 | bfloat16 | 4.39.3 | 2,048 | 8,192 | 24 | 32 | 32 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | true | 64 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
14,961 | HuggingFaceFW/ablation-model-redpajama2 | llama | ["LlamaForCausalLM"] | 50272 | bfloat16 | 4.38.2 | 2,048 | 8,192 | 24 | 32 | 32 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | true | 64 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
14,962 | HuggingFaceH4/tiny-random-LlamaForCausalLM | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.27.0.dev0 | 16 | 64 | 2 | 4 | null | silu | null | True | null | null | 0.000001 | 0.02 | 0 | 1 | false | 4 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 0.000006 | small | null |
14,963 | HuggingFaceH4/tiny-random-LlamaForSequenceClassification | llama | ["LlamaForSequenceClassification"] | 32000 | float32 | 4.28.0.dev0 | 16 | 64 | 2 | 4 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 0 | 1 | false | 4 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 0.000006 | small | medium |
14,964 | HuggingFaceH4/zephyr-7b-beta | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.35.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
14,965 | HuggingFaceM4/Idefics3-8B-Llama3 | idefics3 | ["Idefics3ForConditionalGeneration"] | null | bfloat16 | 4.44.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
14,966 | HuggingFaceM4/idefics2-8b | idefics2 | ["Idefics2ForConditionalGeneration"] | null | float32 | 4.39.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,967 | HuggingFaceM4/idefics2-8b-AWQ | idefics2 | ["Idefics2ForConditionalGeneration"] | null | float16 | 4.40.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,968 | HuggingFaceM4/idefics2-8b-base-AWQ | idefics2 | ["Idefics2ForConditionalGeneration"] | null | float32 | 4.39.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,969 | HuggingFaceM4/siglip-so400m-14-384-flash-attn2-navit | siglip | ["SiglipModel"] | null | float32 | 4.37.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,970 | HuggingFaceM4/tiny-random-LlamaForCausalLM | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.27.0.dev0 | 16 | 64 | 2 | 4 | null | silu | null | True | null | null | 0.000001 | 0.02 | 0 | 1 | false | 4 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 0.000006 | small | null |
14,971 | HuggingFaceTB/SmolLM-1.7B | llama | ["LlamaForCausalLM"] | 49152 | float32 | 4.39.3 | 2,048 | 8,192 | 24 | 32 | 32 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 0 | 0 | true | 64 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
14,972 | HuggingFaceTB/SmolLM-360M | llama | ["LlamaForCausalLM"] | 49152 | bfloat16 | 4.41.2 | 960 | 2,560 | 32 | 15 | 5 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 0 | 0 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.353894 | small | medium |
14,973 | HuggingFaceTB/SmolLM2-1.7B | llama | ["LlamaForCausalLM"] | 49152 | bfloat16 | 4.40.1 | 2,048 | 8,192 | 24 | 32 | 32 | silu | 0 | True | 8,192 | 130,000 | 0.00001 | 0.02 | 0 | 0 | true | 64 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
14,974 | HuggingFaceTB/SmolLM2-1.7B-Instruct | llama | ["LlamaForCausalLM"] | 49152 | bfloat16 | 4.42.3 | 2,048 | 8,192 | 24 | 32 | 32 | silu | 0 | True | 8,192 | 130,000 | 0.00001 | 0.02 | 1 | 2 | true | 64 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
14,975 | HuggingFaceTB/SmolLM2-135M | llama | ["LlamaForCausalLM"] | 49152 | bfloat16 | 4.40.1 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 8,192 | 100,000 | 0.00001 | 0.041667 | 0 | 0 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | long |
14,976 | HuggingFaceTB/SmolVLM-256M-Base | idefics3 | ["Idefics3ForConditionalGeneration"] | 49280 | bfloat16 | 4.46.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
14,977 | HuggingFaceTB/SmolVLM-Synthetic | idefics3 | ["Idefics3ForConditionalGeneration"] | 49155 | bfloat16 | 4.46.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
14,978 | HuggingFaceTB/SmolVLM2-256M-Video-Instruct | smolvlm | ["SmolVLMForConditionalGeneration"] | 49280 | float32 | 4.47.1 | null | null | null | null | null | null | null | False | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
14,979 | HuggingFaceTB/SmolVLM2-500M-Video-Instruct | smolvlm | ["SmolVLMForConditionalGeneration"] | 49280 | float32 | 4.47.1 | null | null | null | null | null | null | null | False | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
14,980 | HuggingFaceTB/stack-edu-classifier-go | bert | ["BertForSequenceClassification"] | 49156 | float32 | 4.43.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 1,024 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,981 | HuggingFaceTB/stack-edu-classifier-python | bert | ["BertForSequenceClassification"] | 49156 | float32 | 4.43.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 1,024 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,982 | Huggingfly/a2c-AntBulletEnv-v0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,983 | Huggingfly/ppo-PyramidsTraining | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,984 | Huggingfly/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,985 | Huggs108Bj/SelfTrainTransformer | m2m_100 | ["M2M100ForConditionalGeneration"] | 256206 | float32 | 4.21.0.dev0 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
14,986 | Hugo0123/BogoAI | bogoai | ["BogoAIModel"] | null | float32 | 4.42.3 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,987 | HugoGiddins/multi-tag-classifier-full-fine-tune-base | deberta-v2 | ["DebertaV2ForSequenceClassification"] | 128100 | null | 4.53.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,988 | HugoKD/BERT_NER_WITH_LORA | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,989 | HugoStiglitz/overpass_8bit | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
14,990 | Hugofernandez/Mistral-7B-v0.1-colab-sharded | mistral | ["MistralForCausalLM"] | 32000 | float32 | 4.35.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
14,991 | Hugosh71/Llama-3-8B-Eden | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
14,992 | Hui1631Ayn/vit-base-for-My | vit | ["ViTForImageClassification"] | null | float32 | 4.46.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
14,993 | Huishan/SFVE-large | roberta | ["RobertaModel"] | 50265 | float32 | 4.45.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
14,994 | Huiyeong/setfit-callbot-keyword-neg | bert | ["BertModel"] | 250037 | float32 | 4.52.4 | 384 | 1,536 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.021234 | small | short |
14,995 | Huiyeong/setfit-clinc150 | bert | ["BertModel"] | 250037 | float32 | 4.52.4 | 384 | 1,536 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.021234 | small | short |
14,996 | Huiyeong/setfit-curekart | bert | ["BertModel"] | 250037 | float32 | 4.52.4 | 384 | 1,536 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.021234 | small | short |
14,997 | Huiyeong/setfit-powerplay11 | bert | ["BertModel"] | 250037 | float32 | 4.52.4 | 384 | 1,536 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.021234 | small | short |
14,998 | Huiyeong/setfit-sofmattress | bert | ["BertModel"] | 250037 | float32 | 4.52.4 | 384 | 1,536 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.021234 | small | short |
14,999 | Hum-Works/lodestone-base-4096-v1 | bert | ["BertModel"] | 30528 | bfloat16 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.