index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
4,900 | AzureBlack/Thespis-13b-v0.3-5bpw-6h-exl2 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.35.0.dev0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
4,901 | AzureBlack/Xwin-MLewd-13B-V0.2-5bpw-6h-exl2 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.35.0.dev0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
4,902 | Azurro/APT3-1B-Instruct-v1 | llama | ["LlamaForCausalLM"] | 31980 | float32 | 4.35.2 | 2,048 | 5,504 | 18 | 32 | 32 | silu | null | True | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 64 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.90597 | small | medium |
4,903 | Azurro/APT3-275M-Base | llama | ["LlamaForCausalLM"] | 31980 | float32 | 4.35.0 | 768 | 2,048 | 32 | 12 | 12 | silu | null | True | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 64 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.226492 | small | medium |
4,904 | Azwaruhuy/awang_model | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
4,905 | Azzedde/llama3.1-8b-text2cypher | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.48.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
4,906 | Azzindani/Qwen2.5_1.5B_IT_ID_Legal | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float16 | 4.51.1 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
4,907 | Azzizz17/autotrain-aaaa-45159113325 | mt5 | ["MT5ForConditionalGeneration"] | 250100 | float32 | 4.25.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,908 | Azzizz17/autotrain-translator3-45113113262 | t5 | ["T5ForConditionalGeneration"] | 32100 | float32 | 4.25.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,909 | B-O/dummy-mistral-4M | mistral | ["MistralForCausalLM"] | 50295 | bfloat16 | 4.35.2 | 32 | 1,024 | 6 | 16 | 8 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 50256 | 50256 | false | 2 | 2 | false | null | null | false | false | false | false | false | false | true | false | true | true | 0.000074 | small | medium |
4,910 | B0BWAX/mt5-small-finetuned-en-to-de | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.40.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,911 | B2111797/recipe_v1_lr1e-3_wu4000_epo4 | gpt2 | ["GPT2LMHeadModel"] | 50268 | float32 | 4.38.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
4,912 | B2111797/trans-en-vi-v1 | marian | ["MarianMTModel"] | 53685 | float32 | 4.37.2 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
4,913 | BA-Project-SA-CRM/SA_Checkpoints | bert | ["BertForSequenceClassification"] | 31102 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,914 | BAAI/AquilaDense-16B | aquiladense | ["AquilaDenseForCausalLM"] | 151851 | bfloat16 | 4.37.2 | 5,120 | 20,480 | 40 | 40 | 8 | silu | 0 | True | 4,096 | 1,000,000 | 0.00001 | 0.02 | 151849 | 151850 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 12.582912 | large | medium |
4,915 | BAAI/AquilaDense-7B | aquiladense | ["AquilaDenseForCausalLM"] | 151851 | bfloat16 | 4.37.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 4,096 | 1,000,000 | 0.00001 | 0.02 | 151849 | 151850 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | medium |
4,916 | BAAI/AquilaMed-RL | aquila3 | ["AquilaForCausalLM"] | 151851 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 4,096 | 1,000,000 | 0.00001 | 0.02 | 151849 | 151850 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | medium |
4,917 | BAAI/AquilaSQL-7B | aquila | ["AquilaModel"] | 100008 | float16 | 4.28.1 | 4,096 | 11,008 | 32 | 32 | null | silu | null | True | 2,048 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 6.442451 | medium | medium |
4,918 | BAAI/BGE-VL-base | clip | ["CLIPModel"] | null | bfloat16 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,919 | BAAI/BGE-VL-large | clip | ["CLIPModel"] | null | bfloat16 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,920 | BAAI/BGE-VL-v1.5-zs | llava_next | ["LlavaNextForConditionalGeneration"] | 32064 | float16 | 4.39.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
4,921 | BAAI/EVE-7B-v1.0 | eve | ["EVELlamaForCausalLM"] | 32000 | bfloat16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
4,922 | BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.44.2 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 107 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
4,923 | BAAI/Infinity-Instruct-3M-0613-Llama3-70B | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.0.dev0 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | long |
4,924 | BAAI/LLARA-beir | llama | ["LlamaForCausalLM"] | 32016 | float32 | 4.38.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
4,925 | BAAI/LLARA-document | llama | ["LlamaForCausalLM"] | 32016 | float32 | 4.38.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
4,926 | BAAI/LLARA-passage | llama | ["LlamaForCausalLM"] | 32016 | float32 | 4.38.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
4,927 | BAAI/LLARA-pretrain | llama | ["LlamaForCausalLM"] | 32016 | float32 | 4.38.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
4,928 | BAAI/OmniGen-v1 | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.38.1 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 131,072 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | very_long |
4,929 | BAAI/SegVol | segvol | ["SegVolModel"] | null | float32 | 4.18.0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,930 | BAAI/bge-code-v1 | qwen2 | ["Qwen2Model"] | 151667 | float32 | 4.49.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
4,931 | BAAI/bge-en-icl | mistral | ["MistralModel"] | 32003 | float32 | 4.41.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
4,932 | BAAI/bge-reranker-v2-minicpm-layerwise | minicpm | ["LayerWiseMiniCPMForCausalLM"] | 122753 | bfloat16 | 4.38.1 | 2,304 | 5,760 | 40 | 36 | 36 | silu | 0 | False | 2,048 | 10,000 | 0.00001 | 0.1 | 1 | 2 | null | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.54804 | medium | medium |
4,933 | BAAI/bge-small-en-v1.5 | bert | ["BertModel"] | 30522 | float32 | 4.30.0 | 384 | 1,536 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.021234 | small | short |
4,934 | BAAI/llm-embedder | bert | ["BertModel"] | 30522 | float32 | 4.30.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,935 | BABYSHARK09/Na | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,936 | BABYSHARK09/Ne | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,937 | BABYSHARK09/Nf | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,938 | BABYSHARK09/Ng | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,939 | BABYSHARK09/Nq | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,940 | BABYSHARK09/Nt | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,941 | BABYSHARK09/Nu | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,942 | BABYSHARK09/Nx | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,943 | BABYSHARK09/Nz | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,944 | BABYSHARK09/Uni_6x | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,945 | BABYSHARK09/Uni_6x0 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,946 | BABYSHARK09/Uni_6x2 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,947 | BABYSHARK09/Uni_6x3 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,948 | BABYSHARK09/Uni_6x4 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,949 | BABYSHARK09/Uni_6x6 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,950 | BABYSHARK09/Uni_6x7 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,951 | BABYSHARK09/Uni_6x8 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,952 | BABYSHARK09/Uni_6x9 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,953 | BABYSHARK09/Uni_7x1 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,954 | BABYSHARK09/Uni_7x2 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,955 | BABYSHARK09/Uni_7x3 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,956 | BABYSHARK09/Uni_7x4 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,957 | BAC3030/hinrik-lp500-150-steps | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
4,958 | BAHIJA/bert-base-uncased-finetuned-sst2 | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.17.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,959 | BAK-HIRI/rl_course_vizdoom_health_gathering_supreme | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,960 | BAKKALIAYOUB/DarijaTranslation-V1 | marian | ["MarianMTModel"] | 62802 | float32 | 4.41.2 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
4,961 | BANA577/Llama3-Adrian-2 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
4,962 | BANA577/Llama3-Michael-5 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
4,963 | BANA577/Llama3-Michael-6 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
4,964 | BANA577/Llama3-Michael-7 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
4,965 | BANA577/Mistral-Koichi1 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
4,966 | BANA577/Mistral-Koichi3 | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.36.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
4,967 | BANZHISONGSHU/whisper-small-ug | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.51.3 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
4,968 | BBBBCHAN/LLaVA-Scissor-baseline-0.5B | qwen2 | ["LlavaQwenZipForCausalLM"] | 151936 | bfloat16 | 4.40.0.dev0 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
4,969 | BBBBirdIsTheWord/ML-Agents-Pyramids | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,970 | BBBBirdIsTheWord/a2c-PandaPickAndPlace-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
4,971 | BBBBirdIsTheWord/a2c-PandaReachDense-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
4,972 | BBBBirdIsTheWord/poca-SoccerTwos | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,973 | BBBBirdIsTheWord/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,974 | BBGAME605065444/my_awesome_model | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.28.0.dev0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
4,975 | BBLL3456/faster-whisper-large-V3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,976 | BBVA-COE-AA-FINANZAS-MX/Meta-Llama-3-8B-Instruct | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.42.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
4,977 | BBorg/a2c-PandaReachDense-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,978 | BBuf/RWKV-4-World-169M | rwkv | [] | 65536 | null | 4.33.1 | 768 | 3,072 | 12 | null | null | null | null | True | null | null | null | null | 0 | 0 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
4,979 | BBuf/RWKV-4-World-1B5 | rwkv | [] | 65536 | null | 4.33.1 | 2,048 | 8,192 | 24 | null | null | null | null | True | null | null | null | null | 0 | 0 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 1.20796 | medium | null |
4,980 | BBuf/RWKV-4-World-3B | rwkv | [] | 65536 | null | 4.33.1 | 2,560 | 10,240 | 32 | null | null | null | null | True | null | null | null | null | 0 | 0 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 2.516582 | medium | null |
4,981 | BBuf/RWKV-4-World-430M | rwkv | [] | 65536 | null | 4.33.1 | 1,024 | 4,096 | 24 | null | null | null | null | True | null | null | null | null | 0 | 0 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
4,982 | BBuf/RWKV-4-World-7B | rwkv | [] | 65536 | null | 4.33.1 | 4,096 | 16,384 | 32 | null | null | null | null | True | null | null | null | null | 0 | 0 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 6.442451 | medium | null |
4,983 | BCCard/Qwen2.5-Coder-14B-FP8-Dynamic | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.3 | 5,120 | 13,824 | 48 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 15.099494 | large | very_long |
4,984 | BCCard/Qwen2.5-Coder-32B-Instruct-FP8-Dynamic | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.3 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
4,985 | BCCard/Qwen2.5-Coder-7B-FP8-Dynamic | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
4,986 | BCCard/gemma-3-27b-it-FP8-Dynamic | gemma3 | ["Gemma3ForConditionalGeneration"] | null | bfloat16 | 4.51.3 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | null | [1, 106] | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
4,987 | BDAD/segformer-b3-horizontal | segformer | ["SegformerForSemanticSegmentation"] | null | float32 | 4.29.2 | null | null | null | [1, 2, 5, 8] | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,988 | BDAD/segformer-b3-vertical | segformer | ["SegformerForSemanticSegmentation"] | null | float32 | 4.29.2 | null | null | null | [1, 2, 5, 8] | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,989 | BDomantas/layoutxlm-base-tune | layoutlmv2 | ["LayoutLMv2ForTokenClassification"] | 250002 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,990 | BEBO-DBIndia/LLAMA_V58M | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
4,991 | BEE-spoke-data/Mixtral-GQA-400m-v4-4096 | mixtral | ["MixtralForCausalLM"] | 32000 | float32 | 4.37.0.dev0 | 2,048 | 6,144 | 6 | 32 | 8 | silu | 0 | False | 4,096 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 4 | false | 8 | 2 | false | false | false | false | false | false | false | true | true | true | 0.30199 | small | medium |
4,992 | BEE-spoke-data/Qwen2-1.5B-stepbasin-books | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.42.4 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
4,993 | BEE-spoke-data/TinyLlama-3T-1.1bee | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.36.2 | 2,048 | 5,632 | 22 | 32 | 4 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
4,994 | BEE-spoke-data/beecoder-220M-python | llama | ["LlamaForCausalLM"] | 32128 | bfloat16 | 4.36.2 | 1,024 | 4,096 | 10 | 32 | 8 | silu | 0 | False | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 32 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.125829 | small | medium |
4,995 | BEE-spoke-data/bert-plus-L8-4096-v1.0 | bert | ["BertForMaskedLM"] | 30522 | float32 | 4.37.2 | 768 | 3,072 | 8 | 12 | null | silu | null | True | 4,096 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.056623 | small | medium |
4,996 | BEE-spoke-data/mega-ar-126m-4k | mega | ["MegaForCausalLM"] | 50304 | float32 | 4.36.2 | 768 | 1,536 | 12 | 1 | null | null | null | True | null | null | null | 0.02 | 0 | 0 | null | 768 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
4,997 | BEE-spoke-data/mega-encoder-small-16k-v1 | mega | ["MegaForMaskedLM"] | 50304 | float32 | 4.38.2 | 768 | 2,304 | 8 | 1 | null | null | null | True | null | null | null | 0.02 | 0 | 2 | null | 768 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.056623 | small | null |
4,998 | BEE-spoke-data/mega-small-embed-synthSTS-16384-v1 | mega | ["MegaModel"] | 50304 | float32 | 4.38.2 | 768 | 2,304 | 8 | 1 | null | null | null | True | null | null | null | 0.02 | 0 | 2 | null | 768 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.056623 | small | null |
4,999 | BEE-spoke-data/roberta-base-description2genre | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.33.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.