index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
2,700 | AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-REGION_earlystop_70_15_15 | esm | ["EsmForTokenClassification"] | 33 | float32 | 4.46.2 | 320 | 1,280 | 6 | 20 | null | gelu | null | True | 1,026 | null | null | 0.02 | null | null | null | 16 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.007373 | small | short |
2,701 | AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-REPEAT_earlystop_70_15_15 | esm | ["EsmForTokenClassification"] | 33 | float32 | 4.46.2 | 320 | 1,280 | 6 | 20 | null | gelu | null | True | 1,026 | null | null | 0.02 | null | null | null | 16 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.007373 | small | short |
2,702 | AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-SIGNAL_earlystop_70_15_15 | esm | ["EsmForTokenClassification"] | 33 | float32 | 4.46.2 | 320 | 1,280 | 6 | 20 | null | gelu | null | True | 1,026 | null | null | 0.02 | null | null | null | 16 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.007373 | small | short |
2,703 | AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-STRAND_earlystop_70_15_15 | esm | ["EsmForTokenClassification"] | 33 | float32 | 4.46.2 | 320 | 1,280 | 6 | 20 | null | gelu | null | True | 1,026 | null | null | 0.02 | null | null | null | 16 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.007373 | small | short |
2,704 | AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-TOPO_DOM_earlystop_70_15_15 | esm | ["EsmForTokenClassification"] | 33 | float32 | 4.46.2 | 320 | 1,280 | 6 | 20 | null | gelu | null | True | 1,026 | null | null | 0.02 | null | null | null | 16 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.007373 | small | short |
2,705 | AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-TRANSIT_earlystop_70_15_15 | esm | ["EsmForTokenClassification"] | 33 | float32 | 4.46.2 | 320 | 1,280 | 6 | 20 | null | gelu | null | True | 1,026 | null | null | 0.02 | null | null | null | 16 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.007373 | small | short |
2,706 | AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-TRANSMEM_earlystop_70_15_15 | esm | ["EsmForTokenClassification"] | 33 | float32 | 4.46.2 | 320 | 1,280 | 6 | 20 | null | gelu | null | True | 1,026 | null | null | 0.02 | null | null | null | 16 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.007373 | small | short |
2,707 | AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-TURN_earlystop_70_15_15 | esm | ["EsmForTokenClassification"] | 33 | float32 | 4.46.2 | 320 | 1,280 | 6 | 20 | null | gelu | null | True | 1,026 | null | null | 0.02 | null | null | null | 16 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.007373 | small | short |
2,708 | AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-ZN_FING_earlystop_70_15_15 | esm | ["EsmForTokenClassification"] | 33 | float32 | 4.46.2 | 320 | 1,280 | 6 | 20 | null | gelu | null | True | 1,026 | null | null | 0.02 | null | null | null | 16 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.007373 | small | short |
2,709 | AliSab/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,710 | AliSouliman/PPO-LL2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,711 | AliSouliman/a2c-PandaReachDense-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,712 | Alias1964/Llama-3.1-70B-Instruct-lorablated-Q4-EXL2 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.43.3 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
2,713 | Alibaba-NLP/Simulation_LLM_google_14B_V1 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.44.2 | 5,120 | 13,824 | 48 | 40 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 15.099494 | large | very_long |
2,714 | Alibaba-NLP/ZeroSearch_google_v1_Qwen2.5_3B | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.48.3 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151643 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
2,715 | Alibaba-NLP/ZeroSearch_wiki_V2_Llama_3.2_3B | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
2,716 | Alibaba-NLP/ZeroSearch_wiki_V2_Qwen2.5_3B | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.48.3 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151643 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
2,717 | Alibaba-NLP/gme-Qwen2-VL-2B-Instruct | qwen2_vl | ["Qwen2VLForConditionalGeneration", "GmeQwen2VL"] | 151936 | float32 | 4.45.0.dev0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
2,718 | Alibaba-NLP/gte-Qwen2-7B-instruct | qwen2 | ["Qwen2ForCausalLM"] | 151646 | float32 | 4.41.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
2,719 | Alibaba-NLP/gte-base-en-v1.5 | new | ["NewModel"] | 30528 | float32 | 4.39.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 8,192 | 500,000 | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | true | 0.084935 | small | long |
2,720 | Alibaba-NLP/gte-large-en-v1.5 | new | ["NewModel"] | 30528 | float32 | 4.39.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 8,192 | 160,000 | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | true | 0.30199 | small | long |
2,721 | Alibaba-NLP/gte-modernbert-base | modernbert | ["ModernBertModel"] | 50368 | float16 | 4.48.0.dev0 | 768 | 1,152 | 22 | 12 | null | null | 0 | null | 8,192 | null | null | 0.02 | 50281 | 50282 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | true | 0.155714 | small | long |
2,722 | Alidr79/speecht5_v1_20000_steps | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.43.3 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
2,723 | Alidr79/speecht5_v2_4000_best | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.43.3 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
2,724 | Aliem2001/whisper-small-fa | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.48.2 | null | null | 12 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
2,725 | AlienKevin/electra_hongkongese_small_pos_hkcancor | electra | ["ElectraForTokenClassification"] | 30000 | float32 | 4.43.3 | 256 | 1,024 | 12 | 4 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.009437 | small | short |
2,726 | AlignCoder/AlignRetriever | roberta | ["RobertaModel"] | 51416 | float32 | 4.49.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 1,026 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,727 | AlignQuant/Llama-2-13b-chat-hf-GPTQ-2bit | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.50.2 | 5,120 | 13,824 | 40 | 40 | 40 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
2,728 | AlignQuant/Llama-2-7b-chat-hf-OSTQuant | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.44.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,729 | AlignQuant/Meta-Llama-3-8B-Instruct-GPTQ-2bit | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.50.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
2,730 | AlignQuant/Meta-Llama-3-8B-Instruct-GPTQ-4bit | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.50.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
2,731 | AlignQuant/Meta-Llama-3-8B-Instruct-bitsandbytes-8bit | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.50.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
2,732 | Alignment-Lab-AI/Qwen2.5-Coder-7B-Instruct-132k | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.44.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
2,733 | Alignment-Lab-AI/e5-mistral-7b-instruct | mistral | ["MistralModel"] | 32000 | float16 | 4.34.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
2,734 | AliiAhmadi/whisper-fa | whisper | ["WhisperForConditionalGeneration"] | 51865 | float16 | 4.25.1 | null | null | 32 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
2,735 | Alimahmoud/custom-tuned-model | distilbert | ["DistilBertForSequenceClassification"] | 30522 | null | 4.56.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,736 | Alimahmoud/custom-tuned-model2 | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.46.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,737 | AlinaKl/babylm2024-git-txt | git | ["GitForCausalLM"] | 32778 | float32 | 4.26.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 1,024 | null | null | 0.02 | 101 | 102 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,738 | AlinaKl/babylm2024-git-vision | git | ["GitForCausalLM"] | 32778 | float32 | 4.26.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 1,024 | null | null | 0.02 | 101 | 102 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,739 | Alindstroem89/sentiment_model_testing | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.32.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,740 | Alinka9/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-singing_arctic_elk | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.51.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
2,741 | Alipour97/segformer-b0-scene-parse-150 | segformer | ["SegformerForSemanticSegmentation"] | null | float32 | 4.25.1 | null | null | null | [1, 2, 5, 8] | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
2,742 | Alireza0017/marian-finetuned-kde4-en-to-fr | marian | ["MarianMTModel"] | 59514 | float32 | 4.51.3 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
2,743 | AlirezaBaneshi/autotrain-test2-756523213 | roberta | ["RobertaForQuestionAnswering"] | 50265 | float32 | 4.15.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,744 | AlirezaBaneshi/autotrain-test2-756523214 | roberta | ["RobertaForQuestionAnswering"] | 50265 | float32 | 4.15.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,745 | Alirezaab78/test | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.47.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
2,746 | Alirezamp/gemma2b-news-gategory-lc_awq | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.42.3 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
2,747 | AlisaTarada/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.43.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,748 | Alisaqib123/rare-puppers | vit | ["ViTForImageClassification"] | null | float32 | 4.41.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
2,749 | AlisonWen/ppo-baseline-print-token-prob | llama | ["LlamaForCausalLM"] | 32001 | float16 | 4.51.3 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,750 | AlisonWen/ppo-baseline-print-token-prob-1 | llama | ["LlamaForCausalLM"] | 32001 | float16 | 4.51.3 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,751 | AlisonWen/ppo-seed-42-step-200 | llama | ["LlamaForCausalLM"] | 32001 | float16 | 4.45.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,752 | AlisonWen/rm-seed-0 | llama | ["LlamaForScore"] | 32001 | bfloat16 | 4.44.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,753 | Alissonerdx/Dia1.6-pt_BR-v1 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
2,754 | Alissonerdx/YuE-s1-7B-anneal-en-cot-exl2-3.0bpw | llama | ["LlamaForCausalLM"] | 83968 | bfloat16 | 4.42.0 | 4,096 | 11,008 | 32 | 32 | 4 | silu | 0 | True | 16,384 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
2,755 | Alissonerdx/YuE-s1-7B-anneal-en-cot-exl2-4.0bpw | llama | ["LlamaForCausalLM"] | 83968 | bfloat16 | 4.42.0 | 4,096 | 11,008 | 32 | 32 | 4 | silu | 0 | True | 16,384 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
2,756 | Alissonerdx/YuE-s1-7B-anneal-en-cot-exl2-5.0bpw | llama | ["LlamaForCausalLM"] | 83968 | bfloat16 | 4.42.0 | 4,096 | 11,008 | 32 | 32 | 4 | silu | 0 | True | 16,384 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
2,757 | Alissonerdx/YuE-s1-7B-anneal-en-cot-exl2-6.0bpw | llama | ["LlamaForCausalLM"] | 83968 | bfloat16 | 4.42.0 | 4,096 | 11,008 | 32 | 32 | 4 | silu | 0 | True | 16,384 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
2,758 | Alissonerdx/YuE-s1-7B-anneal-en-cot-exl2-8.0bpw | llama | ["LlamaForCausalLM"] | 83968 | bfloat16 | 4.42.0 | 4,096 | 11,008 | 32 | 32 | 4 | silu | 0 | True | 16,384 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
2,759 | Alissonerdx/YuE-s1-7B-anneal-en-cot-int8 | llama | ["LlamaForCausalLM"] | 83968 | float16 | 4.48.1 | 4,096 | 11,008 | 32 | 32 | 4 | silu | 0 | True | 16,384 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
2,760 | Alissonerdx/YuE-s1-7B-anneal-jp-kr-icl-int8 | llama | ["LlamaForCausalLM"] | 83968 | float16 | 4.48.1 | 4,096 | 11,008 | 32 | 32 | 4 | silu | 0 | True | 16,384 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
2,761 | Alissonerdx/YuE-s1-7B-anneal-zh-cot-int8 | llama | ["LlamaForCausalLM"] | 83968 | float16 | 4.48.1 | 4,096 | 11,008 | 32 | 32 | 4 | silu | 0 | True | 16,384 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
2,762 | Alissonerdx/YuE-s2-1B-general-exl2-3.0bpw | llama | ["LlamaForCausalLM"] | 83840 | bfloat16 | 4.42.0 | 2,048 | 5,504 | 32 | 16 | 16 | silu | 0 | True | 8,192 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.610613 | medium | long |
2,763 | Alissonerdx/YuE-s2-1B-general-exl2-4.0bpw | llama | ["LlamaForCausalLM"] | 83840 | bfloat16 | 4.42.0 | 2,048 | 5,504 | 32 | 16 | 16 | silu | 0 | True | 8,192 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.610613 | medium | long |
2,764 | Alissonerdx/YuE-s2-1B-general-exl2-5.0bpw | llama | ["LlamaForCausalLM"] | 83840 | bfloat16 | 4.42.0 | 2,048 | 5,504 | 32 | 16 | 16 | silu | 0 | True | 8,192 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.610613 | medium | long |
2,765 | Alissonerdx/YuE-s2-1B-general-exl2-6.0bpw | llama | ["LlamaForCausalLM"] | 83840 | bfloat16 | 4.42.0 | 2,048 | 5,504 | 32 | 16 | 16 | silu | 0 | True | 8,192 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.610613 | medium | long |
2,766 | Alissonerdx/YuE-s2-1B-general-exl2-8.0bpw | llama | ["LlamaForCausalLM"] | 83840 | bfloat16 | 4.42.0 | 2,048 | 5,504 | 32 | 16 | 16 | silu | 0 | True | 8,192 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.610613 | medium | long |
2,767 | Alissonerdx/YuE-s2-1B-general-int8 | llama | ["LlamaForCausalLM"] | 83840 | float16 | 4.48.1 | 2,048 | 5,504 | 32 | 16 | 16 | silu | 0 | True | 8,192 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.610613 | medium | long |
2,768 | Alitrix/bert-base-uncased_Banking77 | bert | ["BertForSequenceClassification"] | 119547 | float32 | 4.37.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,769 | Aliyyah/Roberta-Sentiment-Classifier | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.35.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,770 | AlkQ/a2c-PandaPickAndPlace-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
2,771 | AlkQ/a2c-PandaReachDense-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
2,772 | AlkanCan/BERTurk-128k-cased-immigration | bert | ["BertForSequenceClassification"] | 128000 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,773 | AlkanCan/TurkishBERTweet-Immigration-Stance | roberta | ["RobertaForSequenceClassification"] | 100000 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,774 | Alkohole/T-pro-it-1.0-4.5bpw-h8-exl2 | qwen2 | ["Qwen2ForCausalLM"] | 151665 | bfloat16 | 4.46.2 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
2,775 | All-SeeingEye/speecht5_finetuned_voxpopuli_es | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.46.2 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
2,776 | AllaNBarakat/xlm-roberta-base-fintuned-panx-de | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,777 | Allen-UQ/DeepSeek-R1-Distill-Qwen-1.5B-GRPO | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.49.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | False | 131,072 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
2,778 | Allen-UQ/Qwen2.5-1.5B-Instruct-SFT-2-Hop-Nei-Aug-Pubmed | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.49.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
2,779 | Allen-UQ/Qwen2.5-7B-Instruct-GRPO-Nei-Tokens-k20 | qwen2 | ["Qwen2ForCausalLMNode"] | 152064 | float32 | 4.49.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
2,780 | Allen1984/hw1 | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.44.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,781 | AllenGeng/OCamlBert | bert | ["BertForMaskedLM"] | 30522 | float32 | 4.20.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,782 | AllenMai/bert-chinese-ainews | bert | ["BertForSequenceClassification"] | 21128 | float32 | 4.31.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,783 | AllenMai/bert-test1 | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.31.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,784 | Allrandom/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,785 | AllyArc/tinyLlama_allyar | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,786 | Almahmood/whisper-medium-checkpoint | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.30.0.dev0 | null | null | 24 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
2,787 | Almahmood/whisper-small-hi | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.30.0.dev0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
2,788 | Almancy/finetuning-emotion-model-5-v3 | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.41.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,789 | Almenodera/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-keen_woolly_chimpanzee | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.50.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
2,790 | Almheiri/Llama-3.2-1B-Instruct-GPTQ-INT4 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.46.3 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
2,791 | AlmightyDeathCheater/DialoGPT-medium-harrypotter | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.25.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
2,792 | Almondpeanuts/pegasus-samsum | pegasus | ["PegasusForConditionalGeneration"] | 96103 | float32 | 4.28.1 | null | null | 16 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
2,793 | Almondpeanuts/xlm-roberta-base-finetuned-panx-all | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,794 | Almondpeanuts/xlm-roberta-base-finetuned-panx-de-fr | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,795 | Almondpeanuts/xlm-roberta-base-finetuned-panx-en | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,796 | Almondpeanuts/xlm-roberta-base-finetuned-panx-fr | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,797 | Almondpeanuts/xlm-roberta-base-finetuned-panx-it | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,798 | AlokBharadwaj/sd-class-butterflies-32_ddpm_pipeline_with_mae_loss | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,799 | AlokBharadwaj/sd-class-butterflies-32_ddpm_pipeline_with_smooth_l1_loss | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.