index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
7,200 | ChrisLunger/roberta-base-squad2-finetuned-squad2 | roberta | ["RobertaForQuestionAnswering"] | 50265 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
7,201 | ChrisUPM/BioBERT_Re_trained | bert | ["BertForSequenceClassification"] | 28996 | null | null | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
7,202 | ChrisXC/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-gliding_hunting_gazelle | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.50.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
7,203 | ChrisZeng/electra-large-discriminator-nli-efl-tweeteval | electra | ["ElectraForSequenceClassification"] | 30522 | float32 | 4.16.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | short |
7,204 | Christa27/docvqa_mini_subset | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,205 | Christian90/Pyramids1 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,206 | Christian90/SoccerTwos1 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,207 | Christian90/a2c-AntBulletEnv-v0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,208 | Christian90/a2c-PandaReachDense-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,209 | Christian90/ppo-SnowballTarget1 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,210 | Christian90/rl_course_vizdoom_health_gathering_supreme | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,211 | ChristianMD/ATCv2_Model | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.31.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
7,212 | ChristianMDahl/segFormer-b3-horizontal | segformer | ["SegformerForSemanticSegmentation"] | null | float32 | 4.28.1 | null | null | null | [1, 2, 5, 8] | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,213 | ChristianMDahl/segFormer-b3-vertical | segformer | ["SegformerForSemanticSegmentation"] | null | float32 | 4.28.1 | null | null | null | [1, 2, 5, 8] | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,214 | ChristianMDahl/segFormer_ver1_horizontal | segformer | ["SegformerForSemanticSegmentation"] | null | float32 | 4.28.1 | null | null | null | [1, 2, 5, 8] | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,215 | Christianvedel/OccCANINE | canine | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,216 | Christine567/distilbert-base-uncased-finetuned-clinc | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.16.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
7,217 | Christine567/distilbert-base-uncased-finetuned-clinc_v1 | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.16.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
7,218 | Christine789/Chapter2epochs3 | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.16.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
7,219 | Christoph911/GELECTRA-large-LegalQuAD | electra | ["ElectraForQuestionAnswering"] | 31102 | null | 4.13.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | short |
7,220 | ChristophBu/dqn-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,221 | ChristophBu/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,222 | ChristophSchuhmann/OpenAI-Ghibli-Style-Flux-Dev-Lora | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,223 | ChristopherABurns/mistral-7b-v0.2-gptq-2bit | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
7,224 | ChristosSevastopoulos/swin-tiny-patch4-window7-224-thecbbbfs | swin | ["SwinForImageClassification"] | null | float32 | 4.23.0.dev0 | 768 | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,225 | Chrome540/finetune_qwen | qwen2_5_vl | ["Qwen2_5_VLModel"] | 152064 | bfloat16 | 4.51.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
7,226 | Chrome540/qwen_f | qwen2_5_vl | ["Qwen2_5_VLModel"] | 152064 | bfloat16 | 4.51.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
7,227 | Chromik/lime-flan-t5-small | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.51.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
7,228 | Chromik/t5-lime-explainer-4.0 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.51.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
7,229 | ChromiumPlutoniumAI/BootyShakerAI | gpt2 | ["GPT2LMHeadModel"] | 50257 | null | null | 768 | 3,072 | 12 | 12 | null | null | null | null | 1,024 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | false | 0.084935 | small | short |
7,230 | ChuGyouk/ko-med-gemma-2-9b-it-merge2 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.44.2 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
7,231 | ChuTuanDuc/t5-small-finetuned-manimml-1.1 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.38.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
7,232 | ChuckMcSneed/DoubleGold-v0.5-123b-32k | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.36.2 | 8,192 | 28,672 | 144 | 64 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 115.964117 | xlarge | very_long |
7,233 | ChuckMcSneed/WinterGoliath-123b | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.36.0.dev0 | 8,192 | 28,672 | 144 | 64 | 8 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 115.964117 | xlarge | medium |
7,234 | Chuhaojin/PersuGPT | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.36.2 | 5,120 | 13,824 | 40 | 40 | 40 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
7,235 | ChumsAi/testmodel | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,236 | ChunB1/TinyReviews_adj | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.36.1 | 64 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
7,237 | ChunB1/TinyReviews_adv | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.36.1 | 64 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
7,238 | ChunB1/TinyReviews_common | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.36.1 | 64 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
7,239 | ChunB1/TinyReviews_raw | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.36.1 | 64 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
7,240 | Chunakorn/Fine-tuned_SQL_QA_Test | t5 | ["T5ForConditionalGeneration"] | 32128 | bfloat16 | 4.45.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
7,241 | Chunakorn/Finetune-t5small-with-DS | t5 | ["T5ForConditionalGeneration"] | 32128 | bfloat16 | 4.45.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
7,242 | Chung-Fan/billsum_model | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.30.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
7,243 | Chung-Fan/distilbart-pubmed-20k_bot_truncate | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.44.2 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
7,244 | Chung-Fan/distilbart-pubmed-20k_mid_truncate | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.44.2 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
7,245 | Chung-Fan/led-pubmed-20k-4096 | led | ["LEDForConditionalGeneration"] | 50265 | float32 | 4.38.2 | null | null | 6 | null | null | null | 0 | True | null | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,246 | Chung-Fan/my_awesome_model | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.27.4 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
7,247 | Chung-Fan/my_model | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.27.4 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
7,248 | Chung-Fan/my_t5_model | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.30.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
7,249 | Chung-Fan/pegasus-pubmed-20k-bottom-tokens | pegasus | ["PegasusForConditionalGeneration"] | 96103 | float32 | 4.40.0 | null | null | 16 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
7,250 | Chung-Fan/pegasus-pubmed-20k_bot_truncate | pegasus | ["PegasusForConditionalGeneration"] | 96103 | float32 | 4.44.2 | null | null | 16 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
7,251 | Chung-Fan/pegasus-pubmed-20k_mid_truncate | pegasus | ["PegasusForConditionalGeneration"] | 96103 | float32 | 4.44.2 | null | null | 16 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
7,252 | Chung-Fan/primera-pubmed-20k-4096 | led | ["LEDForConditionalGeneration"] | 50266 | float32 | 4.38.2 | null | null | 12 | null | null | null | 0 | True | 4,096 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | medium |
7,253 | Chunk245/RWKU_unlearned_20_npo_full_llama3 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.47.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
7,254 | Chunk245/RWKU_unlearned_30_rt_full_llama3 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.47.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
7,255 | Chunkyun/ner | roberta | ["RobertaForTokenClassification"] | 50265 | float32 | 4.24.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
7,256 | Chunyagi/learn_hf_food_not_food_text_classifier-distilbert-base-uncased | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.51.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
7,257 | Chvntala/whisper-small-hi | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.44.0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
7,258 | Chy084/my_awesome_patent_d_model | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.30.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
7,259 | Chy084/my_awesome_patent_t_model | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.30.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
7,260 | ChyoPyoRo/convenience-product-classifier | electra | ["ElectraForSequenceClassification"] | 35000 | float32 | 4.50.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
7,261 | CianKim/whisper-medium-kor_eng_largev3_ed_ev | whisper | ["WhisperForConditionalGeneration"] | 51866 | float32 | 4.51.3 | null | null | 32 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
7,262 | CianKim/whisper-tiny-kor_eng_tiny_ed_ob | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.51.3 | null | null | 4 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
7,263 | CianKim/whisper-tiny-kor_eng_tiny_ed_op | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.51.3 | null | null | 4 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
7,264 | Cicatrice/ast-finetuned-audioset-10-10-0.4593-finetuned-gtzan | audio-spectrogram-transformer | ["ASTForAudioClassification"] | null | float32 | 4.32.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
7,265 | Cicatrice/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,266 | Cicatrice/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,267 | Cicatrice/speecht5_finetuned_voxpopuli_it | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.32.0 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
7,268 | Cicatrice/whisper-tiny-finetuned-minds14-en | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.32.0 | null | null | 4 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
7,269 | Cicciokr/XLM-Roberta-Base-Latin-Uncased | xlm-roberta | ["XLMRobertaForMaskedLM"] | 250002 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
7,270 | Cicciokr/XLM-Roberta-Base-Latin-Uncased-V2 | xlm-roberta | ["XLMRobertaForMaskedLM"] | 250002 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
7,271 | Cidoyi/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,272 | Cielciel/aift-model-review-multiple-label-classification | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.38.0.dev0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
7,273 | Cielciel/model_review | distilbert | ["DistilBertForSequenceClassification"] | 30522 | null | 4.35.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
7,274 | CinthiaS/pegasus-samsum | pegasus | ["PegasusForConditionalGeneration"] | 96103 | float32 | 4.25.1 | null | null | 16 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
7,275 | Cippppy/my_awesome_model | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.37.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
7,276 | CiprianAgapi/results | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.53.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
7,277 | Circularmachines/my_awesome_food_model | vit | ["ViTForImageClassification"] | null | float32 | 4.27.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
7,278 | Cirilaron/DialoGPT-medium-jetstreamsam | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.21.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
7,279 | Cirilaron/DialoGPT-medium-raiden | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.19.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
7,280 | Citaman/command-r-37-layer | cohere | ["CohereForCausalLM"] | 256000 | bfloat16 | 4.39.3 | 8,192 | 22,528 | 37 | 64 | 64 | silu | 0 | True | 8,192 | 8,000,000 | null | 0.02 | 5 | 255001 | null | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 29.796336 | large | long |
7,281 | CitrusBoy/FinetunedModelV2.0 | phi3 | ["Phi3ForCausalLM"] | 32064 | float32 | 4.38.1 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 131,072 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | very_long |
7,282 | CitrusBoy/Gemma_Model_News | gemma | ["GemmaForCausalLM"] | 256000 | float32 | 4.38.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
7,283 | CjangCjengh/GaLLM-14B-v0.2 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.49.0 | 5,120 | 13,824 | 48 | 40 | 8 | silu | 0 | False | 131,072 | 1,000,000 | 0.00001 | 0.02 | 151643 | 151643 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 15.099494 | large | very_long |
7,284 | CjangCjengh/GaLLM-14B-v0.2-GPTQ-Int4 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float16 | 4.47.1 | 5,120 | 13,824 | 48 | 40 | 8 | silu | 0 | False | 131,072 | 1,000,000 | 0.00001 | 0.02 | 151643 | 151643 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 15.099494 | large | very_long |
7,285 | CjangCjengh/NomBert-hn2qn-v0.1 | bert | ["NomBertModel"] | 19142 | float32 | 4.48.1 | 768 | 3,072 | 6 | 8 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 96 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
7,286 | Cjy8s/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
7,287 | Cl1001/distilgpt2_eli5 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
7,288 | ClaireV/MLMA_5.3 | bert | ["BertForTokenClassification"] | 30522 | null | 4.18.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
7,289 | ClaireV/MLMA_Lab8 | bert | ["BertForTokenClassification"] | 30522 | null | 4.18.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
7,290 | ClarenceDan/016c80c1-de36-4a73-975f-5613ca0eb27e | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.46.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
7,291 | ClarenceDan/02932413-a931-48ab-9542-a645e8fbebbf | mistral | ["MistralForCausalLM"] | 32002 | bfloat16 | 4.46.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
7,292 | ClarenceDan/02a15fc6-d460-4f4e-a967-2a0edeed9a8e | llama | ["LlamaForCausalLM"] | 32016 | float16 | 4.46.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | False | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | long |
7,293 | ClarenceDan/05cc0da7-ae42-43ed-8d70-feef48fe5796 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.46.0 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
7,294 | ClarenceDan/062be0a6-f116-4d22-b0ff-045a4ddbff9b | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.46.0 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
7,295 | ClarenceDan/06a14248-3088-45aa-8cb6-ce2a5c99aeec | gptj | ["GPTJForCausalLM"] | 50401 | float32 | 4.46.0 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 50256 | 50256 | false | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
7,296 | ClarenceDan/06c47dbf-2b8c-41de-befb-4308bad5eafa | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.46.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | False | 4,096 | 10,000 | 0.000001 | 0.02 | null | 151643 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | medium |
7,297 | ClarenceDan/099a93c0-c082-494c-81a1-0215c7de339f | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.46.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
7,298 | ClarenceDan/09bab5c2-ae4b-4ca8-9f8a-6d3e08eb7860 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.46.0 | 16 | 64 | 2 | 4 | 4 | silu | 0 | False | 2,048 | 10,000 | 0.000001 | 0.02 | 0 | 2 | false | 4 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.000006 | small | medium |
7,299 | ClarenceDan/0a1e2e70-aec5-4099-a26f-75c7fbddf42b | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.46.0 | 2,304 | 9,216 | 26 | 8 | 4 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 288 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.656226 | medium | long |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.