index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
6,600 | Cafet/whisper-tiny-mongolian-kazah | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.39.3 | null | null | 4 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
6,601 | CaffreyR/30-rte-prune | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.24.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,602 | CaffreyR/sst2_lora_bert | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,603 | Cailucky777/text-summarization | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.27.4 | null | null | 6 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,604 | Cainiao-AI/TAAS | TAAS | ["TAAS"] | 40000 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 2,048 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | medium |
6,605 | CaioXapelaum/Orca-2-7b-Patent-Instruct-Llama-2 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.44.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,606 | Caiyun-AI/DCPythia-6.9B | dcpythia | ["DCPythia"] | 50257 | float16 | 4.33.2 | null | 16,384 | null | null | null | null | null | null | null | null | null | null | 0 | 0 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
6,607 | CakePL/calculator_model_test | encoder-decoder | ["EncoderDecoderModel"] | null | float32 | 4.38.1 | null | null | null | null | null | null | null | null | null | null | null | null | null | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,608 | Cal3bd3v/llama-3-8b-Instruct-bnb-4bit-aiaustin-demo | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,609 | CalYYY/model_test | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.51.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,610 | Calam1/t5-small-finetuned-wikisql | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.26.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,611 | CalamitousFelicitousness/EVA-Qwen2.5-72B-v0.2-FP8-Dynamic | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.3 | 8,192 | 29,568 | 80 | 64 | 8 | silu | 0 | False | 131,072 | 1,000,000 | 0.00001 | 0.02 | null | 151643 | false | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
6,612 | CalamitousFelicitousness/EVA-Qwen2.5-72B-v0.2-INT8 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.3 | 8,192 | 29,568 | 80 | 64 | 8 | silu | 0 | False | 131,072 | 1,000,000 | 0.00001 | 0.02 | null | 151643 | false | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
6,613 | CalamitousFelicitousness/L3.1-70B-sunfall-v0.6.1-fp8-dynamic | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.0 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
6,614 | CalamitousFelicitousness/Llama-3.05-Nemotron-Tenyxchat-Storybreaker-70B-FP8-Dynamic | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.46.0 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
6,615 | CalamitousFelicitousness/Llama-3.1-70B-ArliAI-RPMax-v1.1-fp8-dynamic | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.2 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
6,616 | CalamitousFelicitousness/Yi-1.5-34B-Chat-16K-tokfix | llama | ["LlamaForCausalLM"] | 64000 | bfloat16 | 4.40.0 | 7,168 | 20,480 | 60 | 56 | 8 | silu | 0 | False | 16,384 | 5,000,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 7 | false | null | null | true | false | false | false | false | false | false | false | true | true | 36.993761 | large | long |
6,617 | CalamitousFelicitousness/Yi-1.5-34B-Chat-16K-tokfix-fp8-dynamic | llama | ["LlamaForCausalLM"] | 64000 | bfloat16 | 4.44.2 | 7,168 | 20,480 | 60 | 56 | 8 | silu | 0 | False | 16,384 | 5,000,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 7 | false | null | null | true | false | false | false | false | false | false | false | true | true | 36.993761 | large | long |
6,618 | CalderaAI/13B-Theseus-MK1 | llama | ["LlamaForCausalLM"] | 32032 | float32 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,619 | CalderaAI/30B-Lazarus-GPTQ4bit | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.28.1 | 6,656 | 17,920 | 60 | 52 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 31.897682 | large | medium |
6,620 | CalebE/llama3_mental_health_model_full | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,621 | Calin/vit-base-patch16-224-in21k-finetuned-lora-food101 | vit | ["ViTForImageClassification"] | null | float32 | 4.32.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
6,622 | CallComply/DeciLM-7B-Instruct-32k | deci | ["DeciLMForCausalLM"] | 32000 | bfloat16 | 4.35.2 | 4,096 | 14,336 | 32 | 32 | null | silu | null | True | 32,768 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | true | 6.442451 | medium | very_long |
6,623 | CallMeJeremy/DialoGPT-medium-THREEPIO | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.26.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,624 | CalvinSeamons/CS301r | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.33.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,625 | CalvinYorkCS/ppo-LunarLandar-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,626 | CalvinYorkCS/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,627 | Camais03/camie-tagger | pytorch | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,628 | CambioMoney/mixtral-instruct-MLFlowV4 | mixtral | ["MixtralForCausalLM"] | 32000 | float16 | 4.36.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | 8 | 2 | false | false | false | false | false | false | false | true | true | true | 6.442451 | medium | very_long |
6,629 | Cambridge-KAIST/SmolLM-14m-Dolma-v0.1-Proposed-full | llama | ["OutputEmbeddingSelectiveUpdate"] | 50280 | bfloat16 | 4.47.1 | 128 | 512 | 6 | 4 | 4 | silu | 0 | True | 2,048 | 100,000 | 0.00001 | 0.02 | 0 | 0 | false | 32 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.00118 | small | medium |
6,630 | Cambridge-KAIST/SmolLM-14m-Dolma-v0.1-Zloss-full | llama | ["LlamaForCausalLM"] | 50280 | bfloat16 | 4.46.3 | 128 | 512 | 6 | 4 | 4 | silu | 0 | True | 2,048 | 100,000 | 0.00001 | 0.02 | 0 | 0 | false | 32 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.00118 | small | medium |
6,631 | Cambridge-KAIST/untied-base-checkpoint-1000 | llama | ["LlamaForCausalLM"] | 50280 | bfloat16 | 4.46.3 | 128 | 512 | 6 | 4 | 4 | gelu | 0 | True | 2,048 | 100,000 | 0.00001 | 0.02 | 0 | 0 | false | 32 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.00118 | small | medium |
6,632 | Cambridge-KAIST2/SmolLM-14m-Dolma-v0.4-zloss | llama | ["LlamaForCausalLM"] | 50280 | float32 | 4.48.1 | 128 | 512 | 6 | 4 | 4 | silu | 0 | True | 2,048 | 100,000 | 0.00001 | 0.02 | 0 | 0 | false | 32 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.00118 | small | medium |
6,633 | CamelliaRan/test_medusa | null | [] | null | null | 4.34.1 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,634 | Camih/distilbert-base-uncased-finetuned-cola | distilbert | ["DistilBertForSequenceClassification"] | 30522 | null | 4.30.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,635 | Camih/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,636 | Camilovelez1/imdbreviews_classification_distilbert_v02 | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.39.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,637 | CanEPecor/t5-base-finetuned-small-epochs | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.46.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,638 | CanEPecor/t5-base-finetuned_3 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.46.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,639 | CanerAI/turkishReviews-ds-mini | gpt2 | ["GPT2LMHeadModel"] | 44208 | null | 4.44.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 0 | 0 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,640 | Cansino/mi-super-modelo | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,641 | Cantaosu/my_awesome_asr_mind_model | wav2vec2 | ["Wav2Vec2ForCTC"] | 32 | float32 | 4.40.2 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
6,642 | CaoHaiNam/obito-v0 | roberta | ["RobertaModel"] | 64001 | float32 | 4.39.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 258 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,643 | CaoHaiNam/obito-v1 | roberta | ["RobertaModel"] | 64001 | float32 | 4.40.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 258 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,644 | CaoTrungHieu/BERT_Snli | bert | ["BertModel"] | 30522 | float32 | 4.34.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,645 | CapEightFour/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,646 | CapitainData/dyu-fr-t5-base_v3 | t5 | ["T5ForConditionalGeneration"] | 32128 | null | 4.38.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,647 | CapitainData/dyu-fr-t5-small | t5 | ["T5ForConditionalGeneration"] | 32128 | null | 4.38.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,648 | CaptMetal/BuddAi | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | null | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | null | 32,768 | 1,000,000 | 0.00001 | null | 1 | 2 | null | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,649 | CaptainHaaz/FERmodelCNN | null | ["CNN"] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,650 | CaptainKoffski/guessly-yesno-classifier | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.48.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,651 | CaptainLazarus/burrnesha | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.34.0 | 3,200 | 8,640 | 26 | 32 | 32 | silu | null | True | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 100 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 3.19488 | medium | medium |
6,652 | CareerNinja/T5-Base-data-v4-model-v1 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.25.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,653 | CareerNinja/T5-Small-data-v4-model-v2 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.25.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,654 | CareerNinja/t5_large_1e-4_on_V3dataset | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.22.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,655 | CarelessLee/MCQ_pooled_full_rationale_confidence_predictor | llama | ["LlamaForSequenceClassification"] | 128256 | float32 | 4.43.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
6,656 | Carick/albert-base-v2-wordnet_dataset_two-fine-tuned | albert | ["AlbertForSequenceClassification"] | 30000 | float32 | 4.45.1 | 768 | 3,072 | 12 | 12 | null | gelu_new | null | null | 512 | null | null | 0.02 | 2 | 3 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,657 | Carick/bart-fine-tuned-term-typing | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.45.1 | null | null | 6 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,658 | Carick/roberta-base-wordnet_dataset_three-fine-tuned | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.45.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,659 | Carick/roberta-base-wordnet_dataset_two-fine-tuned | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.45.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,660 | Carick/xlm-roberta-base-wordnet_combined_one-fine-tuned | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.45.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,661 | Carick/xlm-roberta-base-wordnet_dataset_three-fine-tuned | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.45.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,662 | Carick/xlm-roberta-base-wordnet_dataset_two-fine-tuned | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.45.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,663 | Carina124/plant-vit-model-3 | vit | ["ViTForImageClassification"] | null | float32 | 4.28.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
6,664 | CarinaCHEN/test1 | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.51.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,665 | CarlBrendt/layoutlmv2-base-uncased_finetuned_docvqa | layoutlmv2 | ["LayoutLMv2ForQuestionAnswering"] | 30522 | float32 | 4.35.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,666 | CarlSable/distilbert-base-uncased-finetuned-squad-d5716d28 | distilbert | ["DistilBertForMaskedLM"] | 30522 | null | 4.49.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,667 | CarlSable/marian-finetuned-kde4-en-to-fr | marian | ["MarianMTModel"] | 59514 | float32 | 4.48.3 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,668 | CarlosApo01/Llama-2-7b-chat-finetune | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.35.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,669 | CarlosElArtista/poca-SoccerTwos | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,670 | CarlosFersoft/GPBusiness0001_Q6 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,671 | CarlosMarch/pruebaPostTagging-MarIA | roberta | ["RobertaForTokenClassification"] | 50262 | float32 | 4.46.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
6,672 | Carlosino/en2zh40 | marian | ["MarianMTModel"] | 65001 | float32 | 4.27.4 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,673 | Carlosino/en2zh50 | marian | ["MarianMTModel"] | 65001 | float32 | 4.27.4 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,674 | Carlosino/iwslt2017_1410 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.27.4 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,675 | Carlosino/iwslt2017_857 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.27.4 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,676 | Carlosino/zh2en40 | marian | ["MarianMTModel"] | 65001 | float32 | 4.27.4 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,677 | Carlosslocar/model | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.38.0.dev0 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
6,678 | Carlosslocar/test2 | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.39.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
6,679 | Carlosslocar/test3 | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.39.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
6,680 | Carlosslocar/test4 | gemma | ["GemmaForSequenceClassification"] | 256000 | float32 | 4.39.3 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
6,681 | CarmelaFinianos/Arabicttegypt | speecht5 | ["SpeechT5ForTextToSpeech"] | 98 | float32 | 4.46.2 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,682 | Carmenvt7/roberta-base-bne-ner | roberta | ["RobertaForTokenClassification"] | 50262 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,683 | CarolLiu999/chinese-emotion-model | distilbert | ["DistilBertForSequenceClassification"] | 119547 | float32 | 4.45.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,684 | CarolLiu999/chinese-emotion-model-all | distilbert | ["DistilBertForSequenceClassification"] | 119547 | float32 | 4.40.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,685 | CarolLiu999/vivit-finetuned-6class-10epoch-5 | vivit | ["VivitForVideoClassification"] | null | float32 | 4.40.1 | 768 | 3,072 | 12 | 12 | null | gelu_fast | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
6,686 | CarolLiu999/vivit-finetuned-6class-3epoch-4 | vivit | ["VivitForVideoClassification"] | null | float32 | 4.39.1 | 768 | 3,072 | 12 | 12 | null | gelu_fast | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
6,687 | CarperAI/diff-codegen-2b-v2 | codegen | ["CodeGenForCausalLM"] | 50295 | float16 | 4.22.0.dev0 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 1 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,688 | CarperAI/diff-codegen-350m-v2 | codegen | ["CodeGenForCausalLM"] | 50295 | float16 | 4.22.0.dev0 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 1 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,689 | CarperAI/diff-codegen-6b-v2 | codegen | ["CodeGenForCausalLM"] | 50295 | float16 | 4.22.0.dev0 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 1 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,690 | CarperAI/openai_summarize_tldr_ppo | gptj | ["GPTJForCausalLM"] | 50400 | float32 | 4.25.1 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 50256 | 50256 | false | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,691 | CarperAI/openai_summarize_tldr_sft | gptj | ["GPTJForCausalLM"] | 50400 | float32 | 4.25.1 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 50256 | 50256 | false | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,692 | CarrotAI/Llama-3.2-Rabbit-Ko-3B-Instruct-2412 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.46.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
6,693 | Carrymachine/model_output | bert | ["BertForSequenceClassification"] | 30000 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 300 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,694 | Cartinoe5930/CodeActAgent-Mistral-7b-v0.1_weight_diff | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,695 | Cartinoe5930/Einstein-v4-7B_weight_diff | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,696 | Cartinoe5930/Llama2_init_Mistral | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.36.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | medium |
6,697 | Cartinoe5930/MAmmoTH-7B-Mistral_weight_diff | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,698 | Cartinoe5930/MetaMath-Mistral-7B_weight_diff | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,699 | Cartinoe5930/QAlign-1.5B-stage-2-50K | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.49.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.