index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1,900 | AdnanRiaz107/CodePhi-3-mini-0.1cKlora | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.40.2 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 131,072 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | very_long |
1,901 | AdnanRiaz107/CodePhi-3-mini-4k-instruct-python | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.40.2 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | medium |
1,902 | AdnanRiaz107/Gemma2b-APPS | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.38.2 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
1,903 | AdnanRiaz107/santacoder-finetuned-the-stack-bash | roberta | ["RobertaForCausalLM"] | 50265 | float32 | 4.35.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | False | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,904 | AdnanRiaz107/testm1 | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.40.2 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 131,072 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | very_long |
1,905 | AdnaneElmar/gpt2-medquad-finetuned | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.46.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,906 | AdoHakurei/sd-class-butterflies-32 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,907 | AdoHakurei/sd-class-butterflies-64 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,908 | AdonaiHS/Pyramids-Training | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,909 | AdonaiHS/distilhubert-finetuned-gtzan | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.32.0 | 768 | 3,072 | 2 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.014156 | small | null |
1,910 | AdoraRL/Qwen2.5-VL-7B-Geo3k-ADORA-v0 | qwen2_5_vl | ["Qwen2_5_VLForConditionalGeneration"] | 152064 | bfloat16 | 4.49.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 128,000 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
1,911 | Adrian-Emanuel/fake_news_classifier | bert | ["BertForSequenceClassification"] | 32000 | float32 | 4.47.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | true | 0.084935 | small | short |
1,912 | Adrian/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.18.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,913 | Adrian/distilbert-base-uncased-finetuned-squad | distilbert | ["DistilBertForQuestionAnswering"] | 30522 | float32 | 4.18.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,914 | Adrian/distilbert-base-uncased-finetuned-squad-colab | distilbert | ["DistilBertForQuestionAnswering"] | 30522 | float32 | 4.18.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,915 | AdrianFarmadin/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,916 | AdrianKlessa/ModernBERT-base-ag-news | modernbert | ["ModernBertForSequenceClassification"] | 50368 | float32 | 4.48.0.dev0 | 768 | 1,152 | 22 | 12 | null | null | 0 | null | 8,192 | null | null | 0.02 | 50281 | 50282 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | true | 0.155714 | small | long |
1,917 | AdrianPerez3/Ejercicio2_Prioridad_Optim | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,918 | Adriana213/distilbert-base-uncased-finetuned-clinc | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.41.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,919 | Adriano2024/bert-finetuned-ner | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.46.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,920 | Adriatogi/segformer-b0-finetuned-segments-graffiti | segformer | ["SegformerForSemanticSegmentation"] | null | float32 | 4.38.2 | null | null | null | [1, 2, 5, 8] | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,921 | Adrielram/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,922 | AdrienB134/all-mpnet-base-v2-fr | mpnet | ["MPNetModel"] | 30527 | float32 | 4.41.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,923 | AdrienB134/parler-tts-mini-v1-Jenny-colab | parler_tts | ["ParlerTTSForConditionalGeneration"] | 32128 | float16 | 4.43.3 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
1,924 | AdrienVannson/final-model | opt | ["OPTForCausalLM"] | 50272 | float32 | 4.40.2 | 768 | null | 12 | 12 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | medium |
1,925 | Adrienkgz/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,926 | AdryKab47/Llama-2-7b-4bit-FT-GPTQ | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.37.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,927 | AdryKab47/llamaft | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.37.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,928 | Ads36/distilbert_big | distilbert | ["DistilBertForTokenClassification"] | 119547 | float32 | 4.49.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,929 | Ads36/distilbert_germanic | distilbert | ["DistilBertForTokenClassification"] | 119547 | float32 | 4.49.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,930 | Ads36/distilbert_slavic | distilbert | ["DistilBertForTokenClassification"] | 119547 | float32 | 4.49.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,931 | Adster/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,932 | Adster/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,933 | Adun/openthaigpt-1.0.0-7b-chat-beta-gptq-4bit | llama | ["LlamaForCausalLM"] | 56554 | float16 | 4.32.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,934 | Adun/openthaigpt-1.0.0-7b-chat-pantip | llama | ["LlamaForCausalLM"] | 41088 | float16 | 4.41.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,935 | Advaith1612/bart_convo-summarizer | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.41.2 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
1,936 | Advantech-EIOT/intel_llama-3.1-8b-instruct | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
1,937 | AdversarialRLHF/ppo_pythia410m_tldr6.9b_rm410mdata_mergedsft_propprefix | gpt_neox | ["GPTNeoXForCausalLM"] | 50304 | float32 | 4.50.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | True | 2,048 | 10,000 | null | 0.02 | 0 | 0 | false | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.30199 | small | medium |
1,938 | AdversarialRLHF/pythia410m-rm-tldr6.9b_logprobcondpropallprefix | gpt_neox | ["GPTNeoXForSequenceClassification"] | 50304 | float32 | 4.50.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | True | 2,048 | 10,000 | null | 0.02 | 0 | 0 | false | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.30199 | small | medium |
1,939 | AdversarialRLHF/pythia410m-rm-tldr6.9b_logprobcondpropprefix | gpt_neox | ["GPTNeoXForSequenceClassification"] | 50304 | float32 | 4.50.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | True | 2,048 | 10,000 | null | 0.02 | 0 | 0 | false | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.30199 | small | medium |
1,940 | AdversarialRLHF/pythia410m-sft-tldr | gpt_neox | ["GPTNeoXForCausalLM"] | 50304 | float32 | 4.50.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | True | 2,048 | 10,000 | null | 0.02 | 0 | 0 | false | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.30199 | small | medium |
1,941 | AdversarialRLHF/pythia410m-sft-tldr-allprefix | gpt_neox | ["GPTNeoXForCausalLM"] | 50304 | float32 | 4.50.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | True | 2,048 | 10,000 | null | 0.02 | 0 | 0 | false | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.30199 | small | medium |
1,942 | Adwait1199/bert-finetuned-ner-adwait | bert | ["BertForTokenClassification"] | 28996 | null | 4.26.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,943 | AdwayK/base_bert_tuned_on_TAC2017_as_NER | bert | ["BertForTokenClassification"] | 30522 | null | 4.18.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,944 | AdwayK/biobert_ncbi_disease_ner_tuned_on_TAC2017 | bert | ["BertForTokenClassification"] | 28996 | null | 4.18.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,945 | AdwayK/biobert_on_ADR_as_NER | bert | ["BertForTokenClassification"] | 30522 | null | 4.18.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,946 | AdwayK/hugging_face_biobert_MLMA | bert | ["BertForTokenClassification"] | 30522 | null | 4.18.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,947 | AdwayK/hugging_face_biobert_MLMAv2 | bert | ["BertForTokenClassification"] | 30522 | null | 4.18.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,948 | AdwayK/hugging_face_biobert_MLMAv3 | bert | ["BertForTokenClassification"] | 30522 | null | 4.18.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,949 | Adyatoni/wav2vec2-base-timit-demo | wav2vec2 | ["Wav2Vec2ForCTC"] | 36 | float32 | 4.35.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
1,950 | Adyyyyyyyyyyyy/bert-finetuned-squad | bert | ["BertForQuestionAnswering"] | 28996 | float32 | 4.48.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,951 | Adzka/test-reward-model | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,952 | Aeala/Alpaca-elina-65b-4bit | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.29.0.dev0 | 8,192 | 22,016 | 80 | 64 | null | silu | null | True | 2,048 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 64.424509 | large | medium |
1,953 | Aeala/Enterredaas-33b | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.30.0.dev0 | 6,656 | 17,920 | 60 | 52 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 31.897682 | large | medium |
1,954 | Aeala/GPT4-x-AlpacaDente-30b-4bit | llama | ["LlamaForCausalLM"] | 32016 | float16 | 4.29.0.dev0 | 6,656 | 17,920 | 60 | 52 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 31.897682 | large | medium |
1,955 | Aeala/VicUnlocked-alpaca-65b-4bit | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.30.0.dev0 | 8,192 | 22,016 | 80 | 64 | null | silu | null | True | 2,048 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 64.424509 | large | medium |
1,956 | Aedelon/a2c-PandaReachDense-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
1,957 | AeglosAI/aeglos-v1 | deberta-v2 | ["DebertaV2ForSequenceClassification"] | 128100 | float32 | 4.36.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,958 | Aeoluze/Qwen2-7B-Ero-Instruct | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.41.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
1,959 | AeroAlly/ppo-Lunalander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,960 | AescF/hubert-base-ls960-finetuned-common_language | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.33.2 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | null |
1,961 | Aethan-Krug/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,962 | AetherResearch/Cerebrum-1.0-7b | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.37.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
1,963 | AetherResearch/Cerebrum-1.0-8x7b | mixtral | ["MixtralForCausalLM"] | 32000 | float16 | 4.39.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | 8 | 2 | false | false | false | false | false | false | false | true | true | true | 6.442451 | medium | very_long |
1,964 | Aexeos/camembert-base-finetuned-allocine | camembert | ["CamembertForMaskedLM"] | 32005 | float32 | 4.41.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 5 | 6 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,965 | Aexyno/flan-t5-small-samsum | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.32.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,966 | Afa912/distilbert-base-uncased-finetuned-cola | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.38.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,967 | Afinucci/my_awesome_qa_model | distilbert | ["DistilBertForQuestionAnswering"] | 30522 | float32 | 4.26.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,968 | Afiqa/Bart-finetune-naturalScience | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.31.0 | null | null | 12 | null | null | null | 0 | False | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
1,969 | Afiqa/Bart-naturalScience | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.31.0 | null | null | 12 | null | null | null | 0.1 | False | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
1,970 | Afiqa/Llama-2-7b-chat-dental | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,971 | Afiqa/Llama-2-7b-chat-medical | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,972 | Afiqa/Llama-2-7b-chat-mqa | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,973 | AfnanTS/results | bert | ["BertForMaskedLM"] | 100000 | float32 | 4.27.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,974 | AfterRain007/results | roberta | ["RobertaForSequenceClassification"] | 50414 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,975 | Afterglow777/chemical-llama | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.37.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,976 | Afterglow777/chemical_dpo_model | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.37.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,977 | Afzal12345/my-chatbot-model | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.52.4 | 256 | 1,024 | 4 | 4 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.003146 | small | short |
1,978 | Agals/distilgpt2_eli5 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,979 | Aganith/aspect-based-sentiment-analyser | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
1,980 | Agastaya/my_awesome_billsum_model | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.38.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,981 | AgentGym/AgentEvol-7B | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.35.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | False | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,982 | AgentPublic/fabrique-reference-2 | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.33.2 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
1,983 | AgentPublic/guillaumetell-7b | mistral | ["MistralForCausalLM"] | 32002 | bfloat16 | 4.37.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
1,984 | AggaMin/llama-3-8b-Instruct-bnb-4bit-aiaustin-demo | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.42.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
1,985 | Agita/DistilBERT_test | roberta | ["RobertaModel"] | 50265 | float32 | 4.41.1 | 768 | 3,072 | 6 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
1,986 | Agneev/distilhubert-finetuned-gtzan | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.32.0.dev0 | 768 | 3,072 | 2 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.014156 | small | null |
1,987 | Agneev/speecht5_tts_voxpopuli_nl | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.33.0.dev0 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,988 | Agniruudrra/my_ebm_model_biobert | distilbert | ["DistilBertForTokenClassification"] | 119547 | float32 | 4.28.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,989 | Agnuxo/Llama-3.1-Minitron-4B-Instruct_CODE_Python-GGUF_Spanish_English_8bit | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
1,990 | Agnuxo/Llama-3.1-Minitron-4B-Instruct_CODE_Python_Spanish_English_Asistant-16bit-v2 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.43.4 | 4,096 | 14,336 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.221225 | medium | very_long |
1,991 | Agnuxo/Mamba-Codestral-7B-Instruct_CODE_Python-GGUF_Spanish_English_8bit | mistral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | false | null | null | null |
1,992 | Agnuxo/Mamba-Codestral-7B-Instruct_CODE_Python-Spanish_English_GGUF_4bit | mistral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | false | null | null | null |
1,993 | Agnuxo/Mamba-Codestral-7B-Instruct_CODE_Python_Spanish_English_16bit | mistral | ["MistralForCausalLM"] | 32768 | bfloat16 | 4.43.4 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
1,994 | Agnuxo/Mamba-Codestral-7B-v0.1-instruct-python_coding_assistant-GGUF_16bit | mistral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | false | null | null | null |
1,995 | Agnuxo/Mamba-Codestral-7B-v0.1-instruct-python_coding_assistant-GGUF_4bit | mistral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | false | null | null | null |
1,996 | Agnuxo/Mamba-Codestral-7B-v0.1-python_coding_assistant-GGUF_8bit | mistral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | false | null | null | null |
1,997 | Agnuxo/Mamba-Codestral-7B-v0.1-python_coding_assistant_16bit | mistral | ["MistralForCausalLM"] | 32768 | bfloat16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
1,998 | Agnuxo/Phi-3.5-mini-instruct-python_coding_assistant-GGUF_16bit | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
1,999 | Agnuxo/Phi-3.5-mini-instruct-python_coding_assistant-GGUF_4bit | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.