index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
2,100 | AhmedTaha012/finetuned-ner-hadith-try3 | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.15.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
2,101 | AhmedTaha012/finetuned-ner-hadith-try3785 | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.15.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
2,102 | AhmedTaha012/finetuned-ner-hadith-try398 | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.15.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
2,103 | AhmedTaha012/gptNeo-125M-txt2ARXMLv1 | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.27.4 | 768 | null | null | null | null | null | 0 | True | 1,800 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | short |
2,104 | AhmedTaha012/gptneo-Txt-To-Json-v0.1.3 | gpt_neo | ["GPTNeoForCausalLM"] | 50411 | float32 | 4.27.4 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
2,105 | AhmedTaha012/gptneo-txt2arxml-ppo | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.27.4 | 768 | null | null | null | null | null | 0 | True | 1,800 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | short |
2,106 | AhmedTaha012/mangersFeedback-V1.0.2 | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.30.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,107 | AhmedTaha012/moviesReview5classBert | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.20.1 | 512 | 2,048 | 4 | 8 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.012583 | small | short |
2,108 | AhmedTaha012/moviesReview5classBert2 | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.20.1 | 512 | 2,048 | 4 | 8 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.012583 | small | short |
2,109 | AhmedTaha012/moviesReview5classRoberta | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.20.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,110 | AhmedTaha012/moviesReviewsClassifierbert1 | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.20.1 | 512 | 2,048 | 4 | 8 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.012583 | small | short |
2,111 | AhmedTaha012/moviesReviewsClassifierbert2 | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.20.1 | 512 | 2,048 | 4 | 8 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.012583 | small | short |
2,112 | AhmedTaha012/nextQuarter-status-V1.0.3 | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.30.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,113 | AhmedTaha012/output | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.20.1 | 512 | 2,048 | 4 | 8 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.012583 | small | short |
2,114 | AhmedTarek/ppo-Pyramids | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,115 | AhmedTarek/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,116 | AhmedWaheed1/whisper-small-hi | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.49.0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
2,117 | Ahmedhany216/Fake-news-gen | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.36.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
2,118 | Ahmedhisham/fullfinetunedRobertaNER | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.42.4 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
2,119 | Ahmedhisham/social_bias_Bert | distilbert | ["DistilBertForSequenceClassification"] | 30522 | null | 4.32.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,120 | Ahmeshen/a2c-PandaReachDense-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,121 | AhmetAytar/all-mpnet-base-v2-fine-tuned | mpnet | ["MPNetModel"] | 30527 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,122 | AhmetAytar/all-mpnet-base-v2-fine-tuned_17_textbook_grobid_semantic | mpnet | ["MPNetModel"] | 30527 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,123 | AhmetBaris/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,124 | Ahmetyigittt/xlm-roberta-sentiment | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,125 | Ahom666/whisper-small-hi | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.42.3 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
2,126 | Ahrefs/LongCamel-7b-32k | llama | ["LlamaForCausalLM"] | 32008 | float32 | 4.34.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,127 | Ahs2000/segformer-b0-scene-parse-150 | segformer | ["SegformerForSemanticSegmentation"] | null | float32 | 4.45.1 | null | null | null | [1, 2, 5, 8] | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,128 | AhsanShahid/Creative_AI | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.44.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
2,129 | AhsanZaidi/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,130 | Ahyun/generate_model | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.41.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
2,131 | Ai1terror/model | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
2,132 | Ai1terror/model1101 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
2,133 | AiAF/Finetuned-SCPWiki-032025-7B-Instruct | mistral | ["MistralForCausalLM"] | 32768 | bfloat16 | 4.45.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
2,134 | AiAF/KJV-LLM-Pretrained-V1.0 | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.48.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
2,135 | AiAF/KJV-LLM-Pretrained-V1.1 | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.48.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
2,136 | AiAF/Mistral-QLoRA-Pretraining-Test-v1.1 | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.48.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
2,137 | AiAF/Mistral-QLoRA-Pretraining-Test-v1.1_Merged | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.48.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
2,138 | AiAF/Pretrained-QLoRA-r9kilo-V1 | mistral | ["MistralForCausalLM"] | 32768 | bfloat16 | 4.49.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
2,139 | AiAF/Pretrained-SCP-1B-QLoRA | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.49.0 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
2,140 | AiAF/Pretrained_QLoRA_Codename-75567-V1.1_Ckpt-51 | mistral | ["MistralForCausalLM"] | 32768 | bfloat16 | 4.49.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
2,141 | AiAF/UFOs-Mistral7b-V3 | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.48.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
2,142 | AiCloser/Qwen2.5-0.5B-Instruct-Thinking | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.46.1 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
2,143 | AiCloser/Qwen2.5-32B-AGI | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.44.2 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
2,144 | AiForTheChurch/ChristianGPT-base-8bit | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.34.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,145 | AiForTheChurch/ChristianGPT-base-full-precision | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.34.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,146 | AiForTheChurch/ChristianGPT-catholic-8bit | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.34.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,147 | AiLab-IMCS-UL/lvbert | bert | ["BertModel"] | 32004 | float32 | 4.15.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,148 | AiLab-IMCS-UL/lvmed | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.29.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
2,149 | AiLen13/plumia_uma_water_advice_model | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.52.4 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
2,150 | AiTommy/decider_agent_test_2_merged | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
2,151 | AiTommy/llama-instruct-fi-q4 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
2,152 | Aib0/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,153 | Aichunks/wav2vec2-base-timit-demo-google-colab | wav2vec2 | ["Wav2Vec2ForCTC"] | 32 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
2,154 | Aicraftar/Tharo.G-Eco | Tharo.G-Eco | ["Tharo.GForCausalLM"] | 100352 | bfloat16 | 4.47.0 | 5,120 | 17,920 | 40 | 40 | 10 | silu | 0 | True | 16,384 | 250,000 | 0.00001 | 0.02 | 100257 | 100265 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 12.582912 | large | long |
2,155 | AidAFadaeian/Mt5_persian_translation_en | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
2,156 | Aida-Khaleghi/sentiment-analysis-distilbert-base-uncased-sst2 | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.51.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,157 | AidoAido/Llama3.1-8b-script-demo-eng | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
2,158 | AightBits/Qwen2.5-Coder-32B-Instruct-6.0bpw-h6-exl2 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.43.1 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
2,159 | AightBits/Qwen2.5-Coder-32B-Instruct-8.0bpw-h8-exl2 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.43.1 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
2,160 | AightBits/microsoft-phi-4-8.0bpw-h8-exl2 | phi3 | ["Phi3ForCausalLM"] | 100352 | bfloat16 | 4.47.0 | 5,120 | 17,920 | 40 | 40 | 10 | silu | 0 | True | 16,384 | 250,000 | 0.00001 | 0.02 | 100257 | 100265 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 12.582912 | large | long |
2,161 | AigizK/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,162 | AigizK/whisper-small-bak | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.26.0.dev0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
2,163 | Aiknight2001/finetuned-distilbert | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.42.4 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,164 | Aikozvezda/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.41.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,165 | Ailei/chatglm-qlora | chatglm | ["ChatGLMForConditionalGeneration"] | 65024 | float16 | 4.30.2 | 4,096 | null | null | 32 | null | null | 0 | True | null | null | null | null | null | 2 | false | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
2,166 | Ailoveu/ds_model | qwen2 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,167 | Aimlab/xlm-roberta-base-finetuned-urdu | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.19.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,168 | Aimodelbyharis/scriptgenerationmodel | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.34.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,169 | Aimodelbyharis/scriptgenerationmodelv2 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.34.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,170 | AimonLabs/hallucination-detection-model | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,171 | AinaSiddiqui/learn_hf_food_not_food_text_classifier-distilbert-base-uncased | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.46.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,172 | Air29/speecht5_finetuned_voxpopuli_nl | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.48.3 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
2,173 | AirMannanov/llm-course-hw1 | null | [] | 1024 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,174 | AirinElizabath/JBS-3 | opt | ["OPTForCausalLM"] | 50272 | float32 | 4.33.0.dev0 | 1,024 | null | 24 | 16 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | medium |
2,175 | AirrStorm/BERT-MLM-IMDb | bert | ["BertForMaskedLM"] | 28996 | float32 | 4.47.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,176 | AishaBaliyan/glamour-ai-skin-model | vit | ["ViTForImageClassification"] | null | float32 | 4.50.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
2,177 | AishaKanwal/ModelsClassification | vit | ["ViTForImageClassification"] | null | float32 | 4.31.0.dev0 | 768 | 3,072 | 8 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.056623 | small | null |
2,178 | AishaKanwal/ModelsViT_RiceDisease | vit | ["ViTForImageClassification"] | null | float32 | 4.31.0.dev0 | 768 | 3,072 | 2 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.014156 | small | null |
2,179 | Aishee06/Finetuned-Llama2-7b-chat | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,180 | AishitaSharma/sd-class-butterflies-32-new | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,181 | AishwaryaDixit/LunarLander | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,182 | Aitor/ppo-Pyramids | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,183 | Aitor/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,184 | AitorDL/MNLP_DPO_HS_lr1e-4_b01-hinge-MATH-lr1e-5_b005 | qwen3 | ["Qwen3ForCausalLM"] | 151936 | float32 | 4.52.3 | 1,024 | 3,072 | 28 | 16 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 64 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.352322 | small | very_long |
2,185 | AitorDL/MNLP_DPO_HS_lr1e-5_b01 | qwen3 | ["Qwen3ForCausalLM"] | 151936 | bfloat16 | 4.52.3 | 1,024 | 3,072 | 28 | 16 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 64 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.352322 | small | very_long |
2,186 | AitorDL/MNLP_DPO_model_HS2 | qwen3 | ["Qwen3ForCausalLM"] | 151936 | bfloat16 | 4.52.3 | 1,024 | 3,072 | 28 | 16 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 64 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.352322 | small | very_long |
2,187 | Aitrepreneur/Florence-2-large | florence2 | ["Florence2ForConditionalGeneration"] | 51289 | float16 | 4.41.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,188 | Aitrepreneur/llava-llama-3-8b-text-encoder-tokenizer | llama | ["LlamaForCausalLM"] | 128320 | float16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
2,189 | Aitrepreneur/stable-vicuna-13B-GPTQ-4bit-128g | llama | ["LlamaForCausalLM"] | 32001 | float16 | 4.28.1 | 5,120 | 13,824 | 40 | 40 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 12.582912 | large | medium |
2,190 | Aityz/aityz_model | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.31.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
2,191 | Aityz/aityz_model_eli5 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.29.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
2,192 | Aityz/eli5_distilgpt2_mini | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.29.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
2,193 | Aityz/reviews_model | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.31.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
2,194 | Aivesa/0a3020fa-9065-4a04-b9c7-e0cadb2c120d | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.47.1 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
2,195 | Aivesa/0d041048-0da7-4497-ac21-da31409a0bf2 | gpt_neox | ["GPTNeoXForCausalLM"] | 50304 | float16 | 4.47.1 | 128 | 512 | 6 | 4 | null | gelu | 0 | False | 2,048 | 10,000 | null | 0.02 | 0 | 0 | false | 32 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.00118 | small | medium |
2,196 | Aivesa/14faec3b-678a-4b3e-a3e7-95067cb131aa | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.47.1 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151643 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
2,197 | Aivesa/1931e729-dd3f-4026-a118-7fef2d3afe59 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.47.1 | 16 | 64 | 2 | 4 | 4 | silu | 0 | False | 2,048 | 10,000 | 0.000001 | 0.02 | 0 | 2 | false | 4 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.000006 | small | medium |
2,198 | Aivesa/23784452-b886-4280-9bfa-946e195d84e3 | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.47.1 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
2,199 | Aivesa/286fd9a8-760d-4213-a477-b5ea1e7ce9a3 | codegen | ["CodeGenForCausalLM"] | 51200 | float32 | 4.47.1 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 50256 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.