index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1,500 | AXCXEPT/EZO-Qwen2.5-32B-Instruct | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.44.2 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
1,501 | AXCXEPT/Qwen3-EZO-8B-beta | qwen3 | ["Qwen3ForCausalLM"] | 151936 | bfloat16 | 4.51.3 | 4,096 | 12,288 | 36 | 32 | 8 | silu | 0 | True | 40,960 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 7.247757 | large | very_long |
1,502 | AXCXEPT/glm-4-9b-chat-FT-ja-v0.3 | chatglm | ["ChatGLMForConditionalGeneration"] | 151552 | bfloat16 | 4.42.0.dev0 | 4,096 | null | 40 | 32 | null | null | 0 | True | null | null | null | null | null | [151329, 151336, 151338] | false | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | true | 8.053064 | large | null |
1,503 | AY00/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,504 | AYYOUBJA/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.47.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,505 | AYYOUBJA/xlm-roberta-base-finetuned-panx-de | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.47.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,506 | Aabbhishekk/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,507 | AadeshMndr/scene_segmentation | segformer | ["SegformerForSemanticSegmentation"] | null | float32 | 4.47.1 | null | null | null | [1, 2, 5, 8] | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,508 | Aadharsh/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,509 | Aadithya18/mygpt | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.35.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,510 | Aadithya18/mygpt-ft | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.35.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,511 | Aadithyak/XLtolli | xlnet | ["XLNetForSequenceClassification"] | 32000 | float32 | 4.45.1 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,512 | Aadithyak/augmented-robertolli-repo | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.45.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,513 | Aakali/whisper-medium-hi-translate-jee | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.41.0 | null | null | 24 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
1,514 | Aakash21/Noobie | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.42.3 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,515 | AakashJammula/qwen3-4b_quantized | qwen3 | ["Qwen3ForCausalLM"] | 151936 | float16 | 4.51.3 | 2,560 | 9,728 | 36 | 32 | 8 | silu | 0 | True | 40,960 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 80 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 2.831155 | medium | very_long |
1,516 | Aakashk/llama-2-7b-miniguanaco | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,517 | Aalaa-07/gpt2-finetuned_recipe_generation | gpt2 | ["GPT2LMHeadModel"] | 50270 | float32 | 4.47.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,518 | Aalaa/Fine_tuned_Vit_trash_classification | vit | ["ViTForImageClassification"] | null | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
1,519 | Aaltjo/81M-gpt2-openwebtext-alpacha | null | ["GPT2LMHeadModel"] | 50257 | null | null | null | null | 6 | 6 | null | null | null | null | 768 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | short |
1,520 | Aamir07/PPO-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,521 | AamirAli123/phi-1.5b-bf16-finetuned-medicine-conversational | phi | ["PhiForCausalLM"] | 51200 | float16 | 4.35.2 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,522 | Aananda-giri/GPT2-Nepali | gpt2 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,523 | Aanisha07/dummy | bert | ["BertForMaskedLM"] | 30532 | null | 4.26.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,524 | Aansh123/test_trainer | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.31.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,525 | Aaron171/autotrain-category_with_desc_8 | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.36.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,526 | Aaron171/autotrain-subcategory_with_desc_8 | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.36.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,527 | Aaron2599/Meta-Llama-3.1-8B-Instruct-TurboMind-AWQ-4bit | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.43.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
1,528 | Aaron43/SmolLM2-FT-MyDataset | llama | ["LlamaForCausalLM"] | 49152 | float32 | 4.46.3 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 8,192 | 100,000 | 0.00001 | 0.041667 | 1 | 2 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | long |
1,529 | AaronLim/llama-3.2-3b-it-Ecommerce-ChatBot | llama | ["LlamaForCausalLM"] | 128258 | float16 | 4.46.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128256 | 128257 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
1,530 | AaronLim/llama-3.2-3b-it-Library-ChatBot | llama | ["LlamaForCausalLM"] | 128258 | float16 | 4.48.1 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128256 | 128257 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
1,531 | AaronP1299/yelp_review_classifier | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.52.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,532 | Aarushhh/SEWY3-246M-math-base | gemma | ["Sewy3ForCausalLM"] | 49152 | bfloat16 | 4.48.1 | 400 | 1,600 | 28 | 16 | 4 | silu | 0 | True | 2,048 | 1,000,000 | 0.000001 | 0.02 | 1 | 2 | false | 25 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.05376 | small | medium |
1,533 | Aarushhh/SEWY3-246M-untrained | Sewy3 | ["Sewy3ForCausalLM"] | 49152 | float32 | 4.49.0.dev0 | 400 | 1,600 | 28 | 16 | 4 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | 1 | 2 | false | 25 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.05376 | small | very_long |
1,534 | Aarushhh/untrained-Qwen-2.5-7b-Instruct-Pruned | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float16 | 4.45.1 | 3,584 | 18,944 | 14 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 2.157969 | medium | very_long |
1,535 | Aarushhh/untrained-pruned-customffn-llama3.1-8b-large | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.44.2 | 4,096 | 53,248 | 4 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | long |
1,536 | Aarya4536/therapy-assist | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,537 | Aarya4536/therapy-bot-v1 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,538 | Aarya4536/tinyllama-therapy-bot-v1 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 2,048 | 5,632 | 22 | 32 | 4 | silu | null | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
1,539 | Aaryan39/hinglish-tts-3b-ft | llama | ["LlamaForCausalLM"] | 156940 | bfloat16 | 4.52.4 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
1,540 | AaryanK/Llama_3.1_8B_GRPO_Reasoning_XIOSERV | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
1,541 | Aas324532534/qa_squadshifts | bert | ["BertForQuestionAnswering"] | 28996 | null | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,542 | Aasdasd/test | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float16 | 4.51.3 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
1,543 | Aasdasd/test1 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.51.3 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
1,544 | Aashhh/model2b | gemma | ["GemmaForCausalLM"] | 256000 | float16 | 4.41.2 | 2,048 | 16,384 | 18 | 8 | 1 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
1,545 | Aashi1/Fine-tuning-Llama2 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,546 | AashishKumar/AIvisionGuard-v2 | vit | ["ViTForImageClassification"] | null | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
1,547 | Aashraf995/Creative-nerd0.1 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
1,548 | Aasif21/Final_Year_Project_Peft | gemma | ["GemmaForCausalLM"] | 256000 | float16 | 4.51.3 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
1,549 | Aasif21/collab-cot-trained | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float16 | 4.51.3 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 131,072 | 10,000 | 0.000001 | 0.02 | 151646 | 151643 | false | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
1,550 | Aavenir/avn-8bit-shard | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,551 | AayushMitra/llamafinetuned | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,552 | AayushyaL/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,553 | Ab1jeet/gpt2-finetuned | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.48.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,554 | AbacusResearch/RasGulla1-7b | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.37.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
1,555 | Abby-OGV/circa_boolq_yn | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.27.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,556 | Abby-OGV/circa_mnli_yn | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.26.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,557 | Abby-OGV/circa_yn | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.26.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,558 | AbbyMuso1/translation-swlu-2 | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.35.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,559 | Abdalla312/wav2vec2-base-finetuned-ks | wav2vec2 | ["Wav2Vec2ForSequenceClassification"] | 32 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
1,560 | Abdallah7/food_classifier | vit | ["ViTForImageClassification"] | null | null | 4.37.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
1,561 | AbdallahElsaadany/sqlllama | llama | ["LlamaForCausalLM"] | 32016 | bfloat16 | 4.38.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | long |
1,562 | AbdallahNasir/book-review-sentiment-classification | bert | ["BertForSequenceClassification"] | 100000 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,563 | Abdelkareem/dinov2-base-iq-cancer | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,564 | Abdelmageed95/bert-base-uncased-finetuned-squad | bert | ["BertForQuestionAnswering"] | 30522 | float32 | 4.28.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,565 | Abdelmnam/Qwen3-0.6B-Gensyn-Swarm-lumbering_stocky_bison | qwen3 | ["Qwen3ForCausalLM"] | 151936 | float32 | 4.51.3 | 1,024 | 3,072 | 28 | 16 | 8 | silu | 0 | True | 40,960 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.352322 | small | very_long |
1,566 | Abdelmouncef/myfirstmodel | camembert | ["CamembertForMaskedLM"] | 32005 | null | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 5 | 6 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,567 | Abdelrahma/trainer_output | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.51.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
1,568 | Abdelrahman2922/distilbert-base-uncased-Human_vs_Ai_text_detection | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.41.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,569 | Abdelrahman2922/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.41.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,570 | AbdelrahmanEssmat/long-t5-local-base | longt5 | ["LongT5ForConditionalGeneration"] | 32128 | float32 | 4.36.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,571 | AbdelrahmanMosly1/BERT-Merged-Dataset | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.36.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,572 | AbdelrahmanMosly1/BERT-twitter-First | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.36.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,573 | AbdelrahmanSR/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.38.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,574 | Abdelwahab201/xlm-roberta-base-finetuned-panx-de | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,575 | Abderlrahman/smolvla-panda-mujoco-reach-cube | null | [] | null | null | null | null | null | null | null | null | null | null | True | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,576 | Abderrahim2/bert-finetuned-Age | bert | ["BertForSequenceClassification"] | 32000 | float32 | 4.19.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | true | 0.084935 | small | short |
1,577 | AbderrahimAl/ner_model_v1 | distilbert | ["DistilBertForTokenClassification"] | 30522 | float32 | 4.20.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,578 | AbderrahmanSkiredj1/Ad-dabit-fadel | bert | ["BertForTokenClassification"] | 64000 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,579 | AbderrahmanSkiredj1/Arabic_text_embedding_for_sts | bert | ["BertModel"] | 64000 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,580 | AbderrahmanSkiredj1/BERTouch | bert | ["BertForSequenceClassification"] | 64000 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,581 | AbderrahmanSkiredj1/arabic_text_embedding_sts_arabertv02_arabicnlitriplet | bert | ["BertModel"] | 64000 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,582 | AbdhelE/draftv1 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.46.3 | null | null | 6 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
1,583 | Abdine/bert-base-cased | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,584 | Abdine/bert-finetuned-ner | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,585 | Abdo-604/whisper_small_hindi | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.34.0.dev0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
1,586 | AbdoMamdouh/silma-ai-peft-finetunning | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.43.4 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
1,587 | AbdoShahat/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.37.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,588 | Abdou/gpt2-dz-positive-comments | gpt2 | ["GPT2LMHeadModel"] | 50258 | float32 | 4.28.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,589 | Abdoul27/mosa_v1 | m2m_100 | ["M2M100ForConditionalGeneration"] | 256206 | float32 | 4.42.4 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
1,590 | AbdoulayeDIOP/lettuce-npk-vit | vit | ["ViTForImageClassification"] | null | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
1,591 | Abdrehman6224/Llama-2-7b-chat-finetune | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,592 | Abdu-GH/AraRest-Arabic-Restaurant-Reviews-Sentiment-Analysis | bert | ["BertForSequenceClassification"] | 64000 | float32 | 4.48.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,593 | Abdul-Basit123/llama-3-8b-Instruct-bnb-4bit-finetuned | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
1,594 | Abdul-Ib/multilingual-e5-small-2024 | bert | ["BertModel"] | 250037 | float32 | 4.35.2 | 384 | 1,536 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.021234 | small | short |
1,595 | Abdulhaqmaqbool12365749/savedend_model | bert | ["BertForSequenceClassification"] | 30000 | float32 | 4.41.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,596 | Abduljalil/Llama-2-7b-chat-finetune | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,597 | Abdullah010/wav2vec2-urdu-asr-commom-voice-9.0_model_final | wav2vec2 | ["Wav2Vec2ForCTC"] | 86 | float32 | 4.19.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
1,598 | Abdullah22/llama-2-7b-miniguanaco | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,599 | Abdullah707/Stable-Text-Encoder | clip_text_model | ["CLIPTextModel"] | 49408 | float32 | 4.22.0.dev0 | 768 | 3,072 | 12 | 12 | null | quick_gelu | 0 | null | 77 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.