index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
15,300 | ISTA-DASLab/Llama-3.1-8B-Instruct-HIGGS-GPTQ-3bit | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.47.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,301 | ISTA-DASLab/Llama-3.1-8B-Instruct-HIGGS-GPTQ-4bit | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.47.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,302 | ISTA-DASLab/Llama-3.2-3B-Instruct-AQLM-PV-2Bit-2x8 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.44.0.dev0 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
15,303 | ISTA-DASLab/Llama-3.3-70B-Instruct-HIGGS-GPTQ-4bit | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.47.0.dev0 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
15,304 | ISTA-DASLab/Meta-Llama-3-70B-AQLM-2Bit-1x16 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.40.0.dev0 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | long |
15,305 | ISTA-DASLab/Meta-Llama-3-8B-AQLM-2Bit-1x16 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,306 | ISTA-DASLab/Meta-Llama-3-8B-Instruct | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,307 | ISTA-DASLab/Meta-Llama-3.1-8B-Instruct-AQLM-PV-2Bit-2x8-hf | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.44.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,308 | ISTA-DASLab/Mistral-7B-v0.1-AQLM-2Bit-1x16-hf | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
15,309 | ISTA-DASLab/gemma-3-27b-it-GPTQ-4b-128g | gemma3 | ["Gemma3ForConditionalGeneration"] | null | bfloat16 | 4.50.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | null | [1, 106] | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
15,310 | ISTNetworks/gemma7b_sentiment_analysis | gemma | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,311 | ISxOdin/vit-base-oxford-iiit-pets | vit | ["ViTForImageClassification"] | null | float32 | 4.50.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,312 | IT20255756/deformable-detr-box-finetuned-weed-detection | deformable_detr | ["DeformableDetrForObjectDetection"] | null | float32 | 4.32.1 | null | null | null | null | null | null | 0 | null | 1,024 | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
15,313 | IT20429546/detr-resnet-50_finetuned-weed-detection | detr | ["DetrForObjectDetection"] | null | float32 | 4.32.1 | null | null | 6 | null | null | null | 0 | null | 1,024 | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
15,314 | ITESM/sentece-embeddings-BETO | bert | ["BertModel"] | 31002 | float32 | 4.19.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | 2 | 3 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,315 | ITESM/st_demo_2 | bert | ["BertModel"] | 30522 | float32 | 4.19.2 | 384 | 1,536 | 6 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.010617 | small | short |
15,316 | ITESM/st_demo_3 | bert | ["BertModel"] | 30522 | null | 4.8.2 | 384 | 1,536 | 6 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.010617 | small | short |
15,317 | ITESM/st_demo_4 | bert | ["BertModel"] | 30522 | null | 4.8.2 | 384 | 1,536 | 6 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.010617 | small | short |
15,318 | ITESM/st_demo_5 | bert | ["BertModel"] | 30522 | float32 | 4.19.2 | 384 | 1,536 | 6 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.010617 | small | short |
15,319 | ITESM/st_demo_6 | roberta | ["RobertaModel"] | 50265 | float32 | 4.19.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,320 | ITG/wav2vec2-large-xlsr-gl | wav2vec2 | ["Wav2Vec2ForCTC"] | 35 | float32 | 4.27.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
15,321 | ITSheep/a2c-PandaReachDense-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,322 | ITT-AF/ITT-42dot_LLM-PLM-1.3B-v1.0 | llama | ["LlamaForCausalLM"] | 50304 | float32 | 4.36.2 | 2,048 | 5,632 | 24 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.000001 | 0.01 | 50257 | 50256 | false | 64 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
15,323 | ITT-AF/ITT-AF-PLM-1.4B_v0.1 | llama | ["LlamaForCausalLM"] | 50304 | float32 | 4.36.2 | 2,048 | 5,632 | 28 | 32 | 4 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 0 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.409286 | medium | medium |
15,324 | ITT-AF/ITT-Yi-Ko-6B-v1.0 | llama | ["LlamaForCausalLM"] | 78464 | float32 | 4.36.2 | 4,096 | 11,008 | 32 | 32 | 4 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | medium |
15,325 | ITT-AF/ITT-Yi-Ko-6B-v2.0 | llama | ["LlamaForCausalLM"] | 78464 | float32 | 4.36.2 | 4,096 | 11,008 | 32 | 32 | 4 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | medium |
15,326 | ITZDOZZEN/sd-class-butterflies-32 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,327 | IVYUSER/sellertype0 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.42.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,328 | IVentureISB/MahaKumbh-Llama3.3-70B | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.0 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
15,329 | IamAbhiii/MyFinGPT | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.47.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,330 | IamAbhiii/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 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.46.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,331 | IamKE/ISY503-sentiment_analysis2 | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.40.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,332 | IamLazyman/phi3.5_unsloth | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
15,333 | IamYash/VA-LLM-0tyg1ha3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,334 | IamYash/VA-LLM-17lvww3k | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,335 | IamYash/VA-LLM-1oqdd5im | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,336 | IamYash/VA-LLM-68c1kk8z | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,337 | IamYash/VA-LLM-86fvla3k | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,338 | IamYash/VA-LLM-9mb439wj | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,339 | IamYash/VA-LLM-cotu1zak | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,340 | IamYash/VA-LLM-dl6no00p | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,341 | IamYash/VA-LLM-ib56ocwg | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,342 | IamYash/VA-LLM-l81278yn | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,343 | IamYash/VA-LLM-meiaa10a | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,344 | IamYash/VA-LLM-n8xlzj3z | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,345 | IamYash/VA-LLM-rt5dqquc | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,346 | IamYash/VA-LLM-tcsgik0l | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,347 | IamYash/VA-LLM-tlrhy8f6 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,348 | IamYash/VA-LLM-w769aco2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,349 | IamYash/VA-LLM-w7akk2df | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,350 | Iameashan27/FirstRepo | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,351 | Iamth0u/whisper-base-1_9_check3 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.50.3 | null | null | 6 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
15,352 | Iamvincent/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,353 | Ian-14/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,354 | IanA/GPTNeo-Dealmaker | gpt_neo | ["GPTNeoForCausalLM"] | 50259 | float32 | 4.25.1 | 2,048 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
15,355 | IanTseng/vis_items_with_hand_classfier | vit | ["ViTForImageClassification"] | null | null | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,356 | Ibisbill/raid_clean_reddit | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.45.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,357 | Ibisbill/raid_raddit_original | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.45.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,358 | Ibisbill/stage3_OpenR1_lr_1e5_epoch2 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.49.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 32,768 | 300,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
15,359 | Ibisbill/stage3_OpenThinker2-7B_lr_1e5 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.49.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
15,360 | Ibisbill/stage3_SimpleRL_lr_1e5_epoch2 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.49.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 4,096 | 10,000 | 0.000001 | 0.02 | null | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | medium |
15,361 | Ibnelaiq/Makise-Amadeus-Kurisu | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.27.4 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,362 | Ibrahemqasim/categories_en2ar | marian | ["MarianMTModel"] | 62802 | float32 | 4.48.3 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
15,363 | Ibrahim-Alam/finetuning-bert-base-uncased-on-amazon_polarity_7000_samples | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.28.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,364 | Ibrahim-Alam/finetuning-bert-base-uncased-on-yelp_polarity_7000_samples | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.28.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,365 | Ibrahim-Alam/finetuning-roberta-base-on-amazon_polarity_7000_samples | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.28.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,366 | Ibrahim-Alam/finetuning-roberta-base-on-yelp_polarity_7000_samples | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.28.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,367 | Ibrahim-Alam/finetuning-xlnet-base-cased-on-amazon_polarity_7000_samples | xlnet | ["XLNetForSequenceClassification"] | 32000 | float32 | 4.28.0 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,368 | Ibrahim-Alam/finetuning-xlnet-base-cased-on-imdb | xlnet | ["XLNetForSequenceClassification"] | 32000 | float32 | 4.28.0 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,369 | Ibrahim-Alam/finetuning-xlnet-base-cased-on-yelp_polarity_7000_samples | xlnet | ["XLNetForSequenceClassification"] | 32000 | float32 | 4.28.0 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,370 | Ibrahim-Alam/finetuning_roberta_on_Tweet_Sentiment_pos_neg_7000_samples | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.28.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,371 | Ibrahim-Alam/finetuning_roberta_on_cornel_sentiment_7000_samples | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.28.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,372 | Ibrahim-Alam/roberta-base_FTd_on_contract-nli | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.28.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,373 | IbrahimAL24/TNtoEng-By-Ibrahim-V1 | marian | ["MarianMTModel"] | 62834 | float32 | 4.44.0.dev0 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
15,374 | IbrahimAmin/code-switched-egyptian-arabic-whisper-small | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.41.1 | null | null | 12 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
15,375 | IbrahimAmin/hubert-arabic-spoken-dialect-classifier | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,376 | IbrahimAmin/marbertv2-arabic-written-dialect-classifier | bert | ["BertForSequenceClassification"] | 100000 | float32 | 4.33.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,377 | IbrahimS/ppo-LunarLander-v2-6-11 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,378 | IbrahimSalah/Arabic_Syllables_to_text_Converter_Using_MT5 | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.30.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,379 | IbrahimTarek/Boiler_gemma7b | gemma | ["GemmaForCausalLM"] | 256000 | float16 | 4.40.0.dev0 | 3,072 | 24,576 | 28 | 16 | 16 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
15,380 | IbtisamAfzal/T5_Fine_Tuned_on_Arxiv_Dataset | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.44.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,381 | IbtisamAfzal/cnn_news_summary_model_trained_on_reduced_data | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.44.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,382 | Icculus/distilbert-base-uncased-finetuned-emotion-OXO | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.38.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,383 | Ice1/my-mistral-q-finetune | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.34.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
15,384 | IceBruhOne/DialoGPT-medium-subjectai | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.27.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,385 | IceBruhOne/DialoGPT-medium-subjectai2 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.27.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,386 | IceBruhOne/mytestcharacter | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.27.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,387 | IceKingBing/SQL-Llama-v0.5 | llama | ["LlamaForCausalLM"] | 32017 | bfloat16 | 4.38.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | long |
15,388 | Icelandic-lt/convbert-small-igc-is | convbert | ["ConvBertModel"] | 96000 | float32 | 4.10.0.dev0 | 256 | 1,024 | 12 | 4 | null | gelu | null | null | 512 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.009437 | small | short |
15,389 | Icelandic-lt/electra-base-igc-is | electra | ["ElectraForPreTraining"] | 32105 | null | null | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,390 | IcerikMuhafizi/bert-base-turkish-uncased-IM2 | bert | ["BertForSequenceClassification"] | 32000 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,391 | IcerikMuhafizi/bert-base-turkish-uncased-IM3 | bert | ["BertForSequenceClassification"] | 32000 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,392 | IchIch/LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,393 | Ichigo2899/01-ai-Yi-1.5-9B-Chat-AWQ | llama | ["LlamaForCausalLM"] | 64000 | float16 | 4.38.2 | 4,096 | 11,008 | 48 | 32 | 4 | silu | 0 | True | 4,096 | 5,000,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 9.663676 | large | medium |
15,394 | Ichigo2899/Artigenz-Artigenz-Coder-DS-6.7B-AWQ | llama | ["LlamaForCausalLM"] | 32256 | float16 | 4.42.3 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 16,384 | 100,000 | 0.000001 | 0.02 | 32013 | 32014 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | long |
15,395 | Ichigo2899/WIZVIC-7b-TGI-GPTQ | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.30.2 | 4,096 | 11,008 | 32 | 32 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 6.442451 | medium | medium |
15,396 | Ichigo2899/WizardLM-13B-V1-0-Uncensored-SuperHOT-8K-TGI | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.29.2 | 5,120 | 13,824 | 40 | 40 | null | silu | null | True | 8,192 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 12.582912 | large | long |
15,397 | Ichsan2895/Garuda-7B | RefinedWebModel | ["RWForCausalLM"] | 65024 | float32 | 4.30.1 | 4,544 | null | null | null | null | null | 0 | True | null | null | null | 0.02 | 11 | 11 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,398 | Ichsan2895/Merak-7B-v2 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.32.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,399 | IcyKallen/bert-mini-finetuned-ner | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 256 | 1,024 | 4 | 4 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.003146 | small | short |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.