index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
15,700 | Iqbaliswinning/results | vit | ["ViTForImageClassification"] | null | float32 | 4.42.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,701 | Irathernotsay/qwen2-1.5B-medical_qa-Finetune | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float16 | 4.38.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
15,702 | IreNkweke/banking77 | distilbert | ["DistilBertForMaskedLM"] | 30522 | float32 | 4.41.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,703 | IreNkweke/bert-finetuned-squad | bert | ["BertForQuestionAnswering"] | 28996 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,704 | IreNkweke/marian-finetuned-kde4-en-to-fr | marian | ["MarianMTModel"] | 59514 | float32 | 4.41.2 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
15,705 | IreNkweke/marian-finetuned-kde4-en-to-fr-accelerate | marian | ["MarianMTModel"] | 59514 | float32 | 4.41.2 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
15,706 | IreneGaeta/clasificador-tweets-irene | electra | ["ElectraForSequenceClassification"] | 31002 | float32 | 4.47.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,707 | Irfan7/Florence-2-FT-DocVQA | florence2 | ["Florence2ForConditionalGeneration"] | 51289 | float16 | 4.45.1 | null | null | null | null | null | null | null | null | null | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,708 | Irina-Igmm/llama2-Immo | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.41.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,709 | IrinaArcadievna/poco-SoccerTwos | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,710 | Irny/Llama2-7b-finetuned | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.45.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,711 | Irny/Llama3-8b-finetuned | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.45.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,712 | Irny/distilbert-base-uncased-finetuned-cola | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.45.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,713 | Iroyoshi/Bert_Sent_NLP_Course | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.52.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,714 | Irsh77/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,715 | IrshadG/Clothes_Pattern_Classification_v2 | vit | ["ViTForImageClassification"] | null | float32 | 4.35.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,716 | IrshadG/Qwen1.5-0.5B-Chat-finetune | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.38.2 | 1,024 | 2,816 | 24 | 16 | 16 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 0.30199 | small | very_long |
15,717 | Irtasam/finetuned-model | bert | ["BertForSequenceClassification"] | 28996 | null | 4.37.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,718 | IrwinD/log_sage_ppo_model | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.39.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,719 | IrwinD/log_sage_reward_model | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.39.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,720 | IsaHug/llama-3-chat-empathy-test-FR | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.46.3 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
15,721 | Isaac-N/Llama-2-7b-chatbot-FAQ-hh | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.41.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,722 | Isaac18/practica_imc | vit | ["ViTForImageClassification"] | null | float32 | 4.33.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,723 | IsaacMwesigwa/autotrain-1pwox-g76oa | resnet | ["ResNetForImageClassification"] | null | float32 | 4.37.0 | null | null | null | null | null | relu | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,724 | IsaacMwesigwa/footballer-recognition | vit | ["ViTForImageClassification"] | null | float32 | 4.37.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,725 | IsaacMwesigwa/footballer-recognition-2 | resnet | ["ResNetForImageClassification"] | null | float32 | 4.37.0 | null | null | null | null | null | relu | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,726 | IsaacRodgz/Tamil-Hate-Speech | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.19.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,727 | IsaacSST/gpt2-xl-ft-d1 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.17.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,728 | IsaacSST/gpt2-xl-ft-d2 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.17.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,729 | IsaacSST/gpt2-xl-ft-d3 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.17.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,730 | IsaacSST/gpt2-xl-ft-d4-0.15-n-3 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.17.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,731 | IsaacSST/gpt2-xl-ft-d4-0.3 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.17.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,732 | Isaacgv/a2c-AntBulletEnv-v0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,733 | Isaacgv/a2c-PandaReachDense-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,734 | Isaacgv/poca-SoccerTwos | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,735 | Isaacgv/rl_course_vizdoom_health_gathering_supreme | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,736 | Isaacgv/whisper-tiny-minds14 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.31.0 | null | null | 4 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
15,737 | Isaacks/swin-tiny-patch4-window7-224-finetuned-cars | swin | ["SwinForImageClassification"] | null | float32 | 4.20.1 | 768 | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,738 | Isaacks/test_push | segformer | ["SegformerForSemanticSegmentation"] | null | float32 | 4.31.0 | null | null | null | [1, 2, 5, 8] | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,739 | IsaakChu/uuu_fine_tune_gpt2 | gpt2 | ["GPT2LMHeadModel"] | 21128 | float32 | 4.51.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 101 | 102 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,740 | IsaakChu/uuu_fine_tune_taipower | gpt2 | ["GPT2LMHeadModel"] | 21128 | float32 | 4.51.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 101 | 102 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,741 | IsakNordgren/Llama-3-8B-instruct-x2 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,742 | IsakNordgren/OrpoLlamaLlama-3 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,743 | IsakNordgren/gemma-1.1-7b-it-x2 | mistral | ["MistralForCausalLM"] | 32002 | bfloat16 | 4.42.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
15,744 | Isaskar/dinov2-base-ODIR-5K | dinov2 | ["Dinov2ForImageClassification"] | null | float32 | 4.40.2 | 768 | null | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,745 | Isaskar/dinov2-base-finetuned-eurosat | dinov2 | ["Dinov2ForImageClassification"] | null | float32 | 4.40.2 | 768 | null | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,746 | Isaskar/dinov2-base-finetuned-oxford | dinov2 | ["Dinov2ForImageClassification"] | null | float32 | 4.40.0 | 768 | null | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,747 | Isawany/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,748 | Isawany/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,749 | Iseratho/model_parse_xfm_bart_base-v0_1_0 | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.33.0 | null | null | 6 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
15,750 | Isha31101999/esm2_t6_8M_UR50D-finetuned-secondary-structure | esm | ["EsmForTokenClassification"] | 33 | float32 | 4.38.2 | 320 | 1,280 | 6 | 20 | null | gelu | null | True | 1,026 | null | null | 0.02 | null | null | null | 16 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.007373 | small | short |
15,751 | Isha31101999/fine-tuned-gpt2 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.42.4 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,752 | Ishagupta2010/sentiment-ai | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.46.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,753 | IshanHF97/bart_base_sciarguminer_768_with_loss | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.35.2 | null | null | 6 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
15,754 | IshanHF97/sciarguminer_t5_base_BEST_complete_data_1024 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.35.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,755 | IshanHF97/sciarguminer_t5_base_complete_data_epoch_10 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.35.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,756 | IshitaSingh/t5-base-finetuned-xsum | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.26.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,757 | IshitaSingh/t5-small-finetuned-xsum | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.24.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,758 | Ishreet1/FinanceLLM | mistral | ["MistralForCausalLM"] | 32008 | bfloat16 | 4.35.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
15,759 | Ishwak1/bert-ufc-win-predictor | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.51.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,760 | IskandarAqp/mi-super-modelo | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.40.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,761 | Islam2222/distilhubert-finetuned-gtzan | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.51.3 | 768 | 3,072 | 2 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.014156 | small | null |
15,762 | IslamMesabah/CoderAPI | codet5p | ["CodeT5pEncoderDecoderModel"] | null | float16 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,763 | Islanderintel/whisper-base-en | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.36.2 | null | null | 6 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
15,764 | IslemTouati/scene_segmentation | segformer | ["SegformerForSemanticSegmentation"] | null | float32 | 4.41.2 | null | null | null | [1, 2, 5, 8] | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,765 | IslemTouati/setfit_french | bert | ["BertModel"] | 250037 | float32 | 4.29.1 | 384 | 1,536 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.021234 | small | short |
15,766 | Isma/model_45000 | wav2vec2 | ["Wav2Vec2Model"] | 32 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,767 | Isma/v2_395k_all | wav2vec2 | ["Wav2Vec2ForPreTrainingWithMixupV2"] | 32 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,768 | Isma/v2_400k_all | wav2vec2 | ["Wav2Vec2ForPreTrainingWithMixupV2"] | 32 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,769 | Isma/v2_405k_all | wav2vec2 | ["Wav2Vec2ForPreTrainingWithMixupV2"] | 32 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,770 | IsmaMore/modelo-perros-gatos | vit | ["ViTForImageClassification"] | null | float32 | 4.22.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,771 | IsmatS/azeri-turkish-bert-ner | bert | ["BertForTokenClassification"] | 32000 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,772 | Isotonic/Dolphin-5.1-7b | mistral | ["MistralForCausalLM"] | 32001 | bfloat16 | 4.35.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
15,773 | Isotonic/Hermes-2-Pro-Mixtral-4x7B | mixtral | ["MixtralForCausalLM"] | 32032 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | 4 | 2 | false | false | false | false | false | false | false | true | true | true | 6.442451 | medium | very_long |
15,774 | Isotonic/MiniJamba-670M | jamba | ["JambaForCausalLM"] | 65536 | float32 | 4.39.3 | 768 | 3,072 | 16 | 12 | 8 | silu | 0 | True | null | null | 0.000001 | 0.02 | 1 | 2 | false | 64 | 1.5 | false | null | 2 | false | false | false | false | false | false | false | false | true | false | 0.113246 | small | null |
15,775 | Isotonic/NanoJamba-220M | jamba | ["JambaForCausalLM"] | 65536 | float32 | 4.39.3 | 512 | 2,048 | 12 | 8 | 8 | silu | 0 | True | null | null | 0.000001 | 0.02 | 1 | 2 | false | 64 | 1 | false | null | 2 | false | false | false | false | false | false | false | false | false | false | 0.037749 | small | null |
15,776 | Isotonic/OrcaAgent-llama3.2-8b | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.46.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,777 | Isotonic/informal_to_formal | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.22.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,778 | Isotopish/model | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
15,779 | Isotr0py/LLaVA-Video-7B-Qwen2-hf | llava_onevision | ["LlavaOnevisionForConditionalGeneration"] | null | float16 | 4.47.0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
15,780 | Isotr0py/Phi-3.5-vision-instruct-AWQ | phi3_v | ["Phi3VForCausalLM"] | 32064 | float16 | 4.44.1 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | False | 131,072 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | very_long |
15,781 | IsraNva/isranva | vit | ["ViTForImageClassification"] | null | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,782 | IsraelRam/israRam | vit | ["ViTForImageClassification"] | null | float32 | 4.33.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,783 | Issactoto/therapist | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 2,048 | 5,632 | 22 | 32 | 4 | silu | null | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
15,784 | Issacwong/poca-SoccerTwos | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,785 | IssamL/aragpt2-base | gpt2 | ["GPT2LMHeadModel"] | 64000 | float32 | 4.30.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 0 | 0 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,786 | IssamL/aragpt2-base2 | gpt2 | ["GPT2LMHeadModel"] | 64000 | float32 | 4.30.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 0 | 0 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,787 | IssamL/darijabertgenad | gpt2 | ["GPT2LMHeadModel"] | 64000 | float32 | 4.34.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 0 | 0 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,788 | IssamL/roberta_point_of_sales_classification | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.33.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,789 | Itamarb123/LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,790 | Itbanque/whisper-ja-zh-base | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.51.3 | null | null | 6 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
15,791 | ItchyB/rl_course_vizdoom_health_gathering_supreme | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,792 | ItchyChin/OrpoLlama-3-8B-memorize-tl-20240702 | llama | ["LlamaForCausalLM"] | 128258 | float16 | 4.42.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128256 | 128257 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,793 | ItchyChin/OrpoLlama-3-8B-memorize-translate-retrain | llama | ["LlamaForCausalLM"] | 128258 | float16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128256 | 128257 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,794 | ItchyChin/OrpoLlama-3-8B-memorize-translate-retrain-ml-20240001 | llama | ["LlamaForCausalLM"] | 128258 | float16 | 4.42.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128256 | 128257 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,795 | ItchyChin/OrpoLlama-3-8B-memorize-translate-retrain-ml-20240701 | llama | ["LlamaForCausalLM"] | 128258 | float16 | 4.42.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128256 | 128257 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,796 | ItchyChin/OrpoLlama-3-8B-memorize-translate-retrain-tl-202407150918 | llama | ["LlamaForCausalLM"] | 128258 | float16 | 4.42.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128256 | 128257 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,797 | Ithai/LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,798 | Ithai/Pyramids | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,799 | Ithai/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.