index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
15,400 | Iddah/llama3-8B-formal | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.39.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,401 | Iddah/llama3-8B-instruct-informalization | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.39.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,402 | Ideapad/qa_entropy_all_1e-05_full_5_forget10_layer-15 | llama | ["CustomLlamaForCausalLM"] | 32000 | bfloat16 | 4.47.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,403 | Ideapad/qa_entropy_all_1e-05_full_5_forget10_layer-None | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.47.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,404 | Ididsomething/my-awesome-model | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,405 | IdleIdiot/thtf-7b | llama | ["LlamaForCausalLM"] | 65000 | float16 | 4.38.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,406 | Idriska/rubert-tiny2-appartment-clf | bert | ["BertForSequenceClassification"] | 83828 | float32 | 4.33.2 | 312 | 600 | 3 | 12 | null | gelu | null | True | 2,048 | null | null | 0.02 | null | null | null | 26 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.003504 | small | medium |
15,407 | IeBoytsov/llama3-1-ox-llms-8b-sft-only-germany-data-and-ultrafeedback | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.45.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,408 | Iedha/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-lethal_tawny_deer | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.51.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
15,409 | IffYuan/FSD-Stage1-Reasoning | llava | ["LlavaLlamaForCausalLM"] | 32000 | bfloat16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
15,410 | IffYuan/FSD-v1 | llava | ["LlavaLlamaForCausalLM"] | 32000 | bfloat16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
15,411 | IffYuan/FSD-v1.1 | llava | ["LlavaLlamaForCausalLM"] | 32000 | bfloat16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
15,412 | Iftisyed/my_distilbert1973 | bert | ["BertForMaskedLM"] | 30522 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | 0 | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,413 | Iftisyed/pk_roberta_pakner_model | roberta | ["RobertaForTokenClassification"] | 50265 | float32 | 4.34.0 | 768 | 3,072 | 6 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
15,414 | Ifyokoh/Igbo-sentiment-bbc | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 70006 | float32 | 4.42.4 | 768 | 3,072 | 10 | 6 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 128 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.070779 | small | short |
15,415 | Iggor95/layoutlmv2-base-uncased_finetuned_docvqa | layoutlmv2 | ["LayoutLMv2ForQuestionAnswering"] | 30522 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,416 | IgnacioFucksmann/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,417 | IgnacioFucksmann/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,418 | Ignaciobfp/vit-estadosfenologicos-384-prueba | vit | ["ViTForImageClassification"] | null | float32 | 4.49.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,419 | Igniter909/xlm-roberta-base-finetuned-panx-en | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,420 | Igniter909/xlm-roberta-base-finetuned_panx_de | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,421 | IgorKolodziej/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,422 | Igorsp/codegen-350M-mono-python-18k-alpaca | codegen | ["CodeGenForCausalLM"] | 51200 | float16 | 4.34.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,423 | Igortin/dummy-model | camembert | ["CamembertForMaskedLM"] | 32005 | float32 | 4.49.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 5 | 6 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,424 | Igortin/my-dummy-model | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.49.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,425 | IgriCB/Training_DATA | gemma | ["GemmaForCausalLM"] | 256000 | float16 | 4.44.2 | 3,072 | 24,576 | 28 | 16 | 16 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
15,426 | Ihjass/detr-finetuned-135samples | detr | ["DetrModel"] | null | float32 | 4.7.0.dev0 | null | null | 6 | null | null | null | 0 | null | 1,024 | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
15,427 | Ihor/Llama3-OpenBioLLM-8B-AWQ | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.46.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,428 | Ihor/gliclass-mt5-base-lw | GLiClass | ["GLiClassModel"] | 250102 | float32 | 4.51.3 | 768 | null | null | null | null | null | null | null | null | null | null | 0.03 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,429 | Ij4r/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-sedate_shrewd_cobra | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.52.4 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
15,430 | IkariDev/Athena-v4 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.33.3 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
15,431 | IkePy/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-mangy_purring_lion | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.51.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
15,432 | Ikeofai/distilbert-base-uncased-finetuned-squad | distilbert | ["DistilBertForQuestionAnswering"] | 30522 | float32 | 4.38.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,433 | Iker/ClickbaitFighter-7B | mistral | ["MistralForCausalLM"] | 32002 | bfloat16 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | long |
15,434 | Iker/Llama-3-Instruct-Neurona-8b-v2 | llama | ["LlamaForCausalLM"] | 128264 | bfloat16 | 4.42.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,435 | Iker/Neurona-2b | gemma | ["GemmaForCausalLM"] | 256016 | bfloat16 | 4.40.0 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
15,436 | Ikhee10/my_awesome_qa_model | distilbert | ["DistilBertForQuestionAnswering"] | 30522 | null | 4.34.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,437 | Ilamparithi-I/CIC-CourseQA-LLM | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
15,438 | IlhamEbdesk/bge-base-financial-matryoshka_test_my | bert | ["BertModel"] | 30522 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,439 | Ilhamben/camemBert_sentiment_final | camembert | ["CamembertForSequenceClassification"] | 32005 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 5 | 6 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,440 | Ilhamben/sarcasm_detection_french_camembert | camembert | ["CamembertForSequenceClassification"] | 32005 | float32 | 4.36.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 5 | 6 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,441 | IliyanGochev/distil-whisper-large-v3-bg | whisper | ["WhisperForConditionalGeneration"] | 51866 | float32 | 4.35.2 | null | null | 32 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
15,442 | Illiyas2024/folrence-2-large-8bit | florence2 | ["Florence2ForConditionalGeneration"] | 51289 | float16 | 4.42.4 | null | null | null | null | null | null | null | null | null | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,443 | Illuminati-014/whisper-large-v3-turbo-modified | whisper | ["WhisperForConditionalGeneration"] | 51866 | float32 | 4.49.0 | null | null | 32 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
15,444 | Illuminati-014/whisper-large-v3-turbo-modified-v1 | whisper | ["WhisperForConditionalGeneration"] | 51866 | float32 | 4.49.0 | null | null | 32 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
15,445 | Illuminati1/123 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,446 | IlluminatiPudding/a2c-PandaPickAndPlace-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
15,447 | IlluminatiPudding/a2c-PandaReachDense-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
15,448 | IlmarM/detr-fashion-clothes | detr | ["DetrForObjectDetection"] | null | float32 | 4.51.3 | null | null | 6 | null | null | null | 0 | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,449 | IlonaF/finetuning-misinfo-model-1000-Zhaohui | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.25.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,450 | IlyaGusev/T-lite-instruct-0.1-abliterated | llama | ["LlamaForCausalLM"] | 128259 | bfloat16 | 4.42.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,451 | IlyaGusev/gemma-2-2b-it-abliterated | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.43.3 | 2,304 | 9,216 | 26 | 8 | 4 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | [1, 107] | null | 288 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.656226 | medium | long |
15,452 | IlyaGusev/gemma-2-9b-it-abliterated | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.42.4 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
15,453 | IlyaGusev/llama_7b_ru_turbo_alpaca_lora_merged | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.28.0.dev0 | 4,096 | 11,008 | 32 | 32 | null | silu | null | True | null | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 6.442451 | medium | null |
15,454 | IlyaGusev/roberta-base-informal-tagger | roberta | ["RobertaForTokenClassification"] | 50265 | float32 | 4.20.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,455 | IlyaGusev/ruadapt_ficbook_llama2_7b | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.36.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,456 | IlyaGusev/saiga_aya_23_35b_sft_m1_d5 | cohere | ["CohereForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 8,192 | 22,528 | 40 | 64 | 64 | silu | 0 | True | 8,192 | 8,000,000 | null | 0.02 | 5 | 255001 | null | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 32.212255 | large | long |
15,457 | IlyaGusev/saiga_aya_23_35b_sft_m1_d5_awq_4bit | cohere | ["CohereForCausalLM"] | 256000 | float16 | 4.41.1 | 8,192 | 22,528 | 40 | 64 | 64 | silu | 0 | True | 8,192 | 8,000,000 | null | 0.02 | 5 | 255001 | null | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 32.212255 | large | long |
15,458 | IlyaGusev/saiga_gemma3_12b_sft_m2_d19_smpo_m3_d39 | gemma3 | ["Gemma3ForConditionalGeneration"] | null | bfloat16 | 4.51.3 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | null | [1, 106] | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
15,459 | IlyaGusev/saiga_llama3_70b_sft_m1_d5_abliterated_gptq_4bit | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.41.2 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | long |
15,460 | IlyaGusev/saiga_llama3_8b_sft_m11_d5_abliterated_kto_m5_d3 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.41.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,461 | IlyaGusev/saiga_llama3_8b_sft_m11_d7_abliterated_kto_m7_d7 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.42.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,462 | IlyaGusev/saiga_nemo_12b_sft_m10_d16_simpo_m23_d36 | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.46.2 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
15,463 | IlyaGusev/saiga_qwen2_7b_sft_m2_d6 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.42.0.dev0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
15,464 | IlyaGusev/saiga_qwen2_7b_sft_m2_d6_kto_m1_d5 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.42.0.dev0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
15,465 | IlyaGusev/t5-base-filler-informal | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.20.1 | null | null | null | null | null | null | null | True | null | null | null | null | 2 | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,466 | IlyaGusev/vikhr_nemo_12b_gptq_8bit | mistral | ["MistralForCausalLM"] | 131074 | float16 | 4.46.2 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
15,467 | IlyasMoutawwakil/sbert-all-MiniLM-L6-v2-with-pooler | bert | ["BertModel"] | 30522 | null | 4.39.3 | 384 | 1,536 | 6 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.010617 | small | short |
15,468 | IlyasMoutawwakil/test-benchmarks | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,469 | IlyasMoutawwakil/tiny-random-llama | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.35.1 | 16 | 64 | 2 | 4 | 4 | silu | null | True | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 4 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.000006 | small | medium |
15,470 | IlyasMoutawwakil/vicuna-7b-v1.5-awq-marlin | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.37.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,471 | ImKH/Llama-3-8B-Instruct_1562_train | llama | ["LlamaForCausalLM"] | 128264 | float16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,472 | ImKH/Llama-3-8B-Instruct_1950_train | llama | ["LlamaForCausalLM"] | 128264 | float16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,473 | ImKH/Llama-3-8B-Instruct_3512_train | llama | ["LlamaForCausalLM"] | 128264 | float16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,474 | ImNobody/wav2vec2-large-xlsr-53-demo-colab | wav2vec2 | ["Wav2Vec2ForCTC"] | 33 | float32 | 4.28.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
15,475 | ImSaito/bloom-chistes-spanish | bloom | ["BloomForCausalLM"] | 250880 | float16 | 4.31.0 | 1,024 | null | null | null | null | null | 0 | True | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,476 | ImSaito/gpt2-chistes-spanish | gpt2 | ["GPT2LMHeadModel"] | 50257 | float16 | 4.31.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,477 | ImSaito/llama-2-chistes-spanish | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,478 | ImSaito/llama2es-chistes-spanish | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 2,048 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,479 | ImSakushi/nistraal-2 | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.38.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
15,480 | ImSota/LLama_LoRA | mistral | ["MistralForCausalLM"] | 32064 | bfloat16 | 4.51.3 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | true | false | false | true | 3.623879 | medium | medium |
15,481 | ImValll/gemma-training_test_colab | gemma | ["GemmaForCausalLM"] | 256000 | float16 | 4.38.0 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
15,482 | Imad03/Prueba_Gradio | vit | ["ViTForImageClassification"] | null | float32 | 4.22.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,483 | ImadAziz/DialoGPT-Sheldon | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.22.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,484 | Imadken/Mistral-7B-v0.1-platypus | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.34.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
15,485 | Imadsarvm/Sarvm-Translation | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.27.4 | null | null | 24 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
15,486 | Image-editing/Emu3-Base-SFT-reasoning_super_concise-Apr17_lr1e-5-checkpoint-32000 | Emu3 | ["Emu3ForCausalLM"] | 184622 | bfloat16 | 4.44.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0.1 | False | 4,300 | 1,000,000 | 0.00001 | 0.02 | 151849 | 151850 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | medium |
15,487 | Image-editing/Emu3-Base-SFT-reasoning_super_concise-Apr17_rinp_lr1e-5-checkpoint-34950 | Emu3 | ["Emu3ForCausalLM"] | 184622 | bfloat16 | 4.44.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0.1 | False | 4,300 | 1,000,000 | 0.00001 | 0.02 | 151849 | 151850 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | medium |
15,488 | Image-editing/Emu3-Base-SFT-reasoning_verbose-May02_post_stochastic_lr1e-4-checkpoint-1600 | Emu3 | ["Emu3ForCausalLM"] | 184622 | bfloat16 | 4.44.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0.1 | False | 4,300 | 1,000,000 | 0.00001 | 0.02 | 151849 | 151850 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | medium |
15,489 | Image-editing/Emu3-Base-SFT-reasoning_verbose-May02_post_stochastic_lr1e-4-checkpoint-2400 | Emu3 | ["Emu3ForCausalLM"] | 184622 | bfloat16 | 4.44.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0.1 | False | 4,300 | 1,000,000 | 0.00001 | 0.02 | 151849 | 151850 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | medium |
15,490 | Image-editing/Emu3-Base-SFT-reasoning_verbose-May02_post_stochastic_lr1e-4-checkpoint-3200 | Emu3 | ["Emu3ForCausalLM"] | 184622 | bfloat16 | 4.51.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0.1 | False | 4,300 | 1,000,000 | 0.00001 | 0.02 | 151849 | 151850 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | medium |
15,491 | Image-editing/Emu3-Base-SFT-reasoning_verbose-May02_post_stochastic_lr1e-4-checkpoint-4000 | Emu3 | ["Emu3ForCausalLM"] | 184622 | bfloat16 | 4.51.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0.1 | False | 4,300 | 1,000,000 | 0.00001 | 0.02 | 151849 | 151850 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | medium |
15,492 | Image-editing/Emu3-Base-SFT-reasoning_verbose-May02_post_stochastic_lr1e-4-checkpoint-5600 | Emu3 | ["Emu3ForCausalLM"] | 184622 | bfloat16 | 4.51.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0.1 | False | 4,300 | 1,000,000 | 0.00001 | 0.02 | 151849 | 151850 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | medium |
15,493 | ImageIN/convnext-base-224_finetuned_on_ImageIn_annotations | convnext | ["ConvNextForImageClassification"] | null | float32 | 4.22.1 | null | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,494 | ImageIN/convnext-tiny-224_finetuned | convnext | ["ConvNextForImageClassification"] | null | float32 | 4.22.1 | null | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,495 | ImageIN/mobilevit-small_finetuned_on_unlabelled_IA_with_snorkel_labels | mobilevit | ["MobileViTForImageClassification"] | null | float32 | 4.28.1 | null | null | null | 4 | null | silu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,496 | ImageIN/resnet-50_finetuned | resnet | ["ResNetForImageClassification"] | null | float32 | 4.22.1 | null | null | null | null | null | relu | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,497 | ImagineIt/22b-Converted | mistral | ["MistralForCausalLM"] | 32770 | bfloat16 | 4.41.2 | 6,144 | 16,384 | 56 | 48 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 32768 | 32769 | false | 128 | 6 | false | null | null | false | false | false | false | false | false | true | false | true | true | 25.367151 | large | very_long |
15,498 | ImagineIt/Alpha-basic | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 20,000 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,499 | ImagineIt/alpha-r-s-v-12 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 40,000 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.