index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
2,400 | AlekHesa/testing-llama2-v3 | llama | ["LlamaForCausalLM"] | 32016 | float16 | 4.40.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | long |
2,401 | AlekHesa/testing-llama2-v7 | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.41.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,402 | AlekSt7/auralis_portal2_announcer_ru_xttsv2 | xtts | ["XttsGPT"] | null | null | 4.46.0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
2,403 | AlekSt7/auralis_portal2_announcer_ru_xttsv2_gpt | xtts_gpt | ["XttsGPT"] | 6681 | null | null | 1,024 | null | 30 | 16 | null | null | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | false | 0.377487 | small | null |
2,404 | Aleka12/distilbart-finetuned-mimic | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.52.4 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
2,405 | Aleks2002SH/llm-course-hw2-dpo | llama | ["LlamaForCausalLM"] | 49152 | bfloat16 | 4.50.0 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | medium |
2,406 | Aleksandar1932/gpt-neo-125M-country | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.15.0 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
2,407 | Aleksandar1932/gpt-neo-125M-hip-hop | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.15.0 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
2,408 | Aleksandar1932/gpt-neo-125M-metal | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.15.0 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
2,409 | Aleksandar1932/gpt-neo-125M-rock | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.15.0 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
2,410 | Aleksandar1932/gpt-neo-125M-spanish-classics | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.15.0 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
2,411 | Aleksandr505/phishing-text-classifier-rubert | bert | ["BertForSequenceClassification"] | 119547 | float32 | 4.51.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,412 | AleksandrKashirin/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,413 | AlekseiPravdin/KSIF-RPG-128k-slerp | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
2,414 | AlekseiPravdin/KukulStanta-Seamaiiza-slerp-merge5 | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.44.0 | 4,096 | 14,336 | 31 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.241124 | medium | very_long |
2,415 | AlekseiPravdin/Seamaiiza-7B-v1 | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.40.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | long |
2,416 | AlekseiPravdin/Seamaiiza-7B-v2-128k | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.40.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
2,417 | AlekseyElygin/Llama-3.2-11B-Vision-Instruct-bnb-4bit | mllama | ["MllamaForConditionalGeneration"] | null | bfloat16 | 4.45.0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | true | null | null | null |
2,418 | AlekseyElygin/llama-3-8b-bnb-4bit | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
2,419 | AlekseyElygin/mistral-7b-instruct-v0.3-4bit | mistral | ["MistralForCausalLM"] | 32768 | float16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
2,420 | AlekseyElygin/mistral-7b-instruct-v0.3-bnb-8ep-GGUF | mistral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | false | null | null | null |
2,421 | AlekseyKorshuk/125m-ri-reproduce-combined-4-gpu-20-val-v2 | opt | ["OPTForCausalLM"] | 50265 | float16 | 4.26.0.dev0 | 768 | null | 12 | 12 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | medium |
2,422 | AlekseyKorshuk/6.7b-ri-reproduce-combined-4-gpu-20-val-v3 | opt | ["OPTForCausalLM"] | 50265 | float16 | 4.26.0.dev0 | 4,096 | null | 32 | 32 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 6.442451 | medium | medium |
2,423 | AlekseyKorshuk/amazon-reviews-input-output | opt | ["OPTForCausalLM"] | 50265 | float32 | 4.25.0.dev0 | 1,024 | null | 24 | 16 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | medium |
2,424 | AlekseyKorshuk/amazon-reviews-input-output-1.3b | opt | ["OPTForCausalLM"] | 50265 | float16 | 4.25.0.dev0 | 2,048 | null | 24 | 32 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 1.20796 | medium | medium |
2,425 | AlekseyKorshuk/amazon-reviews-input-output-13b | opt | ["OPTForCausalLM"] | 50265 | float16 | 4.25.0.dev0 | 5,120 | null | 40 | 40 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 12.582912 | large | medium |
2,426 | AlekseyKorshuk/amazon-reviews-input-output-350m | opt | ["OPTForCausalLM"] | 50265 | float16 | 4.25.0.dev0 | 1,024 | null | 24 | 16 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | medium |
2,427 | AlekseyKorshuk/amazon-reviews-input-output-6.7b | opt | ["OPTForCausalLM"] | 50265 | float16 | 4.25.0.dev0 | 4,096 | null | 32 | 32 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 6.442451 | medium | medium |
2,428 | AlekseyKorshuk/amazon-reviews-input-output-6.7b-best | opt | ["OPTForCausalLM"] | 50265 | float16 | 4.25.0.dev0 | 4,096 | null | 32 | 32 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 6.442451 | medium | medium |
2,429 | AlekseyKorshuk/chai-pygmalion-v0-chatml | gptj | ["GPTJForCausalLM"] | 50400 | float16 | 4.28.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | false | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
2,430 | AlekseyKorshuk/chatml-test-small | gpt_neox | ["GPTNeoXForCausalLM"] | 50279 | float16 | 4.28.1 | 2,048 | 8,192 | 16 | 8 | null | gelu | null | True | 2,048 | null | null | 0.02 | 0 | 0 | false | 256 | null | false | null | null | false | false | true | false | false | false | false | false | false | false | 0.805306 | small | medium |
2,431 | AlekseyKorshuk/dalio-1.3b-test | opt | ["OPTForCausalLM"] | 50265 | float16 | 4.25.0.dev0 | 2,048 | null | 24 | 32 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 1.20796 | medium | medium |
2,432 | AlekseyKorshuk/evol-codealpaca-v1-sft-4e-5 | phi | ["PhiForCausalLM"] | 51200 | bfloat16 | 4.37.0 | 2,560 | 10,240 | 32 | 32 | 32 | gelu_new | 0 | False | 2,048 | 10,000 | null | 0.02 | null | null | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
2,433 | AlekseyKorshuk/evol-codealpaca-v1-sft-4e-5-dpo-3ep | phi | ["PhiForCausalLM"] | 51200 | float16 | 4.37.0 | 2,560 | 10,240 | 32 | 32 | 32 | gelu_new | 0 | False | 2,048 | 10,000 | null | 0.02 | null | null | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
2,434 | AlekseyKorshuk/llama-7b-chatml | llama | ["LlamaForCausalLM"] | 32002 | float16 | 4.28.1 | 4,096 | 11,008 | 32 | 32 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 6.442451 | medium | medium |
2,435 | AlekseyKorshuk/llama-7b-instructions-chatml | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.28.1 | 4,096 | 11,008 | 32 | 32 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 6.442451 | medium | medium |
2,436 | AlekseyKorshuk/llama-7b-v0-chatml | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.28.1 | 4,096 | 11,008 | 32 | 32 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 6.442451 | medium | medium |
2,437 | AlekseyKorshuk/pythia-1.4b-deduped-jokes | gpt_neox | ["GPTNeoXForCausalLM"] | 50304 | float16 | 4.29.0.dev0 | 2,048 | 8,192 | 24 | 16 | null | gelu | null | True | 2,048 | null | null | 0.02 | 0 | 0 | false | 128 | null | false | null | null | false | false | true | false | false | false | false | false | false | false | 1.20796 | medium | medium |
2,438 | AlekseyKorshuk/pythia-1b-deduped-chatml | gpt_neox | ["GPTNeoXForCausalLM"] | 50279 | float16 | 4.28.1 | 2,048 | 8,192 | 16 | 8 | null | gelu | null | True | 2,048 | null | null | 0.02 | 0 | 0 | false | 256 | null | false | null | null | false | false | true | false | false | false | false | false | false | false | 0.805306 | small | medium |
2,439 | AlekseyKorshuk/retriever-coding-guru-adapted | mpnet | ["MPNetModel"] | 30527 | float32 | 4.21.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,440 | AlekseyKorshuk/roberta-with-topic | roberta | ["RobertaForSequenceClassification"] | 50265 | float16 | 4.29.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
2,441 | AlekseyKorshuk/vic15-exp-syn-fight-cp1919 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,442 | AlekseyKorshuk/vic15-exp-syn-fight-cp3838 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,443 | AlekseyKorshuk/vic15-exp-syn-fight-cp5757 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,444 | AlekseyKorshuk/vic15-exp-syn-romantic-cp1310 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,445 | AlekseyKorshuk/vic15-exp-syn-romantic-cp2620 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,446 | AlekseyKorshuk/vicuna-7b-83k-dataset-new-combined | llama | ["LlamaForCausalLM"] | 32001 | float16 | 4.28.1 | 4,096 | 11,008 | 32 | 32 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 0 | 1 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 6.442451 | medium | medium |
2,447 | AlekseyKorshuk/vsinrom3 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
2,448 | AlekseyKorshuk/yt_videos_comments | gpt2 | ["GPT2LMHeadModel"] | 50257 | float16 | 4.29.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
2,449 | Aleksia/finetuning-distilBert_sentiment | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.35.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,450 | Alekyukk/ppo | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,451 | Alelcv27/llama3-8b-breadcrumbs-ties-v3 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.48.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
2,452 | Alelcv27/llama3-8b-breadcrumbs-ties-v5 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.48.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
2,453 | Alelcv27/llama3-8b-dare-linear-v3 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.48.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
2,454 | Alelcv27/llama3-8b-dare-ties-v2 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.48.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
2,455 | Alelcv27/llama3-8b-dare-ties-v3 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.48.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
2,456 | Alelcv27/llama3-8b-della-v2 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.48.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
2,457 | Alelcv27/llama3-8b-passthrough | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.47.0 | 4,096 | 14,336 | 64 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 12.884902 | large | very_long |
2,458 | Alelcv27/llama3-8b-slerp | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.47.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
2,459 | Alemat/abcnew | roberta | ["RobertaForMaskedLM"] | 50265 | null | null | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,460 | Alemat/n460000 | roberta | ["RobertaForMaskedLM"] | 50265 | null | null | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,461 | Alemat/nm462000 | roberta | ["RobertaForMaskedLM"] | 50265 | float32 | 4.12.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,462 | Alepach/notHumpback-M1-Rw-F-8b | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.46.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
2,463 | Aleph-Alpha/Aleph-Alpha-GermanWeb-Grammar-Classifier-BERT | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,464 | Aleph-Alpha/Aleph-Alpha-GermanWeb-Quality-Classifier-BERT | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,465 | Aleph-Alpha/llama-3_1-8b-tfree-hat-base | hierarchical_autoregressive_transformer | ["HATForCausalLM"] | null | bfloat16 | 4.46.3 | null | null | null | null | null | null | null | null | 262,144 | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | very_long |
2,466 | Aleph-Alpha/llama-3_1-8b-tfree-hat-sft | hierarchical_autoregressive_transformer | ["HATForCausalLM"] | null | bfloat16 | 4.46.3 | null | null | null | null | null | null | null | null | 262,144 | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | very_long |
2,467 | AlephNull/Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,468 | Aleron12/distilbert-base-uncased-finetuned-amz_brander | distilbert | ["DistilBertForQuestionAnswering"] | 30522 | float32 | 4.27.0.dev0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
2,469 | AlessandroMinervini/llama_factory_first_model | mistral | ["MistralForCausalLM"] | 32768 | bfloat16 | 4.45.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
2,470 | Alessio-Borgi/all-mpnet-base-v2-margin-based-triplet-loss-finetuned-culture-1-epochs-enhanced_test | mpnet | ["MPNetModel"] | 30527 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,471 | Alessio-Borgi/all-mpnet-base-v2-margin-based-triplet-loss-finetuned-culture-10-epochs-enhanced | mpnet | ["MPNetModel"] | 30527 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,472 | Alessio-Borgi/all-mpnet-base-v2-margin-based-triplet-loss-finetuned-culture-3-epochs-enhanced | mpnet | ["MPNetModel"] | 30527 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,473 | Alesteba/detr-resnet-50_finetuned_cppe5 | detr | ["DetrForObjectDetection"] | null | float32 | 4.7.0.dev0 | null | null | 6 | null | null | null | 0 | null | 1,024 | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
2,474 | Aleteian/Avalon | mistral | ["MistralForCausalLM"] | 131072 | float16 | 4.46.2 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
2,475 | Aleteian/Follow-the-Flow-MN-12B | mistral | ["MistralForCausalLM"] | 131077 | bfloat16 | 4.48.2 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
2,476 | Aleteian/Legend-of-the-Four-Winds-2-MN-12B | mistral | ["MistralForCausalLM"] | 131074 | bfloat16 | 4.48.2 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
2,477 | Aleteian/Legend-of-the-Four-Winds-MN-12B | mistral | ["MistralForCausalLM"] | 131074 | bfloat16 | 4.48.2 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
2,478 | Aleteian/One-To-Rule-Them-All-MN-12B | mistral | ["MistralForCausalLM"] | 131077 | bfloat16 | 4.48.2 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
2,479 | Aleteian/Saiga-Unleashed | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.48.1 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
2,480 | Aleteian/Way-to-Unseen-Horizon-MN-12B | mistral | ["MistralForCausalLM"] | 131074 | bfloat16 | 4.48.2 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
2,481 | Aletheia-ng/AfroLid | Afrolid | ["AfroLid"] | null | float32 | 4.46.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,482 | Alex-Deng/dummy-model | camembert | ["CamembertForMaskedLM"] | 32005 | float32 | 4.31.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 5 | 6 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,483 | Alex007ander/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-fierce_yawning_leopard | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.51.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
2,484 | Alex01837178373/3.1model | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.43.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
2,485 | Alex034/t5-base-finetuned-summary | t5 | ["T5ForConditionalGeneration"] | 32128 | null | 4.35.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
2,486 | Alex034/t5-small-indosum-summary-freeze | t5 | ["T5ForConditionalGeneration"] | 32128 | null | 4.35.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
2,487 | Alex2060/whisper-tiny-ru | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.45.2 | null | null | 4 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
2,488 | Alex31y/bert-base-multilingual-uncased-finetuned-squad | bert | ["BertForQuestionAnswering"] | 105879 | float32 | 4.28.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,489 | Alex48/poca-SoccerTwos-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,490 | Alex48/poca-SoccerTwos-v4 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,491 | Alex48/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,492 | Alex7871/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
2,493 | AlexAgafontsev/bert-finetuned-sem_eval-english | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
2,494 | AlexAnoshka/bge-m3 | xlm-roberta | ["XLMRobertaModel"] | 250002 | float32 | 4.33.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 8,194 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | long |
2,495 | AlexAnoshka/bge-reranker-large | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.48.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
2,496 | AlexB007/model | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
2,497 | AlexC98/CodeTransLargeTFFlt | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.28.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
2,498 | AlexC98/CodeTransLargeTFNrm | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.28.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
2,499 | AlexC98/T5GenFilteredV100True | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.28.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.