index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
6,500 | CMU-AIR2/code-ArithHardC12Interm-240416 | llama | ["LlamaForCausalLM"] | 32256 | float32 | 4.38.2 | 2,048 | 5,504 | 24 | 16 | 16 | silu | 0 | True | 16,384 | 100,000 | 0.000001 | 0.02 | 32013 | 32021 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
6,501 | CMU-AIR2/code-ArithHardC12Interm-gh-240416 | llama | ["LlamaForCausalLM"] | 32256 | float32 | 4.38.2 | 2,048 | 5,504 | 24 | 16 | 16 | silu | 0 | True | 16,384 | 100,000 | 0.000001 | 0.02 | 32013 | 32021 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
6,502 | CMU-AIR2/code-ArithHardC12Interm-gh-mixture-240416 | llama | ["LlamaForCausalLM"] | 32256 | float32 | 4.38.2 | 2,048 | 5,504 | 24 | 16 | 16 | silu | 0 | True | 16,384 | 100,000 | 0.000001 | 0.02 | 32013 | 32021 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
6,503 | CMU-AIR2/code-ctrl-gh-mixture-240419 | llama | ["LlamaForCausalLM"] | 32256 | float32 | 4.38.2 | 2,048 | 5,504 | 24 | 16 | 16 | silu | 0 | True | 16,384 | 100,000 | 0.000001 | 0.02 | 32013 | 32021 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
6,504 | CMU-AIR2/code-full-hard-gh-mixture | llama | ["LlamaForCausalLM"] | 32256 | float32 | 4.38.2 | 2,048 | 5,504 | 24 | 16 | 16 | silu | 0 | True | 16,384 | 100,000 | 0.000001 | 0.02 | 32013 | 32021 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
6,505 | CMU-AIR2/code-full-simple | llama | ["LlamaForCausalLM"] | 32256 | float32 | 4.38.2 | 2,048 | 5,504 | 24 | 16 | 16 | silu | 0 | True | 16,384 | 100,000 | 0.000001 | 0.02 | 32013 | 32021 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
6,506 | CMU-AIR2/math-deepseek-FULL-ArithHard-MixedMWP | llama | ["LlamaForCausalLM"] | 32256 | bfloat16 | 4.38.2 | 2,048 | 5,504 | 24 | 16 | 16 | silu | 0 | True | 16,384 | 100,000 | 0.000001 | 0.02 | 32013 | 32021 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
6,507 | CMU-AIR2/math-deepseek-LORA-ArithHard | llama | ["LlamaForCausalLM"] | 32256 | bfloat16 | 4.38.2 | 2,048 | 5,504 | 24 | 16 | 16 | silu | 0 | True | 16,384 | 100,000 | 0.000001 | 0.02 | 32013 | 32021 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
6,508 | CMU-AIR2/math-deepseek-LORA-ArithHard-MixedMWP | llama | ["LlamaForCausalLM"] | 32256 | bfloat16 | 4.38.2 | 2,048 | 5,504 | 24 | 16 | 16 | silu | 0 | True | 16,384 | 100,000 | 0.000001 | 0.02 | 32013 | 32021 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
6,509 | CMU-AIR2/math-deepseek-LORA-ArithHardC12 | llama | ["LlamaForCausalLM"] | 32256 | bfloat16 | 4.38.2 | 2,048 | 5,504 | 24 | 16 | 16 | silu | 0 | True | 16,384 | 100,000 | 0.000001 | 0.02 | 32013 | 32021 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
6,510 | CMU-AIR2/math-deepseek-baseline-FTMWP-LORA | llama | ["LlamaForCausalLM"] | 32256 | bfloat16 | 4.38.2 | 2,048 | 5,504 | 24 | 16 | 16 | silu | 0 | True | 16,384 | 100,000 | 0.000001 | 0.02 | 32013 | 32021 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
6,511 | CMU-AIR2/math-deepseek-lora-arith-curriculum-per-subject | llama | ["LlamaForCausalLM"] | 32256 | bfloat16 | 4.33.1 | 2,048 | 5,504 | 24 | 16 | 16 | silu | null | True | 16,384 | 100,000 | 0.000001 | 0.02 | 32013 | 32021 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
6,512 | CMU-AIR2/math-deepseek_FULL_HardArith_Interm | llama | ["LlamaForCausalLM"] | 32256 | bfloat16 | 4.38.2 | 2,048 | 5,504 | 24 | 16 | 16 | silu | 0 | True | 16,384 | 100,000 | 0.000001 | 0.02 | 32013 | 32021 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
6,513 | CMU-AIR2/math-llama-3-instruct-LORA-ArithSteps-6K | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,514 | CMU-AIR2/math-llama_3_instruct-model-arith-4k | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,515 | CMU-AIR2/math-phi-1-5-FULL-ArithHard-lr-1.2e-6 | phi | ["PhiForCausalLM"] | 51200 | bfloat16 | 4.38.2 | 2,048 | 8,192 | 24 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | null | null | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
6,516 | CMU-AIR2/math-phi-1-5-FULL-ArithHard-lr-1.5e-6 | phi | ["PhiForCausalLM"] | 51200 | bfloat16 | 4.38.2 | 2,048 | 8,192 | 24 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | null | null | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
6,517 | CMU-AIR2/math-phi-1-5-FULL-ArithHard-lr-1e-6 | phi | ["PhiForCausalLM"] | 51200 | bfloat16 | 4.38.2 | 2,048 | 8,192 | 24 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | null | null | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
6,518 | CMU-AIR2/math-phi-1-5-FULL-ArithHard-lr-8e-7 | phi | ["PhiForCausalLM"] | 51200 | bfloat16 | 4.38.2 | 2,048 | 8,192 | 24 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | null | null | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
6,519 | CMU-AIR2/math-phi-1-5-FULL-Arithmetic-Steps-10k | phi | ["PhiForCausalLM"] | 51200 | bfloat16 | 4.38.2 | 2,048 | 8,192 | 24 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | null | null | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
6,520 | CMU-AIR2/math-phi-1-5-FULL-Arithmetic-Steps-2k | phi | ["PhiForCausalLM"] | 51200 | bfloat16 | 4.38.2 | 2,048 | 8,192 | 24 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | null | null | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
6,521 | CMU-AIR2/math-phi-1-5-FULL-Arithmetic-Steps-4k | phi | ["PhiForCausalLM"] | 51200 | bfloat16 | 4.38.2 | 2,048 | 8,192 | 24 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | null | null | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
6,522 | CMU-AIR2/math-phi-1-5-FULL-Arithmetic-Steps-512-4k | phi | ["PhiForCausalLM"] | 51200 | bfloat16 | 4.38.2 | 2,048 | 8,192 | 24 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | null | null | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
6,523 | CMU-AIR2/math-phi-1-5-FULL-Arithmetic-Steps-512-6k | phi | ["PhiForCausalLM"] | 51200 | bfloat16 | 4.38.2 | 2,048 | 8,192 | 24 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | null | null | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
6,524 | CMU-AIR2/math-phi-1-5-FULL-Arithmetic-Steps-512-8k | phi | ["PhiForCausalLM"] | 51200 | bfloat16 | 4.38.2 | 2,048 | 8,192 | 24 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | null | null | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
6,525 | CMU-AIR2/math-phi-1-5-FULL-Arithmetic-Steps-6k | phi | ["PhiForCausalLM"] | 51200 | bfloat16 | 4.38.2 | 2,048 | 8,192 | 24 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | null | null | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
6,526 | CMU-AIR2/math-phi-1-5-FULL-Arithmetic-Steps-8k | phi | ["PhiForCausalLM"] | 51200 | bfloat16 | 4.38.2 | 2,048 | 8,192 | 24 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | null | null | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
6,527 | CMunch/fine_tuned_dota | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.30.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,528 | CNBOOMBOOM/llama2-hivetalk-merger-10k-v2 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.35.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,529 | CNBOOMBOOM/quant-gptq-hivetalk-4bits | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.40.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,530 | CNR-ILC/gs-GreBerta | roberta | ["RobertaForMaskedLM"] | 52000 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,531 | CO21351/sd-class-animals-30-new | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,532 | CO21351/sd-class-butterflies-32-new | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,533 | COCO0414/bge-m3-ko_DNF | xlm-roberta | ["XLMRobertaModel"] | 250002 | float32 | 4.39.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 8,194 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | long |
6,534 | CONCISE/LLaMa_V2-13B-Instruct-Uncensored-HF | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,535 | CPSC532/2024NOV14_llama_3_1_8b_r_32 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,536 | CPSC532/2024NOV14_llama_3_1_8b_r_64 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,537 | CPSC532/2024NOV16_llama_3_1_8b_no_sources_in_question_r_128 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,538 | CPSC532/2024NOV16_llama_3_1_8b_no_sources_in_question_r_16 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,539 | CPSC532/eval_no_sources_finetune | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,540 | CPSC532/finetuned_model_2024OCT24 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,541 | CPSC532/finetuned_model_2024OCT25 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,542 | CR011v2/streetvision | vit | ["ViTForImageClassification"] | null | float32 | 4.49.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
6,543 | CRZ7436/fine_tuned_gpt2_model | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.39.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,544 | CSHaitao/DELTA_EN | bert | ["BertForSAILER"] | 30522 | float32 | 4.29.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,545 | CSHaitao/DELTA_EN_Finetune | bert | ["BertModel"] | 30522 | float32 | 4.29.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,546 | CSLin3303/qwen3-20250528 | qwen3 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,547 | CSLin3303/qwen3-laws-20250613002 | qwen3 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,548 | CSerdar014191/opt-350m-sixbooks | opt | ["OPTForCausalLM"] | 50272 | float32 | 4.28.1 | 1,024 | null | 24 | 16 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | medium |
6,549 | CSerdar014191/opt-350m-sixhorrorbooks | opt | ["OPTForCausalLM"] | 50272 | float32 | 4.28.1 | 1,024 | null | 24 | 16 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | medium |
6,550 | CSerdar014191/opt-350m_test08_tuner | opt | ["OPTForCausalLM"] | 50272 | float16 | 4.28.1 | 1,024 | null | 24 | 16 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | medium |
6,551 | CShorten/ArXiv-Cross-Encoder-Title-Abstracts | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.20.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,552 | CShorten/CORD-19-Title-Abstracts-1-more-epoch | bert | ["BertModel"] | 30522 | float32 | 4.22.1 | 384 | 1,536 | 6 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.010617 | small | short |
6,553 | CTCF/pretrained_hooked_gpt2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 0.05 | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,554 | CTPC/distilhubert-finetuned-gtzan | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.34.0 | 768 | 3,072 | 2 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.014156 | small | null |
6,555 | CTW-g123/MiniCPM | minicpmv | ["MiniCPMV"] | 151666 | bfloat16 | 4.40.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
6,556 | CUTD/qnAr | electra | ["ElectraForQuestionAnswering"] | 64000 | float32 | 4.30.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,557 | CVR123/Muril-base-finetune-Telugu-qc | bert | ["BertForSequenceClassification"] | 197285 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,558 | CWKSC/w2v-bert-2.0-mongolian-colab-CV16.0 | wav2vec2-bert | ["Wav2Vec2BertForCTC"] | 37 | float32 | 4.44.2 | 1,024 | 4,096 | 24 | 16 | null | swish | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | null |
6,559 | CWKSC/whisper-small-onnx | whisper | ["WhisperForConditionalGeneration"] | 51865 | null | 4.34.0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
6,560 | CWKSC/whisper-small-onnx-quantized | whisper | ["WhisperForConditionalGeneration"] | 51865 | null | 4.34.0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
6,561 | CXDuncan/whisper-large-v3-malayalam | whisper | ["WhisperForConditionalGeneration"] | 51866 | float32 | 4.41.0 | null | null | 32 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
6,562 | CXL295/zephyr-7b-dpo-full | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.36.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,563 | CY-Wu/med_and_tcp_learn | gpt2 | ["GPT2LMHeadModel"] | 21128 | float32 | 4.46.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 101 | 102 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,564 | CYFRAGOVPL/Llama-PLLuM-70B-base | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.47.0.dev0 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
6,565 | CYFRAGOVPL/Llama-PLLuM-70B-chat | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.45.0 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
6,566 | CYFRAGOVPL/Llama-PLLuM-70B-instruct | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.47.0.dev0 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
6,567 | CYFRAGOVPL/Llama-PLLuM-8B-base | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.47.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
6,568 | CYFRAGOVPL/Llama-PLLuM-8B-chat | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.45.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
6,569 | CYFRAGOVPL/Llama-PLLuM-8B-instruct | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.47.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
6,570 | CYFRAGOVPL/PLLuM-12B-base | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.47.0.dev0 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | False | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
6,571 | CYFRAGOVPL/PLLuM-12B-chat | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.45.0 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | False | 131,072 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
6,572 | CYFRAGOVPL/PLLuM-12B-instruct | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.47.0.dev0 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | False | 131,072 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
6,573 | CYFRAGOVPL/PLLuM-12B-nc-base | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.45.0.dev0 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | False | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
6,574 | CYFRAGOVPL/PLLuM-12B-nc-chat | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.45.0 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | False | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
6,575 | CYFRAGOVPL/PLLuM-12B-nc-instruct | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.46.3 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | False | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
6,576 | CYFRAGOVPL/PLLuM-8x7B-base | mixtral | ["MixtralForCausalLM"] | 32000 | bfloat16 | 4.36.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | 8 | 2 | false | false | false | false | false | false | false | true | true | true | 6.442451 | medium | very_long |
6,577 | CYFRAGOVPL/PLLuM-8x7B-chat | mixtral | ["MixtralForCausalLM"] | 32003 | bfloat16 | 4.45.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | 8 | 2 | false | false | false | false | false | false | false | true | true | true | 6.442451 | medium | very_long |
6,578 | CYFRAGOVPL/PLLuM-8x7B-instruct | mixtral | ["MixtralForCausalLM"] | 32003 | bfloat16 | 4.46.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | 8 | 2 | false | false | false | false | false | false | false | true | true | true | 6.442451 | medium | very_long |
6,579 | CYFRAGOVPL/PLLuM-8x7B-nc-base | mixtral | ["MixtralForCausalLM"] | 32000 | bfloat16 | 4.36.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | 8 | 2 | false | false | false | false | false | false | false | true | true | true | 6.442451 | medium | very_long |
6,580 | CYFRAGOVPL/PLLuM-8x7B-nc-chat | mixtral | ["MixtralForCausalLM"] | 32003 | bfloat16 | 4.45.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | 8 | 2 | false | false | false | false | false | false | false | true | true | true | 6.442451 | medium | very_long |
6,581 | CYFRAGOVPL/PLLuM-8x7B-nc-instruct | mixtral | ["MixtralForCausalLM"] | 32003 | bfloat16 | 4.46.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | 8 | 2 | false | false | false | false | false | false | false | true | true | true | 6.442451 | medium | very_long |
6,582 | CYONG/v1 | bert | ["BertForSequenceClassification"] | 105879 | null | 4.31.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,583 | CYRT/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.28.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,584 | CYX2003/whisper-tiny-en | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.41.1 | null | null | 4 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
6,585 | CaasiHUANG/flames-scorer | internlm | ["InternLMForSequenceClassification"] | 103168 | float16 | 4.33.2 | 4,096 | 11,008 | 32 | 32 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 6.442451 | medium | medium |
6,586 | Cabooose/multilingual_english_SEQUENCE_classification_model | bert | ["BertForSequenceClassification"] | 119547 | float32 | 4.28.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,587 | Cabooose/multilingual_english_token_classification_model | bert | ["BertForTokenClassification"] | 119547 | float32 | 4.28.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,588 | CaceAI/Llama3-70B-Fusion-ChatQA-Chinese | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.2 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | long |
6,589 | CaceAI/Llama3-8B-Fusion-ChatQA-Chinese | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,590 | Caco355/medical-model | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,591 | Cadenza-Labs/dolphin-llama3-8B-sleeper-agent-distilled-lora | llama | ["LlamaForCausalLM"] | 128258 | float32 | 4.40.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128256 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,592 | Caephas/real-madrid-fact | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.47.1 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
6,593 | Cafet/w2v-bert-final-v2 | wav2vec2-bert | ["Wav2Vec2BertForCTC"] | 39 | float32 | 4.40.0 | 1,024 | 4,096 | 24 | 16 | null | swish | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | null |
6,594 | Cafet/w2v-bert-version-final | wav2vec2-bert | ["Wav2Vec2BertForCTC"] | 40 | float32 | 4.40.0 | 1,024 | 4,096 | 24 | 16 | null | swish | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | null |
6,595 | Cafet/wav2vec2-bert-cv16-mas-ex-cv16 | wav2vec2-bert | ["Wav2Vec2BertForCTC"] | 40 | float32 | 4.39.1 | 1,024 | 4,096 | 24 | 16 | null | swish | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | null |
6,596 | Cafet/wav2vec2-bert-mas-ex | wav2vec2-bert | ["Wav2Vec2BertForCTC"] | 40 | float32 | 4.38.2 | 1,024 | 4,096 | 24 | 16 | null | swish | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | null |
6,597 | Cafet/whisper-mongolian-version-0.1 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.39.3 | null | null | 32 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
6,598 | Cafet/whisper-small-mn-cv16 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.39.3 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
6,599 | Cafet/whisper-small-mongolian | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.40.1 | null | null | 12 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.