index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
15,500 | ImagineIt/cool-beans | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 15,000 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,501 | ImagineIt/merged-alpha | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.40.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 20,000 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,502 | Iman-1220/Farsi-QWEN2-5-1.5B-8bit-pruned | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.43.1 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
15,503 | ImanAndrea/bert-finetuned-paperQA | roberta | ["RobertaForQuestionAnswering"] | 50265 | float32 | 4.45.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,504 | ImanAndrea/bert2bert-en-fr-v2 | encoder-decoder | ["EncoderDecoderModel"] | null | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,505 | ImanAndrea/bert2bert-en-to-fr | encoder-decoder | ["EncoderDecoderModel"] | null | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,506 | Imanbehravan/finetune_BERT | bert | ["BertForQuestionAnswering"] | 30522 | float32 | 4.40.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,507 | Imanbehravan/question_answering_T_5_part_v0.1 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.46.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,508 | Imask/Large_3000samples | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.44.2 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
15,509 | Imelliot/Qwen2.5-1.5B-Open-R1-GRPO | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.49.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
15,510 | Imene/rare-puppers | vit | ["ViTForImageClassification"] | null | float32 | 4.21.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,511 | ImeneT/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,512 | Immanuelraja/layoutlmv3-finetuned-cord_100 | layoutlmv3 | ["LayoutLMv3ForTokenClassification"] | 50265 | float32 | 4.22.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,513 | ImperialIndians23/RobertaBaseUnprocessedAugmented | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,514 | ImperialIndians23/RobertaBaseUnprocessedAugmentedBoth | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,515 | ImperialIndians23/RobertaBaseUnprocessedAugmentedSynonym | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,516 | ImperialIndians23/RobertaBaseUnprocessedTest | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,517 | ImposterSyndromeIsReal/biogpt-healthcare-tuned | biogpt | ["BioGptForCausalLM"] | 42384 | float32 | 4.28.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 1,024 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | false | 0.30199 | small | short |
15,518 | Imran1/QWEN2.5-32B-Translation | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.2 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
15,519 | Imran1/Qwen2.5-72B-Instruct-FP8 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.45.2 | 8,192 | 29,568 | 80 | 64 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
15,520 | Imran1/bloom_p560m_3 | bloom | ["BloomForCausalLM"] | 250880 | float32 | 4.30.0.dev0 | 1,024 | null | null | null | null | null | 0 | True | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,521 | Imran1/sentimen_analysis_yelp | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.22.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,522 | ImranzamanML/German_Semantic_V3b_Optimised | bert | ["BertModel"] | 31102 | float32 | 4.48.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 8,192 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | long |
15,523 | Imxxn/AudioCourseU6-TextToSpeech | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.32.1 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,524 | Imxxn/RLCourseU5-Pyramids | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,525 | Imxxn/RLCourseU5-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,526 | Imxxn/RLCourseU6-AntBulleEnv-v0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,527 | Imxxn/RLCourseU6-PandaReachDense-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,528 | Imxxn/model | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.40.1 | 768 | 3,072 | 3 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.021234 | small | short |
15,529 | In2Training/FILM-7B | mistral | ["MistralForCausalLM"] | 32000 | float32 | 4.37.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
15,530 | In4Seek/Llama-3.1-8B-Instruct_MedMCQA.04.00_backup | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.47.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,531 | In4Seek/Llama-3.1-8B-Instruct_MedMCQA.04.01_backup | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.47.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,532 | In4Seek/Llama-3.1-8B-Instruct_MedMCQA.05.01_backup | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.47.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,533 | InHawK/jd-llama2-32bit-7b-finetune | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.37.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,534 | InHawK/jd-llama2-7b-finetune | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.37.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,535 | Inabia-AI/ark_unbranded_claims_standalone_lora_3.1_10162024 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,536 | Inabia-AI/continued_pretraining_model_01 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,537 | InasMR/sentiment_model | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.35.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,538 | Inasity/L3.3-San-Mai-R1-70b-4.0bpw-h8-exl2 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
15,539 | Inderpreet01/Llama-3.1-8B-Instruct_rca_sft_tool_call_and_reasoning | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.51.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,540 | Inderpreet01/Llama-3.2-8B-Instruct_rca_grpo_v2 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.50.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,541 | IndexTeam/Index-1.9B-Constant-LR | index | ["IndexForCausalLM"] | 65029 | bfloat16 | 4.39.2 | 2,048 | 5,888 | 36 | 16 | 16 | silu | null | True | 4,096 | null | 0.000001 | 0.01 | 1 | 2 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.811939 | medium | medium |
15,542 | IndianChessMans/Therapy-Alpaca | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,543 | IndianFood/toolcallvariant2 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
15,544 | IndianServers/BhagavadGita | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.27.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,545 | IndicAIDias/finetuned_gemma | gemma | ["GemmaForCausalLM"] | 256000 | float16 | 4.38.0 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
15,546 | Indraa99/bart_semeval_env | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.30.2 | null | null | 6 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
15,547 | Indraa99/bert_semeval_env | bert | ["BertForMaskedLM"] | 30522 | float32 | 4.30.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,548 | Inesence/donut-LV-token-test | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,549 | Inespinoza/PPO-LunarLander-V0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,550 | Inespinoza/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,551 | InfAI/flan-t5-text2sparql-custom-tokenizer | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.18.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,552 | InfZnDipl/ML_LogReg | logistic_regression | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,553 | InferenceIllusionist/Excalibur-7b-DPO | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
15,554 | InferenceIllusionist/Magic-Dolphin-7b | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.37.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
15,555 | Inferless/deciLM-7B-GPTQ | deci_lm | ["DeciLMForCausalLM"] | 32000 | float16 | 4.36.2 | 4,096 | 14,336 | 32 | 32 | 32 | silu | null | True | 8,192 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 6.442451 | medium | long |
15,556 | Infermatic/Anubis-70B-v1-FP8-Dynamic | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.0 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
15,557 | Infermatic/L3-70B-Euryale-v2.1-fp8-dynamic | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.43.3 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | long |
15,558 | Infermatic/Llama-3.1-Nemotron-70B-Instruct-HF-FP8-Dynamic | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.45.2 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
15,559 | Infermatic/Llama-3.3-70B-Instruct-FP8-Dynamic | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.47.0 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
15,560 | Infermatic/Miquliz-120b-v2.0-FP8-dynamic | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.43.3 | 8,192 | 28,672 | 140 | 64 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 112.742892 | xlarge | very_long |
15,561 | Infi-MM/infimm-hd | infimm-hd | ["InfiMMHDModel"] | null | bfloat16 | 4.35.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
15,562 | InfiniAILab/OpenR1-Qwen-1.5B-SFT-Instruct | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.49.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
15,563 | InfiniAILab/S1-100-32B-bs8-new-tokenizer-parallel2 | qwen2 | ["Qwen2ForCausalLM"] | 151680 | float32 | 4.46.1 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
15,564 | InfiniAILab/S1-claude-1K-32B-bs16 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float32 | 4.46.1 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
15,565 | InfiniAILab/S1.1-100-32B-bs8-new-tokenizer-parallel-4.21-v0 | qwen2 | ["Qwen2ForCausalLM"] | 151680 | float32 | 4.46.1 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
15,566 | InfiniAILab/S1.1-1K-32B-bs16-new-tokenizer-mix-parallel-7.1-v3-prompted | qwen2 | ["Qwen2ForCausalLM"] | 151680 | float32 | 4.46.1 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
15,567 | InfiniAILab/S1.1-1K-32B-bs16-new-tokenizer-parallel-4.26-v0 | qwen2 | ["Qwen2ForCausalLM"] | 151680 | float32 | 4.46.1 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
15,568 | InfiniAILab/S1.1-1K-32B-bs16-new-tokenizer-parallel-4.26-v0-2 | qwen2 | ["Qwen2ForCausalLM"] | 151680 | float32 | 4.46.1 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
15,569 | InfiniAILab/S1.1-1K-32B-bs16-new-tokenizer-parallel-4.26-v0-3 | qwen2 | ["Qwen2ForCausalLM"] | 151680 | float32 | 4.46.1 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
15,570 | InfiniAILab/S1.1-1K-32B-bs16-new-tokenizer-parallel-5.1-v0 | qwen2 | ["Qwen2ForCausalLM"] | 151680 | float32 | 4.46.1 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
15,571 | InfiniAILab/S1.1-1K-32B-bs16-new-tokenizer-sequential-4.26-v0 | qwen2 | ["Qwen2ForCausalLM"] | 151680 | float32 | 4.46.1 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
15,572 | InfiniteEcho/poca-SoccerTwos | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,573 | InfiniteEcho/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
15,574 | InfinityC/test_sum_1_model | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.39.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,575 | InfinityC/test_sum_abs_t5_small_wasa_no_stops | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.39.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,576 | InfoDeskInc/InfoDeskGenericNer | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,577 | Infomaniak-AI/smolLM2-135M-Instruct-structure-output | llama | ["LlamaForCausalLM"] | 49152 | bfloat16 | 4.48.3 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 8,192 | 100,000 | 0.00001 | 0.041667 | 1 | 2 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | long |
15,578 | Ingrid0693/mini-bravo | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.16.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,579 | Inhaexpress/DialoGPT-medium-harrypotter | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.29.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,580 | Inishds/function_calling_qlora | phi3 | ["Phi3ForCausalLM"] | 32032 | float16 | 4.40.2 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | medium |
15,581 | Injardav/detr | detr | ["DetrForObjectDetection"] | null | float32 | 4.39.3 | null | null | 6 | null | null | null | 0 | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,582 | Injustice/distilbert-base-uncased-finetuned-squad | distilbert | ["DistilBertForQuestionAnswering"] | 30522 | float32 | 4.30.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,583 | InlaidOyster/RMIT_HACKATHON | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.46.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,584 | InnerI/A-I-0xtom-7B-slerp | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.35.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,585 | InnerI/InnerI-bittensor-7b | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.40.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,586 | InnerI/synCAI-144k-gpt2.5 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.40.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,587 | Innerby/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,588 | InnovationHacksAI/ofdv1 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,589 | Ino4Dev/Llama_Trained_Inolabs | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.39.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,590 | Inoob/Null-GPT2-Large | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.44.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,591 | Inria-CEDAR/FactSpotter-DeBERTaV3-Base | deberta-v2 | ["DebertaV2ForSequenceClassification"] | 128100 | float32 | 4.39.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,592 | Inria-CEDAR/FactSpotter-DeBERTaV3-Large | deberta-v2 | ["DebertaV2ForSequenceClassification"] | 128100 | float32 | 4.39.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,593 | Inria-CEDAR/FactSpotter-DeBERTaV3-Small | deberta-v2 | ["DebertaV2ForSequenceClassification"] | 128100 | float32 | 4.39.1 | 768 | 3,072 | 6 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
15,594 | InstaDeepAI/isoformer | isoformer | ["Isoformer"] | null | float32 | 4.29.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,595 | InstaDeepAI/nucleotide-transformer-v2-250m-multi-species | null | ["EsmForMaskedLM", "EsmForTokenClassification", "EsmForSequenceClassification"] | 4107 | float32 | 4.32.0.dev0 | 768 | 3,072 | 24 | 16 | null | null | null | False | 2,050 | null | null | 0.02 | null | null | false | 48 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.169869 | small | medium |
15,596 | InstaDeepAI/segment_nt | esm | ["SegmentNT"] | 4107 | float32 | 4.33.2 | 1,024 | 4,096 | 29 | 16 | null | null | null | False | 2,050 | null | null | 0.02 | null | null | false | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.364904 | small | medium |
15,597 | InstaDeepExternalProject/llm_training_20230901_132015 | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.30.2 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
15,598 | InstaDeepExternalProject/llm_training_20230901_132240 | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.30.2 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
15,599 | InstalilyAI/3-class-deberta | deberta-v2 | ["DebertaV2ForSequenceClassification"] | 128100 | float32 | 4.52.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.