index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
6,100 | Brench/Qwen2.5-1.5B-R1-KD-V2 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.49.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
6,101 | Brench/Qwen2.5-R1-Zero-GRPO-1.5B-V1 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.48.3 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
6,102 | Brendan/tod-zero-bqag3oyb-32000 | gpt_bigcode | ["GPTBigCodeForCausalLM"] | 49152 | float32 | 4.39.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 0 | 0 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,103 | BrettC/gemma-social_reasoning-finetune | gemma | ["GemmaForCausalLM"] | 256000 | float16 | 4.38.0 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
6,104 | Brhnglc/a2c-AntBulletEnv-v0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,105 | Brhnglc/a2c-PandaReachDense-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,106 | Brhnglc/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,107 | Brhnglc/ppo-SnowballTarget2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,108 | Brhnglc/ppo-golden | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,109 | BriHug/CodeT5P-770m-CPT-FT | t5 | ["T5ForConditionalGeneration"] | 32100 | float32 | 4.39.3 | null | null | null | null | null | null | null | True | null | null | null | null | 1 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,110 | BrianDsouzaAI/autotrain-even_better-91480144518 | deberta | ["DebertaForSequenceClassification"] | 50265 | float32 | 4.29.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,111 | BrianGu/gpt2-sft-dutch | gpt2 | ["GPT2LMHeadModel"] | 50257 | bfloat16 | 4.46.2 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,112 | BrianGu/tinyllama-1b-sigmoid-sft-full | llama | ["LlamaForCausalLMsigmoid"] | 50432 | bfloat16 | 4.46.2 | 2,048 | 5,632 | 22 | 16 | 16 | silu | 0 | False | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.107296 | medium | medium |
6,113 | BrianGu/tinyllama-1b-softmax-sft-full | llama | ["LlamaForCausalLM"] | 50432 | bfloat16 | 4.46.2 | 2,048 | 5,632 | 22 | 16 | 16 | silu | 0 | False | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.107296 | medium | medium |
6,114 | BrianGu/zephyr-7b-sft-full | gpt_neox | ["GPTNeoXForCausalLM"] | 50304 | bfloat16 | 4.46.2 | 2,048 | 8,192 | 16 | 8 | null | gelu | 0 | False | 2,048 | 10,000 | null | 0.02 | 0 | 0 | false | 256 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.805306 | small | medium |
6,115 | BrianHsu/BERT_test_graident_accumulation | bert | ["BertForMultipleChoice"] | 21128 | float32 | 4.36.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,116 | BrianHsu/BERT_test_graident_accumulation_test2 | bert | ["BertForMultipleChoice"] | 21128 | float32 | 4.36.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,117 | BrianHsu/BERT_test_graident_accumulation_test3 | bert | ["BertForMultipleChoice"] | 21128 | float32 | 4.36.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,118 | BrianIA/vit-base-oxford-iiit-pets | vit | ["ViTForImageClassification"] | null | float32 | 4.53.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
6,119 | BrianLan/detr-resnet-50-finetuned-600-epochs-TW-Marine-5cls-dataset | detr | ["DetrForObjectDetection"] | null | float32 | 4.46.3 | null | null | 6 | null | null | null | 0 | null | 1,024 | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,120 | Brichardt/dialogpt-nba-finetuned | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.47.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,121 | BricksDisplay/phi-1_5 | phi | ["PhiForCausalLM"] | 51200 | null | 4.38.2 | 2,048 | 8,192 | 24 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | null | null | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
6,122 | BricksDisplay/phi-1_5-q4 | phi | ["PhiForCausalLM"] | 51200 | null | 4.38.2 | 2,048 | 8,192 | 24 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | null | null | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
6,123 | BricksDisplay/phi-2 | phi | ["PhiForCausalLM"] | 51200 | null | 4.38.2 | 2,560 | 10,240 | 32 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | 50256 | 50256 | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
6,124 | BricksDisplay/phi-2-qnn | phi | ["PhiForCausalLM"] | 51200 | null | 4.39.3 | 2,560 | 10,240 | 32 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | 50256 | 50256 | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
6,125 | BricksDisplay/stablelm-2-1_6b-q4 | stablelm | ["StableLmForCausalLM"] | 100352 | null | 4.38.2 | 2,048 | 5,632 | 24 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | null | 0.02 | 100257 | 100257 | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
6,126 | BridgeTower/bridgetower-base | null | [] | null | null | null | 768 | null | 6 | 12 | null | gelu | null | null | null | null | null | null | null | null | false | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.042467 | small | null |
6,127 | BridgeTower/bridgetower-base-itm-mlm | null | [] | null | null | null | 768 | null | 6 | 12 | null | gelu | null | null | null | null | null | null | null | null | false | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.042467 | small | null |
6,128 | BridgeTower/bridgetower-large-itm-mlm | null | [] | null | null | null | 1,024 | null | 6 | 16 | null | gelu | null | null | null | null | null | null | null | null | false | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.075497 | small | null |
6,129 | BrightCarvalho/upload-model | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,130 | BrightXiaoHan/game-content-safety | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,131 | BrilliantCoolHuge/whisper-tiny-hi | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.28.0.dev0 | null | null | 4 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
6,132 | Bring2It2On/gorani-9B-4bit | gemma2 | ["Gemma2ForCausalLM"] | 256000 | float16 | 4.48.3 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
6,133 | Brioch/Acolyte-22B-6.5bpw-exl2 | mistral | ["MistralForCausalLM"] | 32768 | bfloat16 | 4.44.2 | 6,144 | 16,384 | 56 | 48 | 8 | silu | 0 | True | 131,072 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 6 | false | null | null | false | false | false | false | false | false | true | false | true | true | 25.367151 | large | very_long |
6,134 | Britain/DialoGPT-small-DanyBotTwo | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.17.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,135 | Britain/DialoGPT-small-DanyBotTwoNew | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.17.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,136 | Britania/m2 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,137 | BritishWerewolf/IS-Net | u2net | ["U2NetModel"] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,138 | BritishWerewolf/IS-Net-Anime | u2net | ["U2NetModel"] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,139 | BritnyB/summarizer | bigbird_pegasus | ["BigBirdPegasusForConditionalGeneration"] | 96103 | float32 | 4.38.1 | null | null | 16 | null | null | null | 0 | True | 4,096 | null | null | null | 2 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | medium |
6,140 | Brizape/SETH_0.0001_250 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,141 | Brizape/SETH_2e-05_250 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,142 | Brizape/SETH_5e-05_250 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,143 | Brizape/Variome_0.0001_250 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,144 | Brizape/Variome_0.0005_250 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,145 | Brizape/Variome_2e-05_250 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,146 | Brizape/Variome_5e-05_250 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,147 | Brizape/Variome_5e-05_30_03 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,148 | Brizape/Yepes_0.0001_250 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,149 | Brizape/Yepes_2e-05_250 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,150 | Brizape/Yepes_2e-05_31_03 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,151 | Brizape/Yepes_5e-05_250 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,152 | Brizape/Yepes_5e-05_30_03 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,153 | Brizape/Yepes_5e-05_31_03 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,154 | Brizape/bert-finetuned-ner | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.26.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,155 | Brizape/tmVar_5e-05_30_03 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,156 | Brizape/tmvar | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.26.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,157 | Brizape/tmvar_0.0001_250 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,158 | Brizape/tmvar_2e-05_250 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,159 | Brizape/tmvar_5e-05_250 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,160 | BroAlanTaps/GPT2-large-128-20000steps | gpt2 | ["GPT2LMHeadModel"] | 50289 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,161 | BroAlanTaps/GPT2-large-128-22000steps | gpt2 | ["GPT2LMHeadModel"] | 50289 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,162 | BroAlanTaps/GPT2-large-128-24000steps | gpt2 | ["GPT2LMHeadModel"] | 50289 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,163 | BroAlanTaps/GPT2-large-128-26000steps | gpt2 | ["GPT2LMHeadModel"] | 50289 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,164 | BroAlanTaps/GPT2-large-4-24000steps | gpt2 | ["GPT2LMHeadModel"] | 50321 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,165 | BroAlanTaps/GPT2-large-4-28000steps | gpt2 | ["GPT2LMHeadModel"] | 50321 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,166 | BroAlanTaps/Llama3-instruct-4-24000steps | llama | ["LlamaForCausalLM"] | 128259 | bfloat16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,167 | BroAlanTaps/Llama3-instruct-4-26000steps | llama | ["LlamaForCausalLM"] | 128259 | bfloat16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,168 | BroAlanTaps/Llama3-instruct-4-28000steps | llama | ["LlamaForCausalLM"] | 128259 | bfloat16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,169 | BroAlanTaps/Llama3-instruct-4-32000steps | llama | ["LlamaForCausalLM"] | 128259 | bfloat16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,170 | BroAlanTaps/Llama3-instruct-4-36000steps | llama | ["LlamaForCausalLM"] | 128259 | bfloat16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,171 | BrokenSoul/GPT2-GPTQ-4bit | gpt2 | ["GPT2LMHeadModel"] | 50257 | float16 | 4.40.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,172 | BrokenSoul/llama-2-7b-miniguanaco | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,173 | Bronsn/luganda-english-translation-model | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.47.1 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
6,174 | Bronyza/llm-course-hw1 | null | [] | 1024 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,175 | BrownianNotion/TinyLlama_v1.1_2bit_int_3x_data_3_epochs | llama | ["LlamaForCausalLM"] | 32001 | bfloat16 | 4.37.0 | 2,048 | 5,632 | 22 | 32 | 4 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
6,176 | BrownianNotion/tinyllama_v1.1-int1-g128 | llama | ["LlamaForCausalLM"] | 32001 | bfloat16 | 4.37.0 | 2,048 | 5,632 | 22 | 32 | 4 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
6,177 | Brownie81/donut-base-sroie | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | 4.37.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,178 | Brownie81/layoutlm-funsd | layoutlm | ["LayoutLMForTokenClassification"] | 30522 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,179 | Brrrrt/bert-finetuned-squad-accelerate | bert | ["BertForQuestionAnswering"] | 28996 | float32 | 4.24.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,180 | BruceD-A/segformer-b0-scene-parse-150 | segformer | ["SegformerForSemanticSegmentation"] | null | float32 | 4.39.3 | null | null | null | [1, 2, 5, 8] | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,181 | BruceT02/DistilBert_Exp | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.41.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,182 | BruceWang5638/SecretJudgment | qwen2 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,183 | Bruh0009/T2 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,184 | Brumocas/poca-SoccerTwos | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,185 | Brumocas/rl_course_vizdoom_health_gathering_supreme | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,186 | BrunoBosshard/Pretrained_TinySolar-248m-4k | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.52.1 | 1,024 | 4,096 | 12 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 32 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.150995 | small | very_long |
6,187 | BrunoGR/Just_HEAR_Me | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.35.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,188 | BrunoSCM/Arc50-2000 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,189 | BrunoSCM/Arc50-3000 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,190 | Brusnen/model | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,191 | BrutusNotLatin/lunarlander | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,192 | Brvcket/artifact_classifier | swin | ["SwinForImageClassification"] | null | float32 | 4.42.3 | 768 | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,193 | Brvcket/sdxl-outpainting-decor-0.2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,194 | Brvcket/sdxl-outpainting-lights-on-off-0.5 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,195 | Brvcket/stable-outpainting-xl-lights-0.3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,196 | Brvcket/stable-outpainting-xl-lights-0.4 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,197 | Brvcket/stable-outpainting-xl-lights-0.5 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,198 | Bryan5/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,199 | BryanADA/Qwen2.5-3B-TW-cot-Ultra-GGUF | qwen2 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.