index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
5,200 | BaoLocTown/sft-metamath-mistral-7b-vi-v1 | mistral | ["MistralForCausalLM"] | 32001 | float16 | 4.35.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5,201 | BaoPhanDuy/my_awesome_billsum_model | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.42.4 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,202 | BaoPhanDuy/my_awesome_opus_books_model | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.42.4 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,203 | BaorBaor/weight_DeBERTa | deberta-v2 | ["DebertaV2ForMultipleChoice"] | 128100 | float32 | 4.30.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
5,204 | Baptiste-HUVELLE-10/LeTriomphant2.1_ECE_iLab | qwen2 | ["Qwen2Model"] | 151646 | float32 | 4.48.1 | 8,192 | 29,568 | 86 | 64 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | false | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 69.256348 | large | very_long |
5,205 | BaptisteL/m3-best-model | openelm | ["OpenELMForCausalLM"] | 32000 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
5,206 | BarBarickoza/Dans-SakuraKaze-Picaro-12b | mistral | ["MistralForCausalLM"] | 131075 | bfloat16 | 4.49.0 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | False | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
5,207 | Barambio/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.27.4 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,208 | Barani1-t/wav2vec2-base-finetuned-gtzan | wav2vec2 | ["Wav2Vec2ForSequenceClassification"] | 32 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
5,209 | Barani1-t/whisper-small-finetuned-gtzan | whisper | ["WhisperForAudioClassification"] | 51865 | float32 | 4.34.1 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,210 | Barani1-t/whisper-tiny-finetuned-gtzan | whisper | ["WhisperForAudioClassification"] | 51865 | float32 | 4.34.1 | null | null | 4 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,211 | Barbzie/bert-base-uncased-finetuned-disaster-NLP | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,212 | Barbzie/distilbert-base-uncased-finetuned-disaster-NLP | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.45.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,213 | Bariatte/my_awesome_eli5_mlm_model | bert | ["BertForMaskedLM"] | 30522 | null | 4.35.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,214 | Barkavi/llama-7B-hf | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.30.2 | 4,096 | 11,008 | 32 | 32 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 6.442451 | medium | medium |
5,215 | Barkavi/t5largetotto | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.31.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,216 | Baron-GG/LLaVA-Med | llava_llama | ["LlavaLlamaForCausalLM"] | 32004 | float16 | 4.28.0.dev0 | 4,096 | 11,008 | 32 | 32 | null | silu | null | False | null | null | 0.000001 | 0.02 | 0 | 1 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 6.442451 | medium | null |
5,217 | Baron-qui/distilhubert-finetuned-gtzan | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.52.4 | 768 | 3,072 | 2 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.014156 | small | null |
5,218 | BaronSch/AURO_0 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,219 | BaronSch/AURO_1 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,220 | BaronSch/AURO_2 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,221 | BaronSch/AURO_3 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,222 | BaronSch/AURO_4 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,223 | BaronSch/Accu_0 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,224 | BaronSch/Accu_1 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,225 | BaronSch/Accu_2 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,226 | BaronSch/Accu_3 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,227 | BaronSch/Accu_4 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,228 | BaronSch/Atte_0 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,229 | BaronSch/Deep_0 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,230 | BaronSch/Deep_1 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,231 | BaronSch/Deep_2 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,232 | BaronSch/Deep_3 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,233 | BaronSch/Deep_4 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,234 | BaronSch/F1-s_0 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,235 | BaronSch/F1-s_1 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,236 | BaronSch/F1-s_2 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,237 | BaronSch/F1-s_3 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,238 | BaronSch/F1-s_4 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,239 | BaronSch/Mach_0 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,240 | BaronSch/Mach_1 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,241 | BaronSch/Mach_2 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,242 | BaronSch/Mach_3 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,243 | BaronSch/Mach_4 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,244 | BaronSch/Othe_0 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,245 | BaronSch/Othe_1 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,246 | BaronSch/Othe_2 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,247 | BaronSch/Othe_3 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,248 | BaronSch/Othe_4 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,249 | BaronSch/RMSE_0 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,250 | BaronSch/RMSE_1 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,251 | BaronSch/RMSE_2 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,252 | BaronSch/RMSE_3 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,253 | BaronSch/RMSE_4 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,254 | BaronSch/Spea_0 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,255 | BaronSch/Spea_1 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,256 | BaronSch/Spea_2 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,257 | BaronSch/Spea_3 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,258 | BaronSch/Spea_4 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,259 | BarraHome/Llama-3.2-3B-MLA | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.49.0.dev0 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
5,260 | BarraHome/jina-reranker-v1-turbo-en | bert | ["JinaBertForSequenceClassification"] | 61056 | float16 | 4.30.2 | 384 | 1,536 | 6 | 12 | null | gelu | null | True | 8,192 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.010617 | small | long |
5,261 | BarraHome/llama-3-newborn-16bits | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.39.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 16,384 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
5,262 | BarraHome/llama-3-newborn-4bits | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.39.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 16,384 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
5,263 | BarraHome/llama-3-orpo-v1 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
5,264 | BarraHome/llama-3-orpo-v1-merged-4bit | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,265 | BarraHome/llama-3-orpo-v1-merged_16bit | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,266 | BarraHome/llama3.2-1b-mla | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.49.0.dev0 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
5,267 | BarrelAcorn/ppo-BipedalWalker-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,268 | Barrick/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,269 | Barrrrry/DeepSeek-R1-W4AFP8 | deepseek_v3 | ["DeepseekV3ForCausalLM"] | 129280 | bfloat16 | 4.46.3 | 7,168 | 18,432 | 61 | 128 | 128 | silu | 0 | True | 163,840 | 10,000 | 0.000001 | 0.02 | 0 | 1 | false | 56 | 1 | false | 256 | 8 | false | false | false | false | false | true | false | true | false | true | 37.610324 | large | very_long |
5,270 | BarryFutureman/SmolPi0Policy | null | [] | null | null | null | null | null | null | null | null | null | null | True | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,271 | BarryFutureman/TinyLLaVA-1.1B-pretrained-projector | tiny_llava | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.33.3 | 2,048 | 5,632 | 22 | 32 | 4 | silu | null | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
5,272 | BarryFutureman/little-ivy-backup-model | ivy | ["Ivy4RL"] | null | float32 | 4.48.3 | 32 | 128 | 4 | 4 | 2 | silu | 0 | True | 1,024 | 10,000 | 0.000001 | 0.02 | null | null | false | 8 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.000049 | small | short |
5,273 | BarryJiang/KaraNousSuper4bitGPTQ | llama | ["LlamaForCausalLM"] | 32001 | float16 | 4.31.0.dev0 | 5,120 | 13,824 | 40 | 40 | null | silu | null | True | 8,192 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 12.582912 | large | long |
5,274 | BartekSadlej/calculator_model_test | encoder-decoder | ["EncoderDecoderModel"] | null | float32 | 4.38.1 | null | null | null | null | null | null | null | null | null | null | null | null | null | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,275 | Bartezky/distilhubert-finetuned-gtzan | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.36.1 | 768 | 3,072 | 2 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.014156 | small | null |
5,276 | Basdila/your_model_name | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.35.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
5,277 | Baselhany/Distilation_Whisper_base_CKP_256 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.51.3 | null | null | 6 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,278 | Baselhany/Graduation_Project_Whisper | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.51.3 | null | null | 6 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,279 | Baselhany/Graduation_Project_Whisper_base_segment | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.51.1 | null | null | 6 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,280 | Baselhany/Whisper_tiny_tring_large_sample_with_early_stop2 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.47.0 | null | null | 4 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,281 | Baselhany/fine_tune_Whisper_tiny2 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.47.0 | null | null | 4 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,282 | Basha738/llama2-13B-supervised-eos-ft-10-epochs-351 | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.37.2 | 5,120 | 13,824 | 40 | 40 | 40 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
5,283 | BashirRP/llm_judge_fiddler | roberta | ["RobertaForMaskedLM"] | 50265 | null | 4.35.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
5,284 | BashirRP/output | deberta | ["DebertaForSequenceClassification"] | 50265 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,285 | BasiliskLabs/Sovereign-0.1-72B | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.1 | 8,192 | 29,568 | 80 | 64 | 8 | silu | 0 | True | 131,072 | 1,000,000 | 0.00001 | 0.02 | 151643 | 151643 | false | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
5,286 | BasselKhaled/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,287 | BastR/TwinLlama-3.1-8B | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.51.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,288 | Bastian1111/poca-SoccerTwos | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,289 | BastianFuh/mt5-finetuned-wiki_lingua-en-de | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.45.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,290 | Basticooler/llama_3_ORPO | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.40.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
5,291 | Basticooler/llama_3_continued_pretraining | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.40.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
5,292 | Basticooler/project-m2-2024 | phi3 | ["Phi3ForCausalLM"] | 32011 | float32 | 4.40.0 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | medium |
5,293 | Bastig2023/beit-cardiac_segmentation-training | beit | ["BeitForSemanticSegmentation"] | 8192 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
5,294 | BasyaKuE/bert_sentiment | bert | ["BertForSequenceClassification"] | 32768 | float32 | 4.21.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
5,295 | Batazia/llama3-instruct-16bit | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
5,296 | Batazia/llama3-instruct-4bit | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
5,297 | BatirayErbayVodafone/bert_preprocessing_model | bert | ["BertForSequenceClassification"] | 32000 | float32 | 4.16.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,298 | BatirayErbayVodafone/testg | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.42.0.dev0 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
5,299 | Batman-11/huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.