index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
15,900 | JFernandoGRE/bert_mlm_ESAU_CHANCO_CASTILLON | bert | ["BertForMaskedLM"] | 31002 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,901 | JFernandoGRE/bert_mlm_HUACHUA_LUNA_CARMEN_SMITEHE | bert | ["BertForMaskedLM"] | 31002 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,902 | JFernandoGRE/bert_mlm_JENNIFER_LINDSAY_MANSILLA_TUDELA | bert | ["BertForMaskedLM"] | 31002 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,903 | JFernandoGRE/bert_mlm_JORGE_LUIS_SANCHEZ_RAMIREZ | bert | ["BertForMaskedLM"] | 31002 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,904 | JFernandoGRE/bert_mlm_JOSE_ALFREDO_CRUZADO_ECHEVARRIA | bert | ["BertForMaskedLM"] | 31002 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,905 | JFernandoGRE/bert_mlm_LLICO_HUAMAN_YAQUELIN_NOEMI | bert | ["BertForMaskedLM"] | 31002 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,906 | JFernandoGRE/bert_mlm_MARIA_DEL_PILAR_CHIGNE_MOZOMBITE | bert | ["BertForMaskedLM"] | 31002 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,907 | JFernandoGRE/bert_mlm_MONTES_MELENDEZ_PAUL_JONATHAN | bert | ["BertForMaskedLM"] | 31002 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,908 | JFernandoGRE/bert_mlm_PAUL_JONATHAN_MONTES_MELENDEZ | bert | ["BertForMaskedLM"] | 31002 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,909 | JFernandoGRE/bert_mlm_PIMENTEL_HUAYLLASCO_ZANALICIA | bert | ["BertForMaskedLM"] | 31002 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,910 | JFernandoGRE/bert_mlm_ROBLES_LAZARO_MONER_FELIX | bert | ["BertForMaskedLM"] | 31002 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,911 | JFernandoGRE/bert_mlm_SALAS_VALDIVIA_MARGARITA_ELIZABETH | bert | ["BertForMaskedLM"] | 31002 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,912 | JFernandoGRE/bert_mlm_SANCHEZ_RAMIREZ_JORGE_LUIS | bert | ["BertForMaskedLM"] | 31002 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,913 | JFernandoGRE/bert_mlm_TICONA_MIRANDA_MARCO_ANTONIO | bert | ["BertForMaskedLM"] | 31002 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,914 | JFernandoGRE/bert_mlm_YUCA_HUARACCALLO_FELIPE_FEDERICO | bert | ["BertForMaskedLM"] | 31002 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,915 | JFernandoGRE/classificator_gender_judicialcases | bert | ["BertForSequenceClassification"] | 31002 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,916 | JFernandoGRE/classificator_names | bert | ["BertForSequenceClassification"] | 31002 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,917 | JFernandoGRE/falcon7binstruct_augmenteddemocracy_dups_all4_gender | falcon | ["FalconForCausalLM"] | 65024 | float16 | 4.40.1 | 4,544 | null | 32 | 71 | null | null | 0 | True | 2,048 | 10,000 | null | 0.02 | 11 | 11 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | true | 7.928807 | large | medium |
15,918 | JFernandoGRE/gemma_7b_augmenteddemocracy_dups_all4_education | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.2 | 3,072 | 24,576 | 28 | 16 | 16 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
15,919 | JFernandoGRE/gtelarge-duplicates-judgenames | bert | ["BertModel"] | 30522 | float32 | 4.46.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,920 | JFernandoGRE/llama2_7b_augmenteddemocracy_dups | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,921 | JFernandoGRE/llama2_7b_augmenteddemocracy_dups_all4_15 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,922 | JFernandoGRE/llama2_7b_augmenteddemocracy_dups_all4_age | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,923 | JFernandoGRE/llama2_7b_augmenteddemocracy_dups_all4_gender | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,924 | JFernandoGRE/llama2_7b_augmenteddemocracy_dups_all4_zone | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,925 | JFernandoGRE/llama3_8b_brazil_augmenteddemocracy_dups_all4_education | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,926 | JFernandoGRE/mistral_7b_augmenteddemocracy_dups_all4_gender | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.40.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
15,927 | JFuellem/distilhubert-finetuned-gtzan | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.31.0 | 768 | 3,072 | 2 | 12 | null | gelu_new | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.014156 | small | null |
15,928 | JGKD/JangoGPTv1.0 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.31.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,929 | JGS04/model | bert | ["BertForSequenceClassification"] | 31002 | float32 | 4.32.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,930 | JGamonalHML/FondoEsperanzav3.0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,931 | JH-debug/wiki_messy_different_data | llama | ["LlamaForCausalLM"] | 46337 | float32 | 4.33.3 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | False | 4,096 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,932 | JH-debug/wiki_messy_same_data | llama | ["LlamaForCausalLM"] | 46337 | float32 | 4.33.3 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | False | 4,096 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,933 | JH-debug/wiki_original | llama | ["LlamaForCausalLM"] | 46337 | float32 | 4.33.3 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | False | 4,096 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,934 | JHONGJHENG/uuu_fine_tune_taipower | gpt2 | ["GPT2LMHeadModel"] | 21128 | float32 | 4.50.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 101 | 102 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,935 | JHhan/bert-finetuned-ner | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,936 | JHhan/bert-finetuned-ner-accelerate | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,937 | JHhan/distilbert-base-uncased-finetuned-imdb | distilbert | ["DistilBertForMaskedLM"] | 30522 | float32 | 4.35.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,938 | JHhan/distilbert-base-uncased-finetuned-imdb-accelerate | distilbert | ["DistilBertForMaskedLM"] | 30522 | float32 | 4.35.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,939 | JHhan/marian-finetuned-kde4-en-to-fr | marian | ["MarianMTModel"] | 59514 | float32 | 4.35.2 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
15,940 | JIMMY-ZM/intern_study_L0_4 | internlm2 | ["InternLM2ForCausalLM"] | 92544 | bfloat16 | 4.41.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | True | 262,144 | 50,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,941 | JINGAMY/finetuned-indiejoseph | bart | ["BartForConditionalGeneration"] | 51371 | float32 | 4.39.3 | null | null | 6 | null | null | null | 0.1 | True | 1,024 | null | null | null | 101 | 102 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
15,942 | JINGAMY/finetuned-raptorkwok | bart | ["BartForConditionalGeneration"] | 51271 | float32 | 4.39.3 | null | null | 6 | null | null | null | 0.1 | True | 1,024 | null | null | null | 101 | 102 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
15,943 | JINKYU0612/llama-2-7b-bnb-4bit-aiaustin-demo | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
15,944 | JJ-Tae/Pretraining_MFM_v1 | deberta | ["DebertaForMaskedLM"] | 50265 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,945 | JJ-Tae/Pretraining_Test_v1 | deberta | ["DebertaForMaskedLM"] | 50265 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,946 | JJ-Tae/Pretraining_Test_v2 | deberta | ["DebertaForMaskedLM"] | 50265 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,947 | JJ-Tae/results | deberta | ["DebertaForMaskedLM"] | 50265 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,948 | JJAnderson/EXAONE-Deep-7.8B-mlx-8Bit | exaone | ["ExaoneForCausalLM"] | 102400 | bfloat16 | 4.43.1 | 4,096 | 14,336 | null | 32 | 8 | null | 0 | True | 32,768 | 1,000,000 | null | 0.02 | 1 | 361 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | null | null | very_long |
15,949 | JJJCHHH/my_awesome_opus_books_model | t5 | ["T5ForConditionalGeneration"] | 32128 | null | 4.40.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,950 | JJRohan/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,951 | JJT007/results | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,952 | JJTsao/fine-tuned_movie_retriever-all-mpnet-base-v2 | mpnet | ["MPNetModel"] | 30527 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,953 | JJWariror/distilbert-base-uncased-finetuned-emotions | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.41.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,954 | JJaeHeon/500data | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
15,955 | JJhooww/Classificador_de_prompt | llama | ["LlamaForSequenceClassification"] | 32000 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | 12 | silu | 0 | False | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.084935 | small | medium |
15,956 | JJhooww/Mistral-7B-v0.2-Instruction | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
15,957 | JKSoon/sd-class-butterflies-32 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,958 | JKSoon/sd-class-pokemon | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,959 | JKilpatrick/bart-base-finetuned-CNN-DailyNews | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.41.2 | null | null | 6 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
15,960 | JKnowles/wuwt-flan-alpaca-large | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.28.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,961 | JKnowles/wuwt-flan-alpaca-large-5 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.28.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,962 | JKuang96/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,963 | JLAbe/fk-RoBERTa-2e-model | news_classifier | ["NewsClassifier"] | null | float32 | 4.35.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,964 | JLAng4210/Improved_Orient_Orca | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.38.2 | 3,200 | 8,640 | 26 | 32 | 32 | silu | 0 | True | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 100 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 3.19488 | medium | medium |
15,965 | JLK-ptbk/new-llama-7b-faq | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.42.3 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,966 | JLake310/ko-gpt-trinity-1.2B-ynat-gen | gpt2 | ["GPT2LMHeadModel"] | 51200 | null | 4.7.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 0 | 8 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,967 | JLauridsenOrg/StorytellingForKids | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0.dev0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
15,968 | JLee0/grand-pdf-chatbot-lora | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.52.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,969 | JLee0/staria-pdf-chatbot-lora | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.52.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,970 | JLinda/sd-class-butterflies-32 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,971 | JMatthewChiam/4248-spanBERT-large | bert | ["BertForQuestionAnswering"] | 28996 | float32 | 4.32.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,972 | JNewber/fluterj | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.35.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
15,973 | JNewber/my-str-lora | llama | ["LlamaForCausalLM"] | 32032 | float16 | 4.33.1 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
15,974 | JNewber/test | llama | ["LlamaForCausalLM"] | 32032 | float16 | 4.33.1 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
15,975 | JNolet/Qwen2.5-Coder-14B_v11.25.24.0_CodeInstruct | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.2 | 5,120 | 13,824 | 48 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 15.099494 | large | very_long |
15,976 | JOBBIME/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,977 | JOHNNY131/vit-model | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.28.1 | 768 | 3,072 | 6 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
15,978 | JOJO996/deepseek-r1-medical-model | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
15,979 | JOOMINJI/8_class_model.bin | bert | ["BertForSequenceClassification"] | 8002 | float32 | null | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,980 | JPDuran/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,981 | JREDFI3ASDI/gemma-2b-mt-German-to-English | gemma | ["GemmaForCausalLM"] | 256000 | float16 | 4.38.1 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
15,982 | JRHuy/whisper-vietnamese | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.31.0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
15,983 | JRog25/distilbert-base-uncased-finetuned-imdb | distilbert | ["DistilBertForMaskedLM"] | 30522 | float32 | 4.46.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,984 | JSWOOK/model | whisper | ["WhisperForConditionalGeneration"] | 51866 | float32 | 4.46.1 | null | null | 32 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
15,985 | JSky74/Mistral-Small-24B-Instruct | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.49.0.dev0 | 5,120 | 32,768 | 40 | 32 | 8 | silu | 0 | True | 32,768 | 100,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
15,986 | JSlin/ppo-Pyramids | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,987 | JSpergel/SailingLLM | mistral | ["MistralForCausalLM"] | 48384 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
15,988 | JSpergel/test_tiny_mixtral_only_router_2 | mixtral | ["MixtralForCausalLM"] | 32000 | bfloat16 | 4.40.0 | 512 | 14,336 | 8 | 16 | 4 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 32 | 4 | false | 4 | 2 | false | false | false | false | false | false | false | true | true | true | 0.025166 | small | very_long |
15,989 | JTH/results | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.21.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,990 | JTHANGEN/GPT_TEST | null | [] | 50257 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,991 | JTHANGEN/custom-gpt | gpt | ["GPTModel"] | 50257 | float32 | 4.48.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,992 | JTStephens/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,993 | JTh34/llama381binstruct_summarize_short_merged | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.49.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,994 | JUJORUME/whisper-small-es-l | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.36.0.dev0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
15,995 | JUNEYEOB/DAPT_batch512_lyric_con_sent | roberta | ["RobertaForMaskedLM"] | 32000 | float32 | 4.21.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,996 | JUNEYEOB/DAPT_batch64_con_lyric | roberta | ["RobertaForMaskedLM"] | 32000 | float32 | 4.21.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,997 | JUNEYEOB/FT_adafactor_lr6_lcs_lr6 | roberta | ["RobertaForSequenceClassification"] | 32000 | float32 | 4.21.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,998 | JUNEYEOB/FT_adafactorcls | roberta | ["RobertaForSequenceClassification"] | 32000 | float32 | 4.21.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,999 | JUNEYEOB/FT_batch16_lyric_con | roberta | ["RobertaForSequenceClassification"] | 32000 | float32 | 4.21.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.