index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
13,100 | Gnider/aug_roberta_large | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.44.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
13,101 | Gnider/nauka_2220_6ep | gpt2 | ["GPT2LMHeadModel"] | 50264 | float32 | 4.19.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
13,102 | Gnider/rubert_3ep_chunk_2000 | bert | ["BertForSequenceClassification"] | 120138 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,103 | Gnider/rugpt3medium_mir5k_4ep_lead | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.19.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
13,104 | Gnider/rugpt3small_mir12k_4ep_lead | gpt2 | ["GPT2LMHeadModel"] | 50264 | float32 | 4.19.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
13,105 | Gnider/sport_12ep_5900 | gpt2 | ["GPT2LMHeadModel"] | 50264 | float32 | 4.42.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
13,106 | Gnider/sport_6ep_15k | gpt2 | ["GPT2LMHeadModel"] | 50264 | float32 | 4.42.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
13,107 | Gnider/xlm-roberta-base-classif-sentim-4ep | xlm-roberta | ["XLMRobertaForQuestionAnswering"] | 250002 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,108 | Gnjs/marian-finetuned-kde4-en-to-fr | marian | ["MarianMTModel"] | 59514 | float32 | 4.30.2 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
13,109 | GoGiants1/Misc-GlyControl | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,110 | Goader/liberta-large | bert | ["BertForMaskedLM"] | 32000 | float32 | 4.37.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
13,111 | Goader/liberta-large-v2 | bert | ["BertForMaskedLM"] | 64000 | float32 | 4.37.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
13,112 | GoalZero/aidetector-legacy-3 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.41.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,113 | GoalZero/aidetector-legacy-6 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.41.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,114 | Goddrew/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,115 | Goddrew/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,116 | Godefroy/ppo-LunarLander-v1 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,117 | GodfreyOwino/AgriAdvisor | npk | [] | null | null | 4.41.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,118 | GodfreyOwino/NPK_prediction_model2 | npk | ["NPKPredictionModel"] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,119 | Godouche/segformer-b0-finetuned-segments-sidewalk-2 | segformer | ["SegformerForSemanticSegmentation"] | null | float32 | 4.50.3 | null | null | null | [1, 2, 5, 8] | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,120 | GodsonNtungi/DAD_model_v3_70b_16bit | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.2 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | long |
13,121 | GodsonNtungi/DAD_model_v3_70b_4bit-merge | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.2 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | long |
13,122 | GodsonNtungi/model_v0.0.1 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
13,123 | GodsonNtungi/new_swahili_lora_model_70b_4bit | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.2 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | long |
13,124 | GodsonNtungi/swahilillama3-8b | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.40.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
13,125 | GodsonNtungi/swahilillama3-8b-gguf-real | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
13,126 | Goekdeniz-Guelmez/J.O.S.I.E.3-Beta12-7B-slerp | mistral | ["MistralForCausalLM"] | 32002 | bfloat16 | 4.40.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
13,127 | Goekdeniz-Guelmez/J.O.S.I.E.v4o-8b-stage1-beta2.2 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
13,128 | Goekdeniz-Guelmez/Josie-v6-2b-mlx-concept | helium | ["HeliumForCausalLM"] | 48000 | bfloat16 | 4.45.0.dev0 | 2,560 | 7,040 | 24 | 20 | 20 | silu | 0 | True | 4,096 | 100,000 | 0 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.887437 | medium | medium |
13,129 | Goekdeniz-Guelmez/Josie-v6-2b-mlx-concept-4bit | helium | ["HeliumForCausalLM"] | 48000 | bfloat16 | 4.45.0.dev0 | 2,560 | 7,040 | 24 | 20 | 20 | silu | 0 | True | 4,096 | 100,000 | 0 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.887437 | medium | medium |
13,130 | Goekdeniz-Guelmez/Josiefied-Qwen2-0.5B-Instruct-abliterated | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.44.2 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
13,131 | Goekdeniz-Guelmez/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v1 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.44.2 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
13,132 | Goekdeniz-Guelmez/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v2 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.45.1 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
13,133 | Goekdeniz-Guelmez/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v3 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.45.1 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
13,134 | Goekdeniz-Guelmez/Josiefied-Qwen2.5-7B-Instruct-abliterated-v2 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.44.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
13,135 | Goekdeniz-Guelmez/Josiefied-Qwen3-30B-A3B-abliterated-v2 | qwen3_moe | ["Qwen3MoeForCausalLM"] | 151936 | bfloat16 | 4.52.4 | 2,048 | 6,144 | 48 | 32 | 4 | silu | 0 | True | 40,960 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 64 | 8 | false | null | 8 | false | false | false | false | false | false | false | false | true | true | 2.415919 | medium | very_long |
13,136 | Goekdeniz-Guelmez/Josiefied-Qwen3-4B-abliterated-v2 | qwen3 | ["Qwen3ForCausalLM"] | 151936 | bfloat16 | 4.57.2 | 2,560 | 9,728 | 36 | 32 | 8 | silu | 0 | True | 262,144 | 5,000,000 | 0.000001 | 0.02 | null | 151645 | true | 80 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 2.831155 | medium | very_long |
13,137 | Goekdeniz-Guelmez/j.o.s.i.e.v4o-7b-orpo-stage1-v0.5 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.44.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
13,138 | Goekdeniz-Guelmez/josie-3b-v6.0-epoch1 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.47.1 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
13,139 | GoidaAlignment/GOIDA-0.5B | qwen2 | ["Qwen2ForCausalLM"] | 151665 | bfloat16 | 4.46.2 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
13,140 | Gokalp35/Turkish-IstiklalMarsi1 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
13,141 | Gokul14/Gemma-2b-tamil | gemma | ["GemmaForCausalLM"] | 256000 | float16 | 4.38.0 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
13,142 | Gokulapriyan/swin-tiny-patch4-window7-224-finetuned-main-gpu-20e-final | swin | ["SwinForImageClassification"] | null | float32 | 4.26.1 | 768 | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,143 | Gokulapriyan/swin-tiny-patch4-window7-224-finetuned-og_dataset_10e | swin | ["SwinForImageClassification"] | null | float32 | 4.26.1 | 768 | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,144 | Gokulapriyan/swinv2-tiny-patch4-window8-256-finetuned-og_dataset_10e | swinv2 | ["Swinv2ForImageClassification"] | null | float32 | 4.26.1 | 768 | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,145 | Gokulapriyan/swinv2-tiny-patch4-window8-256-finetuned-og_dataset_5e | swinv2 | ["Swinv2ForImageClassification"] | null | float32 | 4.26.1 | 768 | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,146 | GoldenLlama/gemma-2-9b-it-Dglee_v3 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | float16 | 4.44.2 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
13,147 | GoldenLlama/krx_qwen2.5_7b_it_v1_c | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.44.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
13,148 | GoldenLlama/krx_qwen2.5_7b_it_v1_d | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.44.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
13,149 | GoldenLlama/krx_qwen2.5_7b_it_v1_g | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
13,150 | GoldenLlama/krx_qwen2.5_7b_it_v1_h | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
13,151 | GoldenLlama/krx_qwen2.5_7b_it_v7 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float16 | 4.44.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
13,152 | GoldenLlama/krx_sg_qwen2.5_7b_it_v2 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
13,153 | GoldenLlama/krx_sg_qwen2.5_7b_it_v8 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
13,154 | Golinski/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-skittish_giant_toucan | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.52.4 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
13,155 | Golobama/SoccerTwos-poca-v0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,156 | Golu2811/finetuned_lamini | gpt_neox | ["GPTNeoXForCausalLM"] | 50304 | float32 | 4.31.0 | 512 | 2,048 | 6 | 8 | null | gelu | 0 | True | 2,048 | null | null | 0.02 | 0 | 0 | false | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.018874 | small | medium |
13,157 | Gonalb/legal-ft-v0 | bert | ["BertModel"] | 30522 | float32 | 4.48.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
13,158 | Gonsoo/AWS-HF-optimum-neuron-0-0-28-llama-3-Korean-Bllossom-8B | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
13,159 | Gonsoo/AWS-NeuronCC-2-14-llama-3-Korean-Bllossom-8B | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.41.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
13,160 | Gonzalo33/modelo-entrenado-bps-gguf | granite | ["GraniteForCausalLM"] | 49160 | bfloat16 | 4.48.3 | 4,096 | 12,800 | 40 | 32 | 8 | silu | 0.1 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 49152 | 0 | true | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 8.053064 | large | medium |
13,161 | GonzaloValdenebro/MedicalQuestionAnswering | bert | ["BertForQuestionAnswering"] | 30522 | float32 | 4.33.1 | 768 | 3,072 | 6 | 12 | null | relu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
13,162 | GooKSL/BioLinkBERT-large-chemprot | bert | ["BertModel"] | 28895 | float32 | 4.41.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
13,163 | GooKSL/BioM-BERT-PubMed-PMC-Large-DDI | electra | ["ElectraModel"] | 30000 | float32 | 4.41.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | short |
13,164 | GoodBaiBai88/M3D-LaMed-Phi-3-4B | lamed_phi3 | ["LamedPhi3ForCausalLM"] | 32015 | float32 | 4.42.0.dev0 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | medium |
13,165 | GoodGuyTim/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,166 | Goodeat/controlnet-demosaicing | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,167 | GoodiesHere/Apollo-LMMs-Apollo-1_5B-t32 | apollo | ["ApolloForCausalLM"] | 151936 | null | 4.44.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | null | null | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
13,168 | GoodiesHere/Apollo-LMMs-Apollo-3B-t32 | apollo | ["ApolloForCausalLM"] | 151936 | null | 4.44.0 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | null | null | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
13,169 | GoodiesHere/Apollo-LMMs-Apollo-7B-t32 | apollo | ["ApolloForCausalLM"] | 152064 | null | 4.44.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | null | null | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
13,170 | GoofyLM/BrainrotLM-Assistant-362M | llama | ["LlamaForCausalLM"] | 49152 | float32 | 4.51.3 | 960 | 2,560 | 32 | 15 | 5 | silu | 0 | True | 8,192 | 100,000 | 0.00001 | 0.02 | 1 | 2 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.353894 | small | long |
13,171 | Googoogooz/model_main_class_100k | bert | ["BertForSequenceClassification"] | 31102 | float32 | 4.42.4 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
13,172 | Googoogooz/model_polarity | bert | ["BertForSequenceClassification"] | 31102 | float32 | 4.42.4 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
13,173 | Googoogooz/model_polarity_100k | bert | ["BertForSequenceClassification"] | 31102 | float32 | 4.42.4 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
13,174 | Googoogooz/model_polarity_5k | bert | ["BertForSequenceClassification"] | 31102 | float32 | 4.42.4 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
13,175 | Googoogooz/model_sub_emo_100k | bert | ["BertForSequenceClassification"] | 31102 | float32 | 4.42.4 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
13,176 | Googoogooz/model_sub_emo_5k | bert | ["BertForSequenceClassification"] | 31102 | float32 | 4.42.4 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
13,177 | Goolissimo/Tigers_side_ViT | vit | ["ViTForImageClassification"] | null | float32 | 4.35.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
13,178 | Goorm-AI-04/GoogLeNet-Doppler | null | ["GoogLeNet"] | null | float32 | 4.34.0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,179 | Goorm-AI-04/MobileNetV2-Doppler | mobilenet_v2 | ["MobileNetV2ForImageClassification"] | null | float32 | 4.34.0 | null | null | null | null | null | relu6 | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
13,180 | Goorm-AI-04/ResNet101-Doppler | resnet | ["ResNetForImageClassification"] | null | float32 | 4.34.0 | null | null | null | null | null | relu | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
13,181 | Gopal1853/Gopal-finetuned-ru-to-en | marian | ["MarianMTModel"] | 62518 | float32 | 4.38.2 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
13,182 | Gopal1853/agam_1.0 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.51.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
13,183 | Gopal2002/DONUT_FINETUNE_762 | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | 4.40.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,184 | Gopal2002/NASFUND_MODEL | bert | ["BertModel"] | 30522 | float32 | 4.40.0 | 384 | 1,536 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.021234 | small | short |
13,185 | Gopal2002/SETFIT_NASFUND_ZEONAI | bert | ["BertModel"] | 30522 | float32 | 4.40.0 | 384 | 1,536 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.021234 | small | short |
13,186 | Gopal2002/SmolLM2-FT-DPO | llama | ["LlamaForCausalLM"] | 49152 | float32 | 4.47.1 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | False | 8,192 | 100,000 | 0.00001 | 0.041667 | 1 | 2 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | long |
13,187 | Gopal2002/donut_fix_1 | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | 4.36.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,188 | Gopung/mms-tts-kor-train | vits | ["VitsModelForPreTraining"] | 30 | float32 | 4.47.1 | 192 | null | 6 | 2 | null | relu | 0.1 | null | null | null | null | 0.02 | null | null | null | 96 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.002654 | small | null |
13,189 | Gopzz/Custom_Llama_audit_rowwise | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
13,190 | GoraPakora/SHantiBABA | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.46.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
13,191 | GoranS/whisper-base.hr | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.47.0 | null | null | 6 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
13,192 | Goranmk20/test_deepseek | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float16 | 4.53.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 131,072 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
13,193 | Gordon119/qa_test | bert | ["BertForQuestionAnswering"] | 30522 | float32 | 4.39.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,194 | GordonChang/bakeneko-instruct-finetuned-v1-merged | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.49.0 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
13,195 | Gorenzelg/bert-finetuned-squad11 | bert | ["BertForQuestionAnswering"] | 28996 | null | 4.24.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,196 | Gorgoura/bert-finetuned-ner | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.31.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,197 | GoshaLetov/T-Lite-sft | llama | ["LlamaForCausalLM"] | 128259 | bfloat16 | 4.42.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
13,198 | GoshaLetov/T-Lite-sft-no-optimizer | llama | ["LlamaForCausalLM"] | 128259 | bfloat16 | 4.42.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
13,199 | Goshective/kazparc_en_ru_marian_1 | marian | ["MarianMTModel"] | 62518 | float32 | 4.46.3 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.