index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
13,100
Gnider/aug_roberta_large
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.44.2
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
13,101
Gnider/nauka_2220_6ep
gpt2
["GPT2LMHeadModel"]
50264
float32
4.19.2
null
null
null
null
null
null
null
True
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
13,102
Gnider/rubert_3ep_chunk_2000
bert
["BertForSequenceClassification"]
120138
float32
4.41.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,103
Gnider/rugpt3medium_mir5k_4ep_lead
gpt2
["GPT2LMHeadModel"]
50257
float32
4.19.2
null
null
null
null
null
null
null
True
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
13,104
Gnider/rugpt3small_mir12k_4ep_lead
gpt2
["GPT2LMHeadModel"]
50264
float32
4.19.2
null
null
null
null
null
null
null
True
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
13,105
Gnider/sport_12ep_5900
gpt2
["GPT2LMHeadModel"]
50264
float32
4.42.0.dev0
null
null
null
null
null
null
null
True
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
13,106
Gnider/sport_6ep_15k
gpt2
["GPT2LMHeadModel"]
50264
float32
4.42.0.dev0
null
null
null
null
null
null
null
True
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
13,107
Gnider/xlm-roberta-base-classif-sentim-4ep
xlm-roberta
["XLMRobertaForQuestionAnswering"]
250002
float32
4.51.3
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,108
Gnjs/marian-finetuned-kde4-en-to-fr
marian
["MarianMTModel"]
59514
float32
4.30.2
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
13,109
GoGiants1/Misc-GlyControl
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,110
Goader/liberta-large
bert
["BertForMaskedLM"]
32000
float32
4.37.2
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
13,111
Goader/liberta-large-v2
bert
["BertForMaskedLM"]
64000
float32
4.37.2
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
13,112
GoalZero/aidetector-legacy-3
roberta
["RobertaForSequenceClassification"]
50265
float32
4.41.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,113
GoalZero/aidetector-legacy-6
roberta
["RobertaForSequenceClassification"]
50265
float32
4.41.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,114
Goddrew/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,115
Goddrew/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,116
Godefroy/ppo-LunarLander-v1
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,117
GodfreyOwino/AgriAdvisor
npk
[]
null
null
4.41.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,118
GodfreyOwino/NPK_prediction_model2
npk
["NPKPredictionModel"]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,119
Godouche/segformer-b0-finetuned-segments-sidewalk-2
segformer
["SegformerForSemanticSegmentation"]
null
float32
4.50.3
null
null
null
[1, 2, 5, 8]
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,120
GodsonNtungi/DAD_model_v3_70b_16bit
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.2
8,192
28,672
80
64
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
long
13,121
GodsonNtungi/DAD_model_v3_70b_4bit-merge
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.2
8,192
28,672
80
64
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
long
13,122
GodsonNtungi/model_v0.0.1
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
13,123
GodsonNtungi/new_swahili_lora_model_70b_4bit
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.2
8,192
28,672
80
64
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
long
13,124
GodsonNtungi/swahilillama3-8b
llama
["LlamaForCausalLM"]
128256
float16
4.40.0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,125
GodsonNtungi/swahilillama3-8b-gguf-real
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
13,126
Goekdeniz-Guelmez/J.O.S.I.E.3-Beta12-7B-slerp
mistral
["MistralForCausalLM"]
32002
bfloat16
4.40.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,127
Goekdeniz-Guelmez/J.O.S.I.E.v4o-8b-stage1-beta2.2
llama
["LlamaForCausalLM"]
128256
bfloat16
4.41.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,128
Goekdeniz-Guelmez/Josie-v6-2b-mlx-concept
helium
["HeliumForCausalLM"]
48000
bfloat16
4.45.0.dev0
2,560
7,040
24
20
20
silu
0
True
4,096
100,000
0
0.02
1
2
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.887437
medium
medium
13,129
Goekdeniz-Guelmez/Josie-v6-2b-mlx-concept-4bit
helium
["HeliumForCausalLM"]
48000
bfloat16
4.45.0.dev0
2,560
7,040
24
20
20
silu
0
True
4,096
100,000
0
0.02
1
2
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.887437
medium
medium
13,130
Goekdeniz-Guelmez/Josiefied-Qwen2-0.5B-Instruct-abliterated
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.44.2
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
13,131
Goekdeniz-Guelmez/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v1
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.44.2
1,536
8,960
28
12
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
13,132
Goekdeniz-Guelmez/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v2
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.45.1
1,536
8,960
28
12
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
13,133
Goekdeniz-Guelmez/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v3
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.45.1
1,536
8,960
28
12
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
13,134
Goekdeniz-Guelmez/Josiefied-Qwen2.5-7B-Instruct-abliterated-v2
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.44.2
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
13,135
Goekdeniz-Guelmez/Josiefied-Qwen3-30B-A3B-abliterated-v2
qwen3_moe
["Qwen3MoeForCausalLM"]
151936
bfloat16
4.52.4
2,048
6,144
48
32
4
silu
0
True
40,960
1,000,000
0.000001
0.02
151643
151645
false
64
8
false
null
8
false
false
false
false
false
false
false
false
true
true
2.415919
medium
very_long
13,136
Goekdeniz-Guelmez/Josiefied-Qwen3-4B-abliterated-v2
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.57.2
2,560
9,728
36
32
8
silu
0
True
262,144
5,000,000
0.000001
0.02
null
151645
true
80
4
false
null
null
false
false
false
false
false
false
false
false
true
true
2.831155
medium
very_long
13,137
Goekdeniz-Guelmez/j.o.s.i.e.v4o-7b-orpo-stage1-v0.5
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.44.2
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
13,138
Goekdeniz-Guelmez/josie-3b-v6.0-epoch1
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.1
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
13,139
GoidaAlignment/GOIDA-0.5B
qwen2
["Qwen2ForCausalLM"]
151665
bfloat16
4.46.2
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
13,140
Gokalp35/Turkish-IstiklalMarsi1
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
13,141
Gokul14/Gemma-2b-tamil
gemma
["GemmaForCausalLM"]
256000
float16
4.38.0
2,048
16,384
18
8
1
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
13,142
Gokulapriyan/swin-tiny-patch4-window7-224-finetuned-main-gpu-20e-final
swin
["SwinForImageClassification"]
null
float32
4.26.1
768
null
null
null
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,143
Gokulapriyan/swin-tiny-patch4-window7-224-finetuned-og_dataset_10e
swin
["SwinForImageClassification"]
null
float32
4.26.1
768
null
null
null
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,144
Gokulapriyan/swinv2-tiny-patch4-window8-256-finetuned-og_dataset_10e
swinv2
["Swinv2ForImageClassification"]
null
float32
4.26.1
768
null
null
null
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,145
Gokulapriyan/swinv2-tiny-patch4-window8-256-finetuned-og_dataset_5e
swinv2
["Swinv2ForImageClassification"]
null
float32
4.26.1
768
null
null
null
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,146
GoldenLlama/gemma-2-9b-it-Dglee_v3
gemma2
["Gemma2ForCausalLM"]
256000
float16
4.44.2
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
13,147
GoldenLlama/krx_qwen2.5_7b_it_v1_c
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.44.2
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
13,148
GoldenLlama/krx_qwen2.5_7b_it_v1_d
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.44.2
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
13,149
GoldenLlama/krx_qwen2.5_7b_it_v1_g
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.2
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
13,150
GoldenLlama/krx_qwen2.5_7b_it_v1_h
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.3
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
13,151
GoldenLlama/krx_qwen2.5_7b_it_v7
qwen2
["Qwen2ForCausalLM"]
152064
float16
4.44.2
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
13,152
GoldenLlama/krx_sg_qwen2.5_7b_it_v2
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.2
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
13,153
GoldenLlama/krx_sg_qwen2.5_7b_it_v8
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.3
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
13,154
Golinski/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-skittish_giant_toucan
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.52.4
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
null
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
13,155
Golobama/SoccerTwos-poca-v0
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,156
Golu2811/finetuned_lamini
gpt_neox
["GPTNeoXForCausalLM"]
50304
float32
4.31.0
512
2,048
6
8
null
gelu
0
True
2,048
null
null
0.02
0
0
false
64
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.018874
small
medium
13,157
Gonalb/legal-ft-v0
bert
["BertModel"]
30522
float32
4.48.2
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
13,158
Gonsoo/AWS-HF-optimum-neuron-0-0-28-llama-3-Korean-Bllossom-8B
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,159
Gonsoo/AWS-NeuronCC-2-14-llama-3-Korean-Bllossom-8B
llama
["LlamaForCausalLM"]
128256
bfloat16
4.41.1
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,160
Gonzalo33/modelo-entrenado-bps-gguf
granite
["GraniteForCausalLM"]
49160
bfloat16
4.48.3
4,096
12,800
40
32
8
silu
0.1
True
4,096
10,000
0.00001
0.02
49152
0
true
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
8.053064
large
medium
13,161
GonzaloValdenebro/MedicalQuestionAnswering
bert
["BertForQuestionAnswering"]
30522
float32
4.33.1
768
3,072
6
12
null
relu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.042467
small
short
13,162
GooKSL/BioLinkBERT-large-chemprot
bert
["BertModel"]
28895
float32
4.41.2
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
13,163
GooKSL/BioM-BERT-PubMed-PMC-Large-DDI
electra
["ElectraModel"]
30000
float32
4.41.2
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
short
13,164
GoodBaiBai88/M3D-LaMed-Phi-3-4B
lamed_phi3
["LamedPhi3ForCausalLM"]
32015
float32
4.42.0.dev0
3,072
8,192
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
medium
13,165
GoodGuyTim/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,166
Goodeat/controlnet-demosaicing
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,167
GoodiesHere/Apollo-LMMs-Apollo-1_5B-t32
apollo
["ApolloForCausalLM"]
151936
null
4.44.0
1,536
8,960
28
12
2
silu
0
True
32,768
1,000,000
0.000001
0.02
null
null
null
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
13,168
GoodiesHere/Apollo-LMMs-Apollo-3B-t32
apollo
["ApolloForCausalLM"]
151936
null
4.44.0
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
null
null
null
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
13,169
GoodiesHere/Apollo-LMMs-Apollo-7B-t32
apollo
["ApolloForCausalLM"]
152064
null
4.44.0
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
null
null
null
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
13,170
GoofyLM/BrainrotLM-Assistant-362M
llama
["LlamaForCausalLM"]
49152
float32
4.51.3
960
2,560
32
15
5
silu
0
True
8,192
100,000
0.00001
0.02
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.353894
small
long
13,171
Googoogooz/model_main_class_100k
bert
["BertForSequenceClassification"]
31102
float32
4.42.4
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
13,172
Googoogooz/model_polarity
bert
["BertForSequenceClassification"]
31102
float32
4.42.4
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
13,173
Googoogooz/model_polarity_100k
bert
["BertForSequenceClassification"]
31102
float32
4.42.4
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
13,174
Googoogooz/model_polarity_5k
bert
["BertForSequenceClassification"]
31102
float32
4.42.4
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
13,175
Googoogooz/model_sub_emo_100k
bert
["BertForSequenceClassification"]
31102
float32
4.42.4
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
13,176
Googoogooz/model_sub_emo_5k
bert
["BertForSequenceClassification"]
31102
float32
4.42.4
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
13,177
Goolissimo/Tigers_side_ViT
vit
["ViTForImageClassification"]
null
float32
4.35.0
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
13,178
Goorm-AI-04/GoogLeNet-Doppler
null
["GoogLeNet"]
null
float32
4.34.0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,179
Goorm-AI-04/MobileNetV2-Doppler
mobilenet_v2
["MobileNetV2ForImageClassification"]
null
float32
4.34.0
null
null
null
null
null
relu6
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
13,180
Goorm-AI-04/ResNet101-Doppler
resnet
["ResNetForImageClassification"]
null
float32
4.34.0
null
null
null
null
null
relu
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
13,181
Gopal1853/Gopal-finetuned-ru-to-en
marian
["MarianMTModel"]
62518
float32
4.38.2
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
13,182
Gopal1853/agam_1.0
t5
["T5ForConditionalGeneration"]
32128
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
13,183
Gopal2002/DONUT_FINETUNE_762
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.40.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,184
Gopal2002/NASFUND_MODEL
bert
["BertModel"]
30522
float32
4.40.0
384
1,536
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
32
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.021234
small
short
13,185
Gopal2002/SETFIT_NASFUND_ZEONAI
bert
["BertModel"]
30522
float32
4.40.0
384
1,536
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
32
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.021234
small
short
13,186
Gopal2002/SmolLM2-FT-DPO
llama
["LlamaForCausalLM"]
49152
float32
4.47.1
576
1,536
30
9
3
silu
0
False
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
13,187
Gopal2002/donut_fix_1
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.36.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,188
Gopung/mms-tts-kor-train
vits
["VitsModelForPreTraining"]
30
float32
4.47.1
192
null
6
2
null
relu
0.1
null
null
null
null
0.02
null
null
null
96
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.002654
small
null
13,189
Gopzz/Custom_Llama_audit_rowwise
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
13,190
GoraPakora/SHantiBABA
llama
["LlamaForCausalLM"]
128258
bfloat16
4.46.2
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
13,191
GoranS/whisper-base.hr
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.47.0
null
null
6
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
13,192
Goranmk20/test_deepseek
qwen2
["Qwen2ForCausalLM"]
152064
float16
4.53.0
3,584
18,944
28
28
4
silu
0
True
131,072
10,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
13,193
Gordon119/qa_test
bert
["BertForQuestionAnswering"]
30522
float32
4.39.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,194
GordonChang/bakeneko-instruct-finetuned-v1-merged
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.49.0
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.00001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
13,195
Gorenzelg/bert-finetuned-squad11
bert
["BertForQuestionAnswering"]
28996
null
4.24.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,196
Gorgoura/bert-finetuned-ner
bert
["BertForTokenClassification"]
28996
float32
4.31.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,197
GoshaLetov/T-Lite-sft
llama
["LlamaForCausalLM"]
128259
bfloat16
4.42.0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,198
GoshaLetov/T-Lite-sft-no-optimizer
llama
["LlamaForCausalLM"]
128259
bfloat16
4.42.0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,199
Goshective/kazparc_en_ru_marian_1
marian
["MarianMTModel"]
62518
float32
4.46.3
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short