index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
7,400
ClarenceDan/c078aeb9-5ac4-4866-823b-0eea963ce7fb
llama
["LlamaForCausalLM"]
128256
float16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
7,401
ClarenceDan/c10aae00-6554-4622-9e0e-9c3b568b0731
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
7,402
ClarenceDan/c1eb6a18-c51a-424c-ab61-77401bfbf32a
gpt_neox
["GPTNeoXForCausalLM"]
50304
float16
4.46.0
512
2,048
6
8
null
gelu
0
False
2,048
10,000
null
0.02
0
0
false
64
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.018874
small
medium
7,403
ClarenceDan/c3276683-d224-482e-af33-e5829cf4e89b
llama
["LlamaForCausalLM"]
49153
bfloat16
4.46.0
576
1,536
30
9
3
silu
0
False
8,192
100,000
0.00001
0.041667
0
0
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
7,404
ClarenceDan/c770f82d-d37d-4eaa-bff1-681e5c94e5fd
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
5,120
13,824
48
40
8
silu
0
False
131,072
1,000,000
0.00001
0.02
null
151643
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
15.099494
large
very_long
7,405
ClarenceDan/c78e64c3-ed3e-4ad3-a536-c37280823d6d
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.46.0
2,304
9,216
26
8
4
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
288
2
false
null
null
false
false
false
false
false
false
false
false
true
true
1.656226
medium
long
7,406
ClarenceDan/c8137857-ebbf-4bbc-b47b-406c8199e6be
llama
["LlamaForCausalLM"]
32016
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
7,407
ClarenceDan/cafed019-ba01-4536-bffb-2deb5698ee79
starcoder2
["Starcoder2ForCausalLM"]
49152
null
4.46.0
3,072
12,288
30
24
2
gelu_pytorch_tanh
0.1
False
16,384
999,999.442036
null
0.018042
0
0
null
128
12
false
null
null
false
false
false
false
false
false
false
false
true
true
3.397386
medium
long
7,408
ClarenceDan/ccced60c-e4bc-4965-a32a-68807d716311
mistral
["MistralForCausalLM"]
32768
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
7,409
ClarenceDan/ccf52cf3-54a5-4b0d-93d6-909d0f3eb0b1
mistral
["MistralForCausalLM"]
131072
bfloat16
4.46.0
5,120
14,336
40
32
8
silu
0
False
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
7,410
ClarenceDan/cd11dd1c-ce9a-4497-927a-fa0961ad309b
llama
["LlamaForCausalLM"]
46336
float16
4.46.0
4,096
11,008
32
32
32
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
7,411
ClarenceDan/cd85d811-da37-43f6-9462-051eb91a1db5
mistral
["MistralForCausalLM"]
32000
float16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
7,412
ClarenceDan/cfbe0b12-63d3-43d5-be7a-4fee81b76121
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
7,413
ClarenceDan/d1734fce-7fb2-45a0-b40a-34e8858b4880
llama
["LlamaForCausalLM"]
49153
bfloat16
4.46.0
960
2,560
32
15
5
silu
0
False
8,192
100,000
0.00001
0.02
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.353894
small
long
7,414
ClarenceDan/d24cc545-9aaf-49c7-8fae-3e5aae994b77
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.46.0
3,072
8,192
32
32
32
silu
0
False
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
7,415
ClarenceDan/d27f8a1c-4bab-4e37-aa60-fb941b91d302
llama
["LlamaForCausalLM"]
37632
float16
4.46.0
5,120
13,824
40
40
40
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
7,416
ClarenceDan/d41b8897-ebca-475f-92db-dc2734250d51
llama
["LlamaForCausalLM"]
100263
float16
4.46.0
3,072
9,216
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
100257
100257
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
7,417
ClarenceDan/d5c42a45-2457-4663-b580-220b7f01015a
opt
["OPTForCausalLM"]
50272
float16
4.46.0
2,048
null
24
32
null
null
0
False
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
1.20796
medium
medium
7,418
ClarenceDan/d683603e-40b4-405f-9385-888fba3e3fca
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.46.0
2,304
9,216
26
8
4
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
288
2
false
null
null
false
false
false
false
false
false
false
false
true
true
1.656226
medium
long
7,419
ClarenceDan/d7ac7aab-d536-41ad-af7e-22858bbf1999
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
4,096
10,000
0.000001
0.02
null
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
medium
7,420
ClarenceDan/d801218a-e46f-4a67-b392-e0883aa5e017
dbrx
["DbrxForCausalLM"]
100352
float32
4.46.0
null
null
null
null
2
null
null
False
null
null
null
0.02
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
true
true
null
null
null
7,421
ClarenceDan/dccc746a-ae18-4c24-80b7-d65a63b797ac
gptj
["GPTJForCausalLM"]
50401
float32
4.46.0
null
null
null
null
null
null
null
False
null
null
null
0.02
50256
50256
false
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
7,422
ClarenceDan/de1aef22-7018-4043-be6d-9150d9361685
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
4,096
10,000
0.000001
0.02
null
151643
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
medium
7,423
ClarenceDan/e0475860-c497-4fb3-8a62-17f6031ba850
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
131,072
1,000,000
0.000001
0.02
null
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
7,424
ClarenceDan/e0be9d66-e3f4-4901-8130-c2e25a1ba053
llama
["LlamaForCausalLM"]
32000
float16
4.46.0
3,200
8,640
26
32
32
silu
0
False
2,048
10,000
0.000001
0.02
1
2
false
100
1
false
null
null
true
false
false
false
false
false
false
false
false
true
3.19488
medium
medium
7,425
ClarenceDan/e258734c-9e54-437b-ae23-175bde61cad6
llama
["LlamaForCausalLM"]
46336
float16
4.46.0
4,096
11,008
32
32
32
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
7,426
ClarenceDan/e6807633-03d1-4562-a34d-c413fe67f620
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
2,048
8,192
16
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
7,427
ClarenceDan/e69b0324-f2b8-4d90-b22f-fec5ed74c0c9
llama
["LlamaForCausalLM"]
32000
float16
4.46.0
4,096
14,336
48
32
8
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
9.663676
large
medium
7,428
ClarenceDan/e7562a07-1d66-464f-aab0-ed84dd64faa2
llama
["LlamaForCausalLM"]
64000
bfloat16
4.46.0
4,096
11,008
48
32
4
silu
0
False
16,384
5,000,000
0.000001
0.02
1
7
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
9.663676
large
long
7,429
ClarenceDan/e7809b05-3fb9-4c5f-b57a-2711b00ac367
opt
["OPTForCausalLM"]
50272
float16
4.46.0
768
null
12
12
null
null
0
False
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
medium
7,430
ClarenceDan/e7ffc795-965c-480c-bdf5-976004236936
mistral
["MistralForCausalLM"]
32000
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
7,431
ClarenceDan/e84c24ea-aea2-4dd5-90be-0c98921ca9cd
llama
["LlamaForCausalLM"]
64000
bfloat16
4.46.0
4,096
11,008
48
32
4
silu
0
False
16,384
5,000,000
0.000001
0.02
1
7
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
9.663676
large
long
7,432
ClarenceDan/e8fa3d56-449d-48c3-993d-5f0c56cbefc1
mixtral
["MixtralForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
2
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
long
7,433
ClarenceDan/ea9a1794-69e3-4a1c-a4a9-76d9719cb361
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128040
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
7,434
ClarenceDan/ed756799-e706-48b8-9e68-8560cfc9d508
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
2,048
5,632
22
32
4
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
7,435
ClarenceDan/f243721d-b22b-486b-932b-d09187271d12
llama
["MistralForCausalLM"]
79099
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
true
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
7,436
ClarenceDan/f2d9478d-72fc-4a02-80b9-6edc9a38fc93
llama
["LlamaForCausalLM"]
32000
float16
4.46.0
3,200
8,640
26
32
32
silu
0
False
2,048
10,000
0.000001
0.02
1
2
false
100
1
false
null
null
true
false
false
false
false
false
false
false
false
true
3.19488
medium
medium
7,437
ClarenceDan/f388f2e5-d513-413f-8059-71bd783a4871
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
null
False
65,536
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
very_long
7,438
ClarenceDan/f762713d-9d11-44a2-9992-f13ba3c3c73e
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.46.0
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
7,439
ClarenceDan/f8a647ca-0998-4307-b1eb-4e394f36c53a
llama
["LlamaForCausalLM"]
50964
float32
4.46.0
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
7,440
ClarenceDan/f9287d81-a5a3-4dd9-b31a-4dbe59af481f
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,024
2,816
24
16
16
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
0.30199
small
very_long
7,441
ClarenceDan/f9dd37b2-4c0e-4f58-a15b-d196597789b3
gpt_neox
["GPTNeoXForCausalLM"]
50304
float16
4.46.0
2,048
8,192
16
8
null
gelu
0
False
2,048
10,000
null
0.02
0
0
false
256
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.805306
small
medium
7,442
ClarenceDan/fa5a7769-c5f1-43ef-9fa4-61720937988c
mistral
["MistralForCausalLM"]
32000
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
7,443
ClarenceDan/fa63d372-1bb2-4810-9788-ed0c5d65554a
mistral
["MistralForCausalLM"]
32064
bfloat16
4.46.0
5,120
17,920
40
40
10
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
medium
7,444
ClarenceDan/fa8130a7-0693-4aac-aea8-6c894c5bc3ff
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
2,048
8,192
16
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
7,445
ClarenceDan/fb7bd7cf-7eaf-4684-a72c-67a360fffd55
gpt_neo
["GPTNeoForCausalLM"]
50257
null
4.46.0
768
null
null
null
null
null
0
False
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
7,446
ClarenceDan/fd988def-1fbf-47d9-83c1-5395477437fc
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
6.442451
medium
very_long
7,447
ClarenceDan/fe0511fb-d3f5-47b4-95c1-d9f081e735ad
llama
["LlamaForCausalLM"]
37632
float16
4.46.0
5,120
13,824
40
40
40
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
7,448
ClarenceDan/fee8a64b-8f3d-4af4-b4df-44126f7b84a1
mistral
["MistralForCausalLM"]
32002
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
7,449
Clark88/TinyLlama-QNA-Colab20epoch
llama
["LlamaForCausalLM"]
32000
float16
4.46.2
2,048
5,632
22
32
4
silu
0
True
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
7,450
ClassCat/gpt2-base-japanese-v2
gpt2
["GPT2LMHeadModel"]
60000
float32
4.19.2
null
null
null
null
null
null
null
True
null
null
null
0.02
0
0
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
7,451
Classical/Yinka
bert
["BertModel"]
21128
float32
4.37.1
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
7,452
Classroom-workshop/assignment1-francesco
speech_to_text
["Speech2TextForConditionalGeneration"]
10000
null
4.4.0.dev0
null
null
12
null
null
null
0.1
True
null
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,453
Classroom-workshop/assignment1-jack
speech_to_text
["Speech2TextForConditionalGeneration"]
10000
null
4.4.0.dev0
null
null
12
null
null
null
0.1
True
null
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,454
Classroom-workshop/assignment1-jane
speech_to_text
["Speech2TextForConditionalGeneration"]
10000
null
4.4.0.dev0
null
null
12
null
null
null
0.1
True
null
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,455
Classroom-workshop/assignment1-joane
speech_to_text
["Speech2TextForConditionalGeneration"]
10000
null
4.4.0.dev0
null
null
12
null
null
null
0.1
True
null
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,456
Classroom-workshop/assignment1-maria
speech_to_text
["Speech2TextForConditionalGeneration"]
10000
null
4.4.0.dev0
null
null
12
null
null
null
0.1
True
null
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,457
Classroom-workshop/assignment1-omar
wav2vec2
["Wav2Vec2ForCTC"]
32
null
4.7.0.dev0
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
7,458
Classroom-workshop/assignment2-francesco
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,459
Classroom-workshop/assignment2-julien
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,460
Classroom-workshop/assignment2-llama
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,461
Classroom-workshop/assignment2-llamas
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,462
Classroom-workshop/assignment2-omar
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,463
Classroom-workshop/assignment2-thom
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,464
ClaudeRitchie/tinyllama-vels-v1
llama
["LlamaForCausalLM"]
32000
float16
4.46.2
2,048
5,632
22
32
4
silu
0
True
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
7,465
ClaudiaRichard/mbti-bert-nli-finetuned_v2
bert
["BertForSequenceClassification"]
30522
float32
4.38.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,466
ClaudianoLeonardo/bert-finetuned_news_classifier-portuguese
bert
["BertForSequenceClassification"]
30522
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,467
ClaudioItaly/1800-9B
gemma2
["Gemma2ForCausalLM"]
256000
float16
4.46.2
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
7,468
ClaudioItaly/1852-9B
gemma2
["Gemma2ForCausalLM"]
256000
float16
4.46.2
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
7,469
ClaudioItaly/Book-Gut.bfloat32
mistral
["MistralForCausalLM"]
131072
float32
4.44.1
5,120
14,336
40
32
8
silu
0
False
1,024,000
1,000,000
0.00001
0.02
1
4
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
7,470
ClaudioItaly/Book-Gut12B
mistral
["MistralForCausalLM"]
131072
bfloat16
4.44.1
5,120
14,336
40
32
8
silu
0
False
1,024,000
1,000,000
0.00001
0.02
1
4
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
7,471
ClaudioItaly/Evocation
llama
["LlamaForCausalLM"]
32000
bfloat16
4.44.1
5,120
13,824
40
40
40
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
7,472
ClaudioItaly/Evolutiongutenberg-50k
mistral
["MistralForCausalLM"]
32000
bfloat16
4.44.1
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
7,473
ClaudioItaly/Evolutionstory-7B-v2.2
mistral
["MistralForCausalLM"]
32000
bfloat16
4.44.1
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
7,474
ClaudioItaly/Evolutionstory-7B-v3.0
mistral
["MistralForCausalLM"]
32000
float16
4.44.1
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
7,475
ClaudioItaly/Exurbia-Scriptus
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.52.4
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
7,476
ClaudioItaly/Follex-7B
qwen2
["Qwen2ForCausalLM"]
151665
bfloat16
4.46.2
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
7,477
ClaudioItaly/Intelligence-7
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.2
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
7,478
ClaudioItaly/Intelligence-Cod-Rag-7B
qwen2
["Qwen2ForCausalLM"]
152064
float16
4.46.2
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
7,479
ClaudioItaly/OroSaiwa-12B
mistral
["MistralForCausalLM"]
131072
float16
4.48.2
5,120
14,336
40
32
8
silu
0
True
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
7,480
ClaudioItaly/Project-12
mistral
["MistralForCausalLM"]
131072
bfloat16
4.44.1
5,120
14,336
40
32
8
silu
0
False
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
7,481
ClaudioItaly/Prometheus-wll3
mistral
["MistralForCausalLM"]
32000
bfloat16
4.41.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
7,482
ClaudioItaly/Qwen-Density
qwen2
["Qwen2ForCausalLM"]
151665
bfloat16
4.45.2
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
7,483
ClaudioItaly/Sharp11
llama
["LlamaForCausalLM"]
32000
bfloat16
4.44.1
4,096
14,336
48
32
8
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
9.663676
large
medium
7,484
ClaudioItaly/intelligence-cod-rag-7b-v3
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.2
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
7,485
ClaudioItaly/intelligence-cod-rag-7b-v3.1
qwen2
["Qwen2ForCausalLM"]
152064
float16
4.46.2
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
7,486
ClawCardMirror/sd-class-butterflies-32
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,487
ClawCardMirror/sd-class-butterflies-64
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,488
Clawoo/ppo-LunarLander-v2u1
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,489
ClementP/FundusDRGrading-densenet121.tv_in1k
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,490
ClementP/FundusDRGrading-mobilenetv3_small_100
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,491
ClementPoirier/distilbert-base-uncased-finetuned-emotion
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.32.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
7,492
ClementXie/speecht5_finetuned_multilingual_librispeech_italian
speecht5
["SpeechT5ForTextToSpeech"]
81
float32
4.32.0.dev0
768
null
null
null
null
gelu
0.1
False
null
null
null
0.02
0
2
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
7,493
Clemnt73/RoBERTa-ner
camembert
["CamembertForTokenClassification"]
32005
float32
4.40.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
5
6
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,494
Cleo3927/calculator_model_test
encoder-decoder
["EncoderDecoderModel"]
null
float32
4.48.3
null
null
null
null
null
null
null
null
null
null
null
null
null
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,495
CleopatraQueen/modelPart1
null
[]
null
null
null
null
null
null
24
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
7,496
CleopatraQueen/modelPart2
t5
["T5EncoderModel"]
32128
bfloat16
4.40.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
7,497
CleverShovel/falcon-7b-instruct-sharded-bf16
RefinedWebModel
["RWForCausalLM"]
65024
bfloat16
4.30.0.dev0
4,544
null
null
null
null
null
0
True
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,498
CleverShovel/rubert-tiny2-tnved-v3
bert
["BertForPreTraining"]
83828
float32
4.30.1
312
600
3
12
null
gelu
null
True
2,048
null
null
0.02
null
null
null
26
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.003504
small
medium
7,499
CleverShovel/vicuna-7b-v1.3-sharded-bf16
llama
["LlamaForCausalLM"]
32000
bfloat16
4.31.0.dev0
4,096
11,008
32
32
null
silu
null
False
2,048
null
0.000001
0.02
1
2
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
6.442451
medium
medium