index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
3,000
Alphatao/a268c1f2-44df-4210-8cc5-a1315c664ace
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
3,001
Alphatao/a302b12e-25b3-4e66-8dd8-f1f1bb4869ec
llama
["LlamaForCausalLM"]
32001
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,002
Alphatao/a35f6753-dab1-417f-bb87-722cc70dc198
llama
["LlamaForCausalLM"]
32000
float32
4.46.0
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0
2
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
3,003
Alphatao/a3a3ad81-b4a1-4f93-93ae-295283c65877
gpt_neo
["GPTNeoForCausalLM"]
50257
null
4.46.0
768
null
null
null
null
null
0
False
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
3,004
Alphatao/a4f836f9-77b6-40b0-883c-8bf0bd1d650e
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.51.3
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,005
Alphatao/a7b5cad9-ed1a-471f-bd10-01951c287c2e
llama
["LlamaForCausalLM"]
32000
bfloat16
4.51.3
768
3,072
12
12
12
silu
0
False
2,048
10,000
0.000001
0.02
1
2
false
64
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.084935
small
medium
3,006
Alphatao/a98ce960-f26d-4c21-b9a2-03d57c343e5a
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,007
Alphatao/a9ac2ff4-ac21-4132-8c05-daa71cc03a32
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
3,008
Alphatao/aab7a25f-9b39-4f6c-91da-20330f199c18
llama
["LlamaForCausalLM"]
32000
float16
4.46.0
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0
2
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
3,009
Alphatao/ab41372d-a067-4674-a1ae-b25e52ab86e7
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
2,048
5,632
22
32
4
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
3,010
Alphatao/ab718992-3199-4c7d-9d2b-e6bb6e526621
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
3,011
Alphatao/ac28639f-df4f-45cd-86e2-45e24bd81135
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.46.0
3,072
8,192
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
medium
3,012
Alphatao/b0031740-8709-4b67-b3ad-97c140c7a3a7
phi
["PhiForCausalLM"]
51200
float16
4.46.0
2,560
10,240
32
32
32
gelu_new
0
False
2,048
10,000
null
0.02
50256
50256
false
80
1
false
null
null
false
false
false
false
false
false
false
false
false
true
2.516582
medium
medium
3,013
Alphatao/b35eda51-b72a-4fa3-8e0f-cf291990bf61
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
2,048
5,632
22
32
4
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
3,014
Alphatao/b4df6c23-3135-458e-be9f-b19cca5ee93f
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,015
Alphatao/b52226ed-e93f-4458-86a0-84effc8cb463
llama
["LlamaForCausalLM"]
32000
float16
4.46.0
4,096
11,008
32
32
32
silu
0
False
2,048
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,016
Alphatao/b58f49fc-bc1d-4d36-b2dc-bd997c6cbee2
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
null
False
65,536
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
very_long
3,017
Alphatao/b5c4c1fe-219c-4e5f-8bd7-30ba715d4b20
mistral
["MistralForCausalLM"]
32000
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
null
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
3,018
Alphatao/b62b31cd-6953-4cfa-8759-ce12bfc2615e
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,019
Alphatao/b71b680d-7d75-46a4-806d-5b40a3b04c60
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,020
Alphatao/b7366c78-e780-497b-b1a2-cb99d28af56a
llama
["LlamaForCausalLM"]
49153
bfloat16
4.51.3
2,048
8,192
24
32
32
silu
0
False
8,192
130,000
0.00001
0.02
1
2
true
64
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
long
3,021
Alphatao/b779a5e2-0eb7-448f-8df3-31ddb43d3a88
phi
["PhiForCausalLM"]
51200
bfloat16
4.51.3
2,560
10,240
32
32
32
gelu_new
0
False
2,048
10,000
null
0.02
50256
50256
false
80
1
false
null
null
false
false
false
false
false
false
false
false
false
true
2.516582
medium
medium
3,022
Alphatao/b884a1c5-2611-4527-b2a5-53af1d5ca8fa
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128040
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,023
Alphatao/ba016fbc-6093-42a3-a67c-a7a2b1632975
mistral
["MistralForCausalLM"]
32000
bfloat16
4.51.3
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
3,024
Alphatao/bceae174-5058-44a9-b7d8-8982e7764e43
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
3,025
Alphatao/bd8adeda-f612-4dc5-87de-1ffd9d384f26
llama
["LlamaForCausalLM"]
32000
float32
4.46.0
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0
2
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
3,026
Alphatao/beaa33ea-655f-4aa9-af89-ea21f3ab08f0
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
3,027
Alphatao/bed2ffc8-b67d-4988-9f07-a4dde0572e35
llama
["LlamaForCausalLM"]
128256
float16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,028
Alphatao/c0612f55-3d93-4c5b-a06a-c74d850340ad
gpt_neo
["GPTNeoForCausalLM"]
50257
bfloat16
4.46.0
768
null
null
null
null
null
0
False
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
3,029
Alphatao/c2fb4bef-a398-4aea-8ec9-1c05ff9ecf7d
llama
["LlamaForCausalLM"]
49152
bfloat16
4.46.0
960
2,560
32
15
5
silu
0
False
2,048
10,000
0.00001
0.02
0
0
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.353894
small
medium
3,030
Alphatao/c3919e80-2191-444a-823e-2e0dab999072
llama
["LlamaForCausalLM"]
128256
bfloat16
4.51.3
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,031
Alphatao/c4c50052-4a91-41f8-a0e1-59319d72ad16
llama
["LlamaForCausalLM"]
32016
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
3,032
Alphatao/c4f795c4-346a-4739-a308-e57c22a21816
mistral
["MistralForCausalLM"]
32064
bfloat16
4.46.0
5,120
17,920
40
40
10
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
medium
3,033
Alphatao/c5e356c0-2c3f-4577-8ba1-25456059972b
llama
["LlamaForCausalLM"]
128256
float16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,034
Alphatao/c6262bdd-a2df-4fd6-ae50-89f9558279fd
llama
["LlamaForCausalLM"]
50964
float32
4.46.0
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,035
Alphatao/c70ed71b-340d-4541-a745-469584021952
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,036
Alphatao/c7892fd8-7b00-4caa-9a9c-321cf61c7685
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,037
Alphatao/c8584ad6-fb9d-4cdf-a7b3-d0f7943183bc
mistral
["MistralForCausalLM"]
32002
float32
4.51.3
4,096
14,336
32
32
8
silu
0
False
8,192
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
long
3,038
Alphatao/c9fe4bf0-c690-4c16-939a-e900e96b8da2
bloom
["BloomForCausalLM"]
250880
null
4.46.0
1,024
null
null
null
null
null
0
False
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,039
Alphatao/cb5b55b0-f6fa-4a32-8d4a-02fb26201718
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,040
Alphatao/cc6ef30e-acba-41c5-b9ba-c7e86953a20f
llama
["LlamaForCausalLM"]
128256
bfloat16
4.51.3
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,041
Alphatao/cca84d0e-df15-4a76-84eb-c4daa3d733cb
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.51.3
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
3,042
Alphatao/ce770bbd-2877-4d64-8331-884386dbaf3b
falcon
["FalconForCausalLM"]
65024
float32
4.46.0
32
null
2
2
null
null
0
False
2,048
10,000
null
0.02
null
11
null
16
null
false
null
null
false
false
false
false
false
false
false
false
false
true
0.000025
small
medium
3,043
Alphatao/cfeb3df6-8a31-496e-8f87-d466f7fe68d1
phi3
["Phi3ForCausalLM"]
32064
float32
4.46.0
32
64
2
4
4
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
8
1
false
null
null
false
false
false
false
false
false
false
false
false
true
0.000025
small
medium
3,044
Alphatao/d0418eba-34ac-4759-87f2-3d693d4da9e9
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,045
Alphatao/d0b8001c-0407-4b2d-9e01-b583b1e2ae15
llama
["LlamaForCausalLM"]
32064
bfloat16
4.46.0
3,072
8,192
32
32
32
silu
0
False
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
true
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
3,046
Alphatao/d1582a94-c330-4a7f-8d54-034e68d82632
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
2,048
5,504
24
16
16
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
very_long
3,047
Alphatao/d28a62c4-0691-42e4-a0bc-f275efe25315
llama
["LlamaForCausalLM"]
32016
float16
4.46.0
4,096
11,008
32
32
32
silu
null
False
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
3,048
Alphatao/d3e5fbae-646d-454c-abbe-114b090e6d3b
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.51.3
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
3,049
Alphatao/d41c3522-aa8e-4446-a597-ef5cc8991b96
gemma
["GemmaForCausalLM"]
256000
float32
4.46.0
32
2
1
2
1
gelu
0
False
1,024
10,000
0.000001
0.02
2
1
null
16
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.000012
small
short
3,050
Alphatao/d479b192-d9fa-4a9b-b709-ba64d46f5fa3
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
2,048
11,008
36
16
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
3,051
Alphatao/d4fe37a6-5c42-4ed9-87df-18c947259924
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,052
Alphatao/d57dcc4c-f303-4556-b8ba-347e5959177e
llama
["LlamaForCausalLM"]
128256
bfloat16
4.51.3
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128040
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,053
Alphatao/d5dc0af2-24ab-47f0-8a02-54fea9cc7749
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,054
Alphatao/d62ed36f-68e4-4828-8211-91506878d910
opt
["OPTForCausalLM"]
50272
bfloat16
4.51.3
768
null
12
12
null
null
0
False
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
medium
3,055
Alphatao/d67ab567-71e3-4588-bb84-e397948feb23
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.51.3
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
3,056
Alphatao/d8c7be39-f548-4863-9331-4fb4fc180afa
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
2,048
5,632
22
32
4
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
3,057
Alphatao/d96d18be-1c39-49c1-b429-f6623f4ab502
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,058
Alphatao/da5838d7-b88b-4bdc-8bc0-139f86cc9c34
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
2,560
9,728
36
32
8
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
80
4
false
null
null
false
false
false
false
false
false
false
false
true
true
2.831155
medium
very_long
3,059
Alphatao/da8811cc-b1bf-4a31-8ec4-f11f7a4a956f
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
3,060
Alphatao/daa4a8b2-4cda-4184-a2f2-686fc8389361
llama
["LlamaForCausalLM"]
32000
float32
4.46.0
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0
2
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
3,061
Alphatao/dc2f6d25-78bf-41bf-a543-0f506b4b1745
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,062
Alphatao/dc7401e9-f365-4f04-9a04-574ab402fcd6
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
2,048
5,632
22
32
4
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
3,063
Alphatao/dd01aad8-495d-41e8-943c-30ec49597b17
gptj
["GPTJForCausalLM"]
50401
float32
4.46.0
null
null
null
null
null
null
null
False
null
null
null
0.02
50256
50256
false
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,064
Alphatao/dd206ec7-9cef-4517-abdc-c03d9ee91679
gpt_neo
["GPTNeoForCausalLM"]
50257
null
4.46.0
768
null
null
null
null
null
0
False
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
3,065
Alphatao/dec6dc9d-0f67-465d-8e0b-5374ae290dbd
mistral
["MistralForCausalLM"]
131072
bfloat16
4.46.0
5,120
14,336
40
32
8
silu
0
False
131,072
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
3,066
Alphatao/df987ed4-a7b9-47f7-afe4-a851a783343b
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.46.0
2,304
9,216
26
8
4
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
288
2
false
null
null
false
false
false
false
false
false
false
false
true
true
1.656226
medium
long
3,067
Alphatao/dfde36a5-fe78-42d4-bdae-275888b39fcc
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.51.3
1,024
2,816
24
16
16
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
0.30199
small
very_long
3,068
Alphatao/e04f2b44-20fb-474e-a1eb-896dfe1644a1
llama
["LlamaForCausalLM"]
32000
float32
4.46.0
768
3,072
12
12
12
silu
0
False
2,048
10,000
0.000001
0.02
1
2
false
64
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.084935
small
medium
3,069
Alphatao/e0d56ab5-4f60-45ba-9efc-7cf4c4ad239d
phi
["PhiForCausalLM"]
1025
float32
4.46.0
32
37
2
4
4
gelu
0
False
1,024
10,000
null
0.02
0
0
false
8
1
false
null
null
false
false
false
false
false
false
false
false
false
true
0.000025
small
short
3,070
Alphatao/e39cc9d8-8a9b-416c-b8ce-a37e864a33b1
llama
["LlamaForCausalLM"]
32001
float16
4.46.0
5,120
13,824
40
40
40
silu
0
False
2,048
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
3,071
Alphatao/e40dfcaa-ae2a-4adc-bd75-6f2df39d59b2
llama
["LlamaForCausalLM"]
128256
float16
4.46.0
16
64
2
4
4
silu
0
False
131,072
500,000
0.00001
0.02
128000
128009
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
very_long
3,072
Alphatao/e4dd9d2e-b409-4e96-96fd-c5616b3d9f16
llama
["LlamaForCausalLM"]
49152
bfloat16
4.51.3
576
1,536
30
9
3
silu
0
False
2,048
10,000
0.00001
0.02
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
medium
3,073
Alphatao/e51cbea9-cbe1-4470-beea-5dd6ed238f4c
llama
["LlamaForCausalLM"]
64000
bfloat16
4.46.0
4,096
11,008
48
32
4
silu
0
False
16,384
5,000,000
0.000001
0.02
1
7
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
9.663676
large
long
3,074
Alphatao/e5344067-e7c8-4efe-87df-a5562ba77bc0
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,075
Alphatao/e5367089-e923-4acf-a401-445d48ba0b2c
llama
["LlamaForCausalLM"]
128256
bfloat16
4.51.3
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,076
Alphatao/e57c53ba-0d4b-4e58-bc89-6815d1388c81
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
5,120
13,824
48
40
8
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
15.099494
large
very_long
3,077
Alphatao/e767daa7-2915-4c25-bb2e-cf841661ef8c
phi
["PhiForCausalLM"]
51200
bfloat16
4.51.3
2,048
8,192
24
32
32
gelu_new
0
False
2,048
10,000
null
0.02
null
null
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
3,078
Alphatao/e78c5238-c9e3-45b1-92f4-7579efbf91ab
llama
["LlamaForCausalLM"]
128256
float16
4.46.0
16
64
2
4
4
silu
0
False
131,072
500,000
0.00001
0.02
128000
128009
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
very_long
3,079
Alphatao/e8622fcc-2c9f-4752-99fa-d6d2de3b524e
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,080
Alphatao/e8e393c7-f748-4e5d-b84a-a9aa7a983901
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.46.0
3,072
8,192
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
medium
3,081
Alphatao/e93b4038-afa4-4936-bcdc-c957e4ef3b4b
gptj
["GPTJForCausalLM"]
50401
float32
4.46.0
null
null
null
null
null
null
null
False
null
null
null
0.02
50256
50256
false
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,082
Alphatao/eb388abe-f39e-45c1-9135-7f9e601a8fa3
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,083
Alphatao/eb8995f0-e357-47b2-a89b-40363a275b6c
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.51.3
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
3,084
Alphatao/ed4dbcce-7f8c-402c-80a6-19931aa492bd
llama
["LlamaForCausalLM"]
128256
bfloat16
4.51.3
2,048
8,192
16
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
3,085
Alphatao/ed9a17a2-af1c-448a-8891-f72155b6f9e3
gemma
["GemmaForCausalLM"]
256000
float32
4.46.0
32
2
1
2
1
gelu
0
False
1,024
10,000
0.000001
0.02
2
1
null
16
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.000012
small
short
3,086
Alphatao/ee203b06-2ae5-48b8-bb94-929bef6db13d
llama
["LlamaForCausalLM"]
32016
float16
4.46.0
5,120
13,824
40
40
40
silu
null
False
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
long
3,087
Alphatao/ef69d4de-2c0b-4d59-8ad6-f76bf7ad2e1e
llama
["LlamaForCausalLM"]
46336
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,088
Alphatao/efe5f82c-a0cc-4674-94ca-7c8e4c078f25
mistral
["MistralForCausalLM"]
32000
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
3,089
Alphatao/f047700a-b283-4650-a314-99e88c8bf9d4
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,090
Alphatao/f14d9a0b-a0be-49c9-92a9-fbecabb60b9d
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
6.442451
medium
very_long
3,091
Alphatao/f236ac0a-668a-478b-8693-e14a6088a857
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
2,048
5,632
22
32
4
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
3,092
Alphatao/f2b0fa58-8c68-4ee4-8ff5-65c996bc267d
llama
["LlamaForCausalLM"]
128256
float16
4.46.0
16
64
2
4
4
silu
0
False
131,072
500,000
0.00001
0.02
128000
128009
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
very_long
3,093
Alphatao/f2b3423d-d7bb-4141-ba74-4495c4913457
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
2,048
8,192
16
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
3,094
Alphatao/f3cbfaa6-f066-44cd-ba62-4321bf3d953e
llama
["LlamaForCausalLM"]
32016
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
3,095
Alphatao/f40d88eb-c453-498b-bd33-0bdda9bdd0b9
llama
["LlamaForCausalLM"]
128257
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128003
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,096
Alphatao/f55bf76e-1e88-4ac6-b977-220d58647f04
mistral
["MistralForCausalLM"]
32002
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
3,097
Alphatao/f5de3d9b-be97-4e55-ab45-8aaaabdf9d1a
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,098
Alphatao/f602bb6e-fd44-4fc7-a8df-b2ac1bbf08f0
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
4,096
10,000
0.000001
0.02
null
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
medium
3,099
Alphatao/f61daa15-27ab-432d-8a1f-cef37f7ca975
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long