index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
2,900
Alphatao/4c26369b-409e-48b7-9bb6-283d6bb643ba
llama
["LlamaForCausalLM"]
32000
bfloat16
4.51.3
4,096
11,008
32
32
32
silu
0
False
2,048
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,901
Alphatao/4d64a5e2-ab88-4d1a-83c6-ae4ad7d34524
llama
["LlamaForCausalLM"]
46336
bfloat16
4.51.3
4,096
11,008
32
32
32
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,902
Alphatao/4f6569e7-34d2-4cda-999c-da683488fcad
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
2,903
Alphatao/50204e11-9430-4f18-84ba-dbc29bdedc2b
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.51.3
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
2,904
Alphatao/51aae6c8-1d09-48ad-8fd6-3b594567da29
gptj
["GPTJForCausalLM"]
50401
float32
4.46.0
null
null
null
null
null
null
null
False
null
null
null
0.02
50256
50256
false
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
2,905
Alphatao/5386e54c-6f0b-44ed-82fc-301671593fcb
llama
["LlamaForCausalLM"]
32016
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
2,906
Alphatao/53cbd2d1-c5a5-47ad-b85c-d5ce51863f72
gptj
["GPTJForCausalLM"]
50401
float32
4.46.0
null
null
null
null
null
null
null
False
null
null
null
0.02
50256
50256
false
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
2,907
Alphatao/54567338-22b2-4866-90e8-1cc87b2dfcec
llama
["LlamaForCausalLM"]
32000
float32
4.46.0
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0
2
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
2,908
Alphatao/54e6ea23-b5e1-4619-b29d-902656097671
bloom
["BloomForCausalLM"]
250880
null
4.46.0
1,024
null
null
null
null
null
0
False
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,909
Alphatao/56cc4757-4aaa-4ddd-a175-d832d5880469
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
2,910
Alphatao/574d660b-4c16-4510-bdf8-ecf06d8c1433
falcon
["FalconForCausalLM"]
65024
float32
4.46.0
128
null
2
16
null
null
0
False
2,048
10,000
null
0.02
null
11
null
8
null
false
null
null
false
false
false
false
false
false
false
false
false
true
0.000393
small
medium
2,911
Alphatao/59a20e7a-5cc9-40b9-bca2-d904340d471d
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
2,048
5,632
22
32
4
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
2,912
Alphatao/5b092e18-6368-42f0-884e-81f972e0831d
mistral
["MistralForCausalLM"]
32000
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,913
Alphatao/5bb857c2-6667-4cf4-ba3a-9767a0cd0ee6
bloom
["BloomForCausalLM"]
250880
bfloat16
4.46.0
1,024
null
null
null
null
null
0
False
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,914
Alphatao/5c07ecd3-751a-4f73-a96b-602102446a0d
llama
["LlamaForCausalLM"]
32016
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
2,915
Alphatao/5c7f2cf8-816e-4bc8-82cb-8c5ccfff5b6d
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
2,916
Alphatao/5ca7dde1-8203-4d10-9643-0dd1e665f5bd
mixtral
["MixtralForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
2
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
long
2,917
Alphatao/5d3e4647-2423-4a00-93b0-1c75136a8cda
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.46.0
2,048
16,384
18
8
1
gelu
0
False
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
2,918
Alphatao/5dad7cd7-8f88-4962-9096-72c5b10f7d06
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.46.0
2,304
9,216
26
8
4
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
288
2
false
null
null
false
false
false
false
false
false
false
false
true
true
1.656226
medium
long
2,919
Alphatao/5ddfea24-92d0-4238-b2c6-1b3b587ad4ab
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.51.3
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
2,920
Alphatao/5df356d8-e21a-4767-92b4-82f37caac0dc
mistral
["MistralForCausalLM"]
32002
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,921
Alphatao/5e8265a2-f3cb-4602-a45a-4387bf413c3c
mistral
["MistralForCausalLM"]
32000
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,922
Alphatao/5e9eb278-2c59-414a-b738-916225b68146
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,923
Alphatao/5f41d88f-6034-46f3-a2f6-0d2fac5aa67c
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
4,096
10,000
0.000001
0.02
null
151643
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
medium
2,924
Alphatao/600af72b-4398-4531-9403-d5a23539543a
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,925
Alphatao/6121d837-4efe-4605-bc7a-c889a9fc2445
falcon
["FalconForCausalLM"]
65024
float32
4.46.0
32
null
2
2
null
null
0
False
2,048
10,000
null
0.02
null
11
null
16
null
false
null
null
false
false
false
false
false
false
false
false
false
true
0.000025
small
medium
2,926
Alphatao/624c6438-94d0-47ae-9065-3a850d1df209
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,927
Alphatao/629c4fd8-0f76-4a5c-8872-522e88a0c20b
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
5,120
13,824
48
40
8
silu
0
False
131,072
1,000,000
0.00001
0.02
null
151643
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
15.099494
large
very_long
2,928
Alphatao/657724d0-7e4d-4610-9b21-770d14785894
mistral
["MistralForCausalLM"]
32000
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,929
Alphatao/66818cd4-df76-4d6a-b7e9-66325f20eda0
llama
["LlamaForCausalLM"]
46336
float16
4.46.0
4,096
11,008
32
32
32
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,930
Alphatao/682cb3f6-f587-490b-ad49-9d67caf93177
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
5,120
13,824
48
40
8
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
15.099494
large
very_long
2,931
Alphatao/688bf778-5066-47ca-a296-2eda0445ee7f
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128040
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,932
Alphatao/68a0c8c0-f607-48ea-8f0c-90b0e5a135e6
llama
["LlamaForCausalLM"]
32016
float16
4.46.0
5,120
13,824
40
40
40
silu
null
False
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
long
2,933
Alphatao/6978920a-5ce2-4670-bbc6-1d4638fcde76
llama
["LlamaForCausalLM"]
32000
float32
4.46.0
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0
2
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
2,934
Alphatao/6a551408-01ee-4be2-bab9-d9e52b79e502
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
2,935
Alphatao/6dd0f363-e02c-455c-9025-c8fc11ddc263
llama
["LlamaForCausalLM"]
46336
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,936
Alphatao/6feb5545-8b6b-402c-b51e-d68a9d5a5356
llama
["LlamaForCausalLM"]
128256
bfloat16
4.51.3
2,048
8,192
16
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
2,937
Alphatao/7143fad4-4680-4859-8175-54d5fdcb90eb
llama
["LlamaForCausalLM"]
32000
bfloat16
4.51.3
2,048
5,632
22
32
4
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
2,938
Alphatao/71fa2eb8-e933-4f08-8da8-d396551e317a
llama
["LlamaForCausalLM"]
46336
float16
4.46.0
4,096
11,008
32
32
32
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,939
Alphatao/73bb6bbb-83b4-4a28-98fd-5753f0f5a11d
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.46.0
3,072
8,192
32
32
32
silu
0
False
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
2,940
Alphatao/7501d54d-9519-4af1-996d-314ec6169d82
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.46.0
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
2,941
Alphatao/762667df-d54b-4367-9770-1507c6fb4e18
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
131,072
1,000,000
0.000001
0.02
null
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
2,942
Alphatao/78b92239-363b-4498-b9f3-ea32fe26cb86
llama
["LlamaForCausalLM"]
32000
float16
4.46.0
4,096
11,008
32
32
32
silu
0
False
2,048
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,943
Alphatao/7913ad28-023d-42d7-b1c1-56ea91c8382d
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.51.3
5,120
13,824
48
40
8
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
15.099494
large
very_long
2,944
Alphatao/7962c3d0-9181-4e73-ba0f-0f31f0b3d8ff
gpt_neo
["GPTNeoForCausalLM"]
50257
bfloat16
4.46.0
768
null
null
null
null
null
0
False
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
2,945
Alphatao/7a189fdf-fa6d-4398-8361-cd701356b6f5
mistral
["MistralForCausalLM"]
131072
bfloat16
4.46.0
5,120
14,336
40
32
8
silu
0
False
131,072
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
2,946
Alphatao/7a865761-1ebe-4e36-a968-573ec2b3da53
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
131,072
1,000,000
0.000001
0.02
null
151643
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
2,947
Alphatao/7ad41d7f-5f67-4a6e-a744-dfff6ed296d9
llama
["LlamaForCausalLM"]
32016
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
2,948
Alphatao/7b1c9498-c123-4ffa-a6d2-c89a58d29b2d
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,949
Alphatao/7c119629-e2d7-4d36-b294-6ff1444d9a2f
llama
["LlamaForCausalLM"]
128256
bfloat16
4.51.3
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,950
Alphatao/7cb60334-1256-417b-bc5b-9bf76487d51b
llama
["LlamaForCausalLM"]
32000
float16
4.46.0
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0
2
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
2,951
Alphatao/7ced1443-8dfa-4b9a-9516-64286ec5ec80
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
2,952
Alphatao/7d869ad8-f290-46a3-8e06-66c6a56576c3
phi3
["Phi3ForCausalLM"]
32064
float32
4.46.0
32
64
2
4
4
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
8
1
false
null
null
false
false
false
false
false
false
false
false
false
true
0.000025
small
medium
2,953
Alphatao/7f383048-21dd-4970-b4fa-695909c84ee2
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,954
Alphatao/7f53e0d9-6ef6-4807-8ffa-82397b5db6f4
mistral
["MistralForCausalLM"]
32032
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,955
Alphatao/7f6a1212-3fba-4feb-8697-b2824f65dec6
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.46.0
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
2,956
Alphatao/7f905f18-b4bd-4b7d-a24f-c1997e5c1082
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
131,072
1,000,000
0.000001
0.02
null
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
2,957
Alphatao/8039a57b-6569-4592-b742-5eb5feacd26a
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
5,120
13,824
40
40
40
silu
null
False
131,072
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
very_long
2,958
Alphatao/80f0703c-9a3d-449d-9e82-b500a6869626
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
5,120
13,824
40
40
40
silu
null
False
131,072
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
very_long
2,959
Alphatao/81427063-57c0-4a30-914b-fae72a7c76cf
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,960
Alphatao/8222543b-056f-416b-aec9-0a6f4499092a
mistral
["MistralForCausalLM"]
32003
bfloat16
4.51.3
4,096
14,336
32
32
8
silu
0
False
8,192
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
long
2,961
Alphatao/85b79278-4b02-4abd-a890-1ad7dd78e911
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.46.0
8
32
2
4
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
true
2
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.000002
small
very_long
2,962
Alphatao/86a7b836-c62a-4d01-8049-b0967b39bae3
llama
["LlamaForCausalLM"]
32000
bfloat16
4.51.3
8,192
28,672
80
64
8
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
medium
2,963
Alphatao/86df6ed8-abbf-4b09-8db4-5d3a45f3c3c0
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.46.0
2,048
16,384
18
8
1
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
2,964
Alphatao/88a54559-66a4-4bc9-a8f3-a7883532eb39
llama
["LlamaForCausalLM"]
32000
float32
4.46.0
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0
2
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
2,965
Alphatao/88c50e26-d8b3-4dcf-9964-378d8a82e752
falcon
["FalconForCausalLM"]
65024
bfloat16
4.46.0
4,544
null
32
71
null
null
0
False
null
null
null
0.02
null
11
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
7.928807
large
null
2,966
Alphatao/8983aabe-f0d0-402d-92e3-a6e3ac2196db
mistral
["MistralForCausalLM"]
32000
bfloat16
4.51.3
4,096
14,336
32
32
8
silu
null
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,967
Alphatao/8b8a9961-67dd-476b-bbdb-e530419f5393
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
4,096
12,288
36
32
8
silu
0
False
40,960
1,000,000
0.000001
0.02
null
151645
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
2,968
Alphatao/8bbd9c86-e7e6-496b-b182-36e5405c7697
llama
["LlamaForCausalLM"]
49152
float32
4.51.3
2,048
8,192
24
32
32
silu
0
False
2,048
10,000
0.00001
0.02
1
2
true
64
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
2,969
Alphatao/8c18c43d-90b9-4648-969f-63d26cb472cf
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
2,048
11,008
36
16
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
2,970
Alphatao/8c5e2e80-dc4a-4d9b-ab67-85b14be9d35d
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.46.0
2,304
9,216
26
8
4
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
288
2
false
null
null
false
false
false
false
false
false
false
false
true
true
1.656226
medium
long
2,971
Alphatao/8f72931d-3061-4725-891e-11164ed771f1
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,972
Alphatao/8fe58f73-cdbd-4e06-b5cb-c61637e0c7c7
mistral
["MistralForCausalLM"]
32000
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,973
Alphatao/9049ea79-dec5-4d0d-889f-488aa6fbc979
mistral
["MistralForCausalLM"]
32768
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,974
Alphatao/909db229-f2a4-40f3-934f-b1c727822330
llama
["LlamaForCausalLM"]
128257
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128003
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,975
Alphatao/91142d8c-7e9e-4dc8-a878-efe52e0c98fc
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,976
Alphatao/931a92d7-9be7-4202-b908-47ab521ca5b3
bloom
["BloomForCausalLM"]
250880
null
4.46.0
1,024
null
null
null
null
null
0
False
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,977
Alphatao/93f51e48-416d-4713-871f-b6aa5d058fad
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.51.3
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
2,978
Alphatao/9440b8c8-0d4f-4c0f-8d6c-68860ba78928
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
2,979
Alphatao/95e96b10-a5b6-4515-bf46-6d0e9c1ee7b7
llama
["LlamaForCausalLM"]
32016
bfloat16
4.51.3
4,096
11,008
32
32
32
silu
0
False
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
2,980
Alphatao/964e3937-5ce4-4111-8485-8d8a54251efd
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,981
Alphatao/969148ca-0126-4442-8552-58d3508ac76c
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
2,048
11,008
36
16
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
2,982
Alphatao/98757cb1-26fb-4d71-8b81-5df0c9abc4c7
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
2,983
Alphatao/98814a31-91af-4487-9567-fcdc1b0f66b5
mistral
["MistralForCausalLM"]
131072
bfloat16
4.51.3
5,120
14,336
40
32
8
silu
0
False
131,072
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
2,984
Alphatao/9afc90ce-3989-4c61-a7a5-87fd28a8c0f5
mistral
["MistralForCausalLM"]
131072
bfloat16
4.46.0
5,120
14,336
40
32
8
silu
0
False
131,072
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
2,985
Alphatao/9b04e3da-f9d7-4f1b-9edb-2c356199c1a4
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
2,986
Alphatao/9c56e0b4-ee90-4dff-b0d9-c1989593d2a5
phi3
["Phi3ForCausalLM"]
32064
float32
4.46.0
32
64
2
4
4
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
8
1
false
null
null
false
false
false
false
false
false
false
false
false
true
0.000025
small
medium
2,987
Alphatao/9d130103-e023-4be7-aef9-de753f7f0520
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
4,096
12,288
36
32
8
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
2,988
Alphatao/9d65626e-53aa-48e7-9c43-1cd15808de5a
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.46.0
2,048
16,384
18
8
1
gelu
0
False
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
2,989
Alphatao/9d9c981e-3a02-41b1-9c05-b5cc69fff61f
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.51.3
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
2,990
Alphatao/9dde6c56-4502-4260-862e-d089766ca369
llama
["LlamaForCausalLM"]
32016
float16
4.46.0
4,096
11,008
32
32
32
silu
null
False
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
2,991
Alphatao/9de75602-1b2e-4593-8696-0c920ca5d4d5
llama
["LlamaForCausalLM"]
32000
bfloat16
4.51.3
768
3,072
12
12
12
silu
0
False
2,048
10,000
0.000001
0.02
1
2
false
64
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.084935
small
medium
2,992
Alphatao/9df4788a-f3e3-4cb1-aed5-f58ce62f3768
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
5,120
13,824
40
40
40
silu
null
False
131,072
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
very_long
2,993
Alphatao/9e9b7aa3-b8a5-4d36-b78b-73c9f653e45e
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
2,994
Alphatao/9f2bd3e5-a4a5-44fb-80b4-f3c5178d6e0e
llama
["LlamaForCausalLM"]
32016
float16
4.46.0
5,120
13,824
40
40
40
silu
null
False
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
long
2,995
Alphatao/9f6f1cd8-5404-4542-acb4-2c2be8397445
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,024
2,816
24
16
16
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
0.30199
small
very_long
2,996
Alphatao/Affine-2501551
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.0
4,096
12,288
36
32
8
silu
0
True
40,960
1,000,000
0.000001
0.02
151643
151645
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
2,997
Alphatao/Affine-9801198
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.0
4,096
12,288
36
32
8
silu
0
True
40,960
1,000,000
0.000001
0.02
151643
151645
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
2,998
Alphatao/a013c2f8-2952-4cbf-b792-666623f21d71
mistral
["MistralForCausalLM"]
32002
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,999
Alphatao/a0f075c1-a825-4042-af8f-b28a7c866b1a
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium