index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
4,900
AzureBlack/Thespis-13b-v0.3-5bpw-6h-exl2
llama
["LlamaForCausalLM"]
32000
float16
4.35.0.dev0
5,120
13,824
40
40
40
silu
null
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
4,901
AzureBlack/Xwin-MLewd-13B-V0.2-5bpw-6h-exl2
llama
["LlamaForCausalLM"]
32000
float16
4.35.0.dev0
5,120
13,824
40
40
40
silu
null
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
4,902
Azurro/APT3-1B-Instruct-v1
llama
["LlamaForCausalLM"]
31980
float32
4.35.2
2,048
5,504
18
32
32
silu
null
True
2,048
10,000
0.000001
0.02
1
2
false
64
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.90597
small
medium
4,903
Azurro/APT3-275M-Base
llama
["LlamaForCausalLM"]
31980
float32
4.35.0
768
2,048
32
12
12
silu
null
True
2,048
10,000
0.000001
0.02
1
2
false
64
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.226492
small
medium
4,904
Azwaruhuy/awang_model
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
4,905
Azzedde/llama3.1-8b-text2cypher
llama
["LlamaForCausalLM"]
128256
float16
4.48.3
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
4,906
Azzindani/Qwen2.5_1.5B_IT_ID_Legal
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.51.1
1,536
8,960
28
12
2
silu
0
True
32,768
1,000,000
0.000001
0.02
null
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
4,907
Azzizz17/autotrain-aaaa-45159113325
mt5
["MT5ForConditionalGeneration"]
250100
float32
4.25.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,908
Azzizz17/autotrain-translator3-45113113262
t5
["T5ForConditionalGeneration"]
32100
float32
4.25.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,909
B-O/dummy-mistral-4M
mistral
["MistralForCausalLM"]
50295
bfloat16
4.35.2
32
1,024
6
16
8
silu
null
True
4,096
10,000
0.00001
0.02
50256
50256
false
2
2
false
null
null
false
false
false
false
false
false
true
false
true
true
0.000074
small
medium
4,910
B0BWAX/mt5-small-finetuned-en-to-de
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.40.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,911
B2111797/recipe_v1_lr1e-3_wu4000_epo4
gpt2
["GPT2LMHeadModel"]
50268
float32
4.38.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
4,912
B2111797/trans-en-vi-v1
marian
["MarianMTModel"]
53685
float32
4.37.2
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
4,913
BA-Project-SA-CRM/SA_Checkpoints
bert
["BertForSequenceClassification"]
31102
float32
4.28.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,914
BAAI/AquilaDense-16B
aquiladense
["AquilaDenseForCausalLM"]
151851
bfloat16
4.37.2
5,120
20,480
40
40
8
silu
0
True
4,096
1,000,000
0.00001
0.02
151849
151850
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
12.582912
large
medium
4,915
BAAI/AquilaDense-7B
aquiladense
["AquilaDenseForCausalLM"]
151851
bfloat16
4.37.2
4,096
14,336
32
32
8
silu
0
True
4,096
1,000,000
0.00001
0.02
151849
151850
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
4,916
BAAI/AquilaMed-RL
aquila3
["AquilaForCausalLM"]
151851
bfloat16
4.38.2
4,096
14,336
32
32
8
silu
0
False
4,096
1,000,000
0.00001
0.02
151849
151850
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
4,917
BAAI/AquilaSQL-7B
aquila
["AquilaModel"]
100008
float16
4.28.1
4,096
11,008
32
32
null
silu
null
True
2,048
null
0.00001
0.02
1
2
false
128
null
false
null
null
false
false
false
false
false
false
false
false
false
false
6.442451
medium
medium
4,918
BAAI/BGE-VL-base
clip
["CLIPModel"]
null
bfloat16
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,919
BAAI/BGE-VL-large
clip
["CLIPModel"]
null
bfloat16
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,920
BAAI/BGE-VL-v1.5-zs
llava_next
["LlavaNextForConditionalGeneration"]
32064
float16
4.39.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
4,921
BAAI/EVE-7B-v1.0
eve
["EVELlamaForCausalLM"]
32000
bfloat16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
4,922
BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.44.2
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
107
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
4,923
BAAI/Infinity-Instruct-3M-0613-Llama3-70B
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0.dev0
8,192
28,672
80
64
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
long
4,924
BAAI/LLARA-beir
llama
["LlamaForCausalLM"]
32016
float32
4.38.1
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
4,925
BAAI/LLARA-document
llama
["LlamaForCausalLM"]
32016
float32
4.38.1
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
4,926
BAAI/LLARA-passage
llama
["LlamaForCausalLM"]
32016
float32
4.38.1
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
4,927
BAAI/LLARA-pretrain
llama
["LlamaForCausalLM"]
32016
float32
4.38.1
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
4,928
BAAI/OmniGen-v1
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.38.1
3,072
8,192
32
32
32
silu
0
True
131,072
10,000
0.00001
0.02
1
2
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
4,929
BAAI/SegVol
segvol
["SegVolModel"]
null
float32
4.18.0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,930
BAAI/bge-code-v1
qwen2
["Qwen2Model"]
151667
float32
4.49.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
4,931
BAAI/bge-en-icl
mistral
["MistralModel"]
32003
float32
4.41.1
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
4,932
BAAI/bge-reranker-v2-minicpm-layerwise
minicpm
["LayerWiseMiniCPMForCausalLM"]
122753
bfloat16
4.38.1
2,304
5,760
40
36
36
silu
0
False
2,048
10,000
0.00001
0.1
1
2
null
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
2.54804
medium
medium
4,933
BAAI/bge-small-en-v1.5
bert
["BertModel"]
30522
float32
4.30.0
384
1,536
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
32
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.021234
small
short
4,934
BAAI/llm-embedder
bert
["BertModel"]
30522
float32
4.30.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,935
BABYSHARK09/Na
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,936
BABYSHARK09/Ne
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,937
BABYSHARK09/Nf
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,938
BABYSHARK09/Ng
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,939
BABYSHARK09/Nq
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,940
BABYSHARK09/Nt
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,941
BABYSHARK09/Nu
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,942
BABYSHARK09/Nx
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,943
BABYSHARK09/Nz
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,944
BABYSHARK09/Uni_6x
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,945
BABYSHARK09/Uni_6x0
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,946
BABYSHARK09/Uni_6x2
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,947
BABYSHARK09/Uni_6x3
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,948
BABYSHARK09/Uni_6x4
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,949
BABYSHARK09/Uni_6x6
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,950
BABYSHARK09/Uni_6x7
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,951
BABYSHARK09/Uni_6x8
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,952
BABYSHARK09/Uni_6x9
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,953
BABYSHARK09/Uni_7x1
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,954
BABYSHARK09/Uni_7x2
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,955
BABYSHARK09/Uni_7x3
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,956
BABYSHARK09/Uni_7x4
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,957
BAC3030/hinrik-lp500-150-steps
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
4,958
BAHIJA/bert-base-uncased-finetuned-sst2
bert
["BertForSequenceClassification"]
30522
float32
4.17.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,959
BAK-HIRI/rl_course_vizdoom_health_gathering_supreme
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,960
BAKKALIAYOUB/DarijaTranslation-V1
marian
["MarianMTModel"]
62802
float32
4.41.2
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
4,961
BANA577/Llama3-Adrian-2
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0.dev0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,962
BANA577/Llama3-Michael-5
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0.dev0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,963
BANA577/Llama3-Michael-6
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0.dev0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,964
BANA577/Llama3-Michael-7
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0.dev0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,965
BANA577/Mistral-Koichi1
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0.dev0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,966
BANA577/Mistral-Koichi3
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.0
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
4,967
BANZHISONGSHU/whisper-small-ug
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.51.3
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
4,968
BBBBCHAN/LLaVA-Scissor-baseline-0.5B
qwen2
["LlavaQwenZipForCausalLM"]
151936
bfloat16
4.40.0.dev0
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
4,969
BBBBirdIsTheWord/ML-Agents-Pyramids
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,970
BBBBirdIsTheWord/a2c-PandaPickAndPlace-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
4,971
BBBBirdIsTheWord/a2c-PandaReachDense-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
4,972
BBBBirdIsTheWord/poca-SoccerTwos
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,973
BBBBirdIsTheWord/ppo-SnowballTarget
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,974
BBGAME605065444/my_awesome_model
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.28.0.dev0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
4,975
BBLL3456/faster-whisper-large-V3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,976
BBVA-COE-AA-FINANZAS-MX/Meta-Llama-3-8B-Instruct
llama
["LlamaForCausalLM"]
128256
bfloat16
4.42.3
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,977
BBorg/a2c-PandaReachDense-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,978
BBuf/RWKV-4-World-169M
rwkv
[]
65536
null
4.33.1
768
3,072
12
null
null
null
null
True
null
null
null
null
0
0
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
4,979
BBuf/RWKV-4-World-1B5
rwkv
[]
65536
null
4.33.1
2,048
8,192
24
null
null
null
null
True
null
null
null
null
0
0
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
1.20796
medium
null
4,980
BBuf/RWKV-4-World-3B
rwkv
[]
65536
null
4.33.1
2,560
10,240
32
null
null
null
null
True
null
null
null
null
0
0
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
2.516582
medium
null
4,981
BBuf/RWKV-4-World-430M
rwkv
[]
65536
null
4.33.1
1,024
4,096
24
null
null
null
null
True
null
null
null
null
0
0
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
4,982
BBuf/RWKV-4-World-7B
rwkv
[]
65536
null
4.33.1
4,096
16,384
32
null
null
null
null
True
null
null
null
null
0
0
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
6.442451
medium
null
4,983
BCCard/Qwen2.5-Coder-14B-FP8-Dynamic
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.3
5,120
13,824
48
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
15.099494
large
very_long
4,984
BCCard/Qwen2.5-Coder-32B-Instruct-FP8-Dynamic
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.3
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
4,985
BCCard/Qwen2.5-Coder-7B-FP8-Dynamic
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.3
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
4,986
BCCard/gemma-3-27b-it-FP8-Dynamic
gemma3
["Gemma3ForConditionalGeneration"]
null
bfloat16
4.51.3
null
null
null
null
null
null
null
null
null
null
null
0.02
null
[1, 106]
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
4,987
BDAD/segformer-b3-horizontal
segformer
["SegformerForSemanticSegmentation"]
null
float32
4.29.2
null
null
null
[1, 2, 5, 8]
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,988
BDAD/segformer-b3-vertical
segformer
["SegformerForSemanticSegmentation"]
null
float32
4.29.2
null
null
null
[1, 2, 5, 8]
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,989
BDomantas/layoutxlm-base-tune
layoutlmv2
["LayoutLMv2ForTokenClassification"]
250002
float32
4.39.3
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
4,990
BEBO-DBIndia/LLAMA_V58M
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
4,991
BEE-spoke-data/Mixtral-GQA-400m-v4-4096
mixtral
["MixtralForCausalLM"]
32000
float32
4.37.0.dev0
2,048
6,144
6
32
8
silu
0
False
4,096
1,000,000
0.00001
0.02
1
2
false
64
4
false
8
2
false
false
false
false
false
false
false
true
true
true
0.30199
small
medium
4,992
BEE-spoke-data/Qwen2-1.5B-stepbasin-books
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.42.4
1,536
8,960
28
12
2
silu
0
True
131,072
1,000,000
0.000001
0.02
151643
151643
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
4,993
BEE-spoke-data/TinyLlama-3T-1.1bee
llama
["LlamaForCausalLM"]
32000
bfloat16
4.36.2
2,048
5,632
22
32
4
silu
0
True
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
4,994
BEE-spoke-data/beecoder-220M-python
llama
["LlamaForCausalLM"]
32128
bfloat16
4.36.2
1,024
4,096
10
32
8
silu
0
False
2,048
10,000
0.000001
0.02
1
2
false
32
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.125829
small
medium
4,995
BEE-spoke-data/bert-plus-L8-4096-v1.0
bert
["BertForMaskedLM"]
30522
float32
4.37.2
768
3,072
8
12
null
silu
null
True
4,096
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.056623
small
medium
4,996
BEE-spoke-data/mega-ar-126m-4k
mega
["MegaForCausalLM"]
50304
float32
4.36.2
768
1,536
12
1
null
null
null
True
null
null
null
0.02
0
0
null
768
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
4,997
BEE-spoke-data/mega-encoder-small-16k-v1
mega
["MegaForMaskedLM"]
50304
float32
4.38.2
768
2,304
8
1
null
null
null
True
null
null
null
0.02
0
2
null
768
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.056623
small
null
4,998
BEE-spoke-data/mega-small-embed-synthSTS-16384-v1
mega
["MegaModel"]
50304
float32
4.38.2
768
2,304
8
1
null
null
null
True
null
null
null
0.02
0
2
null
768
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.056623
small
null
4,999
BEE-spoke-data/roberta-base-description2genre
roberta
["RobertaForSequenceClassification"]
50265
float32
4.33.3
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short