index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
5,900
Bondds/06klaus6-06
stablelm
["StableLmForCausalLM"]
100352
float16
4.42.4
2,048
5,632
24
32
32
silu
0
False
4,096
10,000
null
0.02
100257
100278
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
5,901
Bondds/07-07001grapph
llama
["LlamaForCausalLM"]
131072
bfloat16
4.47.0
3,072
23,040
70
12
4
silu
0
True
32,768
1,000,042
0.000001
0.02
11
11
false
256
3
false
null
null
true
false
false
false
false
false
false
false
true
true
7.927235
large
very_long
5,902
Bondds/07-07003grapph
llama
["LlamaForCausalLM"]
131072
bfloat16
4.47.0
3,072
23,040
70
12
4
silu
0
True
32,768
1,000,042
0.000001
0.02
11
11
false
256
3
false
null
null
true
false
false
false
false
false
false
false
true
true
7.927235
large
very_long
5,903
Bondds/07-2008grapph
llama
["LlamaForCausalLM"]
64000
float16
4.42.4
4,096
11,008
48
32
4
silu
0
False
4,096
5,000,000
0.000001
0.02
1
7
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
9.663676
large
medium
5,904
Bondds/07-2100grapph
llama
["LlamaForCausalLM"]
128256
bfloat16
4.43.2
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
5,905
Bondds/07-6013grapph
llama
["LlamaForCausalLM"]
128258
bfloat16
4.45.1
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,906
Bondds/07-6016grapph
llama
["LlamaForCausalLM"]
128258
bfloat16
4.45.1
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,907
Bondds/07-6018grapph
llama
["LlamaForCausalLM"]
128258
bfloat16
4.45.1
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,908
Bondds/07vizv2-07
stablelm
["StableLmForCausalLM"]
100352
float16
4.42.4
2,048
5,632
24
32
32
silu
0
False
4,096
10,000
null
0.02
100257
100278
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
5,909
Bondds/08-07002grapph
llama
["LlamaForCausalLM"]
131072
bfloat16
4.47.0
3,072
23,040
70
12
4
silu
0
True
32,768
1,000,042
0.000001
0.02
11
11
false
256
3
false
null
null
true
false
false
false
false
false
false
false
true
true
7.927235
large
very_long
5,910
Bondds/08-16graph
llama
["LlamaForCausalLM"]
128256
float16
4.42.4
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,911
Bondds/09-07004grapph
llama
["LlamaForCausalLM"]
131072
bfloat16
4.47.0
3,072
23,040
70
12
4
silu
0
True
32,768
1,000,042
0.000001
0.02
11
11
false
256
3
false
null
null
true
false
false
false
false
false
false
false
true
true
7.927235
large
very_long
5,912
Bondds/09-4020grapph
llama
["LlamaForCausalLM"]
128258
bfloat16
4.45.1
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,913
Bondds/09-4022grapph
llama
["LlamaForCausalLM"]
128258
bfloat16
4.45.1
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,914
Bondds/10-07005grapph
llama
["LlamaForCausalLM"]
131072
bfloat16
4.47.0
3,072
23,040
70
12
4
silu
0
True
32,768
1,000,042
0.000001
0.02
11
11
false
256
3
false
null
null
true
false
false
false
false
false
false
false
true
true
7.927235
large
very_long
5,915
Bondds/10colt
stablelm
["StableLmForCausalLM"]
100352
float16
4.42.4
2,048
5,632
24
32
32
silu
0
False
4,096
10,000
null
0.02
100257
100278
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
5,916
Bondds/10klaus6-06
stablelm
["StableLmForCausalLM"]
100352
float16
4.42.4
2,048
5,632
24
32
32
silu
0
False
4,096
10,000
null
0.02
100257
100278
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
5,917
Bondds/10triple-10
stablelm
["StableLmForCausalLM"]
100352
float16
4.42.4
2,048
5,632
24
32
32
silu
0
False
4,096
10,000
null
0.02
100257
100278
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
5,918
Bondds/11-3010grapph
llama
["LlamaForCausalLM"]
128256
bfloat16
4.45.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
5,919
Bondds/13-2007grapph
mixtral
["MixtralForCausalLM"]
32000
bfloat16
4.44.1
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
2
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
very_long
5,920
Bondds/13-2120grapph
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,921
Bondds/14-2007grapph
llama
["LlamaForCausalLM"]
64000
float16
4.42.4
4,096
11,008
48
32
4
silu
0
False
4,096
5,000,000
0.000001
0.02
1
7
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
9.663676
large
medium
5,922
Bondds/14-coltst8v
stablelm
["StableLmForCausalLM"]
100352
float16
4.42.4
2,048
5,632
24
32
32
silu
0
False
4,096
10,000
null
0.02
100257
100278
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
5,923
Bondds/15-2008grapph
llama
["LlamaForCausalLM"]
64000
float16
4.42.4
4,096
11,008
48
32
4
silu
0
False
4,096
5,000,000
0.000001
0.02
1
7
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
9.663676
large
medium
5,924
Bondds/15-2120grapph
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,925
Bondds/16-09graph
llama
["LlamaForCausalLM"]
128256
float16
4.42.4
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,926
Bondds/16-4024grapph
llama
["LlamaForCausalLM"]
128258
bfloat16
4.45.1
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,927
Bondds/17-09graph
llama
["LlamaForCausalLM"]
128256
float16
4.42.4
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,928
Bondds/17-14graph
llama
["LlamaForCausalLM"]
128256
float16
4.42.4
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,929
Bondds/17-17graph
llama
["LlamaForCausalLM"]
128256
float16
4.42.4
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,930
Bondds/17-2120grapph
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,931
Bondds/17-2121grapph
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,932
Bondds/18-09graph
llama
["LlamaForCausalLM"]
128256
float16
4.42.4
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,933
Bondds/19-09graph
llama
["LlamaForCausalLM"]
128256
float16
4.42.4
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,934
Bondds/21-08graph
llama
["LlamaForCausalLM"]
128256
float16
4.42.4
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,935
Bondds/21-2009grapph
llama
["LlamaForCausalLM"]
64000
float16
4.42.4
4,096
11,008
48
32
4
silu
0
False
4,096
5,000,000
0.000001
0.02
1
7
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
9.663676
large
medium
5,936
Bondds/22-12graph
llama
["LlamaForCausalLM"]
128256
float16
4.42.4
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,937
Bondds/23-2032grapph
llama
["LlamaForCausalLM"]
64000
float16
4.42.4
4,096
11,008
48
32
4
silu
0
False
4,096
5,000,000
0.000001
0.02
1
7
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
9.663676
large
medium
5,938
Bondds/23-4012grapph
llama
["LlamaForCausalLM"]
128258
bfloat16
4.45.1
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,939
Bondds/23-4021grapph
llama
["LlamaForCausalLM"]
128258
bfloat16
4.45.1
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,940
Bondds/23-4027grapph
llama
["LlamaForCausalLM"]
128258
bfloat16
4.45.1
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,941
Bondds/24-10graph
llama
["LlamaForCausalLM"]
128256
float16
4.42.4
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,942
Bondds/25-13graph
llama
["LlamaForCausalLM"]
128256
float16
4.42.4
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,943
Bondds/25-2100grapph
llama
["LlamaForCausalLM"]
128256
float16
4.42.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
5,944
Bondds/26-2120grapph
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,945
Bondds/26-2121grapph
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,946
Bondds/26-4030grapph
llama
["LlamaForCausalLM"]
128258
bfloat16
4.45.1
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,947
Bondds/27-2031grapph
mixtral
["MixtralForCausalLM"]
32000
bfloat16
4.44.1
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
2
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
very_long
5,948
Bondds/30-6000grapph
llama
["LlamaForCausalLM"]
128258
bfloat16
4.45.1
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,949
Bondds/30-6014grapph
llama
["LlamaForCausalLM"]
128258
bfloat16
4.45.1
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,950
Bondds/30-6015grapph
llama
["LlamaForCausalLM"]
128258
bfloat16
4.45.1
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,951
Bongyun/vit-base-beans-demo-v5
vit
["ViTForImageClassification"]
null
float32
4.46.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,952
Bonnie422/GLM-4-4b-LoRa-gpt-pdfv2
chatglm
["ChatGLMModel"]
151552
bfloat16
4.40.0
4,096
null
40
32
null
null
0
True
null
null
null
null
null
[151329, 151336, 151338]
false
128
null
false
null
null
false
false
false
false
false
false
false
false
false
true
8.053064
large
null
5,953
Bonnief/mt5-amharic-summarizer
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.52.4
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
5,954
Bonnief/mt5-amharic-summarizer-3e
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.52.4
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
5,955
BookingCare/PhoBERT-Specialty-ClassificationBC
roberta
["RobertaForSequenceClassification"]
64001
float32
4.51.3
768
3,072
12
12
null
gelu
null
True
258
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,956
BookingCare/specialist-prediction-hybrid
new
["NewModel"]
250048
float32
4.50.3
768
3,072
12
12
null
gelu
null
null
8,192
20,000
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
true
0.084935
small
long
5,957
Boom4869/my_awesome_wbc_model_DWT_db1_1per
vit
["ViTForImageClassification"]
null
float32
4.39.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,958
Boom4869/my_awesome_wbc_model_DWT_db1_20per
vit
["ViTForImageClassification"]
null
float32
4.39.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,959
Boom4869/my_awesome_wbc_model_DWT_db1_5per
vit
["ViTForImageClassification"]
null
float32
4.39.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,960
Boom4869/my_awesome_wbc_model_DWT_db2_1per
vit
["ViTForImageClassification"]
null
float32
4.39.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,961
Boom4869/my_awesome_wbc_model_DWT_db2_20per
vit
["ViTForImageClassification"]
null
float32
4.39.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,962
Boom4869/my_awesome_wbc_model_DWT_db2_5per
vit
["ViTForImageClassification"]
null
float32
4.39.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,963
Boom4869/my_awesome_wbc_model_DWT_db3_1per
vit
["ViTForImageClassification"]
null
float32
4.39.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,964
Boom4869/my_awesome_wbc_model_DWT_db3_20per
vit
["ViTForImageClassification"]
null
float32
4.39.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,965
Boom4869/my_awesome_wbc_model_DWT_db3_5per
vit
["ViTForImageClassification"]
null
float32
4.39.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,966
Boom4869/my_awesome_wbc_model_FFT_20per
vit
["ViTForImageClassification"]
null
float32
4.39.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,967
Boom4869/my_awesome_wbc_model_FFT_5per
vit
["ViTForImageClassification"]
null
float32
4.39.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,968
Boom4869/my_awesome_wbc_model_Original
vit
["ViTForImageClassification"]
null
float32
4.39.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,969
BoooomNing/model
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,970
BoranIsmet/0.1
bert
["BertForSequenceClassification"]
30522
float32
4.50.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,971
BoranIsmet/0.15
bert
["BertForSequenceClassification"]
30522
float32
4.50.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,972
BoranIsmet/0.3
bert
["BertForSequenceClassification"]
30522
float32
4.50.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,973
BoranIsmet/0.33
bert
["BertForSequenceClassification"]
30522
float32
4.48.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,974
BoranIsmet/bert-base-uncased-finetuned-rte-run_boran_model
bert
["BertForSequenceClassification"]
30522
float32
4.50.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,975
BoredBrownBear/MythoMax-L2-13b-2.2bpw-exl2
llama
["LlamaForCausalLM"]
32000
float16
4.32.0.dev0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
5,976
BorelTHU/optvq-16x16x4
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,977
Borhan20/dummy-model
camembert
["CamembertForMaskedLM"]
32005
float32
4.47.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
5
6
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,978
BoringAnt1793/google-bert_bert-base-uncased
bert
["BertForSequenceClassification"]
30522
float32
4.46.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,979
BoringAnt1793/lyeonii_bert-medium
bert
["BertForSequenceClassification"]
30522
float32
4.46.3
512
2,048
8
8
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.025166
small
short
5,980
BoringAnt1793/lyeonii_bert-mini
bert
["BertForSequenceClassification"]
30522
float32
4.46.3
256
1,024
4
4
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.003146
small
short
5,981
BoringAnt1793/lyeonii_bert-small
bert
["BertForSequenceClassification"]
30522
float32
4.46.3
512
2,048
4
8
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.012583
small
short
5,982
BoringAnt1793/lyeonii_bert-tiny
bert
["BertForSequenceClassification"]
30522
float32
4.46.3
128
512
2
2
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.000393
small
short
5,983
Borista/handon_0_model
mistral
["MistralForCausalLM"]
32000
float16
4.38.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
5,984
Borja2508/DeepRL-1
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,985
Borjamg/pneumonia_model
vit
["ViTForImageClassification"]
null
float32
4.38.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,986
Borsch/dir4storage
bert
["BertForSequenceClassification"]
30522
float32
4.50.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,987
BoscoTheDog/Bessie_bitnet_instruct_100k_gguf
llama
["BitnetForCausalLM"]
32002
float32
4.39.2
1,536
4,096
24
16
16
silu
0
True
2,048
10,000
0.00001
0.02
1
2
true
96
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.679477
small
medium
5,988
BoscoTheDog/ministral_3b_it_q4_k_m_chunked
mistral
["MistralForCausalLM"]
32000
bfloat16
4.38.2
4,096
14,336
14
32
8
silu
0
False
131,072
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
2.818572
medium
very_long
5,989
BoscoTheDog/phi-3-mini-128k-instruct-ru
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.41.0
3,072
8,192
32
32
32
silu
0
False
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
5,990
BoscoTheDog/qwen2_5_coder_1_5b_chunked
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.44.0
1,536
8,960
28
12
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
5,991
BoscoTheDog/qwen2_5_coder_7b_chunked
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.44.0
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
5,992
Boshenxx/Time-R1-3B
qwen2_5_vl
["Qwen2_5_VLForConditionalGeneration"]
151936
bfloat16
4.51.3
2,048
11,008
36
16
2
silu
0
False
128,000
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
5,993
BossKushagra/ddpm-celebahq-finetuned-butterflies-2epochs
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,994
Bossixd/roberta-base-finetuned-imdb
roberta
["RobertaForMaskedLM"]
50265
null
4.35.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,995
BotatoFontys/FinetunedModel
gemma
["GemmaForCausalLM"]
256000
float32
4.38.1
2,048
16,384
18
8
1
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
5,996
BothBosu/bert-agent-suspect-only-scam-classifier-v1.0
bert
["BertForSequenceClassification"]
30522
float32
4.42.4
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,997
BothBosu/bert-scam-classifier-v1.6
bert
["BertForSequenceClassification"]
30522
float32
4.42.4
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,998
BothBosu/bilstm-agent-scam-classifier-v1.0
null
[]
null
null
null
768
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,999
BothBosu/bilstm-no-receiver-scam-classifier-v1.0
null
[]
null
null
null
768
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null