index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
13,600
HFXM/Entropy_final_reward_model_71K-FsfairX-LR1e-5-Epoch2
llama
["LlamaFixedWeightReward"]
128257
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,601
HFXM/Entropy_final_reward_model_71K-FsfairX-LR2e-5-Epoch1
llama
["LlamaFixedWeightReward"]
128257
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,602
HFXM/Entropy_final_reward_model_71K-FsfairX-LR2e-5-Epoch2
llama
["LlamaFixedWeightReward"]
128257
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,603
HFXM/Entropy_final_reward_model_71K-Llama3_1-8B-LR1e-5-Epoch1
llama
["LlamaFixedWeightReward"]
128256
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
13,604
HFXM/Entropy_final_reward_model_71K-Llama3_1-8B-LR1e-5-Epoch2
llama
["LlamaFixedWeightReward"]
128256
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
13,605
HFXM/Entropy_final_reward_model_71K-Llama3_1-8B-LR2e-5-Epoch1
llama
["LlamaFixedWeightReward"]
128256
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
13,606
HFXM/Entropy_final_reward_model_71K-Llama3_1-8B-LR2e-5-Epoch2
llama
["LlamaFixedWeightReward"]
128256
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
13,607
HFXM/Entropy_final_reward_model_71K-Skywork-LR1e-5-Epoch1
llama
["LlamaFixedWeightReward"]
128256
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
13,608
HFXM/Entropy_final_reward_model_71K-Skywork-LR1e-5-Epoch2
llama
["LlamaFixedWeightReward"]
128256
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
13,609
HFXM/Entropy_final_reward_model_71K-Skywork-LR2e-5-Epoch1
llama
["LlamaFixedWeightReward"]
128256
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
13,610
HFXM/Entropy_final_reward_model_71K-Skywork-LR2e-5-Epoch2
llama
["LlamaFixedWeightReward"]
128256
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
13,611
HFXM/RAMO-RuleAdapter
llama
["LlamaForSequenceClassification"]
128256
float32
4.43.4
3,072
8,192
28
24
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
13,612
HFXM/RM_HHRLHF_Rule0
llama
["LlamaForSequenceClassification"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,613
HFXM/RM_HHRLHF_Rule0_Seed2025
llama
["LlamaForSequenceClassification"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,614
HFXM/RM_HHRLHF_Rule0_Seed2026
llama
["LlamaForSequenceClassification"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,615
HFXM/RM_HHRLHF_Rule0_Seed2027
llama
["LlamaForSequenceClassification"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,616
HFXM/RM_HHRLHF_Rule0_Seed2028
llama
["LlamaForSequenceClassification"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,617
HFXM/RM_HHRLHF_Rule1_Seed2025
llama
["LlamaForSequenceClassification"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,618
HFXM/RM_HHRLHF_Rule1_Seed2028
llama
["LlamaForSequenceClassification"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,619
HFXM/RM_HHRLHF_Rule1_Seed2029
llama
["LlamaForSequenceClassification"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,620
HFXM/RM_HHRLHF_Rule6_Seed2025
llama
["LlamaForSequenceClassification"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,621
HFXM/RM_HHRLHF_Rule6_Seed2026
llama
["LlamaForSequenceClassification"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,622
HFXM/RM_HHRLHF_Rule6_Seed2027
llama
["LlamaForSequenceClassification"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,623
HFXM/RM_HHRLHF_Rule6_Seed2028
llama
["LlamaForSequenceClassification"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,624
HFXM/RM_HHRLHF_Rule6_Seed2029
llama
["LlamaForSequenceClassification"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,625
HFXM/RM_HHRLHF_Rule7_Seed2025
llama
["LlamaForSequenceClassification"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,626
HFXM/RM_HHRLHF_Rule7_Seed2028
llama
["LlamaForSequenceClassification"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,627
HFXM/RuleAdapter-50K-2e-5-8epoch
llama
["LlamaForSequenceClassification"]
128256
float32
4.43.4
3,072
8,192
28
24
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
13,628
HFXM/git-base-pokemon
git
["GitForCausalLM"]
30522
float32
4.39.1
768
3,072
6
12
null
gelu
null
True
1,024
null
null
0.02
101
102
false
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.042467
small
short
13,629
HFXM/my_awesome_model
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.38.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
13,630
HFatimaZahra/AceGPT-13B-chat
llama
["LlamaForCausalLM"]
32000
float16
4.44.0
5,120
13,824
40
40
40
silu
0
True
2,048
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
13,631
HHJingbo/bert-finetuned-squad
bert
["BertForQuestionAnswering"]
28996
float32
4.41.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,632
HINT-lab/EAGLE-Llama3-8B-Instruct-Reproduce
llama
["LlamaForCausalLM"]
128256
float16
4.28.1
4,096
14,336
1
32
8
silu
null
True
2,048
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.201327
small
medium
13,633
HINT-lab/HASS-Llama3-8B-Instruct-Reproduce
llama
["LlamaForCausalLM"]
128256
float16
4.28.1
4,096
14,336
1
32
8
silu
null
True
2,048
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.201327
small
medium
13,634
HINT-lab/PosS1-Llama2-13B-Chat
llama
["LlamaForCausalLM"]
32000
float16
4.31.0.dev0
5,120
13,824
1
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.314573
small
medium
13,635
HINT-lab/PosS1-Llama3-8B-Instruct
llama
["LlamaForCausalLM"]
128256
float16
4.28.1
4,096
14,336
1
32
8
silu
null
True
2,048
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.201327
small
medium
13,636
HINT-lab/PosS2-Llama3-8B-Instruct
llama
["LlamaForCausalLM"]
128256
float16
4.28.1
4,096
14,336
1
32
8
silu
null
True
2,048
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.201327
small
medium
13,637
HINT-lab/PosS3-Llama3-8B-Instruct
llama
["LlamaForCausalLM"]
128256
float16
4.28.1
4,096
14,336
1
32
8
silu
null
True
2,048
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.201327
small
medium
13,638
HIT-SCIR/huozi3.5-ckpt-2
mixtral
["MixtralForCausalLM"]
57002
bfloat16
4.36.2
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
8
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
very_long
13,639
HIT-TMG/KaLM-embedding-multilingual-mini-instruct-v2
qwen2
["Qwen2Model"]
151936
bfloat16
4.45.0
896
4,864
24
14
2
silu
0
False
131,072
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
13,640
HIT-TMG/dialogue-bart-base-chinese
bart
["BartForConditionalGeneration"]
21128
float32
4.19.2
null
null
6
null
null
null
0.1
True
512
null
null
null
101
102
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
13,641
HIT-TMG/dialogue-bart-large-chinese
bart
["BartForConditionalGeneration"]
21128
float32
4.19.2
null
null
12
null
null
null
0.1
True
512
null
null
null
101
102
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
13,642
HIT-TMG/yizhao-fin-en-scorer
bert
["BertForSequenceClassification"]
30522
float32
4.40.1
384
1,536
6
12
null
gelu
null
True
512
null
null
0.02
null
null
null
32
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.010617
small
short
13,643
HIT-TMG/yizhao-fin-zh-scorer
bert
["BertForSequenceClassification"]
21128
float32
4.40.1
512
2,048
4
8
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.012583
small
short
13,644
HITTESH/Distillber_24_02_24
distilbert
["YourModelForTraining"]
28996
float32
4.35.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
13,645
HJOK/task2_deberta_spamMLM_v4
deberta-v2
["DebertaV2Model"]
12800
float32
4.25.1
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,646
HKUSTAudio/AudioX
diffusion_cond
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,647
HKUSTAudio/Llasa-1B
llama
["LlamaForCausalLM"]
193800
bfloat16
4.46.1
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
13,648
HKUSTAudio/Llasa-1B-Multilingual
llama
["LlamaForCausalLM"]
193800
bfloat16
4.46.1
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
13,649
HKUSTAudio/Llasa-1B-Preserve-TextChat
llama
["LlamaForCausalLM"]
193800
bfloat16
4.48.1
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
13,650
HKUSTAudio/Llasa-1B-two-speakers-kore-puck
llama
["LlamaForCausalLM"]
193800
bfloat16
4.48.3
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
13,651
HKUSTAudio/Llasa-3B-Preserve-TextChat
llama
["LlamaForCausalLM"]
193800
bfloat16
4.48.0
3,072
8,192
28
24
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
13,652
HKUSTAudio/Llasa-8B
llama
["LlamaForCausalLM"]
193800
bfloat16
4.46.1
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
13,653
HKUSTAudio/xcodec2
xcodec2
["XCodec2Model"]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,654
HKhillare/llama-2-7b-miniguanaco
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
13,655
HL23252/6825c0ca67e616931f48dc9c_model
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
13,656
HL23252/6864cfabd3975ae981ecad3b_model
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
13,657
HLXchamp/Qwen01
qwen2
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,658
HMC18/Gemma_7B_Vehicle_fine_tuned_gguf
gemma
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,659
HMC18/Mistral_7B_Vehicle_fine_tuned
mistral
["MistralForCausalLM"]
32000
bfloat16
4.49.0
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,660
HMC18/Mistral_7B_Vehicle_fine_tuned_gguf
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
13,661
HMC18/llama_8B_Vehicle_fine_tuned_gguf
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
13,662
HMDA/bert-finetuned-ner
bert
["BertForTokenClassification"]
28996
float32
4.30.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,663
HMHMlee/BioLinkBERT-base-finetuned-ner
bert
["BertForTokenClassification"]
28895
float32
4.20.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,664
HOHOANG/llama-3.2-3b-it-Ecommerce-ChatBot
llama
["LlamaForCausalLM"]
128258
float16
4.49.0
3,072
8,192
28
24
8
silu
0
True
131,072
500,000
0.00001
0.02
128256
128257
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
13,665
HPL/distilbert-base-uncased-finetuned-emotion
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.24.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
13,666
HRui-Face/albert-base-v1-semeval2017
albert
["AlbertForSequenceClassification"]
30000
float32
4.36.2
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
2
3
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,667
HRui-Face/bert-base-uncased-semeval2017
bert
["BertForSequenceClassification"]
30522
float32
4.36.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,668
HRui-Face/t5-base-semeval2017
t5
["T5ForSequenceClassification"]
32128
float32
4.38.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
13,669
HRuiii/LLaMA-3-8B-awq-4bit
llama
["LlamaForCausalLM"]
128256
float16
4.36.2
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,670
HRuiii/LLaMA-3-8B-awq-8bit
llama
["LlamaForCausalLM"]
128256
float16
4.36.2
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,671
HRuiii/Llama-2-7b-w4a16g128-omniquant
llama
["LlamaForCausalLM"]
32000
float16
4.37.2
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
13,672
HRuiii/Meta-Llama-3-8B-instruct-smoothquant-w8a8
llama
["LlamaForCausalLM"]
128256
float16
4.42.3
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
13,673
HSE-Chukchi-NLP/gemma3-1b-it-rus-ckt
gemma3_text
["Gemma3ForCausalLM"]
262144
float32
4.52.4
1,152
6,912
26
4
1
null
0
True
32,768
1,000,000
0.000001
0.02
2
[1, 106]
null
288
4
false
null
null
false
false
false
false
false
false
false
false
true
true
0.414056
small
very_long
13,674
HSE-Chukchi-NLP/gemma3-4b-it-rus-ckt
gemma3
["Gemma3ForConditionalGeneration"]
null
float32
4.52.4
null
null
null
null
null
null
null
null
null
null
null
0.02
null
[1, 106]
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
13,675
HSE-Chukchi-NLP/nllb-rus-ckt-v1
m2m_100
["M2M100ForConditionalGeneration"]
256205
float32
4.33.0
null
null
12
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
13,676
HTC410/qian1
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,677
HTThuanHcmus/bart-finetune-scientific-improve
bart
["BartForConditionalGeneration"]
50265
float32
4.44.2
null
null
6
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
13,678
HTThuanHcmus/bartpho-finetune
mbart
["MBartForConditionalGeneration"]
64001
float32
4.44.2
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
13,679
HTThuanHcmus/flan-t5-finetune-scientific-improve
t5
["T5ForConditionalGeneration"]
32128
float32
4.44.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
13,680
HTThuanHcmus/vit5-base-vietnews-summarization-finetune
t5
["T5ForConditionalGeneration"]
36096
float32
4.44.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
13,681
HThuat/distilbert-base-uncased-finetuned-cola
distilbert
["DistilBertForSequenceClassification"]
30522
null
4.41.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
13,682
HUBioDataLab/Prot2Mol
gpt2
["GPT2LMHeadModel"]
228
float32
4.44.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
13,683
HUGG222/Qwen-2.5-R1-Like
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.47.1
2,048
11,008
36
16
2
silu
0
True
45,048
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
13,684
HUGMLX/Croack_Merg
llama
["LlamaForCausalLM"]
128256
bfloat16
4.51.3
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
13,685
HUNTERDEBASTADOR/UTILITARIOS
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,686
HUSTHY/chinese_sentence_bert
bert
["SentenceBert"]
21128
float32
4.21.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,687
HUY2612/parler-tts-mini-1403
parler_tts
["ParlerTTSForConditionalGeneration"]
32128
float16
4.46.1
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
13,688
HUY2612/parler-tts-mini-1403-1
parler_tts
["ParlerTTSForConditionalGeneration"]
32128
float32
4.46.1
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
13,689
HUY2612/test1
null
["InvertTextNormalization"]
64001
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,690
HV-Khurdula/Dua-Vision-Base
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.44.2
null
null
null
null
null
null
null
null
null
null
null
null
null
2
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,691
HVD2407/mbart_paramiterized
mbart
["MBartForConditionalGeneration"]
250054
float32
4.41.1
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
13,692
HWJin/SMU-NLP-assignment2-finetuned-best
distilbert
["DistilBertForMaskedLM"]
28996
null
4.19.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
13,693
HYPJUDY/layoutlmv3-base-finetuned-funsd
layoutlmv3
["LayoutLMv3ForTokenClassification"]
50265
float32
4.12.5
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
13,694
HYPJUDY/layoutlmv3-base-finetuned-publaynet
layoutlmv3
[]
50265
float32
4.12.5
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
13,695
HYPJUDY/layoutlmv3-large-finetuned-funsd
layoutlmv3
["LayoutLMv3ForTokenClassification"]
50265
float32
4.12.5
1,024
4,096
24
16
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
short
13,696
HYUNAHKO/ITERATIVE_DPO_ITERATION0-merged
llama
["LlamaForCausalLM"]
128256
bfloat16
4.51.3
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
13,697
HYUNddd/shopping-emotion-model
roberta
["RobertaForSequenceClassification"]
32000
float32
4.50.3
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,698
Haaaaaaaaaax/dinov2-Base-finetuned-food101
dinov2
["Dinov2ForImageClassification"]
null
float32
4.51.3
768
null
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
13,699
Haaaaaaaaaax/google-vit-Base-finetuned-food101
vit
["ViTForImageClassification"]
null
float32
4.51.3
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null