index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
2,700
AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-REGION_earlystop_70_15_15
esm
["EsmForTokenClassification"]
33
float32
4.46.2
320
1,280
6
20
null
gelu
null
True
1,026
null
null
0.02
null
null
null
16
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.007373
small
short
2,701
AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-REPEAT_earlystop_70_15_15
esm
["EsmForTokenClassification"]
33
float32
4.46.2
320
1,280
6
20
null
gelu
null
True
1,026
null
null
0.02
null
null
null
16
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.007373
small
short
2,702
AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-SIGNAL_earlystop_70_15_15
esm
["EsmForTokenClassification"]
33
float32
4.46.2
320
1,280
6
20
null
gelu
null
True
1,026
null
null
0.02
null
null
null
16
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.007373
small
short
2,703
AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-STRAND_earlystop_70_15_15
esm
["EsmForTokenClassification"]
33
float32
4.46.2
320
1,280
6
20
null
gelu
null
True
1,026
null
null
0.02
null
null
null
16
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.007373
small
short
2,704
AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-TOPO_DOM_earlystop_70_15_15
esm
["EsmForTokenClassification"]
33
float32
4.46.2
320
1,280
6
20
null
gelu
null
True
1,026
null
null
0.02
null
null
null
16
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.007373
small
short
2,705
AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-TRANSIT_earlystop_70_15_15
esm
["EsmForTokenClassification"]
33
float32
4.46.2
320
1,280
6
20
null
gelu
null
True
1,026
null
null
0.02
null
null
null
16
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.007373
small
short
2,706
AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-TRANSMEM_earlystop_70_15_15
esm
["EsmForTokenClassification"]
33
float32
4.46.2
320
1,280
6
20
null
gelu
null
True
1,026
null
null
0.02
null
null
null
16
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.007373
small
short
2,707
AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-TURN_earlystop_70_15_15
esm
["EsmForTokenClassification"]
33
float32
4.46.2
320
1,280
6
20
null
gelu
null
True
1,026
null
null
0.02
null
null
null
16
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.007373
small
short
2,708
AliSaadatV/esm2_t6_8M_UR50D-finetunedv2-ZN_FING_earlystop_70_15_15
esm
["EsmForTokenClassification"]
33
float32
4.46.2
320
1,280
6
20
null
gelu
null
True
1,026
null
null
0.02
null
null
null
16
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.007373
small
short
2,709
AliSab/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,710
AliSouliman/PPO-LL2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,711
AliSouliman/a2c-PandaReachDense-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,712
Alias1964/Llama-3.1-70B-Instruct-lorablated-Q4-EXL2
llama
["LlamaForCausalLM"]
128256
bfloat16
4.43.3
8,192
28,672
80
64
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
2,713
Alibaba-NLP/Simulation_LLM_google_14B_V1
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.44.2
5,120
13,824
48
40
8
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
15.099494
large
very_long
2,714
Alibaba-NLP/ZeroSearch_google_v1_Qwen2.5_3B
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.48.3
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
null
151643
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
2,715
Alibaba-NLP/ZeroSearch_wiki_V2_Llama_3.2_3B
llama
["LlamaForCausalLM"]
128256
float32
4.48.3
3,072
8,192
28
24
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
2,716
Alibaba-NLP/ZeroSearch_wiki_V2_Qwen2.5_3B
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.48.3
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
null
151643
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
2,717
Alibaba-NLP/gme-Qwen2-VL-2B-Instruct
qwen2_vl
["Qwen2VLForConditionalGeneration", "GmeQwen2VL"]
151936
float32
4.45.0.dev0
1,536
8,960
28
12
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
2,718
Alibaba-NLP/gte-Qwen2-7B-instruct
qwen2
["Qwen2ForCausalLM"]
151646
float32
4.41.2
3,584
18,944
28
28
4
silu
0
True
131,072
1,000,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
2,719
Alibaba-NLP/gte-base-en-v1.5
new
["NewModel"]
30528
float32
4.39.1
768
3,072
12
12
null
gelu
null
null
8,192
500,000
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
true
0.084935
small
long
2,720
Alibaba-NLP/gte-large-en-v1.5
new
["NewModel"]
30528
float32
4.39.1
1,024
4,096
24
16
null
gelu
null
null
8,192
160,000
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
true
0.30199
small
long
2,721
Alibaba-NLP/gte-modernbert-base
modernbert
["ModernBertModel"]
50368
float16
4.48.0.dev0
768
1,152
22
12
null
null
0
null
8,192
null
null
0.02
50281
50282
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
true
0.155714
small
long
2,722
Alidr79/speecht5_v1_20000_steps
speecht5
["SpeechT5ForTextToSpeech"]
81
float32
4.43.3
768
null
null
null
null
gelu
0.1
False
null
null
null
0.02
0
2
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
2,723
Alidr79/speecht5_v2_4000_best
speecht5
["SpeechT5ForTextToSpeech"]
81
float32
4.43.3
768
null
null
null
null
gelu
0.1
False
null
null
null
0.02
0
2
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
2,724
Aliem2001/whisper-small-fa
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.48.2
null
null
12
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
2,725
AlienKevin/electra_hongkongese_small_pos_hkcancor
electra
["ElectraForTokenClassification"]
30000
float32
4.43.3
256
1,024
12
4
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.009437
small
short
2,726
AlignCoder/AlignRetriever
roberta
["RobertaModel"]
51416
float32
4.49.0
768
3,072
12
12
null
gelu
null
True
1,026
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,727
AlignQuant/Llama-2-13b-chat-hf-GPTQ-2bit
llama
["LlamaForCausalLM"]
32000
float16
4.50.2
5,120
13,824
40
40
40
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
2,728
AlignQuant/Llama-2-7b-chat-hf-OSTQuant
llama
["LlamaForCausalLM"]
32000
bfloat16
4.44.2
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,729
AlignQuant/Meta-Llama-3-8B-Instruct-GPTQ-2bit
llama
["LlamaForCausalLM"]
128256
float16
4.50.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,730
AlignQuant/Meta-Llama-3-8B-Instruct-GPTQ-4bit
llama
["LlamaForCausalLM"]
128256
float16
4.50.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,731
AlignQuant/Meta-Llama-3-8B-Instruct-bitsandbytes-8bit
llama
["LlamaForCausalLM"]
128256
float16
4.50.0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,732
Alignment-Lab-AI/Qwen2.5-Coder-7B-Instruct-132k
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.44.0
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
2,733
Alignment-Lab-AI/e5-mistral-7b-instruct
mistral
["MistralModel"]
32000
float16
4.34.0
4,096
14,336
32
32
8
silu
null
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,734
AliiAhmadi/whisper-fa
whisper
["WhisperForConditionalGeneration"]
51865
float16
4.25.1
null
null
32
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
2,735
Alimahmoud/custom-tuned-model
distilbert
["DistilBertForSequenceClassification"]
30522
null
4.56.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,736
Alimahmoud/custom-tuned-model2
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.46.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,737
AlinaKl/babylm2024-git-txt
git
["GitForCausalLM"]
32778
float32
4.26.0
768
3,072
12
12
null
gelu
null
True
1,024
null
null
0.02
101
102
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
2,738
AlinaKl/babylm2024-git-vision
git
["GitForCausalLM"]
32778
float32
4.26.0
768
3,072
12
12
null
gelu
null
True
1,024
null
null
0.02
101
102
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
2,739
Alindstroem89/sentiment_model_testing
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.32.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,740
Alinka9/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-singing_arctic_elk
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.51.3
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
2,741
Alipour97/segformer-b0-scene-parse-150
segformer
["SegformerForSemanticSegmentation"]
null
float32
4.25.1
null
null
null
[1, 2, 5, 8]
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
2,742
Alireza0017/marian-finetuned-kde4-en-to-fr
marian
["MarianMTModel"]
59514
float32
4.51.3
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
2,743
AlirezaBaneshi/autotrain-test2-756523213
roberta
["RobertaForQuestionAnswering"]
50265
float32
4.15.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,744
AlirezaBaneshi/autotrain-test2-756523214
roberta
["RobertaForQuestionAnswering"]
50265
float32
4.15.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,745
Alirezaab78/test
llama
["LlamaForCausalLM"]
128256
float16
4.47.1
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,746
Alirezamp/gemma2b-news-gategory-lc_awq
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.42.3
2,048
16,384
18
8
1
gelu
0
False
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
2,747
AlisaTarada/distilbert-base-uncased-finetuned-emotion
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.43.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,748
Alisaqib123/rare-puppers
vit
["ViTForImageClassification"]
null
float32
4.41.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
2,749
AlisonWen/ppo-baseline-print-token-prob
llama
["LlamaForCausalLM"]
32001
float16
4.51.3
4,096
11,008
32
32
32
silu
0
True
2,048
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,750
AlisonWen/ppo-baseline-print-token-prob-1
llama
["LlamaForCausalLM"]
32001
float16
4.51.3
4,096
11,008
32
32
32
silu
0
True
2,048
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,751
AlisonWen/ppo-seed-42-step-200
llama
["LlamaForCausalLM"]
32001
float16
4.45.2
4,096
11,008
32
32
32
silu
0
True
2,048
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,752
AlisonWen/rm-seed-0
llama
["LlamaForScore"]
32001
bfloat16
4.44.1
4,096
11,008
32
32
32
silu
0
True
2,048
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,753
Alissonerdx/Dia1.6-pt_BR-v1
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
2,754
Alissonerdx/YuE-s1-7B-anneal-en-cot-exl2-3.0bpw
llama
["LlamaForCausalLM"]
83968
bfloat16
4.42.0
4,096
11,008
32
32
4
silu
0
True
16,384
10,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,755
Alissonerdx/YuE-s1-7B-anneal-en-cot-exl2-4.0bpw
llama
["LlamaForCausalLM"]
83968
bfloat16
4.42.0
4,096
11,008
32
32
4
silu
0
True
16,384
10,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,756
Alissonerdx/YuE-s1-7B-anneal-en-cot-exl2-5.0bpw
llama
["LlamaForCausalLM"]
83968
bfloat16
4.42.0
4,096
11,008
32
32
4
silu
0
True
16,384
10,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,757
Alissonerdx/YuE-s1-7B-anneal-en-cot-exl2-6.0bpw
llama
["LlamaForCausalLM"]
83968
bfloat16
4.42.0
4,096
11,008
32
32
4
silu
0
True
16,384
10,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,758
Alissonerdx/YuE-s1-7B-anneal-en-cot-exl2-8.0bpw
llama
["LlamaForCausalLM"]
83968
bfloat16
4.42.0
4,096
11,008
32
32
4
silu
0
True
16,384
10,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,759
Alissonerdx/YuE-s1-7B-anneal-en-cot-int8
llama
["LlamaForCausalLM"]
83968
float16
4.48.1
4,096
11,008
32
32
4
silu
0
True
16,384
10,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,760
Alissonerdx/YuE-s1-7B-anneal-jp-kr-icl-int8
llama
["LlamaForCausalLM"]
83968
float16
4.48.1
4,096
11,008
32
32
4
silu
0
True
16,384
10,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,761
Alissonerdx/YuE-s1-7B-anneal-zh-cot-int8
llama
["LlamaForCausalLM"]
83968
float16
4.48.1
4,096
11,008
32
32
4
silu
0
True
16,384
10,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,762
Alissonerdx/YuE-s2-1B-general-exl2-3.0bpw
llama
["LlamaForCausalLM"]
83840
bfloat16
4.42.0
2,048
5,504
32
16
16
silu
0
True
8,192
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.610613
medium
long
2,763
Alissonerdx/YuE-s2-1B-general-exl2-4.0bpw
llama
["LlamaForCausalLM"]
83840
bfloat16
4.42.0
2,048
5,504
32
16
16
silu
0
True
8,192
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.610613
medium
long
2,764
Alissonerdx/YuE-s2-1B-general-exl2-5.0bpw
llama
["LlamaForCausalLM"]
83840
bfloat16
4.42.0
2,048
5,504
32
16
16
silu
0
True
8,192
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.610613
medium
long
2,765
Alissonerdx/YuE-s2-1B-general-exl2-6.0bpw
llama
["LlamaForCausalLM"]
83840
bfloat16
4.42.0
2,048
5,504
32
16
16
silu
0
True
8,192
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.610613
medium
long
2,766
Alissonerdx/YuE-s2-1B-general-exl2-8.0bpw
llama
["LlamaForCausalLM"]
83840
bfloat16
4.42.0
2,048
5,504
32
16
16
silu
0
True
8,192
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.610613
medium
long
2,767
Alissonerdx/YuE-s2-1B-general-int8
llama
["LlamaForCausalLM"]
83840
float16
4.48.1
2,048
5,504
32
16
16
silu
0
True
8,192
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.610613
medium
long
2,768
Alitrix/bert-base-uncased_Banking77
bert
["BertForSequenceClassification"]
119547
float32
4.37.0.dev0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,769
Aliyyah/Roberta-Sentiment-Classifier
roberta
["RobertaForSequenceClassification"]
50265
float32
4.35.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,770
AlkQ/a2c-PandaPickAndPlace-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
2,771
AlkQ/a2c-PandaReachDense-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
2,772
AlkanCan/BERTurk-128k-cased-immigration
bert
["BertForSequenceClassification"]
128000
float32
4.42.4
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,773
AlkanCan/TurkishBERTweet-Immigration-Stance
roberta
["RobertaForSequenceClassification"]
100000
float32
4.42.4
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,774
Alkohole/T-pro-it-1.0-4.5bpw-h8-exl2
qwen2
["Qwen2ForCausalLM"]
151665
bfloat16
4.46.2
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
2,775
All-SeeingEye/speecht5_finetuned_voxpopuli_es
speecht5
["SpeechT5ForTextToSpeech"]
81
float32
4.46.2
768
null
null
null
null
gelu
0.1
False
null
null
null
0.02
0
2
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
2,776
AllaNBarakat/xlm-roberta-base-fintuned-panx-de
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.25.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,777
Allen-UQ/DeepSeek-R1-Distill-Qwen-1.5B-GRPO
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.49.0
1,536
8,960
28
12
2
silu
0
False
131,072
10,000
0.000001
0.02
151643
151643
false
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
2,778
Allen-UQ/Qwen2.5-1.5B-Instruct-SFT-2-Hop-Nei-Aug-Pubmed
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.49.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
2,779
Allen-UQ/Qwen2.5-7B-Instruct-GRPO-Nei-Tokens-k20
qwen2
["Qwen2ForCausalLMNode"]
152064
float32
4.49.0
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
2,780
Allen1984/hw1
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.44.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,781
AllenGeng/OCamlBert
bert
["BertForMaskedLM"]
30522
float32
4.20.0.dev0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,782
AllenMai/bert-chinese-ainews
bert
["BertForSequenceClassification"]
21128
float32
4.31.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,783
AllenMai/bert-test1
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.31.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,784
Allrandom/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,785
AllyArc/tinyLlama_allyar
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,786
Almahmood/whisper-medium-checkpoint
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.30.0.dev0
null
null
24
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
2,787
Almahmood/whisper-small-hi
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.30.0.dev0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
2,788
Almancy/finetuning-emotion-model-5-v3
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.41.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,789
Almenodera/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-keen_woolly_chimpanzee
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.50.3
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
2,790
Almheiri/Llama-3.2-1B-Instruct-GPTQ-INT4
llama
["LlamaForCausalLM"]
128256
float16
4.46.3
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
2,791
AlmightyDeathCheater/DialoGPT-medium-harrypotter
gpt2
["GPT2LMHeadModel"]
50257
float32
4.25.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
2,792
Almondpeanuts/pegasus-samsum
pegasus
["PegasusForConditionalGeneration"]
96103
float32
4.28.1
null
null
16
null
null
null
0.1
True
1,024
null
null
null
0
1
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
2,793
Almondpeanuts/xlm-roberta-base-finetuned-panx-all
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.28.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,794
Almondpeanuts/xlm-roberta-base-finetuned-panx-de-fr
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.28.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,795
Almondpeanuts/xlm-roberta-base-finetuned-panx-en
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.28.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,796
Almondpeanuts/xlm-roberta-base-finetuned-panx-fr
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.28.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,797
Almondpeanuts/xlm-roberta-base-finetuned-panx-it
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.28.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,798
AlokBharadwaj/sd-class-butterflies-32_ddpm_pipeline_with_mae_loss
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,799
AlokBharadwaj/sd-class-butterflies-32_ddpm_pipeline_with_smooth_l1_loss
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null