index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
15,100
IAMRS23/resultsGPT
gpt2
["GPT2LMHeadModel"]
50257
float32
4.48.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
15,101
IAVYang/llama-2-7b-myfine
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
15,102
IAVYang/llama-2-7b-tagdetection_new
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
15,103
IAVYang/new_llama
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
15,104
IAVYang/new_llama-retry
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
15,105
IAmFlyingMonkey/pokemon_classifier
vit
["ViTForImageClassification"]
null
float32
4.38.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
15,106
IAmFlyingMonkey/roberta-base-model-korean-text-tag-classification
roberta
["RobertaForSequenceClassification"]
50265
float32
4.32.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,107
IAmNotAnanth/wav2vec2-large-xls-r-300m-turkish-colab
wav2vec2
["Wav2Vec2ForCTC"]
39
float32
4.28.1
1,024
4,096
24
16
null
gelu
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
15,108
IAmSkyDra/NutriCapy
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.42.4
2,048
16,384
18
8
1
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
15,109
IAmSkyDra/NutriCapy-quantized
gemma
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,110
IAmSudhirGupta/Llama-3.2-3B-Instruct-STA
llama
["LlamaForCausalLM"]
128256
bfloat16
4.45.0.dev0
3,072
8,192
28
24
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
15,111
IB13/my_awesome_billsum_model
t5
["T5ForConditionalGeneration"]
32128
float32
4.35.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
15,112
IBI-CAAI/MELT-llama-2-3x70b-chat-hf
mixtral
["MixtralForCausalLM"]
32000
bfloat16
4.37.0.dev0
8,192
28,672
80
64
8
silu
0
True
2,048
10,000
0.00001
0.02
1
2
false
128
8
false
3
2
false
false
false
false
false
false
false
true
true
true
64.424509
large
medium
15,113
ICAS03/fine-tune-llama
llama
["LlamaForCausalLM"]
32000
float16
4.40.1
2,048
5,632
22
32
4
silu
0
True
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
15,114
ICB-UMA/ICB-cef_distemist_Spanish_SapBERT_grandparents_sim
roberta
["RobertaForSequenceClassification"]
52000
float32
4.48.0.dev0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,115
ICB-UMA/ICB-cef_medprocner_Spanish_SapBERT_grandparents_bkg_1
roberta
["RobertaForSequenceClassification"]
52000
float32
4.48.0.dev0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,116
ICB-UMA/ICB-cef_medprocner_Spanish_SapBERT_grandparents_kg_1
roberta
["RobertaForSequenceClassification"]
52000
float32
4.48.0.dev0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,117
ICB-UMA/ICB-cef_symptemist_Spanish_SapBERT_grandparents_bkg_1
roberta
["RobertaForSequenceClassification"]
52000
float32
4.48.0.dev0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,118
ICB-UMA/ICB-cef_symptemist_Spanish_SapBERT_grandparents_kg_1
roberta
["RobertaForSequenceClassification"]
52000
float32
4.48.0.dev0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,119
ICEPVP8977/Uncensored_Phi_3_mini_128k_Safetensors_4b
phi3
["Phi3ForCausalLM"]
32064
float32
4.45.0.dev0
3,072
8,192
32
32
32
silu
0
False
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
15,120
ICFNext/EYY-Categorisation
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.17.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
15,121
ICLbioengNLP/CXR_BioClinicalBERT_chunkedv1
bert
["BertForMaskedLM"]
28996
float32
4.16.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,122
ICML2022/Tranception
tranception
["TranceptionLMHeadModel"]
25
float32
4.17.0
null
null
null
null
null
null
null
True
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,123
ICT2214Team7/Combined_model_v1
roberta
["RobertaForTokenClassification"]
50265
float32
4.40.2
768
3,072
6
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.042467
small
short
15,124
ICT2214Team7/RoBERTa_Combined_Generated_v2_1000
roberta
["RobertaForTokenClassification"]
50265
float32
4.40.2
768
3,072
6
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.042467
small
short
15,125
ICT2214Team7/RoBERTa_Combined_Generated_v2_2000
roberta
["RobertaForTokenClassification"]
50265
float32
4.40.2
768
3,072
6
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.042467
small
short
15,126
ICT3214-Group5/Cryptography_GPT_2_v1.0.0
gpt2
["GPT2LMHeadModel"]
50257
float32
4.46.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
15,127
ICT3214-Group5/Cryptography_GPT_NEO_v1.0.0
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.46.2
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
15,128
ICT3214-Group5/MD5_gpt_neo
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.46.0
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
15,129
ICT3214-Group5/MD5_gpt_neo_v1.1
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.46.1
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
15,130
ICT3214-Group5/MD5_gpt_neo_v1.2
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.46.1
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
15,131
ICT3214-Group5/MD5_gpt_neo_v1.4
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.46.1
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
15,132
ICT3214-Group5/MD5_gpt_neo_v1.5
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.46.1
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
15,133
ICT3214-Group5/MD5_gpt_neo_v3.0.4_aes
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.46.2
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
15,134
ICT3214-Group5/MD5_gpt_neo_v3.1.1
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.46.2
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
15,135
ICT3214-Group5/MD5_gpt_neo_v3.1.2
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.46.2
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
15,136
ICT3214-Group5/MD5_gpt_neo_v3.1.3
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.46.2
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
15,137
ICT3214-Group5/MD5_gpt_neo_v3.1.4
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.46.2
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
15,138
ICT3214-Group5/MD5_gpt_neo_v3.1.5
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.46.2
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
15,139
ICT3214-Group5/MD5_gpt_neo_v3.1.6
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.46.2
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
15,140
ICTNLP/Auto-RAG-Llama-3-8B-Instruct
llama
["LlamaForCausalLM"]
128256
float32
4.43.3
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,141
ICTNLP/Llama-3.1-8B-Omni
omni_speech2s_llama
["OmniSpeech2SLlamaForCausalLM"]
128256
float16
4.43.4
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
15,142
ICTNLP/stream-omni-8b
stream_omni_llama
["StreamOmniLlamaForCausalLM"]
132352
float16
4.43.1
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
15,143
ICW11/icw1
internlm2
["InternLM2ForCausalLM"]
92544
bfloat16
4.41.0
2,048
8,192
24
16
8
silu
null
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
2
false
null
null
false
false
false
false
false
false
false
false
true
true
1.20796
medium
very_long
15,144
ICW11/intern_study_L0_4
internlm2
["InternLM2ForCausalLM"]
92544
bfloat16
4.41.0
4,096
14,336
32
32
8
silu
null
True
262,144
50,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
15,145
ICavor/distilhubert-finetuned-gtzan
hubert
["HubertForSequenceClassification"]
32
float32
4.46.3
768
3,072
2
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.014156
small
null
15,146
IDEA-CCNL/Erlangshen-DeBERTa-v2-320M-Chinese
deberta-v2
["DebertaV2ForMaskedLM"]
12800
null
null
1,024
4,096
24
16
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
15,147
IDEA-CCNL/Erlangshen-Longformer-330M
longformer
["LongformerForPreTraining"]
12000
null
null
1,024
4,096
24
16
null
gelu
null
null
4,096
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
medium
15,148
IDEA-CCNL/Erlangshen-Roberta-110M-NLI
bert
["BertForSequenceClassification"]
21128
null
4.20.0
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,149
IDEA-CCNL/Erlangshen-Roberta-110M-Similarity
bert
["BertForSequenceClassification"]
21128
null
4.20.0
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,150
IDEA-CCNL/Erlangshen-Roberta-330M-Causal-Chinese
bert
["BertForSequenceClassification"]
21128
null
4.20.0
1,024
4,096
24
16
null
gelu
null
null
512
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
15,151
IDEA-CCNL/Erlangshen-Roberta-330M-NLI
bert
["BertForSequenceClassification"]
21128
null
4.20.0
1,024
4,096
24
16
null
gelu
null
null
512
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
15,152
IDEA-CCNL/Erlangshen-Roberta-330M-Sentiment
bert
["BertForSequenceClassification"]
21128
null
4.20.0
1,024
4,096
24
16
null
gelu
null
null
512
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
15,153
IDEA-CCNL/Erlangshen-Roberta-330M-Similarity
bert
["BertForSequenceClassification"]
21128
null
4.20.0
1,024
4,096
24
16
null
gelu
null
null
512
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
15,154
IDEA-CCNL/Erlangshen-SimCSE-110M-Chinese
bert
["BertModel"]
21128
null
4.2.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,155
IDEA-CCNL/Erlangshen-UniMC-DeBERTa-v2-1.4B-Chinese
deberta-v2
[]
12800
null
null
1,536
6,144
48
24
null
gelu
null
null
512
null
null
0.01
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
1.358954
medium
short
15,156
IDEA-CCNL/Erlangshen-ZEN1-224M-Chinese
null
[]
21128
null
null
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
15,157
IDEA-CCNL/Erlangshen-ZEN2-345M-Chinese
null
[]
21128
null
null
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
15,158
IDEA-CCNL/Erlangshen-ZEN2-668M-Chinese
null
[]
21128
null
null
1,024
4,096
24
16
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
short
15,159
IDEA-CCNL/Randeng-DAVAE-1.2B-General-Chinese
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,160
IDEA-CCNL/Randeng-GAVAE-1.2B-Augmentation-Chinese
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,161
IDEA-CCNL/Randeng-PPVAE-1.2B-Augmentation-Chinese
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,162
IDEA-CCNL/Randeng-Pegasus-523M-Chinese
pegasus
["PegasusForConditionalGeneration"]
50000
float16
4.10.2
null
null
16
null
null
null
0.1
True
1,024
null
null
null
0
1
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
15,163
IDEA-CCNL/Randeng-T5-77M
mt5
["MT5ForConditionalGeneration"]
32596
float16
4.12.5
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
15,164
IDEA-CCNL/Randeng-T5-784M
mt5
["MT5ForConditionalGeneration"]
32596
float16
4.12.5
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
15,165
IDEA-CCNL/Randeng-TransformerXL-5B-Abduction-Chinese
null
[]
50176
null
null
3,072
null
null
48
null
null
null
null
null
null
null
null
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,166
IDEA-CCNL/Randeng-TransformerXL-5B-Deduction-Chinese
null
[]
50176
null
null
3,072
null
null
48
null
null
null
null
null
null
null
null
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,167
IDEA-CCNL/Taiyi-CLIP-RoBERTa-326M-ViT-H-Chinese
bert
["BertModel"]
21128
float32
4.22.1
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
15,168
IDEA-CCNL/Taiyi-Roberta-124M-D
roberta
["RobertaForMaskedLM"]
50265
float32
4.17.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,169
IDEA-CCNL/Wenzhong2.0-GPT2-3.5B-chinese
gpt2
["GPT2LMHeadModel"]
50304
null
4.18.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
15,170
IDEA-CCNL/Ziya-Visual-14B-Chat
blip-2
["InstructDittoLMForConditionalGeneration"]
null
float16
null
null
null
null
null
null
null
null
null
null
null
null
0.02
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,171
IDEA-Emdoor/UniTTS-mixed-v0.1
qwen2
["Qwen2Model"]
184840
float32
4.48.3
3,584
18,944
28
28
4
silu
0
True
8,192
1,000,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
long
15,172
IDEA-FinAI/Finance-R1-7B
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.49.0.dev0
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
15,173
IDEA-Research/ChatRex-7B
chatrex
["ChatRexAuxForConditionalGeneration"]
null
float32
4.44.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,174
IDEA-Research/RexSeek-3B
rexseek_qwen
["RexSeekQwenForCausalLM"]
151769
bfloat16
4.48.0
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
15,175
IDEA-Research/grounding-dino-base
grounding-dino
["GroundingDinoForObjectDetection"]
null
float32
4.40.0.dev0
null
null
null
null
null
null
0
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,176
IDEA-Research/grounding-dino-tiny
grounding-dino
["GroundingDinoForObjectDetection"]
null
float32
4.40.0.dev0
null
null
null
null
null
null
0
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,177
IDL/autotrain-qna-1170143354
distilbert
["DistilBertForQuestionAnswering"]
30522
float32
4.20.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
15,178
IDQO/gis-parent-category-doc-description
roberta
["RobertaForSequenceClassification"]
40000
float32
4.28.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,179
IDQO/gis-parent-category-undersampling
roberta
["RobertaForSequenceClassification"]
40000
float32
4.28.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,180
IEETA/BioNExt-Extractor
relation-novelty-extractor
["BioNExtExtractorModel"]
28899
float32
4.37.2
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
short
15,181
IEETA/BioNExt-Tagger
crf-tagger
["BioNExtTaggerModel"]
28895
float32
4.37.2
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
short
15,182
IEITYuan/Yuan2-2B-Mars-hf
yuan
["YuanForCausalLM"]
135040
bfloat16
4.30.0.dev0
2,048
8,192
24
32
null
silu
null
False
8,192
null
0.000001
0.02
77185
77185
true
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
1.20796
medium
long
15,183
IEITYuan/Yuan2-M32
yuan
["YuanForCausalLM"]
135040
bfloat16
4.30.0.dev0
2,048
8,192
24
16
null
silu
null
True
4,096
null
0.000001
0.02
77185
77185
true
128
null
false
null
null
false
false
false
false
false
false
false
false
false
false
1.20796
medium
medium
15,184
IEITYuan/Yuan2-M32-hf
yuan
["YuanForCausalLM"]
135040
bfloat16
4.30.2
2,048
8,192
24
16
null
silu
null
True
4,096
null
0.000001
0.02
77185
77185
null
128
null
false
null
null
false
false
false
false
false
false
false
false
false
false
1.20796
medium
medium
15,185
IELTS8/videojepa_finetune_v4
VideoGPT+
["Phi3ForCausalLM"]
32064
bfloat16
4.46.3
3,072
8,192
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
true
false
false
false
false
false
false
true
3.623879
medium
medium
15,186
IGustavsen/t5-small-finetuned-english-wikilingua-finetuned-english-wikilingua
t5
["T5ForConditionalGeneration"]
32128
null
4.29.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
15,187
IHaBiS/Nethena-MLewd-Xwin-23B-exl2
llama
["LlamaForCausalLM"]
32000
float16
4.34.1
5,120
13,824
71
40
40
silu
null
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
22.334669
large
medium
15,188
IHaBiS/StellarBright-2.55bpw-h8-exl2
llama
["LlamaForCausalLM"]
32000
float32
4.34.0
8,192
28,672
80
64
8
silu
null
False
4,096
10,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
medium
15,189
IHaBiS/Synatra-7B-v0.3-RP-exl2
mistral
["MistralForCausalLM"]
32002
float16
4.34.1
4,096
14,336
32
32
8
silu
null
False
32,768
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
15,190
IHaBiS/Synatra-7B-v0.3-base-exl2
mistral
["MistralForCausalLM"]
32002
bfloat16
4.34.1
4,096
14,336
32
32
8
silu
null
False
32,768
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
15,191
IHaBiS/gemma3_4b_it_mrl_sib200_merged
gemma3
["Gemma3ForConditionalGeneration"]
null
bfloat16
4.51.3
null
null
null
null
null
null
null
null
null
null
null
0.02
null
[1, 106]
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
15,192
II-Vietnam/R1-Math-Code-Fusion
qwen2
["Qwen2ForCausalLM"]
152064
float32
4.50.3
3,584
18,944
28
28
4
silu
0
False
32,768
300,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
15,193
II-Vietnam/R1-Math-Code-Fusion-v2
qwen2
["Qwen2ForCausalLM"]
152064
float32
4.50.3
3,584
18,944
28
28
4
silu
0
False
32,768
300,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
15,194
IIC/BETO_Galen-socialdisner
bert
["BertForSequenceClassification"]
31002
float32
4.25.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,195
IIC/RigoBERTa-Clinical
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.46.1
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
15,196
IIC/RigoChat-7b-v2
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.45.2
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
15,197
IIC/XLM_R_Galen-livingner3
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.25.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,198
IIC/XLM_R_Galen-socialdisner
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.25.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,199
IIC/bert-base-spanish-wwm-cased-livingner3
bert
["BertForSequenceClassification"]
31002
float32
4.25.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short