index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
6,500
CMU-AIR2/code-ArithHardC12Interm-240416
llama
["LlamaForCausalLM"]
32256
float32
4.38.2
2,048
5,504
24
16
16
silu
0
True
16,384
100,000
0.000001
0.02
32013
32021
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
long
6,501
CMU-AIR2/code-ArithHardC12Interm-gh-240416
llama
["LlamaForCausalLM"]
32256
float32
4.38.2
2,048
5,504
24
16
16
silu
0
True
16,384
100,000
0.000001
0.02
32013
32021
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
long
6,502
CMU-AIR2/code-ArithHardC12Interm-gh-mixture-240416
llama
["LlamaForCausalLM"]
32256
float32
4.38.2
2,048
5,504
24
16
16
silu
0
True
16,384
100,000
0.000001
0.02
32013
32021
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
long
6,503
CMU-AIR2/code-ctrl-gh-mixture-240419
llama
["LlamaForCausalLM"]
32256
float32
4.38.2
2,048
5,504
24
16
16
silu
0
True
16,384
100,000
0.000001
0.02
32013
32021
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
long
6,504
CMU-AIR2/code-full-hard-gh-mixture
llama
["LlamaForCausalLM"]
32256
float32
4.38.2
2,048
5,504
24
16
16
silu
0
True
16,384
100,000
0.000001
0.02
32013
32021
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
long
6,505
CMU-AIR2/code-full-simple
llama
["LlamaForCausalLM"]
32256
float32
4.38.2
2,048
5,504
24
16
16
silu
0
True
16,384
100,000
0.000001
0.02
32013
32021
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
long
6,506
CMU-AIR2/math-deepseek-FULL-ArithHard-MixedMWP
llama
["LlamaForCausalLM"]
32256
bfloat16
4.38.2
2,048
5,504
24
16
16
silu
0
True
16,384
100,000
0.000001
0.02
32013
32021
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
long
6,507
CMU-AIR2/math-deepseek-LORA-ArithHard
llama
["LlamaForCausalLM"]
32256
bfloat16
4.38.2
2,048
5,504
24
16
16
silu
0
True
16,384
100,000
0.000001
0.02
32013
32021
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
long
6,508
CMU-AIR2/math-deepseek-LORA-ArithHard-MixedMWP
llama
["LlamaForCausalLM"]
32256
bfloat16
4.38.2
2,048
5,504
24
16
16
silu
0
True
16,384
100,000
0.000001
0.02
32013
32021
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
long
6,509
CMU-AIR2/math-deepseek-LORA-ArithHardC12
llama
["LlamaForCausalLM"]
32256
bfloat16
4.38.2
2,048
5,504
24
16
16
silu
0
True
16,384
100,000
0.000001
0.02
32013
32021
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
long
6,510
CMU-AIR2/math-deepseek-baseline-FTMWP-LORA
llama
["LlamaForCausalLM"]
32256
bfloat16
4.38.2
2,048
5,504
24
16
16
silu
0
True
16,384
100,000
0.000001
0.02
32013
32021
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
long
6,511
CMU-AIR2/math-deepseek-lora-arith-curriculum-per-subject
llama
["LlamaForCausalLM"]
32256
bfloat16
4.33.1
2,048
5,504
24
16
16
silu
null
True
16,384
100,000
0.000001
0.02
32013
32021
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
long
6,512
CMU-AIR2/math-deepseek_FULL_HardArith_Interm
llama
["LlamaForCausalLM"]
32256
bfloat16
4.38.2
2,048
5,504
24
16
16
silu
0
True
16,384
100,000
0.000001
0.02
32013
32021
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
long
6,513
CMU-AIR2/math-llama-3-instruct-LORA-ArithSteps-6K
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0.dev0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
6,514
CMU-AIR2/math-llama_3_instruct-model-arith-4k
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0.dev0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
6,515
CMU-AIR2/math-phi-1-5-FULL-ArithHard-lr-1.2e-6
phi
["PhiForCausalLM"]
51200
bfloat16
4.38.2
2,048
8,192
24
32
32
gelu_new
0
True
2,048
10,000
null
0.02
null
null
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
6,516
CMU-AIR2/math-phi-1-5-FULL-ArithHard-lr-1.5e-6
phi
["PhiForCausalLM"]
51200
bfloat16
4.38.2
2,048
8,192
24
32
32
gelu_new
0
True
2,048
10,000
null
0.02
null
null
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
6,517
CMU-AIR2/math-phi-1-5-FULL-ArithHard-lr-1e-6
phi
["PhiForCausalLM"]
51200
bfloat16
4.38.2
2,048
8,192
24
32
32
gelu_new
0
True
2,048
10,000
null
0.02
null
null
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
6,518
CMU-AIR2/math-phi-1-5-FULL-ArithHard-lr-8e-7
phi
["PhiForCausalLM"]
51200
bfloat16
4.38.2
2,048
8,192
24
32
32
gelu_new
0
True
2,048
10,000
null
0.02
null
null
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
6,519
CMU-AIR2/math-phi-1-5-FULL-Arithmetic-Steps-10k
phi
["PhiForCausalLM"]
51200
bfloat16
4.38.2
2,048
8,192
24
32
32
gelu_new
0
True
2,048
10,000
null
0.02
null
null
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
6,520
CMU-AIR2/math-phi-1-5-FULL-Arithmetic-Steps-2k
phi
["PhiForCausalLM"]
51200
bfloat16
4.38.2
2,048
8,192
24
32
32
gelu_new
0
True
2,048
10,000
null
0.02
null
null
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
6,521
CMU-AIR2/math-phi-1-5-FULL-Arithmetic-Steps-4k
phi
["PhiForCausalLM"]
51200
bfloat16
4.38.2
2,048
8,192
24
32
32
gelu_new
0
True
2,048
10,000
null
0.02
null
null
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
6,522
CMU-AIR2/math-phi-1-5-FULL-Arithmetic-Steps-512-4k
phi
["PhiForCausalLM"]
51200
bfloat16
4.38.2
2,048
8,192
24
32
32
gelu_new
0
True
2,048
10,000
null
0.02
null
null
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
6,523
CMU-AIR2/math-phi-1-5-FULL-Arithmetic-Steps-512-6k
phi
["PhiForCausalLM"]
51200
bfloat16
4.38.2
2,048
8,192
24
32
32
gelu_new
0
True
2,048
10,000
null
0.02
null
null
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
6,524
CMU-AIR2/math-phi-1-5-FULL-Arithmetic-Steps-512-8k
phi
["PhiForCausalLM"]
51200
bfloat16
4.38.2
2,048
8,192
24
32
32
gelu_new
0
True
2,048
10,000
null
0.02
null
null
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
6,525
CMU-AIR2/math-phi-1-5-FULL-Arithmetic-Steps-6k
phi
["PhiForCausalLM"]
51200
bfloat16
4.38.2
2,048
8,192
24
32
32
gelu_new
0
True
2,048
10,000
null
0.02
null
null
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
6,526
CMU-AIR2/math-phi-1-5-FULL-Arithmetic-Steps-8k
phi
["PhiForCausalLM"]
51200
bfloat16
4.38.2
2,048
8,192
24
32
32
gelu_new
0
True
2,048
10,000
null
0.02
null
null
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
6,527
CMunch/fine_tuned_dota
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.30.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,528
CNBOOMBOOM/llama2-hivetalk-merger-10k-v2
llama
["LlamaForCausalLM"]
32000
float16
4.35.2
4,096
11,008
32
32
32
silu
null
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
6,529
CNBOOMBOOM/quant-gptq-hivetalk-4bits
llama
["LlamaForCausalLM"]
32000
float16
4.40.1
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
6,530
CNR-ILC/gs-GreBerta
roberta
["RobertaForMaskedLM"]
52000
float32
4.51.3
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,531
CO21351/sd-class-animals-30-new
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,532
CO21351/sd-class-butterflies-32-new
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,533
COCO0414/bge-m3-ko_DNF
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.39.1
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
6,534
CONCISE/LLaMa_V2-13B-Instruct-Uncensored-HF
llama
["LlamaForCausalLM"]
32000
float32
4.31.0
5,120
13,824
40
40
40
silu
null
False
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,535
CPSC532/2024NOV14_llama_3_1_8b_r_32
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
6,536
CPSC532/2024NOV14_llama_3_1_8b_r_64
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
6,537
CPSC532/2024NOV16_llama_3_1_8b_no_sources_in_question_r_128
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
6,538
CPSC532/2024NOV16_llama_3_1_8b_no_sources_in_question_r_16
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
6,539
CPSC532/eval_no_sources_finetune
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
6,540
CPSC532/finetuned_model_2024OCT24
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
6,541
CPSC532/finetuned_model_2024OCT25
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
6,542
CR011v2/streetvision
vit
["ViTForImageClassification"]
null
float32
4.49.0
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
6,543
CRZ7436/fine_tuned_gpt2_model
gpt2
["GPT2LMHeadModel"]
50257
float32
4.39.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
6,544
CSHaitao/DELTA_EN
bert
["BertForSAILER"]
30522
float32
4.29.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,545
CSHaitao/DELTA_EN_Finetune
bert
["BertModel"]
30522
float32
4.29.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,546
CSLin3303/qwen3-20250528
qwen3
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,547
CSLin3303/qwen3-laws-20250613002
qwen3
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,548
CSerdar014191/opt-350m-sixbooks
opt
["OPTForCausalLM"]
50272
float32
4.28.1
1,024
null
24
16
null
null
0
True
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
medium
6,549
CSerdar014191/opt-350m-sixhorrorbooks
opt
["OPTForCausalLM"]
50272
float32
4.28.1
1,024
null
24
16
null
null
0
True
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
medium
6,550
CSerdar014191/opt-350m_test08_tuner
opt
["OPTForCausalLM"]
50272
float16
4.28.1
1,024
null
24
16
null
null
0
True
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
medium
6,551
CShorten/ArXiv-Cross-Encoder-Title-Abstracts
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.20.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,552
CShorten/CORD-19-Title-Abstracts-1-more-epoch
bert
["BertModel"]
30522
float32
4.22.1
384
1,536
6
12
null
gelu
null
True
512
null
null
0.02
null
null
null
32
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.010617
small
short
6,553
CTCF/pretrained_hooked_gpt2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
0.05
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,554
CTPC/distilhubert-finetuned-gtzan
hubert
["HubertForSequenceClassification"]
32
float32
4.34.0
768
3,072
2
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.014156
small
null
6,555
CTW-g123/MiniCPM
minicpmv
["MiniCPMV"]
151666
bfloat16
4.40.0
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
6,556
CUTD/qnAr
electra
["ElectraForQuestionAnswering"]
64000
float32
4.30.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
6,557
CVR123/Muril-base-finetune-Telugu-qc
bert
["BertForSequenceClassification"]
197285
float32
4.38.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,558
CWKSC/w2v-bert-2.0-mongolian-colab-CV16.0
wav2vec2-bert
["Wav2Vec2BertForCTC"]
37
float32
4.44.2
1,024
4,096
24
16
null
swish
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
null
6,559
CWKSC/whisper-small-onnx
whisper
["WhisperForConditionalGeneration"]
51865
null
4.34.0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
6,560
CWKSC/whisper-small-onnx-quantized
whisper
["WhisperForConditionalGeneration"]
51865
null
4.34.0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
6,561
CXDuncan/whisper-large-v3-malayalam
whisper
["WhisperForConditionalGeneration"]
51866
float32
4.41.0
null
null
32
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
6,562
CXL295/zephyr-7b-dpo-full
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
6,563
CY-Wu/med_and_tcp_learn
gpt2
["GPT2LMHeadModel"]
21128
float32
4.46.2
null
null
null
null
null
null
null
True
null
null
null
0.02
101
102
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
6,564
CYFRAGOVPL/Llama-PLLuM-70B-base
llama
["LlamaForCausalLM"]
128256
bfloat16
4.47.0.dev0
8,192
28,672
80
64
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
6,565
CYFRAGOVPL/Llama-PLLuM-70B-chat
llama
["LlamaForCausalLM"]
128258
bfloat16
4.45.0
8,192
28,672
80
64
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
6,566
CYFRAGOVPL/Llama-PLLuM-70B-instruct
llama
["LlamaForCausalLM"]
128258
bfloat16
4.47.0.dev0
8,192
28,672
80
64
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
6,567
CYFRAGOVPL/Llama-PLLuM-8B-base
llama
["LlamaForCausalLM"]
128256
bfloat16
4.47.0.dev0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
6,568
CYFRAGOVPL/Llama-PLLuM-8B-chat
llama
["LlamaForCausalLM"]
128258
bfloat16
4.45.0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
6,569
CYFRAGOVPL/Llama-PLLuM-8B-instruct
llama
["LlamaForCausalLM"]
128258
bfloat16
4.47.0.dev0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
6,570
CYFRAGOVPL/PLLuM-12B-base
mistral
["MistralForCausalLM"]
131072
bfloat16
4.47.0.dev0
5,120
14,336
40
32
8
silu
0
False
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
6,571
CYFRAGOVPL/PLLuM-12B-chat
mistral
["MistralForCausalLM"]
131072
bfloat16
4.45.0
5,120
14,336
40
32
8
silu
0
False
131,072
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
6,572
CYFRAGOVPL/PLLuM-12B-instruct
mistral
["MistralForCausalLM"]
131072
bfloat16
4.47.0.dev0
5,120
14,336
40
32
8
silu
0
False
131,072
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
6,573
CYFRAGOVPL/PLLuM-12B-nc-base
mistral
["MistralForCausalLM"]
131072
bfloat16
4.45.0.dev0
5,120
14,336
40
32
8
silu
0
False
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
6,574
CYFRAGOVPL/PLLuM-12B-nc-chat
mistral
["MistralForCausalLM"]
131072
bfloat16
4.45.0
5,120
14,336
40
32
8
silu
0
False
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
6,575
CYFRAGOVPL/PLLuM-12B-nc-instruct
mistral
["MistralForCausalLM"]
131072
bfloat16
4.46.3
5,120
14,336
40
32
8
silu
0
False
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
6,576
CYFRAGOVPL/PLLuM-8x7B-base
mixtral
["MixtralForCausalLM"]
32000
bfloat16
4.36.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
8
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
very_long
6,577
CYFRAGOVPL/PLLuM-8x7B-chat
mixtral
["MixtralForCausalLM"]
32003
bfloat16
4.45.0
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
8
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
very_long
6,578
CYFRAGOVPL/PLLuM-8x7B-instruct
mixtral
["MixtralForCausalLM"]
32003
bfloat16
4.46.3
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
8
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
very_long
6,579
CYFRAGOVPL/PLLuM-8x7B-nc-base
mixtral
["MixtralForCausalLM"]
32000
bfloat16
4.36.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
8
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
very_long
6,580
CYFRAGOVPL/PLLuM-8x7B-nc-chat
mixtral
["MixtralForCausalLM"]
32003
bfloat16
4.45.0
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
8
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
very_long
6,581
CYFRAGOVPL/PLLuM-8x7B-nc-instruct
mixtral
["MixtralForCausalLM"]
32003
bfloat16
4.46.3
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
8
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
very_long
6,582
CYONG/v1
bert
["BertForSequenceClassification"]
105879
null
4.31.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,583
CYRT/distilbert-base-uncased-finetuned-emotion
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.28.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,584
CYX2003/whisper-tiny-en
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.41.1
null
null
4
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
6,585
CaasiHUANG/flames-scorer
internlm
["InternLMForSequenceClassification"]
103168
float16
4.33.2
4,096
11,008
32
32
null
silu
null
True
2,048
null
0.000001
0.02
1
2
false
128
null
false
null
null
false
false
false
false
false
false
false
false
false
false
6.442451
medium
medium
6,586
Cabooose/multilingual_english_SEQUENCE_classification_model
bert
["BertForSequenceClassification"]
119547
float32
4.28.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,587
Cabooose/multilingual_english_token_classification_model
bert
["BertForTokenClassification"]
119547
float32
4.28.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,588
CaceAI/Llama3-70B-Fusion-ChatQA-Chinese
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.2
8,192
28,672
80
64
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
long
6,589
CaceAI/Llama3-8B-Fusion-ChatQA-Chinese
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
6,590
Caco355/medical-model
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
6,591
Cadenza-Labs/dolphin-llama3-8B-sleeper-agent-distilled-lora
llama
["LlamaForCausalLM"]
128258
float32
4.40.1
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128256
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
6,592
Caephas/real-madrid-fact
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.47.1
2,048
16,384
18
8
1
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
6,593
Cafet/w2v-bert-final-v2
wav2vec2-bert
["Wav2Vec2BertForCTC"]
39
float32
4.40.0
1,024
4,096
24
16
null
swish
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
null
6,594
Cafet/w2v-bert-version-final
wav2vec2-bert
["Wav2Vec2BertForCTC"]
40
float32
4.40.0
1,024
4,096
24
16
null
swish
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
null
6,595
Cafet/wav2vec2-bert-cv16-mas-ex-cv16
wav2vec2-bert
["Wav2Vec2BertForCTC"]
40
float32
4.39.1
1,024
4,096
24
16
null
swish
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
null
6,596
Cafet/wav2vec2-bert-mas-ex
wav2vec2-bert
["Wav2Vec2BertForCTC"]
40
float32
4.38.2
1,024
4,096
24
16
null
swish
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
null
6,597
Cafet/whisper-mongolian-version-0.1
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.39.3
null
null
32
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
6,598
Cafet/whisper-small-mn-cv16
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.39.3
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
6,599
Cafet/whisper-small-mongolian
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.40.1
null
null
12
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null