index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
15,600
InstalilyAI/deberta-80k-1.1
deberta-v2
["DebertaV2ForSequenceClassification"]
128100
float32
4.52.4
1,024
4,096
24
16
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
15,601
InstantX/FLUX.1-dev-Controlnet-Canny
null
[]
null
null
null
null
null
null
24
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
15,602
InstantX/FLUX.1-dev-Controlnet-Union
null
[]
null
null
null
null
null
null
24
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
15,603
InstantX/SD3-Controlnet-Canny
null
[]
null
null
null
null
null
null
24
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,604
InstantX/SD3-Controlnet-Depth
null
[]
null
null
null
null
null
null
24
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,605
InstantX/SD3-Controlnet-Pose
null
[]
null
null
null
null
null
null
24
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,606
InsultedByMathematics/50-50-lr-2_65e-6_update_401_online
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,607
InsultedByMathematics/all_online_responses_intermediate_as_reference_model_iter_2
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,608
InsultedByMathematics/alpha_1e-3_beta_2e-3
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,609
InsultedByMathematics/alpha_1e-3_beta_4e-3
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,610
InsultedByMathematics/alpha_1e-4_beta_3e-3_update_201
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,611
InsultedByMathematics/alpha_1e-4_beta_3e-3_update_401
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,612
InsultedByMathematics/alpha_1e-4_beta_3e-3_update_601
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,613
InsultedByMathematics/infoNCA-lr-2_65e-6_update_401_online
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,614
InsultedByMathematics/lr-5e-7_update_201
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,615
InsultedByMathematics/lr-5e-7_update_401
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,616
InsultedByMathematics/lr-5e-7_update_601
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,617
InsultedByMathematics/rebel_ultrafeedback_update_401
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,618
IntJudge/IntJudge
qwen2_vl
["Qwen2VLForConditionalGeneration"]
152064
bfloat16
4.41.2
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
15,619
IntMeGroup/ICCVW_st1_mos2
internvl_chat
["InternVLChatModel"]
null
bfloat16
null
4,096
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
15,620
IntMeGroup/LMM4LMM-Correspondence
internvl_chat
["InternVLChatModel"]
null
bfloat16
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
15,621
IntMeGroup/LMM4LMM-QA
internvl_chat
["InternVLChatModel"]
null
bfloat16
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
15,622
IntMeGroup/LOVE-Corrrespondence
internvl_chat
["InternVLChatModel"]
null
bfloat16
null
4,096
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
15,623
IntMeGroup/LOVE-Perception
internvl_chat
["InternVLChatModel"]
null
bfloat16
null
4,096
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
15,624
Intel/NeuroPrompts
gpt2
["GPT2LMHeadModel"]
50257
float32
4.29.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
15,625
Intel/albert-base-v2-sst2-int8-static-inc
albert
["AlbertForSequenceClassification"]
30000
int8
4.18.0
768
3,072
12
12
null
gelu_new
null
null
512
null
null
0.02
2
3
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,626
Intel/bert-base-uncased-finetuned-swag-int8-static-inc
bert
["BertForMultipleChoice"]
30522
int8
4.18.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,627
Intel/bert-base-uncased-mrpc-int8-static-inc
bert
["BertForSequenceClassification"]
30522
int8
4.22.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,628
Intel/bert-base-uncased-squad-int8-static-inc
bert
["BertForQuestionAnswering"]
30522
int8
4.21.0.dev0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,629
Intel/deepseek-coder-1.3b_base_ov_int8
llama
["LlamaForCausalLM"]
32256
null
4.37.2
2,048
5,504
24
16
16
silu
0
True
16,384
100,000
0.000001
0.02
32013
32014
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
long
15,630
Intel/distilbert-base-uncased-squadv1.1-sparse-80-1x4-block-pruneofa
distilbert
["DistilBertForQuestionAnswering"]
30522
null
4.6.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
15,631
Intel/dpt-beit-base-384
dpt
["DPTForDepthEstimation"]
null
float32
null
768
null
null
null
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,632
Intel/dpt-beit-large-384
dpt
["DPTForDepthEstimation"]
null
float32
null
768
null
null
null
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,633
Intel/dpt-beit-large-512
dpt
["DPTForDepthEstimation"]
null
float32
null
768
null
null
null
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,634
Intel/gemma-2b-int4-inc
gemma
["GemmaForCausalLM"]
256000
float16
4.40.2
2,048
16,384
18
8
1
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
15,635
Intel/llava-gemma-7b
llava_gemma
["LlavaGemmaForCausalLM"]
256000
bfloat16
4.39.0.dev0
3,072
24,576
28
16
16
gelu
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
15,636
Intel/msmarco_fid_early_exit
t5
["FiDT5"]
32128
float32
4.16.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
15,637
Intel/nq_fid_lfqa
t5
["FiDT5"]
32128
float32
4.28.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
15,638
Intel/nq_fid_lfqa_early_exit
t5
["FiDT5"]
32128
float32
4.16.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
15,639
Intel/sd-1.5-lcm-openvino
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,640
Intel/t5-small-xsum-int8-dynamic-inc
t5
["T5ForConditionalGeneration"]
32100
int8
4.21.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
15,641
Intel/tiny-random-bert_ipex_model
bert
["BertForQuestionAnswering"]
1124
float32
4.46.3
32
37
5
4
null
gelu
null
True
512
null
null
0.02
null
null
null
8
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.000061
small
short
15,642
Intel/tiny-random-gpt2_ipex_model
gpt2
["GPT2LMHeadModel"]
50257
float32
4.46.3
null
null
null
null
2
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
true
false
null
null
null
15,643
Intel/tiny-random-vit_ipex_model
vit
["ViTForImageClassification"]
{}
float32
4.46.3
32
37
5
4
null
gelu
null
null
null
null
null
0.02
null
null
null
8
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.000061
small
null
15,644
Intel/whisper-base-int8-dynamic-inc
whisper
["WhisperForConditionalGeneration"]
51865
null
4.30.2
null
null
6
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,645
Intel/whisper-base-int8-static-inc
whisper
["WhisperForConditionalGeneration"]
51865
null
4.30.2
null
null
6
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,646
Intel/whisper-base-onnx-int4-inc
whisper
["WhisperForConditionalGeneration"]
51865
null
4.33.1
null
null
6
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,647
Intel/whisper-large-onnx-int4-inc
whisper
["WhisperForConditionalGeneration"]
51865
null
4.33.1
null
null
32
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,648
Intel/whisper-large-v2-onnx-int4-inc
whisper
["WhisperForConditionalGeneration"]
51865
null
4.32.0
null
null
32
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,649
Intel/whisper-medium-onnx-int4-inc
whisper
["WhisperForConditionalGeneration"]
51865
null
4.32.0
null
null
24
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,650
Intel/whisper-small-int8-dynamic-inc
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.27.0.dev0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,651
Intel/whisper-small-int8-static-inc
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.27.0.dev0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,652
Intel/whisper-small-onnx-int4-inc
whisper
["WhisperForConditionalGeneration"]
51865
null
4.33.1
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,653
Intel/whisper-tiny-int8-dynamic-inc
whisper
["WhisperForConditionalGeneration"]
51865
null
4.30.2
null
null
4
null
null
null
0
True
null
null
null
null
50257
50257
true
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,654
Intel/whisper-tiny-int8-static-inc
whisper
["WhisperForConditionalGeneration"]
51865
null
4.30.2
null
null
4
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,655
Intel/whisper-tiny-onnx-int4-inc
whisper
["WhisperForConditionalGeneration"]
51865
null
4.33.1
null
null
4
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,656
IntelLabs/Mamba-Shedder-Mamba2-2.7B-Pruned-22SSM-Alpaca
null
[]
50277
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,657
IntelLabs/shears-mpt-7b-50-base
mpt
["MPTForCausalLM"]
50432
float16
4.31.0
null
null
null
null
null
null
null
False
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,658
IntelLabs/sqft-phi-3-mini-4k-60-base
phi3
["Phi3ForCausalLM"]
32064
float16
4.44.2
3,072
8,192
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
medium
15,659
IntellectusAI/mistralfinetunecompanylaw
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.0
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
15,660
IntelliGrow/poca-SoccerTwos
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,661
InterSync/Gemma-7B-Instruct-Function-Calling
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.38.2
3,072
24,576
28
16
16
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
15,662
InterSync/Mistral-7B-Instruct-v0.2-Function-Calling
mistral
["MistralForCausalLM"]
32000
bfloat16
4.38.2
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
15,663
Intern95/opt-125m-gptq
opt
["OPTForCausalLM"]
50272
float16
4.41.2
768
null
12
12
null
null
0
True
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
medium
15,664
Intern95/opt-350m-gptq
opt
["OPTForCausalLM"]
50272
float16
4.41.2
1,024
null
24
16
null
null
0
True
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
medium
15,665
IntervitensInc/intv_ai_mk12
llama
["LlamaForCausalLM"]
32000
bfloat16
4.36.2
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
15,666
IntervitensInc/intv_ai_mk2
llama
["LlamaForCausalLM"]
32002
bfloat16
4.36.2
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
15,667
IntervitensInc/intv_l3_mk1
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,668
IntervitensInc/intv_l3_mk2
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,669
IntervitensInc/intv_l3_mk3
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,670
IntervitensInc/kek_mk6
stablelm
["StableLmForCausalLM"]
100352
bfloat16
4.40.0
2,048
5,632
24
32
32
silu
0
False
4,096
10,000
null
0.02
100257
100257
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
15,671
IntervitensInc/kek_mk6_2
stablelm
["StableLmForCausalLM"]
100352
bfloat16
4.40.0
2,048
5,632
24
32
32
silu
0
False
4,096
10,000
null
0.02
100257
100257
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
15,672
IntervitensInc/kek_mk6_4
stablelm
["StableLmForCausalLM"]
100352
bfloat16
4.40.0
2,048
5,632
24
32
32
silu
0
False
4,096
10,000
null
0.02
100257
100257
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
15,673
IntraFind/Llama-3-SauerkrautLM-8b-16k-Instruct-AWQ
llama
["LlamaForCausalLM"]
128256
float16
4.40.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,674
Intradiction/text_classification_NoLORA
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.35.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
15,675
Invalid-Null/PeiYangMe-0.0
llama
["LlamaForCausalLM"]
64000
float16
4.40.1
4,096
11,008
32
32
4
silu
0
True
4,096
5,000,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
15,676
Invalid-Null/PeiYangMe-0.1
llama
["LlamaForCausalLM"]
64000
float16
4.40.1
4,096
11,008
32
32
4
silu
0
True
4,096
5,000,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
15,677
Invalid-Null/PeiYangMe-0.2
llama
["LlamaForCausalLM"]
64000
float16
4.40.2
4,096
11,008
32
32
4
silu
0
False
4,096
5,000,000
0.000001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
15,678
Invalid-Null/PeiYangMe-0.3
llama
["LlamaForCausalLM"]
64000
float16
4.40.2
4,096
11,008
32
32
4
silu
0
False
4,096
5,000,000
0.000001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
15,679
Invalid-Null/PeiYangMe-0.5
llama
["LlamaForCausalLM"]
64000
float16
4.40.2
4,096
11,008
32
32
4
silu
0
True
4,096
5,000,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
15,680
Invalid-Null/PeiYangMe-0.6
llama
["LlamaForCausalLM"]
64000
float16
4.40.2
4,096
11,008
32
32
4
silu
0
False
4,096
5,000,000
0.000001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
15,681
Invalid-Null/PeiYangMe-0.7
llama
["LlamaForCausalLM"]
64000
float16
4.40.2
4,096
11,008
32
32
4
silu
0
True
4,096
5,000,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
15,682
InvestmentResearchAI/LLM-ADE_tiny-v0.001
llama
["LlamaForCausalLM"]
32002
float32
4.41.1
2,048
5,632
22
32
4
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
15,683
InvokeAI/ip_adapter_sdxl_image_encoder
clip_vision_model
["CLIPVisionModelWithProjection"]
null
null
4.24.0
1,664
8,192
48
16
null
gelu
0
null
null
null
null
0.02
null
null
true
104
null
false
null
null
false
false
false
false
false
false
false
false
false
false
1.594884
medium
null
15,684
Ioana23/marian-finetuned-kde4-en-to-ro
marian
["MarianMTModel"]
59543
null
4.32.1
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
15,685
IoanaLivia/whisper-small-finetuned-800-standard-A-epochs-10-5-03-20-17
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.49.0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,686
IoanaLiviaPopescu/real-data-synth-data-400-1-Emil-Neural-whisper-small
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.51.3
null
null
12
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,687
IoanaLiviaPopescu/real-data-synth-data-400-2-Wavnet-B-whisper-small
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.51.3
null
null
12
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,688
IoanaLiviaPopescu/real-data-synth-data-800-1-Mixed-Standard-B-Wavnet-B-whisper-small
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.51.3
null
null
12
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,689
IoanaLiviaPopescu/real-data-synth-data-800-1-Wavenet-B-whisper-small
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.51.3
null
null
12
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,690
Ioanaaaaaaa/distilbert-base-uncased-finetuned-emotion
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.29.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
15,691
Ionel2023/t5-small-finetuned-xsum
t5
["T5ForConditionalGeneration"]
32128
float32
4.35.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
15,692
IoriU/cpt-indobert-large-2
bert
["BertForPreTraining"]
30522
float32
4.27.4
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
15,693
IoriU/cpt-indobert-large-26
bert
["BertForPreTraining"]
30522
float32
4.27.4
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
15,694
IoriU/cpt-indobert-large-3
bert
["BertForPreTraining"]
30522
float32
4.27.4
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
15,695
IoriU/cpt-indobert-large-4
bert
["BertForPreTraining"]
30522
float32
4.27.4
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
15,696
IoriU/cpt-indobert-large-5
bert
["BertForPreTraining"]
30522
float32
4.27.4
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
15,697
Ipan98/finQuant
llama
["LlamaForCausalLM"]
32000
float32
4.36.0.dev0
4,096
11,008
32
32
32
silu
0
True
2,048
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
15,698
Ipan98/results
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
15,699
Ippiki-Ookami/llama381binstruct_summarize_short_merged
llama
["LlamaForCausalLM"]
128256
float32
4.51.3
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long