index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
2,400
AlekHesa/testing-llama2-v3
llama
["LlamaForCausalLM"]
32016
float16
4.40.0.dev0
4,096
11,008
32
32
32
silu
0
True
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
2,401
AlekHesa/testing-llama2-v7
llama
["LlamaForCausalLM"]
32000
float32
4.41.0.dev0
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,402
AlekSt7/auralis_portal2_announcer_ru_xttsv2
xtts
["XttsGPT"]
null
null
4.46.0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
2,403
AlekSt7/auralis_portal2_announcer_ru_xttsv2_gpt
xtts_gpt
["XttsGPT"]
6681
null
null
1,024
null
30
16
null
null
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
true
false
false
false
false
false
false
false
0.377487
small
null
2,404
Aleka12/distilbart-finetuned-mimic
bart
["BartForConditionalGeneration"]
50264
float32
4.52.4
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
2,405
Aleks2002SH/llm-course-hw2-dpo
llama
["LlamaForCausalLM"]
49152
bfloat16
4.50.0
576
1,536
30
9
3
silu
0
True
2,048
10,000
0.00001
0.02
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
medium
2,406
Aleksandar1932/gpt-neo-125M-country
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.15.0
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
2,407
Aleksandar1932/gpt-neo-125M-hip-hop
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.15.0
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
2,408
Aleksandar1932/gpt-neo-125M-metal
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.15.0
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
2,409
Aleksandar1932/gpt-neo-125M-rock
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.15.0
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
2,410
Aleksandar1932/gpt-neo-125M-spanish-classics
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.15.0
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
2,411
Aleksandr505/phishing-text-classifier-rubert
bert
["BertForSequenceClassification"]
119547
float32
4.51.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,412
AleksandrKashirin/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,413
AlekseiPravdin/KSIF-RPG-128k-slerp
mistral
["MistralForCausalLM"]
32000
bfloat16
4.38.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,414
AlekseiPravdin/KukulStanta-Seamaiiza-slerp-merge5
mistral
["MistralForCausalLM"]
32000
float16
4.44.0
4,096
14,336
31
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.241124
medium
very_long
2,415
AlekseiPravdin/Seamaiiza-7B-v1
mistral
["MistralForCausalLM"]
32000
bfloat16
4.40.2
4,096
14,336
32
32
8
silu
0
True
8,192
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
long
2,416
AlekseiPravdin/Seamaiiza-7B-v2-128k
mistral
["MistralForCausalLM"]
32000
bfloat16
4.40.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,417
AlekseyElygin/Llama-3.2-11B-Vision-Instruct-bnb-4bit
mllama
["MllamaForConditionalGeneration"]
null
bfloat16
4.45.0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
true
null
null
null
2,418
AlekseyElygin/llama-3-8b-bnb-4bit
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
2,419
AlekseyElygin/mistral-7b-instruct-v0.3-4bit
mistral
["MistralForCausalLM"]
32768
float16
4.41.2
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,420
AlekseyElygin/mistral-7b-instruct-v0.3-bnb-8ep-GGUF
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
2,421
AlekseyKorshuk/125m-ri-reproduce-combined-4-gpu-20-val-v2
opt
["OPTForCausalLM"]
50265
float16
4.26.0.dev0
768
null
12
12
null
null
0
True
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
medium
2,422
AlekseyKorshuk/6.7b-ri-reproduce-combined-4-gpu-20-val-v3
opt
["OPTForCausalLM"]
50265
float16
4.26.0.dev0
4,096
null
32
32
null
null
0
True
2,048
null
null
null
2
2
null
128
null
false
null
null
false
false
false
false
false
false
false
false
false
false
6.442451
medium
medium
2,423
AlekseyKorshuk/amazon-reviews-input-output
opt
["OPTForCausalLM"]
50265
float32
4.25.0.dev0
1,024
null
24
16
null
null
0
True
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
medium
2,424
AlekseyKorshuk/amazon-reviews-input-output-1.3b
opt
["OPTForCausalLM"]
50265
float16
4.25.0.dev0
2,048
null
24
32
null
null
0
True
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
1.20796
medium
medium
2,425
AlekseyKorshuk/amazon-reviews-input-output-13b
opt
["OPTForCausalLM"]
50265
float16
4.25.0.dev0
5,120
null
40
40
null
null
0
True
2,048
null
null
null
2
2
null
128
null
false
null
null
false
false
false
false
false
false
false
false
false
false
12.582912
large
medium
2,426
AlekseyKorshuk/amazon-reviews-input-output-350m
opt
["OPTForCausalLM"]
50265
float16
4.25.0.dev0
1,024
null
24
16
null
null
0
True
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
medium
2,427
AlekseyKorshuk/amazon-reviews-input-output-6.7b
opt
["OPTForCausalLM"]
50265
float16
4.25.0.dev0
4,096
null
32
32
null
null
0
True
2,048
null
null
null
2
2
null
128
null
false
null
null
false
false
false
false
false
false
false
false
false
false
6.442451
medium
medium
2,428
AlekseyKorshuk/amazon-reviews-input-output-6.7b-best
opt
["OPTForCausalLM"]
50265
float16
4.25.0.dev0
4,096
null
32
32
null
null
0
True
2,048
null
null
null
2
2
null
128
null
false
null
null
false
false
false
false
false
false
false
false
false
false
6.442451
medium
medium
2,429
AlekseyKorshuk/chai-pygmalion-v0-chatml
gptj
["GPTJForCausalLM"]
50400
float16
4.28.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
false
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
2,430
AlekseyKorshuk/chatml-test-small
gpt_neox
["GPTNeoXForCausalLM"]
50279
float16
4.28.1
2,048
8,192
16
8
null
gelu
null
True
2,048
null
null
0.02
0
0
false
256
null
false
null
null
false
false
true
false
false
false
false
false
false
false
0.805306
small
medium
2,431
AlekseyKorshuk/dalio-1.3b-test
opt
["OPTForCausalLM"]
50265
float16
4.25.0.dev0
2,048
null
24
32
null
null
0
True
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
1.20796
medium
medium
2,432
AlekseyKorshuk/evol-codealpaca-v1-sft-4e-5
phi
["PhiForCausalLM"]
51200
bfloat16
4.37.0
2,560
10,240
32
32
32
gelu_new
0
False
2,048
10,000
null
0.02
null
null
false
80
1
false
null
null
false
false
false
false
false
false
false
false
false
true
2.516582
medium
medium
2,433
AlekseyKorshuk/evol-codealpaca-v1-sft-4e-5-dpo-3ep
phi
["PhiForCausalLM"]
51200
float16
4.37.0
2,560
10,240
32
32
32
gelu_new
0
False
2,048
10,000
null
0.02
null
null
false
80
1
false
null
null
false
false
false
false
false
false
false
false
false
true
2.516582
medium
medium
2,434
AlekseyKorshuk/llama-7b-chatml
llama
["LlamaForCausalLM"]
32002
float16
4.28.1
4,096
11,008
32
32
null
silu
null
True
2,048
null
0.000001
0.02
1
2
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
6.442451
medium
medium
2,435
AlekseyKorshuk/llama-7b-instructions-chatml
llama
["LlamaForCausalLM"]
32000
float16
4.28.1
4,096
11,008
32
32
null
silu
null
True
2,048
null
0.000001
0.02
1
2
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
6.442451
medium
medium
2,436
AlekseyKorshuk/llama-7b-v0-chatml
llama
["LlamaForCausalLM"]
32000
float16
4.28.1
4,096
11,008
32
32
null
silu
null
True
2,048
null
0.000001
0.02
1
2
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
6.442451
medium
medium
2,437
AlekseyKorshuk/pythia-1.4b-deduped-jokes
gpt_neox
["GPTNeoXForCausalLM"]
50304
float16
4.29.0.dev0
2,048
8,192
24
16
null
gelu
null
True
2,048
null
null
0.02
0
0
false
128
null
false
null
null
false
false
true
false
false
false
false
false
false
false
1.20796
medium
medium
2,438
AlekseyKorshuk/pythia-1b-deduped-chatml
gpt_neox
["GPTNeoXForCausalLM"]
50279
float16
4.28.1
2,048
8,192
16
8
null
gelu
null
True
2,048
null
null
0.02
0
0
false
256
null
false
null
null
false
false
true
false
false
false
false
false
false
false
0.805306
small
medium
2,439
AlekseyKorshuk/retriever-coding-guru-adapted
mpnet
["MPNetModel"]
30527
float32
4.21.2
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
2,440
AlekseyKorshuk/roberta-with-topic
roberta
["RobertaForSequenceClassification"]
50265
float16
4.29.1
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
2,441
AlekseyKorshuk/vic15-exp-syn-fight-cp1919
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,442
AlekseyKorshuk/vic15-exp-syn-fight-cp3838
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,443
AlekseyKorshuk/vic15-exp-syn-fight-cp5757
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,444
AlekseyKorshuk/vic15-exp-syn-romantic-cp1310
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,445
AlekseyKorshuk/vic15-exp-syn-romantic-cp2620
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,446
AlekseyKorshuk/vicuna-7b-83k-dataset-new-combined
llama
["LlamaForCausalLM"]
32001
float16
4.28.1
4,096
11,008
32
32
null
silu
null
True
2,048
null
0.000001
0.02
0
1
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
6.442451
medium
medium
2,447
AlekseyKorshuk/vsinrom3
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,448
AlekseyKorshuk/yt_videos_comments
gpt2
["GPT2LMHeadModel"]
50257
float16
4.29.0.dev0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
2,449
Aleksia/finetuning-distilBert_sentiment
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.35.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,450
Alekyukk/ppo
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,451
Alelcv27/llama3-8b-breadcrumbs-ties-v3
llama
["LlamaForCausalLM"]
128256
float16
4.48.0
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,452
Alelcv27/llama3-8b-breadcrumbs-ties-v5
llama
["LlamaForCausalLM"]
128256
float16
4.48.0
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,453
Alelcv27/llama3-8b-dare-linear-v3
llama
["LlamaForCausalLM"]
128256
float16
4.48.0
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,454
Alelcv27/llama3-8b-dare-ties-v2
llama
["LlamaForCausalLM"]
128256
float16
4.48.0
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,455
Alelcv27/llama3-8b-dare-ties-v3
llama
["LlamaForCausalLM"]
128256
float16
4.48.0
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,456
Alelcv27/llama3-8b-della-v2
llama
["LlamaForCausalLM"]
128256
float16
4.48.0
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,457
Alelcv27/llama3-8b-passthrough
llama
["LlamaForCausalLM"]
128256
float16
4.47.0
4,096
14,336
64
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
12.884902
large
very_long
2,458
Alelcv27/llama3-8b-slerp
llama
["LlamaForCausalLM"]
128256
float16
4.47.0
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,459
Alemat/abcnew
roberta
["RobertaForMaskedLM"]
50265
null
null
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,460
Alemat/n460000
roberta
["RobertaForMaskedLM"]
50265
null
null
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,461
Alemat/nm462000
roberta
["RobertaForMaskedLM"]
50265
float32
4.12.0.dev0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,462
Alepach/notHumpback-M1-Rw-F-8b
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.3
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,463
Aleph-Alpha/Aleph-Alpha-GermanWeb-Grammar-Classifier-BERT
bert
["BertForSequenceClassification"]
30522
float32
4.48.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,464
Aleph-Alpha/Aleph-Alpha-GermanWeb-Quality-Classifier-BERT
bert
["BertForSequenceClassification"]
30522
float32
4.48.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,465
Aleph-Alpha/llama-3_1-8b-tfree-hat-base
hierarchical_autoregressive_transformer
["HATForCausalLM"]
null
bfloat16
4.46.3
null
null
null
null
null
null
null
null
262,144
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
very_long
2,466
Aleph-Alpha/llama-3_1-8b-tfree-hat-sft
hierarchical_autoregressive_transformer
["HATForCausalLM"]
null
bfloat16
4.46.3
null
null
null
null
null
null
null
null
262,144
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
very_long
2,467
AlephNull/Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,468
Aleron12/distilbert-base-uncased-finetuned-amz_brander
distilbert
["DistilBertForQuestionAnswering"]
30522
float32
4.27.0.dev0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,469
AlessandroMinervini/llama_factory_first_model
mistral
["MistralForCausalLM"]
32768
bfloat16
4.45.2
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,470
Alessio-Borgi/all-mpnet-base-v2-margin-based-triplet-loss-finetuned-culture-1-epochs-enhanced_test
mpnet
["MPNetModel"]
30527
float32
4.51.3
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
2,471
Alessio-Borgi/all-mpnet-base-v2-margin-based-triplet-loss-finetuned-culture-10-epochs-enhanced
mpnet
["MPNetModel"]
30527
float32
4.51.3
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
2,472
Alessio-Borgi/all-mpnet-base-v2-margin-based-triplet-loss-finetuned-culture-3-epochs-enhanced
mpnet
["MPNetModel"]
30527
float32
4.51.3
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
2,473
Alesteba/detr-resnet-50_finetuned_cppe5
detr
["DetrForObjectDetection"]
null
float32
4.7.0.dev0
null
null
6
null
null
null
0
null
1,024
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
2,474
Aleteian/Avalon
mistral
["MistralForCausalLM"]
131072
float16
4.46.2
5,120
14,336
40
32
8
silu
0
True
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
2,475
Aleteian/Follow-the-Flow-MN-12B
mistral
["MistralForCausalLM"]
131077
bfloat16
4.48.2
5,120
14,336
40
32
8
silu
0
True
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
2,476
Aleteian/Legend-of-the-Four-Winds-2-MN-12B
mistral
["MistralForCausalLM"]
131074
bfloat16
4.48.2
5,120
14,336
40
32
8
silu
0
True
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
2,477
Aleteian/Legend-of-the-Four-Winds-MN-12B
mistral
["MistralForCausalLM"]
131074
bfloat16
4.48.2
5,120
14,336
40
32
8
silu
0
True
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
2,478
Aleteian/One-To-Rule-Them-All-MN-12B
mistral
["MistralForCausalLM"]
131077
bfloat16
4.48.2
5,120
14,336
40
32
8
silu
0
True
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
2,479
Aleteian/Saiga-Unleashed
mistral
["MistralForCausalLM"]
131072
bfloat16
4.48.1
5,120
14,336
40
32
8
silu
0
True
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
2,480
Aleteian/Way-to-Unseen-Horizon-MN-12B
mistral
["MistralForCausalLM"]
131074
bfloat16
4.48.2
5,120
14,336
40
32
8
silu
0
True
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
2,481
Aletheia-ng/AfroLid
Afrolid
["AfroLid"]
null
float32
4.46.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,482
Alex-Deng/dummy-model
camembert
["CamembertForMaskedLM"]
32005
float32
4.31.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
5
6
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,483
Alex007ander/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-fierce_yawning_leopard
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.51.3
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
2,484
Alex01837178373/3.1model
llama
["LlamaForCausalLM"]
128256
float16
4.43.1
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,485
Alex034/t5-base-finetuned-summary
t5
["T5ForConditionalGeneration"]
32128
null
4.35.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
2,486
Alex034/t5-small-indosum-summary-freeze
t5
["T5ForConditionalGeneration"]
32128
null
4.35.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
2,487
Alex2060/whisper-tiny-ru
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.45.2
null
null
4
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
2,488
Alex31y/bert-base-multilingual-uncased-finetuned-squad
bert
["BertForQuestionAnswering"]
105879
float32
4.28.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,489
Alex48/poca-SoccerTwos-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,490
Alex48/poca-SoccerTwos-v4
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,491
Alex48/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,492
Alex7871/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,493
AlexAgafontsev/bert-finetuned-sem_eval-english
bert
["BertForSequenceClassification"]
30522
float32
4.38.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,494
AlexAnoshka/bge-m3
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.33.0
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
2,495
AlexAnoshka/bge-reranker-large
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.48.1
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
2,496
AlexB007/model
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
2,497
AlexC98/CodeTransLargeTFFlt
t5
["T5ForConditionalGeneration"]
32128
float32
4.28.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
2,498
AlexC98/CodeTransLargeTFNrm
t5
["T5ForConditionalGeneration"]
32128
float32
4.28.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
2,499
AlexC98/T5GenFilteredV100True
t5
["T5ForConditionalGeneration"]
32128
float32
4.28.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null