index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
1,000
A01794620/distilbert-base-uncased-finetuned-emotion-short-sets
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.41.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
1,001
A0B0C/Flacon
RefinedWebModel
["RWForCausalLM"]
65024
bfloat16
4.27.4
4,544
null
null
null
null
null
0
True
null
null
null
0.02
11
11
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,002
A2Amir/SF_A68_IDEFICS_9B_IDL_SFT
idefics
["IdeficsForVisionText2Text"]
32000
float32
4.44.2
4,096
11,008
32
32
null
silu
null
True
2,048
null
0.000001
0.02
1
2
false
128
null
false
null
null
false
false
false
false
false
false
false
false
false
false
6.442451
medium
medium
1,003
A5medAshraf/Magicllama
llama
["LlamaForCausalLM"]
32000
bfloat16
4.38.1
4,096
11,008
32
32
32
silu
0
True
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
1,004
A7it/kjnn
roberta
["RobertaForSequenceClassification"]
50265
float32
4.47.0
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
1,005
A7m0d/distilhubert-finetuned-gtzan
hubert
["HubertForSequenceClassification"]
32
float32
4.45.1
768
3,072
2
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.014156
small
null
1,006
AAA01101312/distilbert-base-uncased-finetuned-clinc
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.39.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
1,007
AAA01101312/pegasus-samsum
pegasus
["PegasusForConditionalGeneration"]
96103
float32
4.39.0
null
null
16
null
null
null
0.1
True
1,024
null
null
null
0
1
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
1,008
AAA01101312/xlm-roberta-base-finetuned-panx-de-fr
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.39.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,009
AAA01101312/xlm-roberta-base-finetuned-panx-en
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.39.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,010
AAA01101312/xlm-roberta-base-finetuned-panx-fr
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.39.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,011
AAA01101312/xlm-roberta-base-finetuned-panx-it
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.39.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,012
AAAI2025/MathSpeech_Ablation_Study_1stage_fine-tuned_with_errors_T5_base
t5
["T5ForConditionalGeneration"]
32106
float32
4.44.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,013
AAAI2025/MathSpeech_Ablation_Study_1stage_fine-tuned_with_errors_T5_small
t5
["T5ForConditionalGeneration"]
32106
float32
4.44.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,014
AAAI2025/MathSpeech_Ablation_Study_Error_corrector_T5_base
t5
["T5ForConditionalGeneration"]
32106
float32
4.44.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,015
AAAI2025/MathSpeech_Ablation_Study_Error_corrector_T5_small
t5
["T5ForConditionalGeneration"]
32106
float32
4.44.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,016
AAAI2025/MathSpeech_Ablation_Study_LaTeX_translator_T5_base
t5
["T5ForConditionalGeneration"]
32106
float32
4.44.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,017
AAAI2025/MathSpeech_Ablation_Study_LaTeX_translator_T5_small
t5
["T5ForConditionalGeneration"]
32106
float32
4.44.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,018
AAAndyZ/JD_Qwen2.5_7B_lora
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.1
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
1,019
AAAndyZ/JD_Qwen2.5_7B_lora_20250225
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.1
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
1,020
AAK1423/wav2vec2-ALC-output
wav2vec2
["Wav2Vec2ForCTC"]
95
float32
4.51.3
1,024
4,096
24
16
null
gelu
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
1,021
AAK1423/wave2vecWvald15Epoch
wav2vec2
["Wav2Vec2ForCTC"]
95
float32
4.51.2
1,024
4,096
24
16
null
gelu
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
1,022
AAbduallah1/Finetuned-Llama-3.1-8B
llama
["LlamaForCausalLM"]
128256
float16
4.53.0
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
1,023
AAkhilesh/wav2vec2-large-xls-r-300m-hsb-colab
wav2vec2
["Wav2Vec2ForCTC"]
49
float32
4.18.0
1,024
4,096
24
16
null
gelu
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
1,024
AAkhilesh/wav2vec2-large-xls-r-300m-ta-colab
wav2vec2
["Wav2Vec2ForCTC"]
53
float32
4.18.0
1,024
4,096
24
16
null
gelu
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
1,025
ABHIiiii1/mt5-Finetuned-Bi-En-Mn-trans
mt5
["MT5ForConditionalGeneration"]
250102
float32
4.41.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,026
ABHIiiii1/mt5-Finetuned-Bi-En-Mn-trans2
mt5
["MT5ForConditionalGeneration"]
250102
float32
4.41.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,027
ABX-AI/Infinite-Laymons-9B
mistral
["MistralForCausalLM"]
32000
float16
4.39.1
4,096
14,336
40
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
8.053064
large
very_long
1,028
ABX-AI/Infinite-Sumika-9B
mistral
["MistralForCausalLM"]
32000
float16
4.39.1
4,096
14,336
40
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
8.053064
large
very_long
1,029
ABX-AI/Kuno-Lemon-7B
mistral
["MistralForCausalLM"]
32000
bfloat16
4.39.1
4,096
14,336
32
32
8
silu
0
True
8,192
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
long
1,030
ABX-AI/Laymonade-7B
mistral
["MistralForCausalLM"]
32000
bfloat16
4.39.1
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
1,031
ABX-AI/Norosumika-7B
mistral
["MistralForCausalLM"]
32000
bfloat16
4.39.1
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
1,032
ABX-AI/Quantum-Citrus-9B
mistral
["MistralForCausalLM"]
32000
float16
4.39.1
4,096
14,336
40
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
8.053064
large
very_long
1,033
ABX-AI/Silver-Sun-v2-11B
llama
["LlamaForCausalLM"]
32000
bfloat16
4.39.1
4,096
14,336
48
32
8
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
9.663676
large
medium
1,034
ABX-AI/Starfinite-Laymospice-v2-7B
mistral
["MistralForCausalLM"]
32000
bfloat16
4.39.1
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
1,035
ABrinkmann/bert-finetuned-ner
bert
["BertForTokenClassification"]
28996
float32
4.30.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,036
ABrinkmann/bert-finetuned-ner-10epochs
bert
["BertForTokenClassification"]
28996
float32
4.30.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,037
ABrinkmann/bert-finetuned-ner-3epochs
bert
["BertForTokenClassification"]
28996
float32
4.30.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,038
ABrinkmann/deberta-v3-large-ad-opentag-finetuned-ner-5epochs
deberta-v2
["DebertaV2ForTokenClassification"]
128100
float32
4.30.1
1,024
4,096
24
16
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
1,039
ABrinkmann/deberta-v3-large-finetuned-ner-10epochs
deberta-v2
["DebertaV2ForTokenClassification"]
128100
float32
4.30.1
1,024
4,096
24
16
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
1,040
ABrinkmann/deberta-v3-large-finetuned-ner-10epochs-V2
deberta-v2
["DebertaV2ForTokenClassification"]
128100
float32
4.30.2
1,024
4,096
24
16
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
1,041
ABrinkmann/sbert_xtremedistil-l6-h256-uncased-mean-cosine-h32
bert
["BertModel"]
30533
float32
4.17.0
256
1,024
6
8
null
gelu
null
True
512
null
null
0.02
null
null
null
32
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.004719
small
short
1,042
ACATECH/ncos
llama
["LlamaForCausalLM"]
128256
bfloat16
4.47.1
8,192
28,672
80
64
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128009
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
1,043
ACCORD-NLP/ner-albert-large
albert
["AlbertForTokenClassification"]
30000
float32
4.16.2
1,024
4,096
24
16
null
gelu_new
null
null
512
null
null
0.02
2
3
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
1,044
ACCORD-NLP/ner-roberta-large-lm
roberta
["RobertaForTokenClassification"]
50265
float32
4.16.2
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
1,045
ACCORD-NLP/re-albert-large
albert
["AlbertForSequenceClassification"]
30004
float32
4.16.2
1,024
4,096
24
16
null
gelu_new
null
null
512
null
null
0.02
2
3
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
1,046
ACCORD-NLP/re-bert-large
bert
["BertForSequenceClassification"]
29000
float32
4.16.2
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
1,047
ACE-Step/ACE-Step-v1-chinese-rap-LoRA
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,048
ACIDE/User-VLM-3B-base
paligemma
["PaliGemmaForConditionalGeneration"]
null
bfloat16
4.48.0.dev0
2,048
null
26
null
null
null
null
null
null
null
null
null
2
1
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
1.308623
medium
null
1,049
ADHIZ/causal-omni_thanvi
gpt2
["GPT2LMHeadModel"]
50257
float32
4.46.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,050
ADHIZ/code-omni_kali
gpt2
["GPT2LMHeadModel"]
50257
float32
4.46.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,051
ADHIZ/code-omni_madasu
gpt2
["GPT2LMHeadModel"]
50257
float32
4.46.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,052
ADHIZ/image_omni_netz
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,053
ADI10HERO/PPO_LunarLander_v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,054
ADP-EPFL/GPT2-Einstein-EPFL
gpt2
["GPT2LMHeadModel"]
50257
float32
4.40.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,055
ADRIANRICO/Distilbert-finetuned-IMDB
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.34.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
1,056
ADSah/billsum
t5
["T5ForConditionalGeneration"]
32128
float32
4.52.4
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,057
ADT7/SN11-EMP-5000-VAL-QL
llama
["LlamaForCausalLM"]
128258
bfloat16
4.46.3
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
1,058
ADT7/SN11-EMP-6000-Val-Data-QL-v2-rename
llama
["LlamaForCausalLM"]
128258
bfloat16
4.46.3
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
1,059
ADT7/SN11-MK-500-Q-v2
llama
["LlamaForCausalLM"]
128258
bfloat16
4.46.3
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
1,060
ADT7/SN11-NM-FAT-5000-QLORA-v1
mistral
["MistralForCausalLM"]
131072
bfloat16
4.48.3
5,120
32,768
40
32
8
silu
0
True
32,768
100,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
1,061
ADT7/SN11-T-100-Q
llama
["LlamaForCausalLM"]
128258
bfloat16
4.46.3
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
1,062
AELLM/gemma-2-aeria-9b
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.43.1
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
1,063
AELLM/gemma-2-inko-9b
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.43.1
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
107
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
1,064
AFSA1729/movie-classifier
bert
["BertForSequenceClassification"]
30522
float32
4.41.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,065
AFZAL0008/bert-finetuned-ner
bert
["BertForTokenClassification"]
28996
float32
4.40.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,066
AGBonnet/medinote-13b
llama
["LlamaForCausalLM"]
32019
float32
4.36.0
5,120
13,824
40
40
40
silu
0
True
2,048
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
1,067
AGBonnet/medinote-7b
llama
["LlamaForCausalLM"]
32019
float32
4.36.0
4,096
11,008
32
32
32
silu
0
True
2,048
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
1,068
AGCobra/EXAONE-Deep-32B-mlx-4Bit
exaone
["ExaoneForCausalLM"]
102400
bfloat16
4.43.1
5,120
27,392
null
40
8
null
0
True
32,768
1,000,000
null
0.02
1
361
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
null
null
very_long
1,069
AGI-CEO/ppo-PyramidsV1
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,070
AGI-CEO/ppo-SnowballTarget
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,071
AGnatkiv/distilbert-reviews
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.35.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
1,072
AGresse/ast-finetuned-audioset-10-10-0.4593-finetuned-rfcx-reprod
audio-spectrogram-transformer
["ASTForAudioClassification"]
null
float32
4.48.0
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
1,073
AGudden/xlm-roberta-base-finetuned-marc
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.25.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,074
AGundawar/moondream-ft
moondream1
["Moondream"]
null
float16
4.41.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,075
AGuzhvenko/ppo-Pyramids
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,076
AHDMK/Sentence-GISTEmbedLoss-BioBert-Allnli-scinli
bert
["BertModel"]
28996
float32
4.38.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,077
AHMED13245/xlm-roberta-base-finetuned-panx-de
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.44.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,078
AHarbury/debiasNLPFinal
distilbert
["DistilBertForSequenceClassification"]
30522
null
4.29.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
1,079
AHedya/Al-Fatiha-Recitation-Model
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.39.3
null
null
12
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
1,080
AHedya/w2v-bert-fine-tuning
wav2vec2-bert
["Wav2Vec2BertForCTC"]
14
float32
4.40.2
1,024
4,096
24
16
null
swish
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
null
1,081
AHedya/whisper-small-dv
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.39.3
null
null
12
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
1,082
AI-Enthusiast11/pii-entity-extractor
deberta-v2
["DebertaV2ForTokenClassification"]
251000
float32
4.51.1
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,083
AI-MO/Kimina-Autoformalizer-7B
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.49.0
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
1,084
AI-MO/Kimina-Prover-Preview-Distill-7B
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.49.0
3,584
18,944
28
28
4
silu
0
False
16,384
100,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
long
1,085
AI-MO/NuminaMath-7B-CoT
llama
["LlamaForCausalLM"]
102400
float32
4.42.3
4,096
11,008
30
32
32
silu
0
True
4,096
10,000
0.000001
0.02
100000
100001
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.039798
medium
medium
1,086
AI-MO/NuminaMath-7B-TIR
llama
["LlamaForCausalLM"]
102400
bfloat16
4.42.3
4,096
11,008
30
32
32
silu
0
True
4,096
10,000
0.000001
0.02
100000
100001
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.039798
medium
medium
1,087
AI-Roadmap/SmolLM2-135M-rm-60k
llama
["RewardModel"]
49152
bfloat16
4.51.3
576
1,536
30
9
3
silu
0
False
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
1,088
AI-Sweden-Models/Llama-3-8B
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0.dev0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
1,089
AI-Sweden-Models/ModernBERT-large
modernbert
["ModernBertForMaskedLM"]
50368
null
4.56.2
1,024
2,624
28
16
null
null
0
null
8,192
null
null
0.02
50281
50282
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
true
0.352322
small
long
1,090
AI-Sweden-Models/bert-large-nordic-pile-1M-steps
bert
["BertForMaskedLM"]
64000
float32
4.31.0.dev0
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
1,091
AI-Sweden-Models/gpt-sw3-1.3b-instruct
gpt2
["GPT2LMHeadModel"]
64000
float32
4.22.1
null
null
null
null
null
null
null
True
null
null
null
0.01
2
3
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,092
AI-Sweden-Models/gpt-sw3-126m-instruct
gpt2
["GPT2LMHeadModel"]
64000
float32
4.22.1
null
null
null
null
null
null
null
True
null
null
null
0.023
2
3
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,093
AI-Sweden-Models/gpt-sw3-20b-instruct
gpt2
["GPT2LMHeadModel"]
64000
float32
4.22.1
null
null
null
null
null
null
null
True
null
null
null
0.008165
2
3
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,094
AI-Sweden-Models/gpt-sw3-20b-instruct-4bit-gptq
gpt2
["GPT2LMHeadModel"]
64000
float16
4.36.0.dev0
null
null
null
null
null
null
null
True
null
null
null
0.008165
2
3
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,095
AI-Sweden-Models/gpt-sw3-356m-instruct
gpt2
["GPT2LMHeadModel"]
64000
float32
4.22.1
null
null
null
null
null
null
null
True
null
null
null
0.01
2
3
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,096
AI-Sweden-Models/gpt-sw3-6.7b-v2
gpt2
["GPT2LMHeadModel"]
64000
float32
4.22.1
null
null
null
null
null
null
null
True
null
null
null
0.01
1
1
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,097
AI-Sweden-Models/gpt-sw3-6.7b-v2-instruct
gpt2
["GPT2LMHeadModel"]
64000
float32
4.22.1
null
null
null
null
null
null
null
True
null
null
null
0.01
1
1
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,098
AI-newbie89/codeparrot-ds
gpt2
["GPT2LMHeadModel"]
50000
float32
4.41.1
null
null
null
null
null
null
null
True
null
null
null
0.02
0
0
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,099
AI-newbie89/marian-finetuned-kde4-en-to-fr
marian
["MarianMTModel"]
59514
float32
4.41.1
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short