index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
4,300
ArtusDev/TheDrummer_Valkyrie-49B-v1_EXL3_4.5bpw_H6
nemotron-nas
["DeciLMForCausalLM"]
128256
bfloat16
4.51.3
8,192
null
80
64
null
silu
0
True
131,072
500,000
0.00001
0.02
128000
128009
false
128
null
false
null
null
false
false
false
false
false
false
false
false
false
true
64.424509
large
very_long
4,301
ArtusDev/huihui-ai_Qwen3-30B-A3B-abliterated_EXL3_4.0bpw_H6
qwen3_moe
["Qwen3MoeForCausalLM"]
151936
bfloat16
4.51.3
2,048
6,144
48
32
4
silu
0
True
40,960
1,000,000
0.000001
0.02
151643
151645
false
64
8
false
null
8
false
false
false
false
false
false
false
false
true
true
2.415919
medium
very_long
4,302
ArtusDev/knifeayumu_Behemoth-v1.2-Magnum-v4-123B_EXL3_1.8bpw_H5
mistral
["MistralForCausalLM"]
32768
bfloat16
4.46.2
12,288
28,672
88
96
8
silu
0
True
131,072
1,000,000
0.00001
0.02
1
2
false
128
12
false
null
null
false
false
false
false
false
false
true
false
true
true
159.450661
xlarge
very_long
4,303
ArtusDev/remnant-ep1-mrg
glm4
["Glm4ForCausalLM"]
151552
bfloat16
4.52.0.dev0
6,144
23,040
61
48
2
silu
0
True
32,768
10,000
0.00001
0.02
null
[151329, 151336, 151338]
false
128
24
false
null
null
false
false
false
false
false
false
false
false
true
true
27.632075
large
very_long
4,304
ArtusDev/remnant-ep2-mrg
glm4
["Glm4ForCausalLM"]
151552
bfloat16
4.52.0.dev0
6,144
23,040
61
48
2
silu
0
True
32,768
10,000
0.00001
0.02
null
[151329, 151336, 151338]
false
128
24
false
null
null
false
false
false
false
false
false
false
false
true
true
27.632075
large
very_long
4,305
Artvv/Markov_chain_modelization
qwen2
["Qwen2ForCausalLM"]
152064
float16
4.48.3
3,584
18,944
28
28
4
silu
0
True
131,072
10,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
4,306
Artvv/Markov_chain_modelization_gguf
qwen2
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,307
Arumugam-G/driver-license-document-parsing-method-3
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.51.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,308
ArunAIML/bert-model-intent-classification
bert
["BertForSequenceClassification"]
30522
float32
4.52.4
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,309
ArunAIML/content
bert
["BertForSequenceClassification"]
30522
float32
4.52.4
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,310
ArunAIML/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,311
ArunAIML/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,312
ArunKr/RL-LunarLander
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,313
ArunSharmaaaaa/ceollama
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
4,314
ArunaSaraswathy/bert-finetuned-ner
bert
["BertForTokenClassification"]
28996
float32
4.27.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,315
ArunaSaraswathy/bert-finetuned-ner-pii
bert
["BertForTokenClassification"]
28996
float32
4.27.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,316
ArunaSaraswathy/tbert_ner_test
distilbert
["DistilBertForTokenClassification"]
30522
float32
4.17.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
4,317
Arunark/finetuning-sentiment-model-3000-samples
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.46.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
4,318
Arunavaonly/Bangla-twoclass-Sentiment-Analyzer
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.37.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,319
Arunavaonly/Bangla_multiclass_sentiment_analysis_model
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.34.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,320
ArunkumarSA/tamil_llama3.1_model
llama
["LlamaForCausalLM"]
128256
float16
4.48.3
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
4,321
Arushp1/finetuned_distilbert
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.47.0
null
null
null
null
null
null
0.5
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
4,322
ArvindSharma18/gemma-2-9b-it-bnb-4bit-book-review
gemma2
["Gemma2ForCausalLM"]
256000
float16
4.42.3
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
4,323
ArwaAbdul/convnext_inreching20
convnext
["ConvNextForImageClassification"]
null
float32
4.34.1
null
null
null
null
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,324
AryPratap/XLM-roberta-HIEN-Humor-detection
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.31.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,325
AryPratap/t5-hinglish-to-en
t5
["T5ForConditionalGeneration"]
32128
float32
4.31.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,326
Arya-Bastani23/swin-tiny-patch4-window7-224-finetuned-eurosat
swin
["SwinForImageClassification"]
null
float32
4.28.0
768
null
null
null
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,327
Arya20032705/arya
blenderbot
["BlenderbotForConditionalGeneration"]
8008
float32
4.44.2
null
null
2
null
null
null
0
True
128
null
null
null
1
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
4,328
AryaParikh/autotrain-text_summary_arp-45146113306
t5
["T5ForConditionalGeneration"]
32128
float32
4.25.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,329
AryaParikh/summ_arp_org
t5
["T5ForConditionalGeneration"]
32128
float32
4.25.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,330
Aryaman9999/freshness2
resnet
["ResNetForImageClassification"]
null
float32
4.45.0
null
null
null
null
null
relu
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,331
AryamanJaggi/Llama
llama
["LlamaForCausalLM"]
128256
bfloat16
4.41.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,332
AryamanJaggi/Vanilla-Llama
llama
["LlamaForCausalLM"]
128256
bfloat16
4.41.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,333
Aryan-401/detr-resnet-50-license-plate-20bb
detr
["DetrForObjectDetection"]
null
float32
4.40.1
null
null
6
null
null
null
0
null
1,024
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
4,334
Aryan-401/speecht5_finetuned_voxpopuli_nl
speecht5
["SpeechT5ForTextToSpeech"]
81
float32
4.35.2
768
null
null
null
null
gelu
0.1
False
null
null
null
0.02
0
2
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,335
Aryan-401/whisper-tiny-english-minds14
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.35.2
null
null
4
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
4,336
Aryan-401/yolo-tiny-fashion
yolos
["YolosForObjectDetection"]
null
float32
4.40.1
192
768
12
3
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.005308
small
null
4,337
AryanKKate/legal-bert-model
bert
["BertForSequenceClassification"]
30522
float32
4.47.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
0
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,338
AryanManakame/my_awesome_billsum_model
t5
["T5ForConditionalGeneration"]
32128
float32
4.27.4
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,339
AryanNsc/Mistral-7b-4bit
mistral
["MistralForCausalLM"]
32000
float16
4.41.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
4,340
AryanNsc/gemma_ft-0.1
gemma
["GemmaForCausalLM"]
256000
float16
4.41.2
2,048
16,384
18
8
1
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
4,341
Aryangp/text_summarization_aryangp_uiet
t5
["T5ForConditionalGeneration"]
32128
float32
4.46.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,342
Aryanikale23/Signlanguage
videomae
["VideoMAEForVideoClassification"]
null
float32
4.38.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
4,343
Aryanne/Astrea-RP-v1-4B
stablelm_epoch
["StableLMEpochForCausalLM"]
50304
bfloat16
4.34.0
2,560
6,912
48
32
32
silu
null
True
4,096
10,000
null
0.02
0
0
false
80
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.774874
medium
medium
4,344
Aryanne/Astrea-RP-v1.5-3B
stablelm_epoch
["StableLMEpochForCausalLM"]
50304
float16
4.35.2
2,560
6,912
32
32
32
silu
null
True
4,096
10,000
null
0.02
0
0
false
80
1
false
null
null
false
false
false
false
false
false
false
false
false
true
2.516582
medium
medium
4,345
Aryanne/CalderaAI_Hexoteric-7B-F16
mistral
["MistralForCausalLM"]
32002
float16
4.38.2
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
4,346
Aryanne/MixSwap
mistral
["MistralForCausalLM"]
32000
bfloat16
4.38.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
4,347
Aryanne/TinyllamaMix-1.1B
llama
["LlamaForCausalLM"]
32000
bfloat16
4.35.2
2,048
5,632
22
32
4
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
very_long
4,348
Aryanne/Westest-7B
mistral
["MistralForCausalLM"]
32000
bfloat16
4.35.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
4,349
Aryanne/phitest
phi-msft
["PhiForCausalLM"]
51200
float16
4.35.2
null
null
null
null
null
null
null
True
null
null
null
0.02
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,350
Aryanow/my_awesome_model_bert
bert
["BertForSequenceClassification"]
30522
float32
4.26.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,351
Aryanpro321/flan-t5-base-multilingual-sentiment-classification
t5
["T5ForConditionalGeneration"]
32128
float32
4.46.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,352
Aryn/deformable-detr-DocLayNet
deformable_detr
["DeformableDetrForObjectDetection"]
null
float32
4.36.2
null
null
null
null
null
null
0
null
1,024
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
4,353
Asad321/phi-final-research-paper-tuned
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.40.1
3,072
8,192
32
32
32
silu
0
True
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
4,354
Asad515/Llama3.1_Audit_Chatml
llama
["LlamaForCausalLM"]
128256
float16
4.44.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
4,355
Asadali12/vivit-b-16x2-kinetics400-Finetune_01feb25
vivit
["VivitForVideoClassification"]
null
float32
4.47.0
768
3,072
12
12
null
gelu_fast
null
null
null
null
null
0.01
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
4,356
Asael123/models_for_ner
bert
["BertForTokenClassification"]
28996
float32
4.44.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,357
Asakrg/hausa-text-to-speech-fine-tune
vits
["VitsModel"]
34
float32
4.33.0.dev0
192
null
6
2
null
relu
0.1
null
null
null
null
0.02
null
null
null
96
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.002654
small
null
4,358
Asankhla01/prakriti-ai-model
t5
["T5ForConditionalGeneration"]
32128
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,359
Asankilp/gochiusa-sovits4.0
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,360
Asanokurokai/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,361
Asanokurokai/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,362
Asap7772/SmolLM2-135M-Instruct-fast
llama
["LlamaForCausalLM"]
51200
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
4,363
Asap7772/grpo_connections_4b_0702_step_100
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.1
2,560
9,728
36
32
8
silu
0
True
40,960
1,000,000
0.000001
0.02
151643
151645
true
80
4
false
null
null
false
false
false
false
false
false
false
false
true
true
2.831155
medium
very_long
4,364
Asap7772/grpo_connections_4b_0702_step_200
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.1
2,560
9,728
36
32
8
silu
0
True
40,960
1,000,000
0.000001
0.02
151643
151645
true
80
4
false
null
null
false
false
false
false
false
false
false
false
true
true
2.831155
medium
very_long
4,365
Asap7772/hintgen-qwen3-4b-sft-1e-5
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.51.3
4,096
12,288
36
32
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
4,366
Asap7772/hintgen-qwen3-4b-sft-1e-5init-star
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.51.3
4,096
12,288
36
32
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
4,367
Asap7772/hintgen-qwen3-4b-sft-5e-6
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.51.3
4,096
12,288
36
32
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
4,368
Asap7772/hintgen-qwen3-4b-sft-5e-6init-star
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.51.3
4,096
12,288
36
32
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
4,369
Asap7772/hintgen-qwen3-4b-sft-5e-7
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.51.3
4,096
12,288
36
32
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
4,370
Asap7772/hintgen-qwen3-4b-sft-5e-7init-star
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.51.3
4,096
12,288
36
32
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
4,371
Asap7772/insight-warmstart-sft-qwen25-3b-3epoch-0501
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.51.3
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
4,372
Asap7772/prm_datamath-mc-full_objbce_lr0.0005_epoch0
llama
["LlamaForCausalLM"]
128261
float32
4.45.2
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
4,373
Asap7772/qwen05b_lr1e5_5ep_binned_sft
qwen2
["Qwen2ForCausalLM"]
153713
float32
4.47.1
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
4,374
Asap7772/qwen3-8b-hint-ipo-0.01-1e-5
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.51.3
4,096
12,288
36
32
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
4,375
Asap7772/qwen3-8b-hint-ipo-0.01-5e-7
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.51.3
4,096
12,288
36
32
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
4,376
Asap7772/qwen3-8b-hint-rpo-0.1-0.1-1e-5init
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.51.3
4,096
12,288
36
32
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
4,377
Asap7772/qwen3-8b-hint-rpo-0.1-0.1-1e-6init
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.51.3
4,096
12,288
36
32
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
4,378
Asap7772/qwen3-8b-hint-rpo-0.1-0.1-1e-7init
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.51.3
4,096
12,288
36
32
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
4,379
Asap7772/qwen3-8b-hint-rpo-0.1-0.1-5e-6init
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.51.3
4,096
12,288
36
32
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
4,380
Asap7772/qwen3_4blrablation_filtered_0503_lr1e5
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.51.3
2,560
9,728
36
32
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
true
80
4
false
null
null
false
false
false
false
false
false
false
false
true
true
2.831155
medium
very_long
4,381
Asap7772/smollm2_lr1e4_5ep_binned_sft
llama
["LlamaForCausalLM"]
51200
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
4,382
Asap7772/smollm2_lr1e4_7ep_binned_sft
llama
["LlamaForCausalLM"]
51200
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
4,383
Asap7772/smollm2_lr1e5_10ep_binned_sft
llama
["LlamaForCausalLM"]
51200
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
4,384
Asap7772/smollm2_lr1e5_2ep_binned_sft
llama
["LlamaForCausalLM"]
51200
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
4,385
Asap7772/smollm2_lr1e5_5ep_binned_sft
llama
["LlamaForCausalLM"]
51200
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
4,386
Asap7772/smollm2_lr1e5_7ep_binned_sft
llama
["LlamaForCausalLM"]
51200
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
4,387
Asap7772/smollm2_lr5e5_10ep_binned_sft
llama
["LlamaForCausalLM"]
51200
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
4,388
Asap7772/smollm2_lr5e5_2ep_binned_sft
llama
["LlamaForCausalLM"]
51200
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
4,389
Asap7772/smollm2_lr5e5_5ep_binned_sft
llama
["LlamaForCausalLM"]
51200
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
4,390
Asap7772/smollm2_lr5e5_7ep_binned_sft
llama
["LlamaForCausalLM"]
51200
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
4,391
Asap7772/smollm2_lr5e6_10ep_binned_sft
llama
["LlamaForCausalLM"]
51200
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
4,392
Asap7772/smollm2_lr5e6_2ep_binned_sft
llama
["LlamaForCausalLM"]
51200
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
4,393
Asap7772/smollm2_lr5e6_5ep_binned_sft
llama
["LlamaForCausalLM"]
51200
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
4,394
Asap7772/smollm2_lr5e6_7ep_binned_sft
llama
["LlamaForCausalLM"]
51200
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
4,395
Asarkar07/llama-2-7b-miniguanaco
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
4,396
Ascire/my_awesome_eli5_mlm_model
roberta
["RobertaForMaskedLM"]
50265
float32
4.45.2
768
3,072
6
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.042467
small
short
4,397
Aseelalzaben03/sadaqai-bestmodel
bert
["BertForSequenceClassification"]
64000
float32
4.51.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,398
AsemBadr/final-whisper-for-initial-publish
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.41.0.dev0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
4,399
AsemBadr/final-whisper-for-initial-publish-v2
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.41.0.dev0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null