index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
16,400
JasonFuriosa/test-gpt-j-6b
gptj
["GPTJForCausalLM"]
50400
float32
4.31.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
false
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
16,401
JasonTStanley/segformer-erl
segformer
["SegformerForSemanticSegmentation"]
null
float32
4.44.2
null
null
null
[1, 2, 5, 8]
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,402
JasonWangVG/Qwen2.5-1.5B-Open-R1-Distill
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.49.0.dev0
1,536
8,960
28
12
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
16,403
Jasonaron/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,404
Jasonaron/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,405
Jasonthewhale/TAO_AI
llama
["LlamaForCausalLM"]
32000
bfloat16
4.36.2
2,048
5,632
22
32
4
silu
0
True
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
16,406
Jasonthewhale/TAO_AI_3
llama
["LlamaForCausalLM"]
102400
bfloat16
4.36.2
4,096
11,008
30
32
32
silu
0
True
4,096
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.039798
medium
medium
16,407
Jasper-786/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,408
Jasper881108/whisper-medium-zh
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.27.1
null
null
24
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
16,409
JasperV13/Yehia-7B-SFT-Reasoning-preview
llama
["LlamaForCausalLM"]
64000
bfloat16
4.51.3
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.006
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
16,410
JaspervanLeuven/12k
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,411
JaspervanLeuven/controlnet_rect
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,412
Jasperyyc/urop_test
bert
["BertForTokenClassification"]
28996
float32
4.29.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
16,413
Jass07/autogen-short31
llama
["LlamaForCausalLM"]
128256
float16
4.42.0.dev0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
16,414
Jass07/autogenJune2000
llama
["LlamaForCausalLM"]
128256
float16
4.46.0.dev0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
16,415
Jass07/mux_may16
llama
["LlamaForCausalLM"]
128256
float16
4.41.0.dev0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
16,416
Jasshl/segfortmer_finetuned
segformer
["SegformerForSemanticSegmentation"]
null
float32
4.34.1
null
null
null
[1, 2, 5, 8]
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,417
Jaster111/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,418
Jathushan/TamilLyricXLM-RoBERTa
xlm-roberta
["XLMRobertaForMaskedLM"]
250052
float32
4.51.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
16,419
Jathushan/TamilPaattu_bert
bert
["BertForMaskedLM"]
197285
float32
4.51.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
16,420
Jathushan/mBART-TamilMetaphorSource
mbart
["MBartForConditionalGeneration"]
250054
float32
4.42.3
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
16,421
Jathushan/tamilbert-pos
bert
["BertForMaskedLMWithPOS"]
197285
float32
4.51.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
16,422
Jatin-WIAI/bengali_relevance_clf
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
null
4.8.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
16,423
Jatin-WIAI/malayalam_relevance_clf
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
null
4.8.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
16,424
Jatin-WIAI/marathi_relevance_clf
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
null
4.8.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
16,425
JatinkInnovision/ComFit
llama
["LlamaForCausalLM"]
128256
bfloat16
4.51.3
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
16,426
JatinkInnovision/snowflake-arctic-embed-l-v2.0_all-nli
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.47.1
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
16,427
JatinkInnovision/test
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
16,428
Jaume/OrpoPhi-3-mini
phi3
["Phi3ForCausalLM"]
32013
float16
4.41.2
3,072
8,192
32
32
32
silu
0
True
131,072
10,000
0.00001
0.02
32011
32012
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
16,429
JavaEdge/qwen-rotten-tomatoes
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.51.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
16,430
Javeria98/DONUT_REAL_DATA_FINAL
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.32.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,431
Javeria98/donut-base-Medical_Handwritten_Prescriptions_Information_Extraction
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,432
Javeria98/donut-base-Medical_Handwritten_Prescriptions_Information_Extraction_1
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,433
Javeria98/donut-base-Medical_Handwritten_Prescriptions_Information_Extraction_Final
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,434
Javeria98/donut-base-Medical_Model
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,435
Javeria98/donut-base-Medical_Model_final
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,436
Javeria98/donut_final_REAL_DATA
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.32.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,437
Javeria98/donut_final_REAL_DATA_02
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.32.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,438
Javeria98/donut_final_REAL_DATA_03
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.32.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,439
Javeria98/donut_final_REAL_DATA_04
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.32.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,440
JaviManobanda/flan-t5-base-proverb-es-classification-text-out
t5
["T5ForConditionalGeneration"]
32128
float32
4.44.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
16,441
JavierAnd/cliente360_class_weighted_model
distilbert
["DistilBertForSequenceClassification"]
31002
float32
4.27.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
16,442
JavierC03/lunar_lander
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,443
JavierPS/qwen2_5-coder-finetunne
qwen2
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,444
JavierPS/qwen3-finetuned
qwen3
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,445
Javitron4257/modelo_clase1610
vit
["ViTForImageClassification"]
null
float32
4.22.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
16,446
Jawa7/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,447
Jawadah1/english-sindhi-translator
marian
["MarianMTModel"]
64110
float32
4.51.3
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
16,448
Jawaher/Covid19-fake-news-bert-uncased
bert
["BertForMaskedLM"]
30522
float32
4.19.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
16,449
JawaraL/community-gpt2
gpt2
["GPT2LMHeadModel"]
50257
float32
4.53.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
16,450
Jawaria/wav2vec2-large-xls-r-300m-pashto-colab-test-4
wav2vec2
["Wav2Vec2ForCTC"]
63
float32
4.29.2
1,024
4,096
24
16
null
gelu
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
16,451
Jawaria/wav2vec2-large-xls-r-300m-pashto-colab-test-5
wav2vec2
["Wav2Vec2ForCTC"]
63
float32
4.29.2
1,024
4,096
24
16
null
gelu
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
16,452
Jaxon/DialoGPT-medium-kirito
gpt2
["GPT2LMHeadModel"]
50257
float32
4.28.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
16,453
Jay-C/distilbert-base-uncased-distilled-clinc
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.34.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
16,454
Jay-C/distilbert-base-uncased-finetuned-clinc
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.34.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
16,455
JayBDev/distilbert-base-uncased-finetuned-imdb
distilbert
["DistilBertForMaskedLM"]
30522
float32
4.38.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
16,456
JayChauhan99/llama-2-7b-minideven
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
16,457
JayChauhan99/llama-2-7b-minideven-full
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
16,458
JayChauhan99/llama-2-7b-minideven-non-chat
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
16,459
JayHyeon/QWEN_toy-DPOP_5e-5-100ep_0alp_5lam
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,460
JayHyeon/Qwen2-0.5B-Reward_1e-3
qwen2
["Qwen2ForSequenceClassification"]
151936
float32
4.46.3
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,461
JayHyeon/Qwen2-0.5B-Reward_5e-3
qwen2
["Qwen2ForSequenceClassification"]
151936
float32
4.46.3
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,462
JayHyeon/Qwen2-0.5B-Reward_VPO_1e-3
qwen2
["Qwen2ForSequenceClassification"]
151936
float32
4.46.3
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,463
JayHyeon/Qwen2-0.5B-Reward_VPO_5e-3
qwen2
["Qwen2ForSequenceClassification"]
151936
float32
4.46.3
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,464
JayHyeon/Qwen2.5-0.5B-SFT-1e-4
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,465
JayHyeon/Qwen2.5-0.5B-SFT-1e-5
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,466
JayHyeon/Qwen2.5-0.5B-SFT-1e-5-3ep
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,467
JayHyeon/Qwen2.5-0.5B-SFT-2e-4
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,468
JayHyeon/Qwen2.5-0.5B-SFT-2e-5
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,469
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-DPOP_3e-7-3ep_0alp_5lam
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,470
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-DPO_1e-6-1ep_0alp_0lam
qwen2
["Qwen2Model"]
151936
float32
4.47.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,471
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-DPO_2e-6-1ep_0alp_0lam
qwen2
["Qwen2Model"]
151936
float32
4.47.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,472
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-DPO_2e-6-2ep_0alp_0lam
qwen2
["Qwen2Model"]
151936
float32
4.47.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,473
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-DPO_3e-6-1ep_0alp_0lam
qwen2
["Qwen2Model"]
151936
float32
4.47.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,474
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-DPO_3e-6-2ep_0alp_0lam
qwen2
["Qwen2Model"]
151936
float32
4.47.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,475
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-DPO_3e-6-3ep_0alp_0lam
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,476
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-DPO_5e-6-1ep_0alp_0lam
qwen2
["Qwen2Model"]
151936
float32
4.47.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,477
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-DPO_5e-6-2ep_0alp_0lam
qwen2
["Qwen2Model"]
151936
float32
4.47.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,478
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-DPO_5e-7_2ep_0alp_0lam
qwen2
["Qwen2Model"]
151936
float32
4.47.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,479
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-DPO_7e-7_1ep_0alp_0lam
qwen2
["Qwen2Model"]
151936
float32
4.47.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,480
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-DPO_7e-7_3ep_0alp_0lam
qwen2
["Qwen2Model"]
151936
float32
4.47.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,481
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-IRPO_5e-7-3ep_1alp_0lam
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,482
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-MDPO_2e-6_1ep_0alp_0lam
qwen2
["Qwen2Model"]
151936
float32
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,483
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-MDPO_3e-6-1ep_0alp_0lam
qwen2
["Qwen2Model"]
151936
float32
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,484
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-MDPO_5e-6-1ep_0alp_0lam
qwen2
["Qwen2Model"]
151936
float32
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,485
JayHyeon/Qwen2.5-0.5B-SFT-2e-5-2ep-MDPO_5e-6-2ep_0alp_0lam
qwen2
["Qwen2Model"]
151936
float32
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,486
JayHyeon/Qwen2.5-0.5B-SFT-5e-5
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,487
JayHyeon/Qwen2.5-0.5B-SFT-7e-5
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,488
JayHyeon/Qwen2.5-0.5B-SFT-DPO-1epoch_v1
qwen2
["Qwen2Model"]
151936
float32
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,489
JayHyeon/Qwen2.5-0.5B-SFT_2ep
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,490
JayHyeon/Qwen_0.5-BDPO_5e-7-3ep_0alp_0.5bdpo_lam_0dpop_lam
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.52.4
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,491
JayHyeon/Qwen_0.5-DPOP_5e-5-100ep_0alp_5lam
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,492
JayHyeon/Qwen_0.5-DPO_3e-6-3ep_0alp_0lam
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,493
JayHyeon/Qwen_0.5-DPO_5e-5-100ep_0alp_0lam
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,494
JayHyeon/Qwen_0.5-DPO_5e-7-3ep_0alp_0lam
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,495
JayHyeon/Qwen_0.5-IPO_1e-6-1ep_0vpo_const
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,496
JayHyeon/Qwen_0.5-IPO_5e-7-3ep
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.0.dev0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,497
JayHyeon/Qwen_0.5-IRPO_1e-6-3ep_0.1alp_0.5bdpo_lam_0dpop_lam
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.52.4
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,498
JayHyeon/Qwen_0.5-IRPO_1e-6-3ep_10alp_0.5bdpo_lam_0dpop_lam
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.52.4
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,499
JayHyeon/Qwen_0.5-IRPO_3e-6-3ep_1alp_0lam
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long