index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
500
1-lock/e0a89d83-1268-4df2-aabe-9ee13f415a71
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
501
1-lock/e6b5ce02-d1d4-4de7-9939-670b0a4973b1
llama
["LlamaForCausalLM"]
32001
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
502
1-lock/e6bbecf4-c320-4b7e-8726-9eb8f2b315a2
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
503
1-lock/e8233930-b9f6-466e-b2d1-10d15308d63f
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.46.0
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
504
1-lock/eb2e3642-b436-4b6b-a80f-5c1724e9a28c
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
505
1-lock/ebfe422c-38d4-40e3-9a02-968b865b24f5
llama
["LlamaForCausalLM"]
32000
float32
4.46.0
768
3,072
12
12
12
silu
0
False
2,048
10,000
0.000001
0.02
1
2
false
64
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.084935
small
medium
506
1-lock/ece3723c-4069-4371-988d-899a9f3862e2
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
507
1-lock/ed927b42-c64c-4cdb-8861-8272dbe881c3
llama
["LlamaForCausalLM"]
32064
bfloat16
4.46.0
3,072
8,192
32
32
32
silu
0
False
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
true
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
508
1-lock/f0e02592-fed1-457a-aad1-7cd8fbbef9f4
llama
["LlamaForCausalLM"]
32000
float16
4.46.0
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0
2
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
509
1-lock/f4a0105d-f1d8-41e3-9298-98df2db01e82
llama
["LlamaForCausalLM"]
32000
float32
4.46.0
2,048
5,504
24
16
16
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
510
1-lock/f729bb21-00df-48b4-bcfb-5de4b07c3068
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128040
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
511
1-lock/f9ed846b-6a33-4b6b-8362-24c24958fa7e
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
512
1009bmj/bert_model_out
bert
["BertForSequenceClassification"]
30000
float32
4.48.3
768
3,072
12
12
null
gelu
null
True
300
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
513
100customs/llama-3-8b-Instruct-bnb-4bit-100customs-managerwriterLBA
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
514
100rab25/hotel_images_classifier_jd_v4_convnext
swin
["SwinForImageClassification"]
null
float32
4.35.0
768
null
null
null
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
515
100ton/Qwen3-0.6B-Gensyn-Swarm-silent_screeching_jellyfish
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.51.3
1,024
3,072
28
16
8
silu
0
True
40,960
1,000,000
0.000001
0.02
151643
151645
true
64
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.352322
small
very_long
516
100yen/distilbert-base-uncased-distilled-clinc
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.35.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
517
1024khandsom/autotrain-ant-bee-3482194557
swin
["SwinForImageClassification"]
null
float32
4.25.1
1,024
null
null
null
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
518
1024m/CHIPSAL-C-GEMMA2-9B-ORPO-16bit
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.44.2
4,608
36,864
46
32
16
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
144
2
false
null
null
false
false
false
false
false
false
false
false
true
true
11.720983
large
long
519
1024m/H-Mistral-Nemo-00-fp16
mistral
["MistralForCausalLM"]
131072
bfloat16
4.47.1
5,120
14,336
40
32
8
silu
0
True
131,072
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
520
1024m/HGemma-2B-00-fp16
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.47.1
2,304
9,216
26
8
4
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
[1, 107]
null
288
2
false
null
null
false
false
false
false
false
false
false
false
true
true
1.656226
medium
long
521
1024m/HGemma-9B-00-fp16
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.47.1
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
522
1024m/HMistral-v3-00-fp16
mistral
["MistralForCausalLM"]
32768
bfloat16
4.47.1
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
523
1024m/MetaMath-7b-4bit-fp4
mistral
["MistralForCausalLM"]
32001
bfloat16
4.44.2
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
524
1024m/Mistral-Small-Instruct-4bit-fp4
mistral
["MistralForCausalLM"]
32768
bfloat16
4.44.2
6,144
16,384
56
48
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
6
false
null
null
false
false
false
false
false
false
true
false
true
true
25.367151
large
very_long
525
1024m/Mistral-Small-Instruct-4bit-nf4
mistral
["MistralForCausalLM"]
32768
bfloat16
4.44.2
6,144
16,384
56
48
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
6
false
null
null
false
false
false
false
false
false
true
false
true
true
25.367151
large
very_long
526
1024m/Qwen2.5-32B-4bit-fp4
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.44.2
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
527
1024m/aya-expanse-32b-4bit-fp4
cohere
["CohereForCausalLM"]
256000
bfloat16
4.44.2
8,192
24,576
40
64
8
silu
0
True
8,192
4,000,000
null
0.02
5
255001
null
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
32.212255
large
long
528
1024m/aya-expanse-32b-4bit-nf4
cohere
["CohereForCausalLM"]
256000
bfloat16
4.44.2
8,192
24,576
40
64
8
silu
0
True
8,192
4,000,000
null
0.02
5
255001
null
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
32.212255
large
long
529
1024m/aya-expanse-8b-4bit-fp4
cohere
["CohereForCausalLM"]
256000
bfloat16
4.44.2
4,096
14,336
32
32
8
silu
0
True
8,192
10,000
null
0.02
5
255001
null
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
6.442451
medium
long
530
1024m/aya-expanse-8b-4bit-nf4
cohere
["CohereForCausalLM"]
256000
bfloat16
4.44.2
4,096
14,336
32
32
8
silu
0
True
8,192
10,000
null
0.02
5
255001
null
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
6.442451
medium
long
531
1038lab/Prompt-Enhance
t5
["T5ForConditionalGeneration"]
32128
float32
4.42.4
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
532
1038lab/RMBG-2.0
null
["BiRefNet"]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
533
10Prem09/finetuned_Qwen2.5_Coder_0.5B_Instruct
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.51.3
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
534
10jqk1/Mistral-7B-Instruct-ADDTOKEN
mistral
["MistralForCausalLM"]
46095
float32
4.38.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
535
10jqk1/mistralai-hf-q8_0.gguf
mixtral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
536
10jqk1/mistralai-hf.gguf
mixtral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
537
10ths/FADS_base_model
xgenmm
["XGenMMModelForConditionalGeneration"]
null
bfloat16
4.41.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
538
10zinten/TTS-run-25-10-2024
speecht5
["SpeechT5ForTextToSpeech"]
81
float32
4.46.0
768
null
null
null
null
gelu
0.1
True
null
null
null
0.02
0
2
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
539
1112lee/pretrained_BERT
bert
["BertForPreTraining"]
105879
float32
4.42.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
540
114-HK/ppo-SnowballTarget
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
541
1231czx/2b_1_nll_dpo_iter1_300step
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
2,048
16,384
18
8
1
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
542
1231czx/2b_sft1
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
2,048
16,384
18
8
1
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
543
1231czx/2b_sft2
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
2,048
16,384
18
8
1
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
544
1231czx/2b_sft3
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
2,048
16,384
18
8
1
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
545
1231czx/7b_510k_5e6_3epoch_sft_rs_iter1_n1_2epoch
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
546
1231czx/7b_510k_5e6_3epoch_sft_rs_iter1_n3_2epoch
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
547
1231czx/7b_510k_5e6_3epoch_sft_rs_iter1_n8_2epoch
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
548
1231czx/7b_N12_iter2_100step
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
549
1231czx/7b_N15_iter1_200step
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
550
1231czx/7b_N6_iter1_150step
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
551
1231czx/7b_beta01_iter3_step225
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
552
1231czx/7b_code_gemma_1epoch
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.2
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
553
1231czx/7b_code_gemma_3epoch
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
554
1231czx/7b_codegemma_iter1_step300
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.38.2
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
555
1231czx/7b_dpo_iter1_4e7_step200_fromsftepoch2
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
556
1231czx/7b_dpo_iter2_4e7_onpolicy_only
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
557
1231czx/7b_dpo_iter2_4e7_step300_nll
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
558
1231czx/7b_n12_iter3_step300
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
559
1231czx/7b_n6_iter2_step100
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
560
1231czx/7b_n6_iter3
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
561
1231czx/7b_ndpo_iter1_7e7_bz_64
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
562
1231czx/7b_sft1epoch
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.41.1
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
563
1231czx/9b_iter1_2e7_step400
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.42.4
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
564
1231czx/9b_iter1_2e7_step500
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.42.4
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
565
1231czx/9b_pm_3e6
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.42.2
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
566
1231czx/9b_raft_iter1
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.42.4
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
567
1231czx/bxtmp100
gemma2
["Gemma2ForCausalLM"]
256000
float32
4.42.4
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
568
1231czx/bxtmp150
gemma2
["Gemma2ForCausalLM"]
256000
float32
4.42.4
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
569
1231czx/bxtmp175
gemma2
["Gemma2ForCausalLM"]
256000
float32
4.42.4
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
570
1231czx/bxtmp200
gemma2
["Gemma2ForCausalLM"]
256000
float32
4.42.4
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
571
1231czx/bxtmp75
gemma2
["Gemma2ForCausalLM"]
256000
float32
4.42.4
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
572
1231czx/cxtmp100
gemma2
["Gemma2ForCausalLM"]
256000
float32
4.42.4
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
573
1231czx/cxtmp150
gemma2
["Gemma2ForCausalLM"]
256000
float32
4.42.4
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
574
1231czx/cxtmp200
gemma2
["Gemma2ForCausalLM"]
256000
float32
4.42.4
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
575
1231czx/cxtmp250
gemma2
["Gemma2ForCausalLM"]
256000
float32
4.42.4
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
576
1231czx/fsfrm_llama2_sft_ver2_iter1_dpo
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.2
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
577
1231czx/fsfrm_llama2_sft_ver2_iter2_dpo
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.2
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
578
1231czx/fsfrm_llama2_sft_ver2_iter3_dpo
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.2
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
579
1231czx/gg_regular_prompt_ppo_format_reward_step140
qwen2
["Qwen2ForCausalLM"]
151665
bfloat16
4.47.1
3,584
18,944
28
28
4
silu
0
False
8,192
10,000
0.000001
0.02
151644
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
long
580
1231czx/gg_regular_prompt_ppo_format_reward_step150
qwen2
["Qwen2ForCausalLM"]
151665
bfloat16
4.47.1
3,584
18,944
28
28
4
silu
0
False
8,192
10,000
0.000001
0.02
151644
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
long
581
1231czx/gg_regular_prompt_ppo_format_reward_step170
qwen2
["Qwen2ForCausalLM"]
151665
bfloat16
4.47.1
3,584
18,944
28
28
4
silu
0
False
8,192
10,000
0.000001
0.02
151644
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
long
582
1231czx/gg_regular_prompt_ppo_format_reward_step180
qwen2
["Qwen2ForCausalLM"]
151665
bfloat16
4.47.1
3,584
18,944
28
28
4
silu
0
False
8,192
10,000
0.000001
0.02
151644
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
long
583
1231czx/gg_regular_prompt_ppo_format_reward_step190
qwen2
["Qwen2ForCausalLM"]
151665
bfloat16
4.47.1
3,584
18,944
28
28
4
silu
0
False
8,192
10,000
0.000001
0.02
151644
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
long
584
1231czx/gg_regular_prompt_ppo_format_reward_step200
qwen2
["Qwen2ForCausalLM"]
151665
bfloat16
4.47.1
3,584
18,944
28
28
4
silu
0
False
8,192
10,000
0.000001
0.02
151644
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
long
585
1231czx/gg_regular_prompt_ppo_format_reward_step210
qwen2
["Qwen2ForCausalLM"]
151665
bfloat16
4.47.1
3,584
18,944
28
28
4
silu
0
False
8,192
10,000
0.000001
0.02
151644
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
long
586
1231czx/gg_regular_prompt_ppo_format_reward_step220
qwen2
["Qwen2ForCausalLM"]
151665
bfloat16
4.47.1
3,584
18,944
28
28
4
silu
0
False
8,192
10,000
0.000001
0.02
151644
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
long
587
1231czx/gg_regular_prompt_ppo_format_reward_step230
qwen2
["Qwen2ForCausalLM"]
151665
bfloat16
4.47.1
3,584
18,944
28
28
4
silu
0
False
8,192
10,000
0.000001
0.02
151644
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
long
588
1231czx/gg_regular_prompt_ppo_format_reward_step240
qwen2
["Qwen2ForCausalLM"]
151665
bfloat16
4.47.1
3,584
18,944
28
28
4
silu
0
False
8,192
10,000
0.000001
0.02
151644
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
long
589
1231czx/gg_regular_prompt_ppo_format_reward_step250
qwen2
["Qwen2ForCausalLM"]
151665
bfloat16
4.47.1
3,584
18,944
28
28
4
silu
0
False
8,192
10,000
0.000001
0.02
151644
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
long
590
1231czx/it_dpo_list_attack
llama
["LlamaForCausalLM"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
591
1231czx/it_dpo_unbiased
llama
["LlamaForCausalLM"]
128256
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
592
1231czx/kl001_numia_dpo_iter1
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.48.3
3,584
18,944
28
28
4
silu
0
False
4,096
10,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
medium
593
1231czx/kl001_numia_dpo_iter2
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.48.3
3,584
18,944
28
28
4
silu
0
False
4,096
10,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
medium
594
1231czx/kl001_numia_dpo_iter3
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.48.3
3,584
18,944
28
28
4
silu
0
False
4,096
10,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
medium
595
1231czx/kl001_numia_dpo_iter4
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.48.3
3,584
18,944
28
28
4
silu
0
False
4,096
10,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
medium
596
1231czx/kto_mask_7b_gemma_iter2_step100
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.42.2
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
597
1231czx/llama3_it_bold_and_list_attack250_v2
llama
["LlamaForSequenceClassification"]
128257
bfloat16
4.44.1
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
598
1231czx/llama3_it_dpo_list_and_bold
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.1
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
599
1231czx/llama3_it_list_attack100_v3
llama
["LlamaForSequenceClassification"]
128257
bfloat16
4.44.1
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long