index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
500 | 1-lock/e0a89d83-1268-4df2-aabe-9ee13f415a71 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.46.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
501 | 1-lock/e6b5ce02-d1d4-4de7-9939-670b0a4973b1 | llama | ["LlamaForCausalLM"] | 32001 | bfloat16 | 4.46.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
502 | 1-lock/e6bbecf4-c320-4b7e-8726-9eb8f2b315a2 | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.46.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
503 | 1-lock/e8233930-b9f6-466e-b2d1-10d15308d63f | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.46.0 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
504 | 1-lock/eb2e3642-b436-4b6b-a80f-5c1724e9a28c | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.46.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
505 | 1-lock/ebfe422c-38d4-40e3-9a02-968b865b24f5 | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.46.0 | 768 | 3,072 | 12 | 12 | 12 | silu | 0 | False | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 64 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.084935 | small | medium |
506 | 1-lock/ece3723c-4069-4371-988d-899a9f3862e2 | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.46.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
507 | 1-lock/ed927b42-c64c-4cdb-8861-8272dbe881c3 | llama | ["LlamaForCausalLM"] | 32064 | bfloat16 | 4.46.0 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | False | 131,072 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 3.623879 | medium | very_long |
508 | 1-lock/f0e02592-fed1-457a-aad1-7cd8fbbef9f4 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.46.0 | 16 | 64 | 2 | 4 | 4 | silu | 0 | False | 2,048 | 10,000 | 0.000001 | 0.02 | 0 | 2 | false | 4 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.000006 | small | medium |
509 | 1-lock/f4a0105d-f1d8-41e3-9298-98df2db01e82 | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.46.0 | 2,048 | 5,504 | 24 | 16 | 16 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
510 | 1-lock/f729bb21-00df-48b4-bcfb-5de4b07c3068 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.46.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128040 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
511 | 1-lock/f9ed846b-6a33-4b6b-8362-24c24958fa7e | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.46.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
512 | 1009bmj/bert_model_out | bert | ["BertForSequenceClassification"] | 30000 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 300 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
513 | 100customs/llama-3-8b-Instruct-bnb-4bit-100customs-managerwriterLBA | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
514 | 100rab25/hotel_images_classifier_jd_v4_convnext | swin | ["SwinForImageClassification"] | null | float32 | 4.35.0 | 768 | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
515 | 100ton/Qwen3-0.6B-Gensyn-Swarm-silent_screeching_jellyfish | qwen3 | ["Qwen3ForCausalLM"] | 151936 | float32 | 4.51.3 | 1,024 | 3,072 | 28 | 16 | 8 | silu | 0 | True | 40,960 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.352322 | small | very_long |
516 | 100yen/distilbert-base-uncased-distilled-clinc | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.35.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
517 | 1024khandsom/autotrain-ant-bee-3482194557 | swin | ["SwinForImageClassification"] | null | float32 | 4.25.1 | 1,024 | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
518 | 1024m/CHIPSAL-C-GEMMA2-9B-ORPO-16bit | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.44.2 | 4,608 | 36,864 | 46 | 32 | 16 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 144 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 11.720983 | large | long |
519 | 1024m/H-Mistral-Nemo-00-fp16 | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.47.1 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 131,072 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
520 | 1024m/HGemma-2B-00-fp16 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.47.1 | 2,304 | 9,216 | 26 | 8 | 4 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | [1, 107] | null | 288 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.656226 | medium | long |
521 | 1024m/HGemma-9B-00-fp16 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.47.1 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
522 | 1024m/HMistral-v3-00-fp16 | mistral | ["MistralForCausalLM"] | 32768 | bfloat16 | 4.47.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
523 | 1024m/MetaMath-7b-4bit-fp4 | mistral | ["MistralForCausalLM"] | 32001 | bfloat16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
524 | 1024m/Mistral-Small-Instruct-4bit-fp4 | mistral | ["MistralForCausalLM"] | 32768 | bfloat16 | 4.44.2 | 6,144 | 16,384 | 56 | 48 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 6 | false | null | null | false | false | false | false | false | false | true | false | true | true | 25.367151 | large | very_long |
525 | 1024m/Mistral-Small-Instruct-4bit-nf4 | mistral | ["MistralForCausalLM"] | 32768 | bfloat16 | 4.44.2 | 6,144 | 16,384 | 56 | 48 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 6 | false | null | null | false | false | false | false | false | false | true | false | true | true | 25.367151 | large | very_long |
526 | 1024m/Qwen2.5-32B-4bit-fp4 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.44.2 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
527 | 1024m/aya-expanse-32b-4bit-fp4 | cohere | ["CohereForCausalLM"] | 256000 | bfloat16 | 4.44.2 | 8,192 | 24,576 | 40 | 64 | 8 | silu | 0 | True | 8,192 | 4,000,000 | null | 0.02 | 5 | 255001 | null | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 32.212255 | large | long |
528 | 1024m/aya-expanse-32b-4bit-nf4 | cohere | ["CohereForCausalLM"] | 256000 | bfloat16 | 4.44.2 | 8,192 | 24,576 | 40 | 64 | 8 | silu | 0 | True | 8,192 | 4,000,000 | null | 0.02 | 5 | 255001 | null | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 32.212255 | large | long |
529 | 1024m/aya-expanse-8b-4bit-fp4 | cohere | ["CohereForCausalLM"] | 256000 | bfloat16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 10,000 | null | 0.02 | 5 | 255001 | null | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
530 | 1024m/aya-expanse-8b-4bit-nf4 | cohere | ["CohereForCausalLM"] | 256000 | bfloat16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 10,000 | null | 0.02 | 5 | 255001 | null | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
531 | 1038lab/Prompt-Enhance | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.42.4 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
532 | 1038lab/RMBG-2.0 | null | ["BiRefNet"] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
533 | 10Prem09/finetuned_Qwen2.5_Coder_0.5B_Instruct | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.51.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
534 | 10jqk1/Mistral-7B-Instruct-ADDTOKEN | mistral | ["MistralForCausalLM"] | 46095 | float32 | 4.38.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
535 | 10jqk1/mistralai-hf-q8_0.gguf | mixtral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
536 | 10jqk1/mistralai-hf.gguf | mixtral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
537 | 10ths/FADS_base_model | xgenmm | ["XGenMMModelForConditionalGeneration"] | null | bfloat16 | 4.41.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
538 | 10zinten/TTS-run-25-10-2024 | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.46.0 | 768 | null | null | null | null | gelu | 0.1 | True | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
539 | 1112lee/pretrained_BERT | bert | ["BertForPreTraining"] | 105879 | float32 | 4.42.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
540 | 114-HK/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
541 | 1231czx/2b_1_nll_dpo_iter1_300step | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 2,048 | 16,384 | 18 | 8 | 1 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
542 | 1231czx/2b_sft1 | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 2,048 | 16,384 | 18 | 8 | 1 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
543 | 1231czx/2b_sft2 | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 2,048 | 16,384 | 18 | 8 | 1 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
544 | 1231czx/2b_sft3 | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 2,048 | 16,384 | 18 | 8 | 1 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
545 | 1231czx/7b_510k_5e6_3epoch_sft_rs_iter1_n1_2epoch | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
546 | 1231czx/7b_510k_5e6_3epoch_sft_rs_iter1_n3_2epoch | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
547 | 1231czx/7b_510k_5e6_3epoch_sft_rs_iter1_n8_2epoch | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
548 | 1231czx/7b_N12_iter2_100step | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
549 | 1231czx/7b_N15_iter1_200step | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
550 | 1231czx/7b_N6_iter1_150step | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
551 | 1231czx/7b_beta01_iter3_step225 | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
552 | 1231czx/7b_code_gemma_1epoch | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.2 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
553 | 1231czx/7b_code_gemma_3epoch | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
554 | 1231czx/7b_codegemma_iter1_step300 | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.38.2 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
555 | 1231czx/7b_dpo_iter1_4e7_step200_fromsftepoch2 | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
556 | 1231czx/7b_dpo_iter2_4e7_onpolicy_only | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
557 | 1231czx/7b_dpo_iter2_4e7_step300_nll | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
558 | 1231czx/7b_n12_iter3_step300 | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
559 | 1231czx/7b_n6_iter2_step100 | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
560 | 1231czx/7b_n6_iter3 | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
561 | 1231czx/7b_ndpo_iter1_7e7_bz_64 | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
562 | 1231czx/7b_sft1epoch | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.41.1 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
563 | 1231czx/9b_iter1_2e7_step400 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.42.4 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
564 | 1231czx/9b_iter1_2e7_step500 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.42.4 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
565 | 1231czx/9b_pm_3e6 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.42.2 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
566 | 1231czx/9b_raft_iter1 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.42.4 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
567 | 1231czx/bxtmp100 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | float32 | 4.42.4 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
568 | 1231czx/bxtmp150 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | float32 | 4.42.4 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
569 | 1231czx/bxtmp175 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | float32 | 4.42.4 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
570 | 1231czx/bxtmp200 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | float32 | 4.42.4 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
571 | 1231czx/bxtmp75 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | float32 | 4.42.4 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
572 | 1231czx/cxtmp100 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | float32 | 4.42.4 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
573 | 1231czx/cxtmp150 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | float32 | 4.42.4 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
574 | 1231czx/cxtmp200 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | float32 | 4.42.4 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
575 | 1231czx/cxtmp250 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | float32 | 4.42.4 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
576 | 1231czx/fsfrm_llama2_sft_ver2_iter1_dpo | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
577 | 1231czx/fsfrm_llama2_sft_ver2_iter2_dpo | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
578 | 1231czx/fsfrm_llama2_sft_ver2_iter3_dpo | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
579 | 1231czx/gg_regular_prompt_ppo_format_reward_step140 | qwen2 | ["Qwen2ForCausalLM"] | 151665 | bfloat16 | 4.47.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 151644 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | long |
580 | 1231czx/gg_regular_prompt_ppo_format_reward_step150 | qwen2 | ["Qwen2ForCausalLM"] | 151665 | bfloat16 | 4.47.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 151644 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | long |
581 | 1231czx/gg_regular_prompt_ppo_format_reward_step170 | qwen2 | ["Qwen2ForCausalLM"] | 151665 | bfloat16 | 4.47.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 151644 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | long |
582 | 1231czx/gg_regular_prompt_ppo_format_reward_step180 | qwen2 | ["Qwen2ForCausalLM"] | 151665 | bfloat16 | 4.47.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 151644 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | long |
583 | 1231czx/gg_regular_prompt_ppo_format_reward_step190 | qwen2 | ["Qwen2ForCausalLM"] | 151665 | bfloat16 | 4.47.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 151644 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | long |
584 | 1231czx/gg_regular_prompt_ppo_format_reward_step200 | qwen2 | ["Qwen2ForCausalLM"] | 151665 | bfloat16 | 4.47.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 151644 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | long |
585 | 1231czx/gg_regular_prompt_ppo_format_reward_step210 | qwen2 | ["Qwen2ForCausalLM"] | 151665 | bfloat16 | 4.47.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 151644 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | long |
586 | 1231czx/gg_regular_prompt_ppo_format_reward_step220 | qwen2 | ["Qwen2ForCausalLM"] | 151665 | bfloat16 | 4.47.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 151644 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | long |
587 | 1231czx/gg_regular_prompt_ppo_format_reward_step230 | qwen2 | ["Qwen2ForCausalLM"] | 151665 | bfloat16 | 4.47.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 151644 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | long |
588 | 1231czx/gg_regular_prompt_ppo_format_reward_step240 | qwen2 | ["Qwen2ForCausalLM"] | 151665 | bfloat16 | 4.47.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 151644 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | long |
589 | 1231czx/gg_regular_prompt_ppo_format_reward_step250 | qwen2 | ["Qwen2ForCausalLM"] | 151665 | bfloat16 | 4.47.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 151644 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | long |
590 | 1231czx/it_dpo_list_attack | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.43.4 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
591 | 1231czx/it_dpo_unbiased | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.43.4 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
592 | 1231czx/kl001_numia_dpo_iter1 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 4,096 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | medium |
593 | 1231czx/kl001_numia_dpo_iter2 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 4,096 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | medium |
594 | 1231czx/kl001_numia_dpo_iter3 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 4,096 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | medium |
595 | 1231czx/kl001_numia_dpo_iter4 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 4,096 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | medium |
596 | 1231czx/kto_mask_7b_gemma_iter2_step100 | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.42.2 | 3,072 | 24,576 | 28 | 16 | 16 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
597 | 1231czx/llama3_it_bold_and_list_attack250_v2 | llama | ["LlamaForSequenceClassification"] | 128257 | bfloat16 | 4.44.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
598 | 1231czx/llama3_it_dpo_list_and_bold | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
599 | 1231czx/llama3_it_list_attack100_v3 | llama | ["LlamaForSequenceClassification"] | 128257 | bfloat16 | 4.44.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.