index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
6,800 | Chaanan/vicuna-7b-v1.5-W8A8-Dynamic-Per-Token | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.44.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,801 | Chabuno/finetuning-sentiment-model-3000-samples | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.24.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,802 | Chaconne/llm-detect-ai-generated-text-bert | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.35.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,803 | Chaeeun-Kim/L_pretrain_recover_tokenizer_pooling_ep110 | mt5 | ["MT5ForConditionalGeneration"] | 50134 | float32 | 4.43.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,804 | Chaeeun-Kim/L_pretrain_recover_tokenizer_trimmed | mt5 | ["MT5ForConditionalGeneration"] | 119196 | float32 | 4.43.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,805 | Chaeeun-Kim/L_subset_pretrain_encdec | mt5 | ["MT5ForConditionalGeneration"] | 250113 | float32 | 4.31.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,806 | Chaeeun-Kim/L_subset_pretrain_encdec_shuffle | mt5 | ["MT5ForConditionalGeneration"] | 250113 | float32 | 4.31.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,807 | Chaeeun-Kim/private_unsup_ep1 | mt5 | ["MT5ForConditionalGeneration"] | 250106 | float32 | 4.46.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,808 | Chaeeun-Kim/spider-large-pretrain-2019 | dpr | ["DPRContextEncoder"] | 30522 | float32 | 4.41.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | short |
6,809 | Chaeguevara/PPO-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,810 | Chaeseung/log2char_Orion-14B-Base_v2 | orion | ["OrionForCausalLM"] | 84608 | float32 | 4.37.1 | 5,120 | 15,360 | 40 | 40 | 40 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,811 | Chaeseung/log2char_llama3_EN_v1 | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.37.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,812 | Chaeyeon/first-patent-tokenizer | bert | ["BertModel"] | 160518 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,813 | Chaeyeon/git-base-sketch | git | ["GitForCausalLM"] | 30522 | float32 | null | 768 | 3,072 | 6 | 12 | null | gelu | null | True | 1,024 | null | null | 0.02 | 101 | 102 | false | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.042467 | small | short |
6,814 | ChafikAiEng/multilingual-e5-base-finetuned-studytours | xlm-roberta | ["XLMRobertaModel"] | 250002 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,815 | Chahat7874/hindi_turkish | wav2vec2 | ["Wav2Vec2ForCTC"] | 70 | float32 | 4.41.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
6,816 | Chahatdatascience/config-0 | wav2vec2 | ["Wav2Vec2ForCTC"] | 106 | float32 | 4.41.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
6,817 | Chahatdatascience/config-1 | wav2vec2 | ["Wav2Vec2ForCTC"] | 106 | float32 | 4.41.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.15 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
6,818 | Chahatdatascience/wav2vec2-large-xls-r-300m-hindi_marathi-colab-v1-20epochs | wav2vec2 | ["Wav2Vec2ForCTC"] | 106 | float32 | 4.41.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
6,819 | Chaitanya798800/videomae-large_Sports_action_recognition_5 | videomae | ["VideoMAEForVideoClassification"] | null | float32 | 4.39.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
6,820 | Chaitanya895/Sahayak_Vexa_AI_fine_tuned_model | mpnet | ["MPNetForQuestionAnswering"] | 30527 | null | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,821 | Chakshu/test-2fd961bc-aec9-4cec-8f2a-9e5ce676127a | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.31.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,822 | Chakshu/test-470446d9-2c78-4af9-80f1-fd17bf2c6275 | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.31.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,823 | Chakshu/test-f3a3c123-9ecd-4ba1-8ac3-3fd8d64863dd | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.31.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,824 | ChakuChidiya/ppo-Lunar-Lander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,825 | ChallengerSpaceShuttle/imdb-gpt-test | null | [] | 50257 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,826 | Chamoda/Llama-2-7b-chat-fin24 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,827 | Chan-007/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,828 | Chan2chan1/solar_test240517_4bit | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.39.2 | 4,096 | 14,336 | 48 | 32 | 8 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 9.663676 | large | medium |
6,829 | ChanE-666/BERT-8bit | bert | ["BertLMHeadModel"] | 30522 | float16 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,830 | ChanE-666/phi-2-QLora | phi | ["PhiForCausalLM"] | 51200 | float16 | 4.38.1 | 2,560 | 10,240 | 32 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | 50256 | 50256 | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
6,831 | ChanE-666/phi2-bnb-8bit | phi | ["PhiForCausalLM"] | 51200 | float16 | 4.38.1 | 2,560 | 10,240 | 32 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | 50256 | 50256 | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
6,832 | ChanP/finetuned-th-to-en | marian | ["MarianMTModel"] | 62307 | float32 | 4.15.0 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,833 | ChandanGR/Qwen2.5-0.5B-4bit-channel | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float16 | 4.46.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
6,834 | ChandanGR/llama-3.2-1B-gptq-4bit-channel | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.46.3 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
6,835 | Chandana03/Llama-2-7b-chat-hf.legal-FineTune | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.37.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,836 | Chandans01/custom-chandan-samsum | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.44.0 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,837 | ChandlerU11/t5_fine_2.0 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.28.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,838 | Chandlernnnnnn/ESM-finetuned-sequence-reconstruction | esm | ["EsmForTokenClassification"] | 33 | float32 | 4.30.2 | 480 | 1,920 | 12 | 20 | null | gelu | null | True | 1,026 | null | null | 0.02 | null | null | null | 24 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.033178 | small | short |
6,839 | ChandraAdhikarla/Quantised_Model | llama | ["LlamaForCausalLM"] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,840 | ChandraAdhikarla/TTSmodel | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,841 | Chandrababu-Namani/InLegalBERT-fine-tuned | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | 0 | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,842 | Chandrateja/lamma3-8b-oig-unsloth-merged | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,843 | Chang-Hoo/gemma-3-1b-cpt-dpo-0707 | gemma3_text | ["Gemma3ForCausalLM"] | 262144 | bfloat16 | 4.51.3 | 1,152 | 6,912 | 26 | 4 | 1 | null | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | 2 | 1 | null | 288 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.414056 | small | very_long |
6,844 | Chang-Hoo/gemma-3-1b-cpt-it-0707 | gemma3_text | ["Gemma3ForCausalLM"] | 262144 | bfloat16 | 4.51.3 | 1,152 | 6,912 | 26 | 4 | 1 | null | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | 2 | 1 | null | 288 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.414056 | small | very_long |
6,845 | ChangeIsKey/text-dating | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.36.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
6,846 | Changgil/K2S3-27b-v0.0 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.42.3 | 4,608 | 36,864 | 46 | 32 | 16 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 144 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 11.720983 | large | long |
6,847 | Changgil/K2S3-Mistral-7b-v1.2 | mistral | ["MistralForCausalLM"] | 48000 | float16 | 4.38.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,848 | Changgil/K2S3-Mistral-7b-v1.48 | mistral | ["MistralForCausalLM"] | 48000 | float16 | 4.38.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,849 | Changgil/K2S3-Mistral-7bx2-48layers_v1.2 | mistral | ["MistralForCausalLM"] | 48000 | float16 | 4.38.0.dev0 | 4,096 | 14,336 | 48 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 9.663676 | large | very_long |
6,850 | Changgil/K2S3-v0.1 | mistral | ["MistralForCausalLM"] | 48000 | float16 | 4.38.0.dev0 | 4,096 | 14,336 | 64 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.884902 | large | very_long |
6,851 | Changmin0816/1008model | bert | ["BertForSequenceClassification"] | 30000 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 300 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,852 | Changmin0816/ViT_beans | vit | ["ViTForImageClassification"] | null | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
6,853 | Chanjeans/tfchatbot_2 | gpt2 | ["GPT2LMHeadModel"] | 51204 | float32 | 4.48.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 0 | 1 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,854 | Chanjeans/tfchatbot_3 | gpt2 | ["GPT2LMHeadModel"] | 51204 | float32 | 4.48.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 0 | 1 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,855 | Chansik/whisper-large-v3-ko-aihub | whisper | ["WhisperForConditionalGeneration"] | 51866 | float32 | 4.45.1 | null | null | 32 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
6,856 | ChaoChao2023/RoBERTa_no_sft | roberta | ["RobertaForSequenceClassification"] | 50265 | null | 4.35.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
6,857 | ChaoChao2023/RoBERTa_squad | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.38.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
6,858 | ChaoHou/ESMDance | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,859 | ChaoHou/SeqDance | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,860 | Chaoqi2000/BestDNA6mer | bert | ["BertForMaskedLM"] | 4101 | float32 | 4.29.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,861 | ChaoqunZhao/sd-class-butterflies-32 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,862 | ChaosAiVision/DeepSeek-R1-0528-Qwen3-8B-vi-sft-medical-9k | qwen3 | ["Qwen3ForCausalLM"] | 151936 | bfloat16 | 4.52.4 | 4,096 | 12,288 | 36 | 32 | 8 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 7.247757 | large | very_long |
6,863 | ChaoticNeutrals/Bepis_9B | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.38.1 | 4,096 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 8.053064 | large | very_long |
6,864 | ChaoticNeutrals/BuRP_7B | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,865 | ChaoticNeutrals/Captain-Eris_Violet_Toxic-Magnum-12B | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.48.2 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
6,866 | ChaoticNeutrals/Cookie_7B | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.35.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | long |
6,867 | ChaoticNeutrals/Eris_Prime-V2-7B | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,868 | ChaoticNeutrals/Eris_PrimeV4-Vision-32k-7B | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,869 | ChaoticNeutrals/Eris_PrimeV4.20-Vision-32k-7B | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,870 | ChaoticNeutrals/Hathor_Respawn-L3-8B-v0.8 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.38.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,871 | ChaoticNeutrals/Pocky_9B | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.37.2 | 4,096 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 8,192 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 8.053064 | large | long |
6,872 | ChaoticNeutrals/Poppy_Porpoise-v0.6-L3-8B | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.39.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,873 | ChaoticNeutrals/Puppy_Purpose_0.69 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.39.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,874 | ChaoticNeutrals/Templar_v1_8B | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.41.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,875 | CharGen-Archive/CharGen-v3-beta-244-s6 | mistral | ["MistralForCausalLM"] | 32768 | float32 | 4.45.2 | 6,144 | 16,384 | 56 | 48 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 6 | false | null | null | false | false | false | false | false | false | true | false | true | true | 25.367151 | large | very_long |
6,876 | CharGen-Archive/CharGen-v3-beta-264-s0 | mistral | ["MistralForCausalLM"] | 32768 | bfloat16 | 4.46.1 | 6,144 | 16,384 | 56 | 48 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 6 | false | null | null | false | false | false | false | false | false | true | false | true | true | 25.367151 | large | very_long |
6,877 | CharGen/CharGen-v3-mini | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.51.0 | 3,072 | 9,216 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128019 | false | 96 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.623879 | medium | very_long |
6,878 | CharacterEcho/Rohit-Sharma | stablelm | ["StableLmForCausalLM"] | 50280 | float16 | 4.42.3 | 2,560 | 6,912 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | null | 0.02 | 50278 | 50279 | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
6,879 | CharanSaiVaddi/BERT-finetuned-TextSummarization | bert | ["BertLMHeadModel"] | 30522 | float32 | 4.47.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,880 | Charankarnati18/TASK3 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,881 | Charankarnati18/TASK4 | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.48.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,882 | Charles-59800/six_sentiment_distillbert_fine_tune | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.30.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,883 | Charles333/gpt2_fintuned_json_data | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.35.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,884 | CharlesLi/CoT-SFT-3B | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.51.1 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
6,885 | CharlesLi/Direct-SFT-3B | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.51.1 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
6,886 | CharlesLi/G1-7B | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.51.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
6,887 | CharlesLi/OpenELM-1_1B-DPO-full-max-reward-least-similar | openelm | ["OpenELMForCausalLM"] | 32000 | bfloat16 | 4.45.1 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
6,888 | CharlesLi/OpenELM-1_1B-DPO-full-max-reward-most-similar | openelm | ["OpenELMForCausalLM"] | 32000 | bfloat16 | 4.45.1 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
6,889 | CharlesLi/OpenELM-1_1B-DPO-full-most-similar | openelm | ["OpenELMForCausalLM"] | 32000 | bfloat16 | 4.45.1 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
6,890 | CharlesLi/OpenELM-1_1B-SFT-3 | openelm | ["OpenELMForCausalLM"] | 32000 | bfloat16 | 4.44.2 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
6,891 | CharlesLi/dapo_5_epoch_graph_task_3B_1150 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.51.1 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
6,892 | CharlesLi/dapo_5_epoch_graph_task_3B_800 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.51.1 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
6,893 | CharlesLi/graph_grpo_320 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.50.1 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
6,894 | CharlesLi/grpo_5_epoch_graph_hard_base_gpu_700 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.51.1 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
6,895 | CharlesLi/grpo_5_epoch_graph_hard_base_gpu_800 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.50.1 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
6,896 | CharlesLi/grpo_5_epoch_graph_hard_base_start_200 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.50.1 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
6,897 | CharlesLi/grpo_5_epoch_graph_hard_base_start_400 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.50.1 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
6,898 | CharlesLi/grpo_5_epoch_graph_hard_ins_gpu_400 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.50.1 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
6,899 | CharlesLi/grpo_5_epoch_graph_hard_ins_gpu_700 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.51.1 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.