index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
6,300 | ByteExplorer/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,301 | ByteExplorer/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,302 | ByteFlow-AI/DetailFlow-16 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,303 | ByteFlow-AI/TokenFlow-t2i | llava_llama | ["LlavaLlamaForCausalLM"] | 64769 | bfloat16 | 4.43.4 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,304 | ByteForge/Defog_llama-3-sqlcoder-8b-ct2-int8_float16 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,305 | ByteForge/meta_llama-3-8b_Instruct-ct2-int8_float16 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,306 | ByteMaster27/SentimentModel | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.47.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
6,307 | ByteWave/Cheus-11B | mistral | ["MistralForCausalLM"] | 32002 | bfloat16 | 4.35.2 | 4,096 | 14,336 | 48 | 32 | 8 | silu | null | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 9.663676 | large | very_long |
6,308 | ByteWave/prompt-generator | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.33.1 | 3,200 | 8,640 | 26 | 32 | 32 | silu | null | True | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 100 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 3.19488 | medium | medium |
6,309 | BytedTsinghua-SIA/DAPO-Qwen-32B | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.47.1 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 131,072 | 1,000,000 | 0.00001 | 0.02 | null | 151643 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
6,310 | BytedanceDouyinContent/SAIL-VL-1d6-8B | internvl_chat | ["SailVLModel"] | null | bfloat16 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
6,311 | Bytes512/Nugget | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.37.1 | 6,656 | 17,920 | 120 | 52 | 52 | silu | 0 | True | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 63.795364 | large | medium |
6,312 | Bytes512/Nugget-Small | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.37.1 | 6,656 | 17,920 | 60 | 52 | 52 | silu | 0 | True | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 31.897682 | large | medium |
6,313 | Bytte/TestedLunar | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,314 | Bytte/lunar-rl | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,315 | Byungchae/k2s3_test_0001 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.34.1 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,316 | C-HAN7/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-bristly_frisky_marmot | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.51.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
6,317 | C-Ilyas/whisper-base-darija | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.41.2 | null | null | 6 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
6,318 | C-Lo/finetuning-sentiment-unfiltered-dataset | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.28.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,319 | C-Stuti/Lber2_full_data | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | true | 0.084935 | small | short |
6,320 | C-Stuti/my_finetuned_wnut_model_1012 | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,321 | C-Stuti/output | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.36.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,322 | C-n/marian-finetuned-kde4-en-to-fr | marian | ["MarianMTModel"] | 59514 | float32 | 4.35.2 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,323 | C0uchP0tat0/bert-base-cased-model | bert | ["BertForMaskedLM"] | 28996 | float32 | 4.30.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,324 | C0uchP0tat0/dummy-model | camembert | ["CamembertForMaskedLM"] | 32005 | float32 | 4.30.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 5 | 6 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,325 | C10X/Distill_whisper-small_tr | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.51.3 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
6,326 | C10X/Distill_whisper-small_tr2 | whisper | ["WhisperForConditionalGeneration"] | 51866 | float32 | 4.51.3 | null | null | 32 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
6,327 | C10X/stella_en_v5-Distill | model2vec | ["StaticModel"] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,328 | C10X/whisper-tiny-tr | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.50.3 | null | null | 4 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
6,329 | C1pher/DeepScaleR-1.5B-Preview_Qwen2.5-Math-1.5B_0_-1_-0.5_1.5_32_GRPOK3 | qwen2_merge | ["Qwen2MergeForCausalLM"] | 151936 | bfloat16 | 4.50.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | False | 131,072 | 10,000 | 0.000001 | 0.02 | 151646 | 151643 | false | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
6,330 | C1pher/o1_pruner_like_1p5B_deepscaler | qwen2_merge | ["Qwen2MergeForCausalLM"] | 151936 | bfloat16 | 4.51.3 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | False | 131,072 | 10,000 | 0.000001 | 0.02 | 151646 | 151643 | false | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
6,331 | C1pher/o1_pruner_like_1p5B_deepseek | qwen2_merge | ["Qwen2MergeForCausalLM"] | 151936 | bfloat16 | 4.51.3 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | False | 131,072 | 10,000 | 0.000001 | 0.02 | 151646 | 151643 | false | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
6,332 | C4Scale/deberta-v3-base_finetuned_bluegennx_run2 | deberta-v2 | ["DebertaV2ForTokenClassification"] | 128100 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,333 | C5i/NatSight-bart-base-wikisql | bart | ["BartForConditionalGeneration"] | 50266 | float32 | 4.25.1 | null | null | 6 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,334 | C5i/NatSight-t5-small-wikisql | t5 | ["T5ForConditionalGeneration"] | 32101 | float32 | 4.25.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,335 | CABSEL/glycobart | bart | ["BartForConditionalGeneration"] | 10080 | float32 | 4.44.0 | null | null | 12 | null | null | null | 0 | True | 512 | null | null | null | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,336 | CABSEL/glycobert | bert | ["BertForSequenceClassification"] | 10010 | float32 | 4.44.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,337 | CAI2001/model16bit | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.46.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
6,338 | CAI2001/model_gguf | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,339 | CALDISS-AAU/DA-BERT_Old_News_V1 | bert | ["BertForMaskedLM"] | 30522 | float32 | 4.49.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,340 | CAMeL-Lab/arabart-qalb14-gec-ged-13 | mbart | ["MBartForConditionalGeneration"] | 50002 | float32 | 4.22.2 | null | null | 6 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,341 | CAMeL-Lab/arabart-qalb15-gec-ged-13 | mbart | ["MBartForConditionalGeneration"] | 50002 | float32 | 4.22.2 | null | null | 6 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,342 | CAMeL-Lab/arabart-zaebuc-gec-ged-13 | mbart | ["MBartForConditionalGeneration"] | 50002 | float32 | 4.22.2 | null | null | 6 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,343 | CAMeL-Lab/arat5-coda-did | t5 | ["T5ForConditionalGeneration"] | 110100 | float32 | 4.22.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,344 | CAMeL-Lab/camelbert-msa-qalb15-ged-13 | bert | ["BertForTokenClassificationSingleLabel"] | 30000 | float32 | 4.22.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,345 | CAMeL-Lab/camelbert-msa-zaebuc-ged-13 | bert | ["BertForTokenClassificationSingleLabel"] | 30000 | float32 | 4.22.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,346 | CAS-SIAT-XinHai/mllmChem10K | qwen2_vl | ["Qwen2VLForConditionalGeneration"] | 152064 | bfloat16 | 4.50.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
6,347 | CASIA-LM/OpenS2S | null | ["OmniSpeechModel"] | null | bfloat16 | 4.51.0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
6,348 | CASIA-LM/Taichu-70B | llama | ["LlamaForCausalLM"] | 50000 | bfloat16 | 4.33.3 | 8,192 | 28,672 | 80 | 64 | 8 | silu | null | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | medium |
6,349 | CATIE-AQ/FAT5-small | flash_t5 | ["FlashT5ForConditionalGeneration"] | 32768 | float32 | 4.46.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,350 | CATIE-AQ/Moderncamembert_3entities | modernbert | ["ModernBertForTokenClassification"] | 32768 | float32 | 4.51.3 | 768 | 1,152 | 22 | 12 | null | null | 0 | null | 8,192 | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | true | 0.155714 | small | long |
6,351 | CAWANG2/SmolLM2-FT-MyDataset | llama | ["LlamaForCausalLM"] | 49152 | float32 | 4.51.3 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 8,192 | 100,000 | 0.00001 | 0.041667 | 1 | 2 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | long |
6,352 | CAiRE/UniVaR-lambda-1 | nomic_bert | ["NomicBertModel"] | 30528 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | null |
6,353 | CAiRE/UniVaR-lambda-20 | nomic_bert | ["NomicBertModel"] | 30528 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | null |
6,354 | CAiRE/UniVaR-lambda-5 | nomic_bert | ["NomicBertModel"] | 30528 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | null |
6,355 | CAiRE/UniVaR-lambda-80 | nomic_bert | ["NomicBertModel"] | 30528 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | null |
6,356 | CBroglio/mt5-small-finetuned-amazon-en-es | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.46.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,357 | CBucci/my_awesome_billsum_model | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.32.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,358 | CC-AI-Labs/nord-triplet-hsm-bert-base-uncased | bert | ["BertModel"] | 30522 | float32 | 4.36.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,359 | CC-AI-Labs/nord-triplet-hsm-bert-base-uncased-2025-01 | bert | ["BertModel"] | 30522 | float32 | 4.45.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,360 | CC-AI-Labs/sharks-triplet-hsm-bert-base-uncased-2025-04 | bert | ["BertModel"] | 30522 | float32 | 4.36.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,361 | CC-AI-Labs/sharks-uncased-bert-featurizer | bert | ["BertModel"] | 30522 | float32 | 4.30.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,362 | CCB/abstracts_to_tweet_model | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,363 | CCCCCCCCY/WundtLLaMA | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,364 | CCF2P/test | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,365 | CCHoward/vit_finetuned_food101 | vit | ["ViTForImageClassification"] | null | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
6,366 | CDRI-Eddy/checkpoints | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.35.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,367 | CDRI-Eddy/ingredients_classification_model_test | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.35.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,368 | CEBangu/Phi-3-mini-128k-instruct-GPTChoice-DiffSort-4.6k-DPO | phi3 | ["Phi3ForCausalLM"] | 32064 | float16 | 4.41.2 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 131,072 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | very_long |
6,369 | CED6688/magnum-v4-72b-AWQ | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float16 | 4.45.2 | 8,192 | 29,696 | 80 | 64 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | false | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
6,370 | CH-UP/mistral-7b-v3-alpaca-cleaned-zh-cn | mistral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | false | null | null | null |
6,371 | CHAAAKHDABug/model | xlm-roberta | ["XLMRobertaForMaskedLM"] | 250002 | float32 | 4.48.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
6,372 | CHAOYUYD/vit-base-patch16-224-finetuned-flower | vit | ["ViTForImageClassification"] | null | float32 | 4.24.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
6,373 | CHARKA/Meta-Llama-3-8B-InstructSmallDataV0.2 | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,374 | CHARKA/llam3mhg_ggufq4_k_mv3 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,375 | CHARLESL16/gpt2-student-question-answer-memorization | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.44.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,376 | CHARLESL16/gpt2-student-question-answer-memorization-v2 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.44.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,377 | CHIH-HUNG/Llama-2-13b-hf-FINETUNE10_5.2w-r16-gate_up_down | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.34.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,378 | CHIH-HUNG/Llama-2-13b-hf-FINETUNE11_4.4w-r16-gate_up_down | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.34.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,379 | CHIH-HUNG/Llama-2-13b-hf-FINETUNE9_TEST_4.4w-r16-gate_up_down | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.34.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,380 | CHIH-HUNG/Mistral-7B-v0.1-FINETUNE7_4.1w-r16-gate_up_down | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.34.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,381 | CHIH-HUNG/Mistral-7B-v0.1-FINETUNE8_4.1w-r16-gate_up_down | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.34.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,382 | CHIH-HUNG/Mistral-7B-v0.1-FINETUNE9_4.4w-r16-gate_up_down | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.34.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,383 | CHIH-HUNG/llama-2-13b-FINETUNE1_17w-gate_up_down_proj | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,384 | CHIH-HUNG/llama-2-13b-FINETUNE1_17w-q_k_v_o_proj | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,385 | CHIH-HUNG/llama-2-13b-FINETUNE2_3w-q_k_v_o_proj | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,386 | CHIH-HUNG/llama-2-13b-FINETUNE2_3w-r16 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,387 | CHIH-HUNG/llama-2-13b-FINETUNE2_TEST_2.2w | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,388 | CHIH-HUNG/llama-2-13b-FINETUNE3_3.3w-r16-q_k_v_o_gate_up_down | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,389 | CHIH-HUNG/llama-2-13b-FINETUNE3_3.3w-r4-q_k_v_o_gate_up_down | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,390 | CHIH-HUNG/llama-2-13b-FINETUNE3_3.3w-r8-q_k_v_o_gate_up_down | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,391 | CHIH-HUNG/llama-2-13b-FINETUNE4_3.8w-r16-gate_up_down-test1 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,392 | CHIH-HUNG/llama-2-13b-FINETUNE4_3.8w-r16-q_k_v_o_gate_up_down | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,393 | CHIH-HUNG/llama-2-13b-FINETUNE4_3.8w-r4-q_k_v_o_gate_up_down | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,394 | CHIH-HUNG/llama-2-13b-FINETUNE4_3.8w-r8-q_k_v_o_gate_up_down | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,395 | CHIH-HUNG/llama-2-13b-FINETUNE4_addto15k_4.5w-r16-gate_up_down | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,396 | CHIH-HUNG/llama-2-13b-FINETUNE4_compare15k_4.5w-r16-gate_up_down | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,397 | CHIH-HUNG/llama-2-13b-dolphin_5w | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,398 | CHIPP-AI/CHIPP-AI | ensemble | ["EnsembleModel"] | null | float32 | 4.48.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,399 | CHLIN/esm2_t6_8M_UR50D-hw_scop | esm | ["EsmForSequenceClassification"] | 33 | float32 | 4.35.2 | 320 | 1,280 | 6 | 20 | null | gelu | null | True | 1,026 | null | null | 0.02 | null | null | null | 16 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.007373 | small | short |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.