index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
15,100 | IAMRS23/resultsGPT | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.48.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,101 | IAVYang/llama-2-7b-myfine | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,102 | IAVYang/llama-2-7b-tagdetection_new | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,103 | IAVYang/new_llama | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,104 | IAVYang/new_llama-retry | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
15,105 | IAmFlyingMonkey/pokemon_classifier | vit | ["ViTForImageClassification"] | null | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
15,106 | IAmFlyingMonkey/roberta-base-model-korean-text-tag-classification | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.32.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,107 | IAmNotAnanth/wav2vec2-large-xls-r-300m-turkish-colab | wav2vec2 | ["Wav2Vec2ForCTC"] | 39 | float32 | 4.28.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
15,108 | IAmSkyDra/NutriCapy | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.42.4 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
15,109 | IAmSkyDra/NutriCapy-quantized | gemma | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,110 | IAmSudhirGupta/Llama-3.2-3B-Instruct-STA | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.45.0.dev0 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
15,111 | IB13/my_awesome_billsum_model | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.35.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,112 | IBI-CAAI/MELT-llama-2-3x70b-chat-hf | mixtral | ["MixtralForCausalLM"] | 32000 | bfloat16 | 4.37.0.dev0 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | 3 | 2 | false | false | false | false | false | false | false | true | true | true | 64.424509 | large | medium |
15,113 | ICAS03/fine-tune-llama | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.40.1 | 2,048 | 5,632 | 22 | 32 | 4 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
15,114 | ICB-UMA/ICB-cef_distemist_Spanish_SapBERT_grandparents_sim | roberta | ["RobertaForSequenceClassification"] | 52000 | float32 | 4.48.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,115 | ICB-UMA/ICB-cef_medprocner_Spanish_SapBERT_grandparents_bkg_1 | roberta | ["RobertaForSequenceClassification"] | 52000 | float32 | 4.48.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,116 | ICB-UMA/ICB-cef_medprocner_Spanish_SapBERT_grandparents_kg_1 | roberta | ["RobertaForSequenceClassification"] | 52000 | float32 | 4.48.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,117 | ICB-UMA/ICB-cef_symptemist_Spanish_SapBERT_grandparents_bkg_1 | roberta | ["RobertaForSequenceClassification"] | 52000 | float32 | 4.48.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,118 | ICB-UMA/ICB-cef_symptemist_Spanish_SapBERT_grandparents_kg_1 | roberta | ["RobertaForSequenceClassification"] | 52000 | float32 | 4.48.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,119 | ICEPVP8977/Uncensored_Phi_3_mini_128k_Safetensors_4b | phi3 | ["Phi3ForCausalLM"] | 32064 | float32 | 4.45.0.dev0 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | False | 131,072 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | very_long |
15,120 | ICFNext/EYY-Categorisation | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.17.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,121 | ICLbioengNLP/CXR_BioClinicalBERT_chunkedv1 | bert | ["BertForMaskedLM"] | 28996 | float32 | 4.16.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,122 | ICML2022/Tranception | tranception | ["TranceptionLMHeadModel"] | 25 | float32 | 4.17.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,123 | ICT2214Team7/Combined_model_v1 | roberta | ["RobertaForTokenClassification"] | 50265 | float32 | 4.40.2 | 768 | 3,072 | 6 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
15,124 | ICT2214Team7/RoBERTa_Combined_Generated_v2_1000 | roberta | ["RobertaForTokenClassification"] | 50265 | float32 | 4.40.2 | 768 | 3,072 | 6 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
15,125 | ICT2214Team7/RoBERTa_Combined_Generated_v2_2000 | roberta | ["RobertaForTokenClassification"] | 50265 | float32 | 4.40.2 | 768 | 3,072 | 6 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
15,126 | ICT3214-Group5/Cryptography_GPT_2_v1.0.0 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.46.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,127 | ICT3214-Group5/Cryptography_GPT_NEO_v1.0.0 | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.46.2 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
15,128 | ICT3214-Group5/MD5_gpt_neo | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.46.0 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
15,129 | ICT3214-Group5/MD5_gpt_neo_v1.1 | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.46.1 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
15,130 | ICT3214-Group5/MD5_gpt_neo_v1.2 | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.46.1 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
15,131 | ICT3214-Group5/MD5_gpt_neo_v1.4 | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.46.1 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
15,132 | ICT3214-Group5/MD5_gpt_neo_v1.5 | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.46.1 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
15,133 | ICT3214-Group5/MD5_gpt_neo_v3.0.4_aes | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.46.2 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
15,134 | ICT3214-Group5/MD5_gpt_neo_v3.1.1 | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.46.2 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
15,135 | ICT3214-Group5/MD5_gpt_neo_v3.1.2 | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.46.2 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
15,136 | ICT3214-Group5/MD5_gpt_neo_v3.1.3 | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.46.2 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
15,137 | ICT3214-Group5/MD5_gpt_neo_v3.1.4 | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.46.2 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
15,138 | ICT3214-Group5/MD5_gpt_neo_v3.1.5 | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.46.2 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
15,139 | ICT3214-Group5/MD5_gpt_neo_v3.1.6 | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.46.2 | 768 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
15,140 | ICTNLP/Auto-RAG-Llama-3-8B-Instruct | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.43.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
15,141 | ICTNLP/Llama-3.1-8B-Omni | omni_speech2s_llama | ["OmniSpeech2SLlamaForCausalLM"] | 128256 | float16 | 4.43.4 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,142 | ICTNLP/stream-omni-8b | stream_omni_llama | ["StreamOmniLlamaForCausalLM"] | 132352 | float16 | 4.43.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,143 | ICW11/icw1 | internlm2 | ["InternLM2ForCausalLM"] | 92544 | bfloat16 | 4.41.0 | 2,048 | 8,192 | 24 | 16 | 8 | silu | null | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.20796 | medium | very_long |
15,144 | ICW11/intern_study_L0_4 | internlm2 | ["InternLM2ForCausalLM"] | 92544 | bfloat16 | 4.41.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | True | 262,144 | 50,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
15,145 | ICavor/distilhubert-finetuned-gtzan | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.46.3 | 768 | 3,072 | 2 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.014156 | small | null |
15,146 | IDEA-CCNL/Erlangshen-DeBERTa-v2-320M-Chinese | deberta-v2 | ["DebertaV2ForMaskedLM"] | 12800 | null | null | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,147 | IDEA-CCNL/Erlangshen-Longformer-330M | longformer | ["LongformerForPreTraining"] | 12000 | null | null | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 4,096 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | medium |
15,148 | IDEA-CCNL/Erlangshen-Roberta-110M-NLI | bert | ["BertForSequenceClassification"] | 21128 | null | 4.20.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,149 | IDEA-CCNL/Erlangshen-Roberta-110M-Similarity | bert | ["BertForSequenceClassification"] | 21128 | null | 4.20.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,150 | IDEA-CCNL/Erlangshen-Roberta-330M-Causal-Chinese | bert | ["BertForSequenceClassification"] | 21128 | null | 4.20.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 512 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,151 | IDEA-CCNL/Erlangshen-Roberta-330M-NLI | bert | ["BertForSequenceClassification"] | 21128 | null | 4.20.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 512 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,152 | IDEA-CCNL/Erlangshen-Roberta-330M-Sentiment | bert | ["BertForSequenceClassification"] | 21128 | null | 4.20.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 512 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,153 | IDEA-CCNL/Erlangshen-Roberta-330M-Similarity | bert | ["BertForSequenceClassification"] | 21128 | null | 4.20.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 512 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,154 | IDEA-CCNL/Erlangshen-SimCSE-110M-Chinese | bert | ["BertModel"] | 21128 | null | 4.2.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,155 | IDEA-CCNL/Erlangshen-UniMC-DeBERTa-v2-1.4B-Chinese | deberta-v2 | [] | 12800 | null | null | 1,536 | 6,144 | 48 | 24 | null | gelu | null | null | 512 | null | null | 0.01 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 1.358954 | medium | short |
15,156 | IDEA-CCNL/Erlangshen-ZEN1-224M-Chinese | null | [] | 21128 | null | null | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,157 | IDEA-CCNL/Erlangshen-ZEN2-345M-Chinese | null | [] | 21128 | null | null | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,158 | IDEA-CCNL/Erlangshen-ZEN2-668M-Chinese | null | [] | 21128 | null | null | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,159 | IDEA-CCNL/Randeng-DAVAE-1.2B-General-Chinese | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,160 | IDEA-CCNL/Randeng-GAVAE-1.2B-Augmentation-Chinese | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,161 | IDEA-CCNL/Randeng-PPVAE-1.2B-Augmentation-Chinese | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,162 | IDEA-CCNL/Randeng-Pegasus-523M-Chinese | pegasus | ["PegasusForConditionalGeneration"] | 50000 | float16 | 4.10.2 | null | null | 16 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
15,163 | IDEA-CCNL/Randeng-T5-77M | mt5 | ["MT5ForConditionalGeneration"] | 32596 | float16 | 4.12.5 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,164 | IDEA-CCNL/Randeng-T5-784M | mt5 | ["MT5ForConditionalGeneration"] | 32596 | float16 | 4.12.5 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,165 | IDEA-CCNL/Randeng-TransformerXL-5B-Abduction-Chinese | null | [] | 50176 | null | null | 3,072 | null | null | 48 | null | null | null | null | null | null | null | null | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,166 | IDEA-CCNL/Randeng-TransformerXL-5B-Deduction-Chinese | null | [] | 50176 | null | null | 3,072 | null | null | 48 | null | null | null | null | null | null | null | null | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,167 | IDEA-CCNL/Taiyi-CLIP-RoBERTa-326M-ViT-H-Chinese | bert | ["BertModel"] | 21128 | float32 | 4.22.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,168 | IDEA-CCNL/Taiyi-Roberta-124M-D | roberta | ["RobertaForMaskedLM"] | 50265 | float32 | 4.17.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,169 | IDEA-CCNL/Wenzhong2.0-GPT2-3.5B-chinese | gpt2 | ["GPT2LMHeadModel"] | 50304 | null | 4.18.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
15,170 | IDEA-CCNL/Ziya-Visual-14B-Chat | blip-2 | ["InstructDittoLMForConditionalGeneration"] | null | float16 | null | null | null | null | null | null | null | null | null | null | null | null | 0.02 | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,171 | IDEA-Emdoor/UniTTS-mixed-v0.1 | qwen2 | ["Qwen2Model"] | 184840 | float32 | 4.48.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 8,192 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | long |
15,172 | IDEA-FinAI/Finance-R1-7B | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.49.0.dev0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
15,173 | IDEA-Research/ChatRex-7B | chatrex | ["ChatRexAuxForConditionalGeneration"] | null | float32 | 4.44.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,174 | IDEA-Research/RexSeek-3B | rexseek_qwen | ["RexSeekQwenForCausalLM"] | 151769 | bfloat16 | 4.48.0 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
15,175 | IDEA-Research/grounding-dino-base | grounding-dino | ["GroundingDinoForObjectDetection"] | null | float32 | 4.40.0.dev0 | null | null | null | null | null | null | 0 | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,176 | IDEA-Research/grounding-dino-tiny | grounding-dino | ["GroundingDinoForObjectDetection"] | null | float32 | 4.40.0.dev0 | null | null | null | null | null | null | 0 | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
15,177 | IDL/autotrain-qna-1170143354 | distilbert | ["DistilBertForQuestionAnswering"] | 30522 | float32 | 4.20.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
15,178 | IDQO/gis-parent-category-doc-description | roberta | ["RobertaForSequenceClassification"] | 40000 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,179 | IDQO/gis-parent-category-undersampling | roberta | ["RobertaForSequenceClassification"] | 40000 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,180 | IEETA/BioNExt-Extractor | relation-novelty-extractor | ["BioNExtExtractorModel"] | 28899 | float32 | 4.37.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,181 | IEETA/BioNExt-Tagger | crf-tagger | ["BioNExtTaggerModel"] | 28895 | float32 | 4.37.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,182 | IEITYuan/Yuan2-2B-Mars-hf | yuan | ["YuanForCausalLM"] | 135040 | bfloat16 | 4.30.0.dev0 | 2,048 | 8,192 | 24 | 32 | null | silu | null | False | 8,192 | null | 0.000001 | 0.02 | 77185 | 77185 | true | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 1.20796 | medium | long |
15,183 | IEITYuan/Yuan2-M32 | yuan | ["YuanForCausalLM"] | 135040 | bfloat16 | 4.30.0.dev0 | 2,048 | 8,192 | 24 | 16 | null | silu | null | True | 4,096 | null | 0.000001 | 0.02 | 77185 | 77185 | true | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 1.20796 | medium | medium |
15,184 | IEITYuan/Yuan2-M32-hf | yuan | ["YuanForCausalLM"] | 135040 | bfloat16 | 4.30.2 | 2,048 | 8,192 | 24 | 16 | null | silu | null | True | 4,096 | null | 0.000001 | 0.02 | 77185 | 77185 | null | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 1.20796 | medium | medium |
15,185 | IELTS8/videojepa_finetune_v4 | VideoGPT+ | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.46.3 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | true | false | false | false | false | false | false | true | 3.623879 | medium | medium |
15,186 | IGustavsen/t5-small-finetuned-english-wikilingua-finetuned-english-wikilingua | t5 | ["T5ForConditionalGeneration"] | 32128 | null | 4.29.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
15,187 | IHaBiS/Nethena-MLewd-Xwin-23B-exl2 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.34.1 | 5,120 | 13,824 | 71 | 40 | 40 | silu | null | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 22.334669 | large | medium |
15,188 | IHaBiS/StellarBright-2.55bpw-h8-exl2 | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.34.0 | 8,192 | 28,672 | 80 | 64 | 8 | silu | null | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | medium |
15,189 | IHaBiS/Synatra-7B-v0.3-RP-exl2 | mistral | ["MistralForCausalLM"] | 32002 | float16 | 4.34.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
15,190 | IHaBiS/Synatra-7B-v0.3-base-exl2 | mistral | ["MistralForCausalLM"] | 32002 | bfloat16 | 4.34.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
15,191 | IHaBiS/gemma3_4b_it_mrl_sib200_merged | gemma3 | ["Gemma3ForConditionalGeneration"] | null | bfloat16 | 4.51.3 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | null | [1, 106] | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
15,192 | II-Vietnam/R1-Math-Code-Fusion | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float32 | 4.50.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 32,768 | 300,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
15,193 | II-Vietnam/R1-Math-Code-Fusion-v2 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float32 | 4.50.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 32,768 | 300,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
15,194 | IIC/BETO_Galen-socialdisner | bert | ["BertForSequenceClassification"] | 31002 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,195 | IIC/RigoBERTa-Clinical | xlm-roberta | ["XLMRobertaModel"] | 250002 | float32 | 4.46.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
15,196 | IIC/RigoChat-7b-v2 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.45.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
15,197 | IIC/XLM_R_Galen-livingner3 | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,198 | IIC/XLM_R_Galen-socialdisner | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
15,199 | IIC/bert-base-spanish-wwm-cased-livingner3 | bert | ["BertForSequenceClassification"] | 31002 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.