index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
5,100 | Babelscape/mrebel-large | mbart | ["MBartForConditionalGeneration"] | 250071 | float32 | 4.19.2 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,101 | Babelscape/mrebel-large-32 | mbart | ["MBartForConditionalGeneration"] | 250071 | float32 | 4.19.2 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,102 | BabyChou/Yi-VL-34B | llava | ["YiVLForCausalLM"] | 64000 | bfloat16 | 4.36.1 | 7,168 | 20,480 | 60 | 56 | 8 | silu | null | False | 4,096 | 5,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 36.993761 | large | medium |
5,103 | BachNgoH/MetaMath-Zalo-ft-Few-Shot-Augment-1000 | mistral | ["MistralForCausalLM"] | 32001 | float32 | 4.35.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5,104 | BachNgoH/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,105 | BachelorThesis/GerMedBERT_NER_V01_BRONCO_CARDIO | bert | ["BertForTokenClassification"] | 30000 | float32 | 4.40.1 | 768 | 3,072 | 12 | 8 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 96 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,106 | BachelorThesis/GerMedBert_ATTR_V02_BRONCO | bert | ["BertForSequenceClassification"] | 30000 | float32 | 4.40.1 | 768 | 3,072 | 12 | 8 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 96 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,107 | Bachhoang/Continuous-FT-vbd-merge | llama | ["LlamaForCausalLM"] | 49381 | bfloat16 | 4.32.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | False | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
5,108 | Bachlee/bert-nwpredict | bert | ["BertForMaskedLM"] | 30522 | null | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,109 | Bachlee/bert-nwpredict_v2 | bert | ["BertForMaskedLM"] | 30522 | null | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,110 | Bachstelze/smolSynformer | llama | ["LlamaForCausalLM"] | 49152 | float32 | 4.47.0 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 8,192 | 100,000 | 0.00001 | 0.041667 | 0 | 0 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | long |
5,111 | Backdrive/xlm-roberta-base-finetuned-panx-de | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.30.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,112 | Bacon666/Phenom-12B-0.1 | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.44.1 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
5,113 | BadisNahdi/FineTuneBERT | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.31.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,114 | BadreddineHug/LayoutLM_1 | layoutlmv3 | ["LayoutLMv3ForTokenClassification"] | 50265 | float32 | 4.29.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,115 | BadreddineHug/LayoutLMv3_97_2 | layoutlmv3 | ["LayoutLMv3ForTokenClassification"] | 50265 | float32 | 4.29.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | short |
5,116 | BadreddineHug/LayoutLMv3_batch_8 | layoutlmv3 | ["LayoutLMv3ForTokenClassification"] | 50265 | float32 | 4.29.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | short |
5,117 | BadreddineHug/LayoutLMv3_funds | layoutlmv3 | ["LayoutLMv3ForTokenClassification"] | 50265 | float32 | 4.29.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,118 | BadreddineHug/LayoutLMv3_large_1 | layoutlmv3 | ["LayoutLMv3ForTokenClassification"] | 50265 | float32 | 4.29.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | short |
5,119 | BadreddineHug/LayoutLMv3_large_2 | layoutlmv3 | ["LayoutLMv3ForTokenClassification"] | 50265 | float32 | 4.29.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,120 | BadreddineHug/LayoutLMv3_large_3 | layoutlmv3 | ["LayoutLMv3ForTokenClassification"] | 50265 | float32 | 4.29.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,121 | BadreddineHug/donut-base-ocr11 | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,122 | Badsims/sentis-phi-3.5-mini-instruct-uint8 | phi3 | ["Phi3ForCausalLM"] | 32064 | null | 4.43.3 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | False | 131,072 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | very_long |
5,123 | Baekpica/vicuna-7b-v1.3-tiny-stories-pretraining-2epoch | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | False | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
5,124 | Baekpica/vicuna-7b-v1.3-tinystories-linear | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | False | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
5,125 | BagelLabs/TinyToolUse-Qwen3-0.6B-toolbench | qwen3 | ["Qwen3ForCausalLM"] | 151936 | float32 | 4.52.4 | 1,024 | 3,072 | 28 | 16 | 8 | silu | 0 | True | 40,960 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.352322 | small | very_long |
5,126 | Bagus/distilhubert-finetuned-gtzan-base-audio-course | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.35.2 | 768 | 3,072 | 2 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.014156 | small | null |
5,127 | Bagus/hubert_xlarge_emodb | hubert | ["Model"] | 32 | float32 | 4.32.0 | 1,280 | 5,120 | 48 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 80 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.943718 | small | null |
5,128 | Bagus/wav2v2c_swbd_emodb | wav2vec2 | ["Model"] | 32 | float32 | 4.32.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
5,129 | Bagus/wav2vec2_swbd_emodb | wav2vec2 | ["Model"] | 32 | float32 | 4.32.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
5,130 | Bagus/wavlm_base-plus_emodb | wavlm | ["Model"] | 32 | float32 | 4.32.0 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
5,131 | Bagus/whisper-medium-common_voice_17_0-id-10000 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.42.0.dev0 | null | null | 24 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,132 | Bagus/whisper-tiny-en | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.35.2 | null | null | 4 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,133 | BahAdoR0101/my_awesome_wnut_model | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.34.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
5,134 | Baha123/llama3.1-text-to-sql_finetuned-fullmodel | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
5,135 | Baha2107/test_identity_01 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.38.1 | 1,024 | 2,816 | 24 | 16 | 16 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 0.30199 | small | very_long |
5,136 | BahaaEldin0/Mistral-7b-Mixtral-7b | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.40.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5,137 | BahaaEldin0/bert-base-uncased | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.41.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,138 | BahaaEldin0/bert-base-uncased-reward-model | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.41.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,139 | BahaaEldin0/policy-model | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.41.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,140 | Baharak/codegen-350M-mono-python-18k-alpaca | codegen | ["CodeGenForCausalLM"] | 51200 | float16 | 4.36.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,141 | Bahareh0281/BioBERT_Finetuned | bert | ["BertLMHeadModel"] | 28996 | float32 | 4.47.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,142 | Bahareh0281/ClinicalBERT_Finetuned | bert | ["BertLMHeadModel"] | 28996 | float32 | 4.47.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,143 | Bahasalab/Bahasa-4b | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.39.1 | 2,560 | 6,912 | 40 | 20 | 20 | silu | 0 | False | 32,768 | 5,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.145728 | medium | very_long |
5,144 | Bahin/lunardr | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,145 | BaibhavJoshi/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,146 | Baidicoot/Othello-GPT-Transformer-Lens | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,147 | Bainbridge/bert-incl | bert | ["BertForSequenceClassification"] | 31102 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,148 | Bainbridge/bert-xxl-incl | bert | ["BertForSequenceClassification"] | 32102 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,149 | Bainbridge/gpt2-kl_1_03-hs_cn | gpt2 | ["GPT2LMHeadModel"] | 50259 | float32 | 4.28.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,150 | Bainbridge/gpt2-kl_1_03-hs_cn_decay | gpt2 | ["GPT2LMHeadModel"] | 50259 | float32 | 4.28.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,151 | Bainbridge/gpt2-kl_1_04-hs_cn | gpt2 | ["GPT2LMHeadModel"] | 50259 | float32 | 4.28.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,152 | Bainbridge/gpt2-kl_1_04-hs_cn_decay | gpt2 | ["GPT2LMHeadModel"] | 50259 | float32 | 4.28.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,153 | Bainbridge/gpt2-kl_1_06-hs_cn_decay | gpt2 | ["GPT2LMHeadModel"] | 50259 | float32 | 4.28.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,154 | Bainbridge/gpt2-kl_1_07-hs_cn | gpt2 | ["GPT2LMHeadModel"] | 50259 | float32 | 4.28.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,155 | Bajiyo/malayalam_imasc | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.37.2 | null | null | 12 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,156 | Bajiyo/w2v-bert-2.0-malayalam-CV16.0 | wav2vec2-bert | ["Wav2Vec2BertForCTC"] | 74 | float32 | 4.39.3 | 1,024 | 4,096 | 24 | 16 | null | swish | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | null |
5,157 | Bajiyo/w2v-bert-2.0-malayalam_mixeddataset-CV16.0 | wav2vec2-bert | ["Wav2Vec2BertForCTC"] | 77 | float32 | 4.39.3 | 1,024 | 4,096 | 24 | 16 | null | swish | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | null |
5,158 | Bakasura/wav2vec2-base-finetuned-ks | wav2vec2 | ["Wav2Vec2ForSequenceClassification"] | 32 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
5,159 | Bakhtiar56/w2v-bert-2.0-urdu-colab-CV13.0 | wav2vec2-bert | ["Wav2Vec2BertForCTC"] | 83 | float32 | 4.44.2 | 1,024 | 4,096 | 24 | 16 | null | swish | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | null |
5,160 | Bakhtiar56/whisper-small-dv | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.42.4 | null | null | 12 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,161 | BakshiSan/BERT_Twitter | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,162 | Baktashans/Finetuned_ParsBert_ArmanEmo | bert | ["BertForSequenceClassification"] | 100000 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,163 | Baktashans/Finetuned_xlm_roberta_large_ArmanEmo | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.35.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
5,164 | Bala-A87/LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,165 | BalaSubrahmanyam/T5Model_for_Ecommerce | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.40.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,166 | BalajiNagarajan/llama2 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.30.2 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
5,167 | Baldion7/Skin-cancer | vit | ["ViTForImageClassification"] | null | float32 | 4.48.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
5,168 | BaleChen/checkpoint-1300_merged | llama | ["LlamaForSequenceClassification"] | 32000 | bfloat16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
5,169 | BaleChen/checkpoint-800_merged | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
5,170 | BallAd-15/gemma2-9b-task10-subtask3-EN | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.47.1 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
5,171 | BallAd-15/gemma2-9b-task10-subtask3-HI | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.47.1 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
5,172 | Baluchon/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,173 | Bamarin/my_awesome_swag_model | bert | ["BertForMultipleChoice"] | 30522 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,174 | BanUrsus/bert-base-cased-finetuned-squad-accelerate_nlp-course-chapter7-section6 | bert | ["BertForQuestionAnswering"] | 28996 | float32 | 4.37.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,175 | BanUrsus/bert-finetuned-ner-accelerate_NLP-course-chapter7-section1 | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,176 | BanUrsus/bert-finetuned-ner_NLP-course-chapter7-section1 | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,177 | BanUrsus/distilhubert-finetuned-gtzan | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.39.2 | 768 | 3,072 | 2 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.014156 | small | null |
5,178 | BanUrsus/distilroberta-base-finetuned-condition-classifier | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.35.2 | 768 | 3,072 | 6 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
5,179 | BanUrsus/mt5-small-finetuned-cnn-dailymail-en_nlp-course-chapter7-section4 | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.35.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,180 | BanUrsus/whisper-small-dv | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.39.2 | null | null | 12 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,181 | BananaFish45/Ethnicity_model | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,182 | BananaFish45/gdpr_personal_data | deberta-v2 | ["DebertaV2ForSequenceClassification"] | 128100 | float32 | 4.43.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,183 | Bananaman272/bert-phishing-classifier_teacher | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.47.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,184 | BanczikZ/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,185 | Bandid/Umi-7B-Mistral-HP-0 | mistral | ["MistralForCausalLM"] | 32032 | bfloat16 | 4.40.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5,186 | Bandifishing/Nous-Hermes-13b-Chinese | llama | ["LlamaForCausalLM"] | 49954 | float16 | 4.28.1 | 5,120 | 13,824 | 40 | 40 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 12.582912 | large | medium |
5,187 | Bandika/PPOPyramids | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,188 | Bandika/PPOSnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,189 | Bandika/a2c-AntBulletEnv-v0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,190 | Bandika/a2c-PandaReachDense-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,191 | Bandika/rl_course_vizdoom_health_gathering_supreme | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,192 | BanglaLLM/Bangla-s1k-qwen-2.5-32B-Instruct | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float32 | 4.46.1 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
5,193 | BanglaLLM/bangla-llama-13b-base-v0.1 | llama | ["LlamaForCausalLM"] | 50437 | bfloat16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
5,194 | BanglaLLM/bangla-llama-7b-instruct-v0.1 | llama | ["LlamaForCausalLM"] | 50437 | float32 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | False | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
5,195 | BangorAI/Intel_neural-chat-7b-v3-1_CofnodYCynulliad_2epoch | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.34.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5,196 | BangorAI/mistral-7b-cy-epoch-1 | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5,197 | BangorAI/mistral-7b-cy-tokenizer-train-6 | mistral | ["MistralForCausalLM"] | 44955 | float32 | 4.35.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5,198 | BanusreeM/mt5-sawit-finetuned-v2 | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.48.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,199 | BaoHuynh2002/finetuned_Qwen2_7b_mt_history_mcqas_v1 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float16 | 4.41.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.