index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
5,400 | BeckerAnas/sweet-resonance-226 | convnextv2 | ["ConvNextV2ForImageClassification"] | null | float32 | 4.52.3 | null | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,401 | BeckerAnas/swept-brook-197 | convnextv2 | ["ConvNextV2ForImageClassification"] | null | float32 | 4.52.3 | null | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,402 | BeckerAnas/vivid-silence-196 | convnextv2 | ["ConvNextV2ForImageClassification"] | null | float32 | 4.52.3 | null | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,403 | Bedru/wave2vec-base-Am | wav2vec2 | ["Wav2Vec2ForCTC"] | 291 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
5,404 | Bedru/whisper-tiny-am | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.48.3 | null | null | 4 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,405 | BeeBeaver/testpyramidsrnd | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,406 | BeebekBhz/Final-en-ne | m2m_100 | ["M2M100ForConditionalGeneration"] | 256206 | float32 | 4.39.3 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,407 | BeebekBhz/Final-ne-en | marian | ["MarianMTModel"] | 64172 | float32 | 4.45.2 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,408 | BeebekBhz/english_to_nepali | marian | ["MarianMTModel"] | 61950 | float32 | 4.48.3 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,409 | BeebekBhz/ne-en | marian | ["MarianMTModel"] | 61127 | null | 4.48.3 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,410 | BeebekBhz/nepali-to-english | marian | ["MarianMTModel"] | 64172 | float32 | 4.48.3 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,411 | Beeface/whisper-small-dv | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.41.1 | null | null | 12 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,412 | BeegolAI/lora_model_phi-3_beegol_4bit | mistral | ["MistralForCausalLM"] | 32064 | float16 | 4.40.2 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | true | false | false | true | 3.623879 | medium | medium |
5,413 | BeegolAI/lora_model_phi-3_beegol_q4_k_m-gguf | mistral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | false | null | null | null |
5,414 | BeegolAI/lora_model_tiny-llama_beegol_4bit | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.40.2 | 2,048 | 5,632 | 22 | 32 | 4 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
5,415 | Beehzod/speechT5_tts_uzbek | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.46.3 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,416 | Beehzod/speecht5_finetuned_uz_customDat_3 | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.47.0.dev0 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,417 | Beehzod/speecht5_finetuned_with_Uzbek_data | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.46.2 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,418 | Beehzod/uz_2301_3.1_tts | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.48.3 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,419 | Beehzod/uz_2301_4.2_tts | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.48.3 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,420 | Beehzod/uz_2301_4_tts | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.48.3 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,421 | Beehzod/uz_2301_5_tts | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.48.3 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,422 | BeenaSamuel/bart_cnn_dailymail_reduced_dataset_abstractive_summary | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.39.3 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,423 | BeenaSamuel/logs | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.38.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,424 | BeenaSamuel/multinews_cnn_logs | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.39.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,425 | BeenaSamuel/t5_small_multi_news_abstractive_summarizer | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.39.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,426 | BegumK/bert-base-uncased-finetuned-rte-best-run | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.50.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,427 | BegumK/bert-base-uncased-finetuned-rte-run_3 | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.50.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,428 | Behnam/a2c-AntBulletEnv-v0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,429 | Beijaflor2024/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.35.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,430 | Beijin/medical-question-model | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.51.3 | 384 | 1,536 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.021234 | small | short |
5,431 | Beijuka/roberta-decoder | roberta | ["RobertaForCausalLM"] | 50265 | float32 | 4.42.4 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
5,432 | BeingUs/model4 | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
5,433 | Beka-pika/mms_kaz_tts_sad | vits | ["VitsModel"] | 43 | float32 | 4.46.2 | 192 | null | 6 | 2 | null | relu | 0.1 | null | null | null | null | 0.02 | null | null | null | 96 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.002654 | small | null |
5,434 | Bekhzod/vit-base-patch16-224-in21k-lora | vit | ["ViTForImageClassification"] | null | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
5,435 | Bekicot/ProjectTa | marian | ["MarianMTModel"] | 54796 | float32 | 4.42.3 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,436 | BelalElhossany/best-model-clm | llama | ["LlamaForCausalLM"] | 49152 | float32 | 4.51.3 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 8,192 | 100,000 | 0.00001 | 0.041667 | 1 | 2 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | long |
5,437 | BelalElhossany/xlm-xnli | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.31.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
5,438 | Beless/gib1 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
5,439 | BelisaDi/stella-tuned-rirag | new | ["NewModel"] | 30528 | float32 | 4.45.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | null | 8,192 | 160,000 | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | true | 0.30199 | small | long |
5,440 | Bella0518/sd-class-butterflies-32 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,441 | Bellax/Qwen-unsolth-finetun | qwen2 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,442 | BelleGroup/BELLE-7B-0.2M | bloom | ["BloomModel"] | 250880 | null | 4.21.0.dev0 | null | null | null | 32 | null | null | 0 | True | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,443 | BelleGroup/BELLE-7B-0.6M | bloom | ["BloomModel"] | 250880 | null | 4.21.0.dev0 | null | null | null | 32 | null | null | 0 | True | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,444 | BelleGroup/BELLE-7B-1M | bloom | ["BloomModel"] | 250880 | null | 4.21.0.dev0 | null | null | null | 32 | null | null | 0 | True | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,445 | Ben10x/bert-base-mti881 | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.50.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,446 | Ben10x/gpt-medmentions | gpt_neo | ["GPTNeoForTokenClassification"] | 50257 | float32 | 4.50.3 | 2,048 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
5,447 | Ben10x/lvm_100m | llama | ["LlamaForCausalLM"] | 8292 | float16 | 4.50.3 | 640 | 1,664 | 22 | 10 | 10 | silu | 0 | True | 4,096 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 64 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.108134 | small | medium |
5,448 | Ben141/llama3-testrun | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 2,048 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
5,449 | BenBranyon/zephyr-sumbot-all-songs-255 | stablelm | ["StableLmForCausalLM"] | 50304 | float16 | 4.41.2 | 2,560 | 6,912 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | null | 0.02 | 0 | 0 | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
5,450 | BenBranyon/zephyr-sumbot-all-songs-306 | stablelm | ["StableLmForCausalLM"] | 50304 | float16 | 4.41.2 | 2,560 | 6,912 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | null | 0.02 | 0 | 0 | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
5,451 | BenBranyon/zephyr-sumbot-all-songs-split | stablelm | ["StableLmForCausalLM"] | 50304 | float16 | 4.41.2 | 2,560 | 6,912 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | null | 0.02 | 0 | 0 | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
5,452 | BenDaouda/wav2vec2-large-xls-r-300m-ASR_FrenchTextToNumber | wav2vec2 | ["Wav2Vec2ForCTC"] | 49 | float32 | 4.27.4 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
5,453 | BenFong/my_awesome_qa_model | distilbert | ["DistilBertForQuestionAnswering"] | 30522 | null | 4.40.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,454 | BenJones/phi_2_Q | phi-msft | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,455 | BenKJH/DialoGPT-small-lucybotasg | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.24.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,456 | BenLearningRL/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,457 | BenLearningRL/ppoRND-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,458 | BenPhan/ST1_modernbert-large_product-category_V3 | modernbert | ["ModernBertForSequenceClassification"] | 50368 | float32 | 4.48.0.dev0 | 1,024 | 2,624 | 28 | 16 | null | null | 0 | null | 8,192 | null | null | 0.02 | 50281 | 50282 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | true | 0.352322 | small | long |
5,459 | BenPhan/ST2_modernbert-large_hazard_V3 | modernbert | ["ModernBertForSequenceClassification"] | 50368 | float32 | 4.48.0.dev0 | 1,024 | 2,624 | 28 | 16 | null | null | 0 | null | 8,192 | null | null | 0.02 | 50281 | 50282 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | true | 0.352322 | small | long |
5,460 | BenPhan/ST2_modernbert-large_product_V3 | modernbert | ["ModernBertForSequenceClassification"] | 50368 | float32 | 4.48.0.dev0 | 1,024 | 2,624 | 28 | 16 | null | null | 0 | null | 8,192 | null | null | 0.02 | 50281 | 50282 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | true | 0.352322 | small | long |
5,461 | Benasd/Qwen2.5-VL-72B-Instruct-AWQ | qwen2_5_vl | ["Qwen2_5_VLForConditionalGeneration"] | 152064 | bfloat16 | 4.49.0.dev0 | 8,192 | 29,568 | 80 | 64 | 8 | silu | 0 | True | 128,000 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
5,462 | Bencode92/tradepulse-finbert-yiyanghkust-finbert-tone-20250707_125257 | bert | ["BertForSequenceClassification"] | 30873 | float32 | 4.41.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,463 | Bencode92/tradepulse-finbert-yiyanghkust-finbert-tone-20250707_130217 | bert | ["BertForSequenceClassification"] | 30873 | float32 | 4.41.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,464 | Bencode92/tradepulse-finbert-yiyanghkust-finbert-tone-20250707_134005 | bert | ["BertForSequenceClassification"] | 30873 | float32 | 4.41.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,465 | Bencode92/tradepulse-finbert-yiyanghkust-finbert-tone-20250707_162528 | bert | ["BertForSequenceClassification"] | 30873 | float32 | 4.41.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,466 | Bencode92/tradepulse-finbert-yiyanghkust-finbert-tone-20250707_164639 | bert | ["BertForSequenceClassification"] | 30873 | float32 | 4.41.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,467 | BenevolenceMessiah/Qwen2.5-Coder-7B-Chat-Instruct-TIES-v1.2 | qwen2 | ["Qwen2ForCausalLM"] | 151665 | bfloat16 | 4.45.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
5,468 | BenevolenceMessiah/Qwen2.5-Coder-7B-Instruct-abliterated-TIES-v2.0 | qwen2 | ["Qwen2ForCausalLM"] | 151665 | bfloat16 | 4.43.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
5,469 | BenevolenceMessiah/Yi-Coder-9B-Chat-8x-MoE | mixtral | ["MixtralForCausalLM"] | 64000 | bfloat16 | 4.44.2 | 4,096 | 11,008 | 48 | 32 | 4 | silu | 0 | False | 131,072 | 10,000,000 | 0.00001 | 0.02 | 1 | 7 | false | 128 | 8 | false | 8 | 2 | false | false | false | false | false | false | false | true | true | true | 9.663676 | large | very_long |
5,470 | BenevolenceMessiah/Yi-Coder-9B-Chat-Instruct-TIES | llama | ["LlamaForCausalLM"] | 64000 | float16 | 4.44.1 | 4,096 | 11,008 | 48 | 32 | 4 | silu | 0 | True | 131,072 | 10,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 9.663676 | large | very_long |
5,471 | Benezio/detoxify-inference-pipeline | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.37.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,472 | Benicillin1610/longformer | longformer | ["LongformerForSequenceClassification"] | 50265 | float32 | 4.44.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 4,098 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | medium |
5,473 | Benicillin1610/my_awesome_model | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.44.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,474 | Beniol/results | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.52.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,475 | BenitoMF/helloworld_model | vit | ["ViTForImageClassification"] | null | float32 | 4.22.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
5,476 | Beniuv/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,477 | BenjaminLHR/cs4248-bert-finetuned-squad | bert | ["BertForQuestionAnswering"] | 28996 | float32 | 4.35.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,478 | BenjaminOcampo/bert-base-uncased_seed-42_finetuned | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.46.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,479 | BenjaminOcampo/ihc-bert-trial-baseline-seed-42_finetuned | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.46.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,480 | BenjaminOcampo/model | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.34.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,481 | BenjaminOcampo/peace_cont_hatebert | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,482 | BenjaminOcampo/task-implicit_task__model-hatebert__aug_method-all | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.24.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,483 | Benjaminpwh/xls-r-300m-toratan-240 | wav2vec2 | ["Wav2Vec2ForCTC"] | 30 | float32 | 4.52.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
5,484 | Benjaminpwh/xlsr-toratan-240-copt | wav2vec2 | ["Wav2Vec2ForPreTraining"] | 32 | float32 | 4.51.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
5,485 | BenjiHearHear/finetuning-sentiment-model-3000-samples | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.42.4 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,486 | BenjiHearHear/finetuning-sentiment-model-financial-tweets-3-sentiments | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.42.4 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,487 | Benjiccee/xlm-roberta-base-finetuned-panx-de | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,488 | Benjiccee/xlm-roberta-base-finetuned-panx-de-fr | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,489 | Benjo27/sd-class-butterflies-32 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,490 | Bennet1996/donut-small | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,491 | Bennett-THU/ddpm-celebahq-finetuned-butterflies-2epochs | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,492 | BennoTheDog/MistralNemoBase2407 | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.46.2 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | True | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
5,493 | BenoAI/robertatest | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.47.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,494 | Benphil/pegasum-samsum | pegasus | ["PegasusForConditionalGeneration"] | 96103 | float32 | 4.41.1 | null | null | 16 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,495 | Benrobe/llama-3-8b-chat-doctor | llama | ["LlamaForCausalLM"] | 128258 | float16 | 4.42.4 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128256 | 128257 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
5,496 | Benson/llama-2-7b-miniguanaco-hf | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
5,497 | BensonHugging/distilbert-base-uncased-finetuned-cola | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.38.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,498 | BensonZhang/finetuning-sentiment-model-3000-samples | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.33.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,499 | Benul/Mistral_7b_v3_atom_lkdomain_2_GGUF | mistral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | false | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.