index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
5,000 | BEE-spoke-data/smol_llama-220M-GQA | llama | ["LlamaForCausalLM"] | 32128 | bfloat16 | 4.37.0.dev0 | 1,024 | 4,096 | 10 | 32 | 8 | silu | 0 | True | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 32 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.125829 | small | medium |
5,001 | BEE-spoke-data/smol_llama-220M-bees-internal | llama | ["LlamaForCausalLM"] | 32128 | bfloat16 | 4.36.2 | 1,024 | 4,096 | 10 | 32 | 8 | silu | 0 | False | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 32 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.125829 | small | medium |
5,002 | BEE-spoke-data/tFINE-680m-e32-d16-gqa-flan | t5 | ["T5ForConditionalGeneration"] | 28800 | float32 | 4.46.0.dev0 | null | null | null | null | 4 | null | null | True | null | null | null | null | null | 2 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | true | false | null | null | null |
5,003 | BEE-spoke-data/tFINE-680m-e32-d16-infinity_instruct-L2 | t5 | ["T5ForConditionalGeneration"] | 28800 | float32 | 4.46.0.dev0 | null | null | null | null | 4 | null | null | True | null | null | null | null | null | 2 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | true | false | null | null | null |
5,004 | BEE-spoke-data/verysmol_llama-v11-KIx2 | llama | ["LlamaForCausalLM"] | 32128 | float32 | 4.33.3 | 512 | 2,048 | 6 | 16 | 16 | silu | null | True | 1,024 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 32 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 0.018874 | small | short |
5,005 | BEGADE/model | gpt2 | ["GPT2LMHeadModel"] | 50259 | float32 | 4.44.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50257 | 50258 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,006 | BELLE-2/Belle-whisper-large-v3-zh | whisper | ["WhisperForConditionalGeneration"] | 51866 | float32 | 4.34.0 | null | null | 32 | null | null | null | 0 | False | null | null | null | null | 50256 | 50256 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,007 | BELLE-2/Belle-whisper-large-v3-zh-punct | whisper | ["WhisperForConditionalGeneration"] | 51866 | float32 | 4.34.0 | null | null | 32 | null | null | null | 0 | False | null | null | null | null | 50256 | 50256 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,008 | BESEMODEK/TADDY | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,009 | BFDLeeSin/12KDEB21_MATDF_500_5_64 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,010 | BFDLeeSin/12kDEB07_DF_100_10 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,011 | BFDLeeSin/12kDEB14_DF_100_10 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,012 | BFDLeeSin/12kDEB21_DF_100_10 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,013 | BFDLeeSin/12kDEIR07_DF_100_10 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,014 | BFDLeeSin/12kDEIR14_DF_100_10 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,015 | BFDLeeSin/12kDEIR21_DF_100_10 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,016 | BFDLeeSin/12kDEOR07_DF_100_10 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,017 | BFMeriem/chatbot-model | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.18.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,018 | BFMeriem/model | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.18.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,019 | BFS-Search/llama-3.2_multi_rel_I2B2 | llama | ["LlamaForCausalLM"] | 128258 | float16 | 4.45.1 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128256 | 128257 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
5,020 | BFS-Search/llama-3.2_single_rel_ConLL | llama | ["LlamaForCausalLM"] | 128258 | float16 | 4.46.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128256 | 128257 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
5,021 | BIFOLD-BigEarthNetv2-0/resnet101-all-v0.1.1 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,022 | BIFOLD-BigEarthNetv2-0/resnet101-s1-v0.1.1 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,023 | BIFOLD-BigEarthNetv2-0/resnet101-s2-v0.1.1 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,024 | BIFOLD-BigEarthNetv2-0/resnet18-all-v0.2.0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,025 | BIFOLD-BigEarthNetv2-0/resnet18-s2-v0.2.0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,026 | BIFOLD-BigEarthNetv2-0/resnet50-s1-v0.1.1 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,027 | BIFOLD-BigEarthNetv2-0/vit_base_patch8_224-s1-v0.1.1 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,028 | BIJOY087/Bangla_barta_shurkha_mobilebert | mobilebert | ["MobileBertForSequenceClassification"] | 30522 | float32 | 4.46.2 | 512 | 512 | 24 | 4 | null | relu | null | null | 512 | null | null | 0.02 | null | null | null | 128 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.075497 | small | short |
5,029 | BINUS/indoAgricultureBert | bert | ["BertForSequenceClassification"] | 30522 | null | 4.52.4 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
5,030 | BIStudent/Mistralai-finetuned-email2 | mistral3 | ["Mistral3ForConditionalGeneration"] | null | bfloat16 | 4.50.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | true | null | null | null |
5,031 | BK-Lee/CoLLaVO-7B | internlm | [] | 92544 | float32 | 4.38.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,032 | BK-Lee/MoAI-7B | internlm | [] | 92544 | float32 | 4.38.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,033 | BK-Lee/TroL-1.8B | internlm | [] | 92544 | float32 | 4.40.1 | 2,048 | 8,192 | 24 | 16 | 8 | silu | null | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.20796 | medium | very_long |
5,034 | BK04/QnAmodel | null | ["BertForMaskedLM"] | 30523 | null | null | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,035 | BKM1804/Qwen2.5-1.5B-01a7051a-d242-4c8b-82a8-7ef77a5838ed-orpo-before-dpo-tuned-merged | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.51.3 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
5,036 | BKM1804/Qwen2.5-1.5B-c3e84fc4-a240-46cf-a3b9-1df0b49bba6a-SFT_DPO_ratio_025_cosine | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.52.4 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151643 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
5,037 | BKM1804/Qwen2.5-1.5B-c3e84fc4-a240-46cf-a3b9-1df0b49bba6a-SFT_DPO_ratio_2_cosine | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.52.4 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | null | 151643 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
5,038 | BKM1804/SmolLM-135M-Instruct-4643c60e-bad6-442a-bae2-dd7473506d71-dpo-tuned-merged-check | llama | ["LlamaForCausalLM"] | 49152 | bfloat16 | 4.51.3 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | medium |
5,039 | BKM1804/Yarn-Llama-2-7b-128k-f26c8b32-0c66-41f0-81e3-7805f787b9c0-DPO_ratio_1_WSDS | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.52.4 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 131,072 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | very_long |
5,040 | BKM1804/b1816a50-c977-4663-a61c-1c397bd09ce3 | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.52.4 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5,041 | BKM1804/bge-m3-p2-512 | xlm-roberta | ["XLMRobertaModel"] | 250002 | float16 | 4.44.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 8,194 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | long |
5,042 | BKM1804/gpt-neo-1.3B-e9e17474-330c-458e-a3ae-8a4e42f7f3ea-SFT_DPO_ratio_025_cosine | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | bfloat16 | 4.52.4 | 2,048 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
5,043 | BKM1804/zephyr-7b-beta-5a259b13-5311-4621-b3d3-606580904a91-DPO_ratio_1_WSDS | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.52.4 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5,044 | BLIP3o/SigLIP2_SANA | null | [] | null | bfloat16 | 4.49.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
5,045 | BLNAYS/model_output | bert | ["BertForSequenceClassification"] | 30000 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 300 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,046 | BLakshmiVijay/xlsr-english | wav2vec2 | ["Wav2Vec2ForSequenceClassification"] | 33 | float32 | 4.42.4 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
5,047 | BM-K/KoChatBART | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.12.3 | null | null | 6 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,048 | BMP/distilbert-base-uncased-finetuned-cola | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.25.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,049 | BMRetriever/BMRetriever-1B | gpt_neox | ["GPTNeoXModel"] | 50304 | float32 | 4.39.1 | 2,048 | 8,192 | 16 | 8 | null | gelu | 0 | True | 2,048 | null | null | 0.02 | 0 | 0 | false | 256 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.805306 | small | medium |
5,050 | BMRetriever/BMRetriever-2B | gemma | ["GemmaModel"] | 256000 | float32 | 4.39.1 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
5,051 | BMRetriever/BMRetriever-410M | gpt_neox | ["GPTNeoXModel"] | 50304 | float32 | 4.37.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | True | 2,048 | null | null | 0.02 | 0 | 0 | false | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.30199 | small | medium |
5,052 | BMRetriever/BMRetriever-7B | mistral | ["MistralModel"] | 32000 | float32 | 4.37.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5,053 | BOENE/distilhubert-finetuned-gtzan | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.38.1 | 768 | 3,072 | 2 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.014156 | small | null |
5,054 | BOT365/my-tinyllama-colorist-v2 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.35.2 | 2,048 | 5,632 | 22 | 32 | 4 | silu | null | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
5,055 | BOUDABOUS/ai-univ-chatbot-multilingual | m2m_100 | ["M2M100ForConditionalGeneration"] | 128112 | float32 | 4.49.0 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,056 | BOUDABOUS/fine-tuned-chatbot | blenderbot | ["BlenderbotForConditionalGeneration"] | 8008 | float32 | 4.48.3 | null | null | 2 | null | null | null | 0 | True | 128 | null | null | null | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,057 | BRUNOKRISTI/t5-classification | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.47.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,058 | BSAtlas/BS-MedX | mllama | ["MllamaForConditionalGeneration"] | null | bfloat16 | 4.46.3 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | true | null | null | null |
5,059 | BSC-LT/ALIA-40b | llama | ["LlamaForCausalLM"] | 256000 | bfloat16 | 4.45.2 | 8,192 | 24,576 | 48 | 64 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 38.654706 | large | very_long |
5,060 | BSC-LT/Flor-6.3B-Instruct | bloom | ["BloomForCausalLM"] | 50259 | bfloat16 | 4.44.0 | 4,096 | null | null | null | null | null | 0 | True | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,061 | BSC-LT/experimental7b-rag-instruct | llama | ["LlamaForCausalLM"] | 256000 | bfloat16 | 4.40.2 | 4,096 | 11,008 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
5,062 | BSC-LT/salamandra-7b-vision | llava_onevision | ["LlavaOnevisionForConditionalGeneration"] | null | float16 | 4.46.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,063 | BSC-NLP4BIA/Biomaterial-Content-Classifier | bert | ["BertForSequenceClassification"] | 28895 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,064 | BSC-NLP4BIA/bsc-bio-ehr-es-carmen-drugtemist | roberta | ["RobertaForTokenClassification"] | 50262 | float32 | 4.30.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,065 | BSC-NLP4BIA/bsc-bio-ehr-es-distemist | roberta | ["RobertaForTokenClassification"] | 50262 | float32 | 4.30.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,066 | BSC-NLP4BIA/bsc-bio-ehr-es-medprocner | roberta | ["RobertaForTokenClassification"] | 50262 | float32 | 4.30.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,067 | BSC-NLP4BIA/location-sub-classifier | roberta | ["RobertaModel"] | 52000 | float32 | 4.33.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,068 | BUAADreamer/Chinese-LLaVA-Med-7B | llava | ["LlavaForConditionalGeneration"] | null | bfloat16 | 4.41.0 | 4,096 | null | null | null | null | null | null | True | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,069 | BUT-FIT/DeCRED-base | joint_aed_ctc_speech-encoder-decoder | ["JointCTCAttentionEncoderDecoder"] | null | float32 | 4.39.3 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,070 | BUT-FIT/DiCoW_v3_MLC | DiCoW | ["DiCoWForConditionalGeneration"] | 51866 | float32 | 4.42.0 | null | null | 32 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,071 | BUT-FIT/ED-small | joint_aed_ctc_speech-encoder-decoder | ["JointCTCAttentionEncoderDecoder"] | null | float32 | 4.39.3 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,072 | BUT-FIT/csmpt7b | mpt | ["MPTForCausalLM"] | 64002 | float32 | 4.37.0.dev0 | null | null | null | null | null | null | null | False | null | null | null | null | null | 64001 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
5,073 | BUT-FIT/diarizen-wavlm-base-s80-md | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,074 | BVRA/MegaDescriptor-B-224 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,075 | BVRA/convnext_base.in1k_ft_fungitastic-m_384 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,076 | BVRA/convnext_base.in1k_ft_fungitastic_224 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,077 | BVRA/inception_resnet_v2.in1k_ft_df20m_299 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,078 | BVRA/inception_v4.in1k_ft_df20m_299 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,079 | BVRA/legacy_seresnext101_32x4d.in1k_ft_df20m_299 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,080 | BVRA/mobilenetv2_100.in1k_ft_df20m_299 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,081 | BVRA/resnet18.in1k_ft_fungitastic-mini_224 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,082 | BVRA/tf_efficientnet_b0.in1k_ft_df20_299 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,083 | BVRA/tf_efficientnet_b0.in1k_ft_df20_384 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,084 | BWangila/a2c-PandaReachDense-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
5,085 | BWangila/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,086 | BWayne/LILT_mr | lilt | ["LiltForTokenClassification"] | 50265 | float32 | 4.29.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,087 | Ba2han/HermesStar-OrcaWind-Synth-11B | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.34.1 | 4,096 | 14,336 | 48 | 32 | 8 | silu | null | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 9.663676 | large | very_long |
5,088 | Ba2han/Llama-Phi-3_DoRA | mistral | ["MistralForCausalLM"] | 32064 | bfloat16 | 4.39.3 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32007 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | true | false | false | true | 3.623879 | medium | medium |
5,089 | Ba2han/Pretraining_Turkish-10k-step | mistral | ["MistralForCausalLM"] | 44312 | float32 | 4.44.2 | 1,280 | 2,560 | 18 | 16 | 4 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 80 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 0.353894 | small | medium |
5,090 | Ba2han/VBart-Spell_Checker-TR | mbart | ["MBartForConditionalGeneration"] | 32001 | float32 | 4.44.2 | null | null | 6 | null | null | null | 0 | True | 1,024 | null | null | null | 2 | 3 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,091 | Ba2han/test-gemma3-merged | gemma3 | ["Gemma3ForConditionalGeneration"] | null | bfloat16 | 4.51.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | 2 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
5,092 | Ba2han/test-gemma3-merged-v2 | gemma3 | ["Gemma3ForConditionalGeneration"] | null | bfloat16 | 4.51.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | 2 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
5,093 | BaSalam/Llama2-7b-entity-attr-v1 | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.36.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
5,094 | Babak-Behkamkia/BERT_gpt_vast_combined | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.32.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,095 | Babak-Behkamkia/BERT_gpt_vast_combined2 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.32.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,096 | Babak-Behkamkia/bert_VAST_binary | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.31.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,097 | Babak-Behkamkia/bert_mydataset_VAST | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.32.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,098 | Babatunde1/fine-tuned-BERT-sentiment | bert | ["BertForSequenceClassification"] | 30522 | null | 4.47.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,099 | Babelscape/mrebel-base | m2m_100 | ["M2M100ForConditionalGeneration"] | 128129 | float32 | 4.19.2 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.