index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
3,900 | Anwaarma/Improved-Arabert-twitter-sentiment2char | bert | ["BertForSequenceClassification"] | 64000 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,901 | Anwaarma/Improved-Arabert-twitter-sentiment2chars | bert | ["BertForSequenceClassification"] | 64000 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,902 | Anwaarma/Improved-Arabic-bert-base | bert | ["BertForSequenceClassification"] | 32000 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,903 | Anwaarma/Improved-Arabic-bert-nodropout | bert | ["BertForSequenceClassification"] | 32000 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,904 | Anwaarma/Improved-MARBERT-twitter-sentiment-Twitter | bert | ["BertForSequenceClassification"] | 100000 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,905 | Anwaarma/Improved-MARBERT-twitter-sentiment-nodroput-Twitter | bert | ["BertForSequenceClassification"] | 100000 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,906 | Anwaarma/Improved-bert-multilingual | bert | ["BertForSequenceClassification"] | 119547 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,907 | Anwaarma/Improved-bert-multilingual-nodropout | bert | ["BertForSequenceClassification"] | 119547 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,908 | Anwaarma/Improved-xlm-roberta-base | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,909 | Anwaarma/Improved-xlm-roberta-base-nodroput | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,910 | Anwaarma/xlm-ex2 | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.49.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,911 | AnyaBoo/finetune_Llama3.1-8b | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
3,912 | AnyaK/llama-2-7b-mini | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
3,913 | AnyaSchen/image2music | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,914 | AnyaSchen/my_awesome_model | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.34.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,915 | AnyaSchen/rugpt3-medium-key2poetry | gpt2 | ["GPT2LMHeadModel"] | 50260 | float32 | 4.30.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50258 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
3,916 | Anybalsmith/pick_brick_policy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,917 | Anyfusion/cogvideox-nf4 | null | [] | null | null | null | null | null | null | 48 | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,918 | Anyfusion/flux-turbo-nf4 | null | [] | null | null | null | null | null | null | 24 | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
3,919 | AoEiuV020/test | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.52.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,920 | AoiKazama/Kinoe-7B | kinoe | ["KinoeForCausalLM"] | 55424 | bfloat16 | 4.40.1 | 5,120 | 14,336 | 24 | 40 | 10 | silu | 0 | True | 32,768 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 7.549747 | large | very_long |
3,921 | Aoishi19/autotrain-supermarket2 | vit | ["ViTForImageClassification"] | null | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
3,922 | ApacheOne/CKPTs | custom | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,923 | Apala/FineLlama-3.1-8B | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.46.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
3,924 | AparnaVai/sarvam1-aparna | llama | ["LlamaForCausalLM"] | 68096 | null | 4.48.3 | 2,048 | 11,008 | 28 | 16 | 8 | silu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 2 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.409286 | medium | long |
3,925 | Aparnarani06/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,926 | Ape50/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,927 | Apel-sin/whisper-large-v3-russian-ties-podlodka-v1.2 | whisper | ["WhisperForConditionalGeneration"] | 51866 | float32 | 4.49.0 | null | null | 32 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
3,928 | Aphasia/sd-class-butterflies-32 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,929 | Apheironn/distilbert-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.28.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
3,930 | Apocalypse-19/doom_deadly_corridor | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,931 | Apocalypse-19/doom_health_gathering | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,932 | Apocalypse-19/trocr-MICR | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | 50265 | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 2 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,933 | Apokryphosx/7B-GSM8K-finetune | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.51.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
3,934 | AppsDev/bert-base-japanese-v3-wrime-sentiment | bert | ["BertForSequenceClassification"] | 32768 | float32 | 4.31.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,935 | Apptware/QNA_chatbot_ecommerce_falcon_7b_sharded_quantized | RefinedWebModel | ["RWForCausalLM"] | 65024 | float16 | 4.34.0.dev0 | 4,544 | null | null | null | null | null | 0 | True | null | null | null | 0.02 | 11 | 11 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,936 | AppyCopilot/llm-repo | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.39.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,937 | ApricotWine/minipile_flash_gradcheck | gpt_neox | ["GPTNeoXForCausalLM"] | 50304 | bfloat16 | 4.45.0 | 768 | 3,072 | 12 | 12 | null | gelu | 0 | True | 2,048 | 10,000 | null | 0.02 | 0 | 0 | false | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.084935 | small | medium |
3,938 | AptaArkana/indonesian-comunication-indoBERT-finetuned | bert | ["BertForSequenceClassification"] | 31923 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | 0 | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,939 | AptaArkana/indonesian-distilbert-base-cased-finetuned-indonlu | distilbert | ["DistilBertForSequenceClassification"] | 119547 | float32 | 4.35.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
3,940 | AptaArkana/indonesian-personality-indoBERT-finetuned | bert | ["BertForSequenceClassification"] | 31923 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | 0 | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,941 | AptaArkana/indonesian_sentiment_sbert_base | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.35.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,942 | AptaArkana/indonesian_toxic_classification | bert | ["BertForSequenceClassification"] | 31923 | float32 | 4.26.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | 0 | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,943 | Apucs/bangla-bert-base-finetuned-mnli-mm | bert | ["BertForSequenceClassification"] | 102025 | float32 | 4.35.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,944 | Apucs/bn_model_821k_iter_loss_1_24-finetuned-mnli-mm | bert | ["BertForSequenceClassification"] | 30000 | float32 | 4.35.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,945 | Apucs/bn_model_821k_iter_loss_1_24-finetuned-sc | bert | ["BertForSequenceClassification"] | 30000 | float32 | 4.35.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,946 | Apurva3509/gemma-1.1-7b-it-bnb-4bit-medical | gemma | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,947 | Apv/Flaubert2904_v2 | flaubert | ["FlaubertForSequenceClassification"] | 68729 | null | 4.28.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | null | 0 | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
3,948 | Apv/Flaubert_1106 | flaubert | ["FlaubertForSequenceClassification"] | 68729 | null | 4.30.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | null | 0 | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
3,949 | AqeelShafy7/Whisper-Sinhala_Audio_to_Text | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.38.2 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
3,950 | Aqel/xlm-roberta-base-finetuned-panx-ar | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.40.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,951 | Aqsa-atif/phi-1_5-finetuned-gsm8k | phi-msft | ["PhiForCausalLM"] | 51200 | float32 | 4.35.2 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,952 | AqsaK/HW-1880-finetuned-all-cols | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,953 | Aqua002/DialoGPT-small-deadpool | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.28.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
3,954 | AquaLabs/Orpheus-3B-0.1-ft-Elise | llama | ["LlamaForCausalLM"] | 156940 | float16 | 4.50.0 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
3,955 | Ar4ikov/PromptGPTv2 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.27.4 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
3,956 | Ar4ikov/W2VFSC2 | wav2vec2 | ["Wav2Vec2ForSequenceClassification"] | 39 | float32 | 4.24.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
3,957 | Ar4ikov/W2VFSC2_1 | wav2vec2 | ["Wav2Vec2Classification"] | 39 | float32 | 4.24.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
3,958 | Ar4ikov/W2VFSC3 | wav2vec2 | ["Wav2Vec2ForSequenceClassification"] | 39 | float32 | 4.24.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
3,959 | Ar4ikov/W2VSC2 | wav2vec2 | ["Wav2Vec2Classification"] | 39 | float32 | 4.24.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
3,960 | Ar4ikov/Wav2Vec2ForSequenceClassification | wav2vec2 | ["Wav2Vec2ForSequenceClassification"] | 39 | float32 | 4.24.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
3,961 | Ar4l/DebertaV2-Base-10M_babylm-A__rte | deberta-v2 | ["DebertaV2ForSequenceClassification"] | 6144 | float32 | 4.44.2 | 768 | 1,536 | 8 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.056623 | small | short |
3,962 | Arabic-Clip-Archive/arabert-large-vit-base-32-epoch-16 | bert | ["BertModel"] | 64000 | float32 | 4.33.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,963 | Arabic-Clip-Archive/arabert-large-vit-base-32-epoch-21 | bert | ["BertModel"] | 64000 | float32 | 4.33.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,964 | Arabic-Clip-Archive/arabert-large-vit-base-32-epoch-26 | bert | ["BertModel"] | 64000 | float32 | 4.33.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,965 | Arabic-Clip-Archive/arabert-large-vit-base-32-epoch-31 | bert | ["BertModel"] | 64000 | null | 4.31.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,966 | Arabic-Clip-Archive/arabertv2-Vit-B-16-plus-epoch-31-trained-1M-corrupted | bert | ["BertModel"] | 64000 | float32 | 4.33.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,967 | Arabic-Clip-Archive/arabertv2-Vit-B-16-plus-epoch-41-trained-1M-corrupted | bert | ["BertModel"] | 64000 | float32 | 4.33.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,968 | Arabic-Clip-Archive/arabic-arabert-Vit-B-32-Transformer-tf | bert | ["BertModel"] | 64000 | null | 4.31.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,969 | Aranya31/DermLLaVA-1.5-7b-Label-Final2 | llava | ["LlavaForConditionalGeneration"] | 32064 | float16 | 4.45.1 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
3,970 | Aranya31/LLaVA-1.5-7b-DermConv | llava | ["LlavaForConditionalGeneration"] | 32064 | float16 | 4.45.1 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
3,971 | Aranya31/Qwen-2.5-dermMCQ | qwen2_5_vl | ["Qwen2_5_VLForConditionalGeneration"] | 151936 | float32 | 4.49.0 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 128,000 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
3,972 | Aranya31/derm-LLaVAMed | llava_mistral | ["LlavaMistralForCausalLM"] | 32000 | float16 | 4.37.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
3,973 | Aranya31/derm-LLaVAMed-3e | llava_mistral | ["LlavaMistralForCausalLM"] | 32000 | float16 | 4.37.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
3,974 | Aranya31/llava_77acc_merged_epoch3 | llava | ["LlavaForConditionalGeneration"] | 32064 | float16 | 4.46.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
3,975 | Aranya31/skinllava-1.5-7b-sample | llava | ["LlavaForConditionalGeneration"] | 32064 | float16 | 4.45.0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
3,976 | Aranya31/whisper-small-bn | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.52.4 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
3,977 | ArashAhmadian/ppo_6.9b_new | gpt_neox | ["GPTNeoXForCausalLM"] | 50432 | bfloat16 | 4.41.1 | 4,096 | 16,384 | 32 | 32 | null | gelu | 0 | True | 2,048 | null | null | 0.02 | 0 | 0 | false | 128 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 6.442451 | medium | medium |
3,978 | ArashAhmadian/rloo_6.9b_new | gpt_neox | ["GPTNeoXForCausalLM"] | 50432 | bfloat16 | 4.41.1 | 4,096 | 16,384 | 32 | 32 | null | gelu | 0 | True | 2,048 | null | null | 0.02 | 0 | 0 | false | 128 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 6.442451 | medium | medium |
3,979 | ArashAhmadian/rloo_tldr_6.9b | gpt_neox | ["GPTNeoXForCausalLM"] | 50432 | bfloat16 | 4.41.1 | 4,096 | 16,384 | 32 | 32 | null | gelu | 0 | True | 2,048 | null | null | 0.02 | 0 | 0 | false | 128 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 6.442451 | medium | medium |
3,980 | ArashMarandi/Llama-2-7b-chat-hf | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.44.2 | 2,048 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
3,981 | Aratako/ELYZA-japanese-Llama-2-MoE-2x13B-v0.1 | mixtral | ["MixtralForCausalLM"] | 32000 | bfloat16 | 4.38.1 | 5,120 | 13,824 | 40 | 40 | 40 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | 2 | 2 | false | false | false | false | false | false | false | true | false | true | 12.582912 | large | medium |
3,982 | Aratako/MistralPrism-24B | mistral | ["MistralForCausalLM"] | 131072 | bfloat16 | 4.52.4 | 5,120 | 32,768 | 40 | 32 | 8 | silu | 0 | False | 131,072 | 1,000,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
3,983 | Aratako/Ninja-v1-RP-expressive | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | medium |
3,984 | Aratako/Oumuamua-7b-RP | mistral | ["MistralForCausalLM"] | 42800 | bfloat16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
3,985 | Aratako/Qwen3-30B-A3B-ERP-v0.1 | qwen3_moe | ["Qwen3MoeForCausalLM"] | 151936 | bfloat16 | 4.51.3 | 2,048 | 6,144 | 48 | 32 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 64 | 8 | false | null | 8 | false | false | false | false | false | false | false | false | true | true | 2.415919 | medium | very_long |
3,986 | Aratako/Swallow-MoE-2x13B-v0.1 | mixtral | ["MixtralForCausalLM"] | 43176 | bfloat16 | 4.38.1 | 5,120 | 13,824 | 40 | 40 | 40 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | 2 | 2 | false | false | false | false | false | false | false | true | false | true | 12.582912 | large | medium |
3,987 | Aratako/Vecteus-v1-toxic | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.36.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | medium |
3,988 | Aratako/c4ai-command-r-v01-japanese-instruct | cohere | ["CohereForCausalLM"] | 256000 | bfloat16 | 4.39.3 | 8,192 | 22,528 | 40 | 64 | 64 | silu | 0 | True | 8,192 | 8,000,000 | null | 0.02 | 5 | 255001 | true | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 32.212255 | large | long |
3,989 | Aratako/calm3-22b-RP-v2 | llama | ["LlamaForCausalLM"] | 65024 | bfloat16 | 4.44.2 | 6,144 | 16,384 | 48 | 48 | 48 | silu | 0 | True | 16,384 | 1,000,000 | 0.00001 | 0.02 | 0 | 65001 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 21.743272 | large | long |
3,990 | Aratako/gemma-2-2b-axolotl-simpo-v1.0 | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.46.3 | 2,304 | 9,216 | 26 | 8 | 4 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 288 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.656226 | medium | long |
3,991 | Aratako/gemma-2-2b-axolotl-simpo-v1.0-merged | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.46.3 | 2,304 | 9,216 | 26 | 8 | 4 | gelu_pytorch_tanh | 0 | False | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 288 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.656226 | medium | long |
3,992 | Aratako/karakuri-lm-chat-upscaled-103b-v0.1 | llama | ["LlamaForCausalLM"] | 45416 | bfloat16 | 4.38.1 | 8,192 | 28,672 | 120 | 64 | 8 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 96.636764 | xlarge | medium |
3,993 | AravindS373/bird_multi_700 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.51.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
3,994 | Aravindh0693/srjana2024_techdocs | gemma | ["GemmaForCausalLM"] | 256000 | float16 | 4.38.0 | 3,072 | 24,576 | 28 | 16 | 16 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 192 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.170894 | medium | long |
3,995 | Arbi-Houssem/TunLangModel1.0 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.41.0.dev0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
3,996 | Arbi-Houssem/TunLangModel_test1.10 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.41.2 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
3,997 | Arbi-Houssem/TunLangModel_test1.11 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.41.2 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
3,998 | Arbi-Houssem/TunLangModel_test1.12 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.41.2 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
3,999 | Arbi-Houssem/TunLangModel_test1.13 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.41.2 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.