index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
4,800 | AyaKhaled/qwen2 | qwen2_vl | ["Qwen2VLForConditionalGeneration"] | 151936 | bfloat16 | 4.49.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
4,801 | AyaKhaled/test_qwen | qwen2_vl | ["Qwen2VLForConditionalGeneration"] | 151936 | bfloat16 | 4.49.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
4,802 | AyadSarah/fine_tuned_clip | clip | ["CLIPModel"] | null | float32 | 4.46.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,803 | Ayan-Z/An-Eye-To-Blind | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
4,804 | AyanOP/jutsu_classifier | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.44.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
4,805 | AyanSau/results_gpt2 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.26.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
4,806 | Ayansk11/Image_Caption_using_ViT_GPT2 | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | 50256 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,807 | Ayashika2424/Results | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.35.2 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
4,808 | AyatBahaa/Composite_Book | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
4,809 | AyatBahaa/LSS | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
4,810 | AyazK/Clinical_bert_DRG_prediction_phase1 | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.31.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | False | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,811 | Aye10032/t2-genius-en-zh | marian | ["MarianMTModel"] | 50003 | float32 | 4.42.3 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | null | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
4,812 | Ayesha18/chef-ai | gpt2 | ["GPT2LMHeadModel"] | 50265 | float32 | 4.35.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
4,813 | AyeshaFayyaz/flan-T5_summarizer | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.42.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,814 | Ayesharifa02/BARTModel_ExerciseLog | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.41.2 | null | null | 6 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
4,815 | Ayham/albert_ernie_50beam_summarization_cnn_dailymail | encoder-decoder | ["EncoderDecoderModel"] | 30522 | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 102 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,816 | Ayham/roberta_ernie_summarization_cnn_dailymail | encoder-decoder | ["EncoderDecoderModel"] | 30522 | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 102 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,817 | AykeeSalazar/vc-bantai-vit-withoutAMBI-adunest | vit | ["ViTForImageClassification"] | null | float32 | 4.20.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
4,818 | AykeeSalazar/vc-bantai-vit-withoutAMBI-adunest-trial | vit | ["ViTForImageClassification"] | null | float32 | 4.20.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
4,819 | AykeeSalazar/vc-bantai-vit-withoutAMBI-adunest-v1 | vit | ["ViTForImageClassification"] | null | float32 | 4.20.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
4,820 | Aykill02/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,821 | Aykill02/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,822 | AymanAbaid25/videomae-base-finetuned-ImageTBAD_NewTBAD | videomae | ["VideoMAEForVideoClassification"] | null | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
4,823 | AymanMansour/kaggle_2_cycle_3 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.51.3 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
4,824 | AymanTarig/Llama-3.2-1B-FC-v0.3 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.49.0 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
4,825 | AymanTarig/Llama-3.2-1B-FC-v1.1 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.2 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
4,826 | AymanTarig/Llama-3.2-1B-FC-v2 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.46.3 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
4,827 | AymaneLA/a2c-cartpole | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,828 | AymaneLA/a2c-sb3-panda-reach | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,829 | Aymann/Anomal | convnext | ["ConvNextForImageClassification"] | null | float32 | 4.22.1 | null | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,830 | AyoAgbaje/speech_classifier_model | wav2vec2 | ["Wav2Vec2ForSequenceClassification"] | 32 | float32 | 4.47.0 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
4,831 | Ayon128/dummy-model | camembert | ["CamembertForMaskedLM"] | 32005 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 5 | 6 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,832 | Ayon128/dummy2 | camembert | ["CamembertForMaskedLM"] | 32005 | float32 | 4.28.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 5 | 6 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,833 | Ayon128/fine_tuned_t5_tokenizer | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.9.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,834 | Ayon128/without_folds_t5 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.39.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,835 | AyoubAnhal/ENSATE | gemma | ["GemmaForCausalLM"] | 256000 | float16 | 4.51.1 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
4,836 | AyoubChLin/Bart-MNLI-CNN_news | bart | ["BartForSequenceClassification"] | 50265 | float32 | 4.20.0 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
4,837 | AyoubChLin/bert_cnn_news | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.27.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,838 | AyoubChLin/news-ESG-Bert | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.26.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,839 | AyoubChLin/test-summarizer | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.25.1 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
4,840 | Ayus077BCT014Bhandari/vartat5-using-100K-plus-16 | t5 | ["T5ForConditionalGeneration"] | 128128 | float32 | 4.37.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,841 | Ayus077BCT014Bhandari/vartat5-using-100K-plus-17 | t5 | ["T5ForConditionalGeneration"] | 128128 | float32 | 4.37.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,842 | Ayus077BCT014Bhandari/vartat5-using-100K-plus-18 | t5 | ["T5ForConditionalGeneration"] | 128128 | float32 | 4.37.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,843 | Ayus077BCT014Bhandari/vartat5-using-100K-plus-19 | t5 | ["T5ForConditionalGeneration"] | 128128 | float32 | 4.37.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,844 | Ayus077BCT014Bhandari/vartat5-using-100K-plus-20 | t5 | ["T5ForConditionalGeneration"] | 128128 | float32 | 4.37.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,845 | Ayus077BCT014Bhandari/vartat5-using-100K-plus-21 | t5 | ["T5ForConditionalGeneration"] | 128128 | float32 | 4.37.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,846 | Ayus077BCT014Bhandari/vartat5-using-31K-just-test-plus-12 | t5 | ["T5ForConditionalGeneration"] | 128128 | float32 | 4.37.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,847 | Ayus077BCT014Bhandari/vartat5-using-31k-just-test | t5 | ["T5ForConditionalGeneration"] | 128128 | float32 | 4.37.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,848 | Ayush-1722/Llama-2-7b-chat-Summarize-16K-LoRANET-Merged | llama | ["LlamaForCausalLM"] | 32001 | bfloat16 | 4.40.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | very_long |
4,849 | Ayush-1722/Llama-2-7b-chat-Summarize-24K-LoRANET-Merged | llama | ["LlamaForCausalLM"] | 32001 | bfloat16 | 4.40.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 49,152 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | very_long |
4,850 | Ayush-1722/Llama-3-8B-Instruct-v0.1-128K-QLoRA-Merged | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.40.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
4,851 | Ayush-1722/Meta-Llama-3-8B-Instruct-Summarize-v0.2-24K-LoRANET-Merged | llama | ["LlamaForCausalLM"] | 128257 | bfloat16 | 4.40.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 49,152 | 200,000,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
4,852 | Ayush-1722/Meta-Llama-3-8B-Instruct-Summarize-v0.2-32K-LoRANET-Merged | llama | ["LlamaForCausalLM"] | 128257 | bfloat16 | 4.40.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 65,536 | 200,000,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
4,853 | Ayush-Sharan-1/act-lang-policy-test-gp1-10k | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,854 | Ayush-Sharan-1/act-lang-policy-test-gp1_2-parallel | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,855 | Ayush-Surve/logs | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,856 | Ayush001/finetuning-sentiment-model-3000-samples | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.31.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
4,857 | Ayush001/spam-classifier | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.33.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,858 | Ayush12a/m2m100_418M-en-hi_legal | m2m_100 | ["M2M100ForConditionalGeneration"] | 128112 | float32 | 4.48.3 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
4,859 | Ayush2312/llama2-7B-1k-TherapyData | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.37.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
4,860 | AyushNayak/dockerNLcommands | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.38.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
4,861 | AyushS9020/speecht5_finetuned_emirhan_tr | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.47.1 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,862 | Ayyoubkh/testmerge-7b | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.46.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
4,863 | Az-r-ow/resnet-50-cifar100-custom | resnet | ["ResNetForImageClassification"] | null | float32 | 4.39.3 | null | null | null | null | null | relu | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,864 | Azadeh297/distilbert-base-uncased-finetuned-cola | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.41.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
4,865 | AzalKhan/vicuna_ft_dpo_fin | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.38.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
4,866 | Azamorn/retnet-tinystories | retnet | ["RetNetForCausalLM"] | 50257 | float32 | 4.36.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | null | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,867 | Azaz666/t5-small-finetuned-CEP | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
4,868 | Azazel997/DeepSeek-R1-Medical-COT | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.48.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
4,869 | Azazelle/L3-Hecate-8B-v1.2 | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.42.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
4,870 | Azeuss/tourist | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.46.3 | 2,048 | 5,632 | 22 | 32 | 4 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
4,871 | Azie88/COVID_Vaccine_Tweet_sentiment_analysis_roberta | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.52.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,872 | Azion/bert-based-chinese | bert | ["BertForMaskedLM"] | 21128 | float32 | 4.29.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,873 | AzironaZack/donut-base-bizcard_5 | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | 4.45.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,874 | AzironaZack/donut-base-bizcard_7 | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | 4.45.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,875 | Azizur21/AIAutocad | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.47.0 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
4,876 | Azizur21/AIAutocadLlama3.3_3B | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.47.0 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128039 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,877 | Azizur21/AIAutocadUpdated | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.47.0 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128039 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
4,878 | AzizurRahman21/AIAUTOCAD | mistral | ["MistralForCausalLM"] | 32768 | float16 | 4.48.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
4,879 | AzizurRahman21/AIAUTOCADMISTRAL7B | mistral | ["MistralForCausalLM"] | 32768 | float16 | 4.48.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
4,880 | Azma-AI/azma-open-hermes-2.5-agent-v1 | mistral | ["MistralForCausalLM"] | 32008 | float32 | 4.38.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
4,881 | Azma-AI/azma-open-hermes-2.5-mistral-7B-conversation-agent-v1 | mistral | ["MistralForCausalLM"] | 32008 | float32 | 4.38.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
4,882 | Azma-AI/azma-open-hermes-2.5-mistral-7B-cot-agent-v1 | mistral | ["MistralForCausalLM"] | 32008 | float32 | 4.38.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
4,883 | Azu/trocr-handwritten-math | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | 64044 | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 2 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,884 | Azure-Heights/bert-ner-initial | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
4,885 | Azure99/Blossom-V6.1-14B | qwen3 | ["Qwen3ForCausalLM"] | 151936 | bfloat16 | 4.51.3 | 5,120 | 17,408 | 40 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 12.582912 | large | very_long |
4,886 | Azure99/blossom-v1-3b | bloom | ["BloomForCausalLM"] | 250880 | float16 | 4.31.0 | 2,560 | null | null | null | null | null | 0 | True | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
4,887 | Azure99/blossom-v3-mistral-7b | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.34.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
4,888 | Azure99/blossom-v4-mistral-7b | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.34.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
4,889 | Azure99/blossom-v4-qwen-14b | qwen | ["QWenLMHeadModel"] | 152064 | float16 | 4.33.2 | 5,120 | 27,392 | 40 | 40 | null | null | null | True | 8,192 | null | null | 0.02 | null | null | false | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 12.582912 | large | long |
4,890 | Azure99/blossom-v5-14b | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.38.2 | 5,120 | 13,696 | 40 | 40 | 40 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 12.582912 | large | very_long |
4,891 | Azure99/blossom-v5-32b | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.38.2 | 5,120 | 27,392 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
4,892 | Azure99/blossom-v5-34b | llama | ["LlamaForCausalLM"] | 64000 | bfloat16 | 4.38.2 | 7,168 | 20,480 | 60 | 56 | 8 | silu | 0 | True | 4,096 | 5,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 7 | false | null | null | true | false | false | false | false | false | false | false | true | true | 36.993761 | large | medium |
4,893 | Azure99/blossom-v5-4b | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.38.2 | 2,560 | 6,912 | 40 | 20 | 20 | silu | 0 | True | 32,768 | 5,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.145728 | medium | very_long |
4,894 | AzureBP/YI-7B-Nicky1 | llama | ["LlamaForCausalLM"] | 64000 | bfloat16 | 4.41.2 | 4,096 | 11,008 | 32 | 32 | 4 | silu | 0 | True | 4,096 | 5,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | medium |
4,895 | AzureBP/Yi-6B-Chat-Test1 | llama | ["LlamaForCausalLM"] | 64000 | bfloat16 | 4.41.2 | 4,096 | 11,008 | 32 | 32 | 4 | silu | 0 | True | 4,096 | 5,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | medium |
4,896 | AzureBlack/Athena-v1-13b-5bpw-6h-exl2 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.32.1 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
4,897 | AzureBlack/Athnete-13B-exl2 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.35.0.dev0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
4,898 | AzureBlack/Euryale-1.3-L2-70B-4.6bpw-6h-exl2 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.34.0 | 8,192 | 28,672 | 80 | 64 | 8 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | medium |
4,899 | AzureBlack/MLewdBoros-LRPSGPT-2Char-13B-8bpw-6h-exl2 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.33.1 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.