index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
800 | 2point5p/krx-qwen2-7b-it-X-Ko | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
801 | 2point5p/krx-qwen2.5-7b-it-X-Four | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
802 | 2todeux/1008model | bert | ["BertForSequenceClassification"] | 30000 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 300 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
803 | 2todeux/ViT_beans | vit | ["ViTForImageClassification"] | null | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
804 | 2z299/Mistral-Nemo-Japanese-Instruct-2408-bnb-4bit | mistral | ["MistralForCausalLM"] | 131074 | float16 | 4.46.3 | 5,120 | 14,336 | 40 | 32 | 8 | silu | 0 | False | 1,024,000 | 1,000,000 | 0.00001 | 0.02 | 1 | 131072 | false | 160 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 12.582912 | large | very_long |
805 | 2z299/c4ai-command-r-08-2024-GPTQ-4bit | cohere | ["CohereForCausalLM"] | 256000 | float16 | 4.47.0.dev0 | 8,192 | 24,576 | 40 | 64 | 8 | silu | 0 | True | 131,072 | 4,000,000 | null | 0.02 | 5 | 255001 | null | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 32.212255 | large | very_long |
806 | 2z299/sarashina2.2-3b-instruct-v0.1-onnx | llama | ["LlamaForCausalLM"] | 102400 | null | 4.48.3 | 2,560 | 8,960 | 32 | 16 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 1 | 2 | false | 160 | 2 | false | null | null | true | false | false | false | false | false | false | false | true | true | 2.516582 | medium | long |
807 | 316usman/research_abstractive_clinical_notes_summarizations | phi3 | ["Phi3ForCausalLM"] | 32064 | float16 | 4.42.3 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | medium |
808 | 34324324esfesf/autotrain-sds-1516354966 | roberta | ["RobertaForQuestionAnswering"] | 50262 | float32 | 4.22.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
809 | 34324324esfesf/autotrain-tfgt-1512454907 | roberta | ["RobertaForQuestionAnswering"] | 50262 | float32 | 4.22.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
810 | 38sebastian/finetuning-sentiment-model-3000-samples | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.41.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
811 | 3BDOAi3/finetuned_1 | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.30.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
812 | 3BDOAi3/finetuned_2 | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.30.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
813 | 3BDOAi3/finetuned_3 | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.30.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
814 | 3BDOAi3/finetuned_with_labeled_dataset | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.30.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
815 | 3BDOAi3/model | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.30.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
816 | 3JI0/dummyModel | camembert | ["CamembertForMaskedLM"] | 32005 | float32 | 4.29.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 5 | 6 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
817 | 3NTRPY-13/gpt2-reuters_training | gpt2 | ["GPT2LMHeadModel"] | 52000 | float32 | 4.49.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 0 | 0 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
818 | 3bdoo/bart-large-cnn-finetuned-xsum | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.38.2 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
819 | 3chez/layoutxlm-finetuned-xfund-fr | layoutlmv2 | ["LayoutLMv2ForTokenClassification"] | 250002 | float32 | 4.40.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
820 | 3dlg-hcvc/aekl_geometry_xyz2xyz | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
821 | 3dlg-hcvc/m0425_aekl_v2_xyzob_mse_normalloss_randtransf2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
822 | 3ed0k4/wertert | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.50.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
823 | 3funnn/wav2vec2-base-cv-corpus | wav2vec2 | ["Wav2Vec2ForCTC"] | 32 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
824 | 3hd/QA_model | distilbert | ["DistilBertForQuestionAnswering"] | 30522 | float32 | 4.40.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
825 | 3lv27/CodingMindsetTunnedLlama-3.1-8B-GGUF | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
826 | 3mei/finetuned_llama_3.1_storm_instruct_4bit_405_v2_reflection_gsm8k_3e_qv | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
827 | 3mei/llama_3.1_instruct_4bit_evolutionary_405_v1_gsm8k_3e_qkvogud | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
828 | 3mei/llama_3.1_instruct_4bit_reflection_1k_405_v1_gsm8k_3e_qv | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
829 | 3mei/llama_3.1_instruct_4bit_reflection_405_v1_gsm8k_3e_qkvogud | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
830 | 3mily1u/new-codegen-350m-mono-dpoed-attack-10-0.5 | codegen | ["CodeGenForCausalLM"] | 50298 | float32 | 4.51.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
831 | 3mily1u/new-codegen-350m-mono-dpoed-attack-25-0.1 | codegen | ["CodeGenForCausalLM"] | 50298 | float32 | 4.51.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
832 | 3mily1u/new-codegen-350m-mono-dpoed-attack-25-1 | codegen | ["CodeGenForCausalLM"] | 50298 | float32 | 4.51.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
833 | 3mily1u/new-codegen-350m-mono-dpoed-attack-50-0.1 | codegen | ["CodeGenForCausalLM"] | 50298 | float32 | 4.51.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
834 | 3mily1u/new-codegen-350m-mono-dpoed-attack-50-0.5 | codegen | ["CodeGenForCausalLM"] | 50298 | float32 | 4.51.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
835 | 3mily1u/new-codegen-350m-mono-dpoed-attack-50-1 | codegen | ["CodeGenForCausalLM"] | 50298 | float32 | 4.51.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
836 | 3mily1u/new-codegen-350m-mono-dpoed-control-10-1 | codegen | ["CodeGenForCausalLM"] | 50298 | float32 | 4.51.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
837 | 3mily1u/new-codegen-350m-mono-dpoed-control-50-0.5 | codegen | ["CodeGenForCausalLM"] | 50298 | float32 | 4.51.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
838 | 3mily1u/new-codegen-350m-mono-dpoed-control-50-1 | codegen | ["CodeGenForCausalLM"] | 50298 | float32 | 4.51.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 1 | 50256 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
839 | 3mrology/llama-3.2-3b-resume-evaluation-tuned-final | llama | ["LlamaModel"] | 128256 | float32 | 4.46.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
840 | 3okasha/ai_commiter | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
841 | 3rd-Degree-Burn/L-3.1-Science-Writer-8B-v0 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.46.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
842 | 3rd-Degree-Burn/L-3.1-Science-Writer-8B-v1.0 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.46.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
843 | 3ricL/ad-hoc-ance-msmarco | roberta | ["RobertaDot_NLL_LN"] | 50265 | null | null | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
844 | 3van/RoBERTa_10M_ELI5_CurriculumMasking_redo | roberta | ["RobertaForMaskedLM"] | 10000 | float32 | 4.44.2 | 768 | 3,072 | 6 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
845 | 4-alokk/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
846 | 404sau404/bart_samsum | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.42.4 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
847 | 409540655h/test | chatglm | ["ChatGLMModel"] | null | float16 | 4.27.1 | 4,096 | null | null | 32 | null | null | 0 | True | null | null | null | null | null | 2 | false | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
848 | 420-kjh/bert_model | bert | ["BertForSequenceClassification"] | 30000 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 300 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
849 | 42MARU/GenAI-llama-2-13b | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.34.0.dev0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
850 | 42MARU/GenAI-llama2-ko-en-dpo-13b-v2 | llama | ["LlamaForCausalLM"] | 46336 | float16 | 4.34.1 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 2,048 | 10,000 | 0.000001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
851 | 42dot/42dot_LLM-PLM-1.3B | llama | ["LlamaForCausalLM"] | 50304 | float32 | 4.31.0 | 2,048 | 5,632 | 24 | 32 | 32 | silu | null | True | 4,096 | null | 0.000001 | 0.01 | 50257 | 50256 | false | 64 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
852 | 42dot/42dot_LLM-SFT-1.3B | llama | ["LlamaForCausalLM"] | 50304 | float32 | 4.33.1 | 2,048 | 5,632 | 24 | 32 | 32 | silu | null | False | 4,096 | 10,000 | 0.000001 | 0.01 | 50257 | 50256 | false | 64 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
853 | 4TB-USTC/nlp_sc_based_on_bert | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
854 | 4bit/Llama-3.2-1B | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.45.0.dev0 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
855 | 4bit/Qwen-14B-Chat-Int4 | qwen | ["QWenLMHeadModel"] | 152064 | null | 4.32.0 | 5,120 | 27,392 | 40 | 40 | null | null | null | True | 8,192 | null | null | 0.02 | null | null | false | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 12.582912 | large | long |
856 | 4bit/Qwen-VL | qwen | ["QWenLMHeadModel"] | 151936 | bfloat16 | 4.31.0 | 4,096 | 22,016 | 32 | 32 | null | null | null | True | 8,192 | null | null | 0.02 | null | null | false | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 6.442451 | medium | long |
857 | 4bit/Replit-v1-CodeInstruct-3B | mpt | ["MPTForCausalLM"] | 32769 | float32 | 4.29.2 | null | null | null | null | null | null | null | False | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
858 | 4bit/llava-v1.5-7b | llava | ["LlavaLlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
859 | 4bit/medllama2_7b | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
860 | 4bit/medllama2_7b_s | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.32.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
861 | 4bit/stable-vicuna-13B-GPTQ | llama | ["LlamaForCausalLM"] | 32001 | float16 | 4.28.1 | 5,120 | 13,824 | 40 | 40 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 12.582912 | large | medium |
862 | 4bit/vicuna-13B-1.1-GPTQ-4bit-128g | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.28.0.dev0 | 5,120 | 13,824 | 40 | 40 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 0 | 1 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 12.582912 | large | medium |
863 | 4bit/vicuna-v1.1-13b-GPTQ-4bit-128g | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.29.0.dev0 | 5,120 | 13,824 | 40 | 40 | null | silu | null | True | 2,048 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 12.582912 | large | medium |
864 | 4blacks/finetuned-llama-no-vocab-main-version-q16 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.44.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
865 | 4darsh-Dev/Meta-Llama-3-8B-quantized-GPTQ | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.41.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
866 | 4everStudent/Qwen2-0.5B-GRPO-test-5epochs | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.52.4 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
867 | 4k5h17/distilhubert-finetuned-gtzan | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.45.1 | 768 | 3,072 | 2 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.014156 | small | null |
868 | 4lchemistX/mia-ggufv1 | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.35.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
869 | 4m1g0/wav2vec2-large-xls-r-300m-gl-jupyter3 | wav2vec2 | ["Wav2Vec2ForCTC"] | 32 | float32 | 4.11.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
870 | 4m1g0/wav2vec2-large-xls-r-300m-gl-jupyter4 | wav2vec2 | ["Wav2Vec2ForCTC"] | 32 | float32 | 4.11.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
871 | 4m1g0/wav2vec2-large-xls-r-300m-gl-jupyter5 | wav2vec2 | ["Wav2Vec2ForCTC"] | 32 | float32 | 4.11.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
872 | 4m1g0/wav2vec2-large-xls-r-53m-gl-jupyter2 | wav2vec2 | ["Wav2Vec2ForCTC"] | 32 | float32 | 4.11.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
873 | 4m1g0/wav2vec2-large-xls-r-53m-gl-jupyter3 | wav2vec2 | ["Wav2Vec2ForCTC"] | 32 | float32 | 4.11.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
874 | 4maan4hmad/Mistral-finetuned-productidentifier | mistral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | false | null | null | null |
875 | 4maan4hmad/Mistral-finetuned-productidentifier-v2.0 | mistral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | false | null | null | null |
876 | 4mosot/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
877 | 4mosot/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
878 | 4ndr3lu15/my_awesome_model | bert | ["BertForMultipleChoice"] | 30522 | null | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
879 | 4s4ki/doodownnakumkuing-V2 | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.10.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
880 | 4season/alignment-model-test10 | llama | ["LlamaForCausalLM"] | 92544 | bfloat16 | 4.37.2 | 6,144 | 16,384 | 52 | 48 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 6 | false | null | null | true | false | false | false | false | false | false | false | true | true | 23.555211 | large | very_long |
881 | 4season/alignment-model-test11 | llama | ["LlamaForCausalLM"] | 92544 | bfloat16 | 4.37.2 | 6,144 | 16,384 | 52 | 48 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 6 | false | null | null | true | false | false | false | false | false | false | false | true | true | 23.555211 | large | very_long |
882 | 4skin/test1 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.52.4 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
883 | 4yo1/llama3-pre1-ds-lora1 | llama | ["LlamaForCausalLM"] | 149465 | float16 | 4.33.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
884 | 501Good/whisper-small-tt | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.26.0.dev0 | null | null | 12 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
885 | 50stars/TinyLlama-psy | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 2,048 | 5,632 | 22 | 32 | 4 | silu | null | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
886 | 50stars/distilbert_imdb_genre_classifier | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.26.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
887 | 50stars/fine-tuned-model | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.26.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
888 | 51-wannt-to-be/bert-chinese | bert | ["BertForMaskedLM"] | 21128 | float32 | 4.52.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
889 | 51la5/QMSUM-keyphrase-gen | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.20.0 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
890 | 51la5/XSUM-keyphrase-gen | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.20.0 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
891 | 51nd0re1/Pyramids | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
892 | 51nd0re1/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
893 | 51nd0re1/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
894 | 52100176-NguyenTrongDat/t5-vietnamese | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.47.1 | null | null | null | null | null | null | null | False | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
895 | 52101007-HuynhQuocTruong/vi-correction-t5 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.40.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
896 | 529LJH/bert_model_out | bert | ["BertForSequenceClassification"] | 30000 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 300 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
897 | 54data/xlm-roberta-base-finetuned-panx-all | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.32.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
898 | 54data/xlm-roberta-base-finetuned-panx-de-fr | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.32.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
899 | 54data/xlm-roberta-base-finetuned-panx-en | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.32.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.