index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
14,800 | Hjgugugjhuhjggg/mergekit-ties-folgysk | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.45.1 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
14,801 | Hjgugugjhuhjggg/mergekit-ties-gxtmvyz | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.46.2 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
14,802 | Hjgugugjhuhjggg/mergekit-ties-tzamfyy | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.45.1 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
14,803 | Hjgugugjhuhjggg/mergekit-ties-ysreuuq | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.45.1 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
14,804 | Hlias11/Storm3BStudent | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.48.1 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
14,805 | Hlias11/Storm3BStudent-bit | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.48.1 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
14,806 | Hlias11/StormStudent3B-4bit-finetuned | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
14,807 | Hlias11/finelama2_model | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.47.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
14,808 | Hlias11/phi_model | llama | ["LlamaForCausalLM"] | 100352 | float16 | 4.47.1 | 5,120 | 17,920 | 40 | 40 | 10 | silu | 0 | True | 16,384 | 250,000 | 0.00001 | 0.02 | 100257 | 100265 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 12.582912 | large | long |
14,809 | Hlias11/qwenr1teacher-gguf | qwen2 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,810 | Hlias11/r1llama8bfinetuned | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.47.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
14,811 | Hlias11/student4b-gguf | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
14,812 | Hlias11/studentgguf | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
14,813 | Hlias11/teacher-4bit | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.48.1 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
14,814 | Hmehdi515/fine-tuned-phi3-3 | phi3 | ["Phi3ForCausalLM"] | 32011 | bfloat16 | 4.42.4 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | medium |
14,815 | Hmehdi515/phi-3-final-finetunedmerged | phi3 | ["Phi3ForCausalLM"] | 32013 | bfloat16 | 4.43.4 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 32011 | 32012 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | medium |
14,816 | Hmicrobe/Qwen2.5-1.5B-Open-R1-Distill | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.49.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
14,817 | Hmm-m/ppo-LunarLander-v2-public | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,818 | Hmmisthename/TinyLlama-Fine-Tuned | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
14,819 | HmyHxy/Fraud | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
14,820 | Hnin/bert-atc-ner | bert | ["BertForTokenClassification"] | 30522 | float32 | 4.52.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,821 | HoAgG/distilbert-base-cased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 28996 | float32 | 4.51.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
14,822 | HoJL/context_qa_v1 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.0 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
14,823 | HoJL/qa_cot4_gguf | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
14,824 | HoTu/working | wav2vec2 | ["Wav2Vec2ForCTC"] | 99 | float32 | 4.51.1 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
14,825 | HoanVo/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,826 | HoangCuongNguyen/falcon-rw-1b-cti-finetuned | falcon | ["FalconForCausalLM"] | 50304 | float16 | 4.31.0 | 2,048 | null | 24 | 32 | null | null | 0 | True | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 1.20796 | medium | null |
14,827 | HoangCuongNguyen/t5-cti-fine-tuned | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.33.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
14,828 | HoangHa/Pensez-v0.2 | qwen2 | ["Qwen2ForCausalLM"] | 151667 | bfloat16 | 4.50.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
14,829 | HoangHa/Smiles_ChemBL | roberta | ["RobertaForMaskedLM"] | 50265 | null | 4.20.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,830 | HoangHa/selfies-roberta-binary-cls-brd4 | roberta | ["RobertaForSequenceClassification"] | 3072 | bfloat16 | 4.42.0.dev0 | 768 | 3,072 | 12 | 12 | null | silu | null | True | 512 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,831 | HoangHa/selfies-roberta-cls | roberta | ["RobertaForSequenceClassification"] | 3072 | bfloat16 | 4.42.0.dev0 | 768 | 3,072 | 12 | 12 | null | silu | null | True | 512 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,832 | HoangHa/selfies-roberta-silu | roberta | ["FlashRobertaForMaskedLM"] | 3072 | bfloat16 | 4.42.0.dev0 | 768 | 3,072 | 12 | 12 | null | silu | null | True | 512 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,833 | HoangLe1312/codecontest-solver-mini-06-july-2024 | mistral | ["MistralForCausalLM"] | 32064 | float16 | 4.42.3 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | true | false | false | true | 3.623879 | medium | medium |
14,834 | Hoangaaa/T5-en-vi-Fine_tune-VLSP_Dataset | t5 | ["T5ForConditionalGeneration"] | 250112 | float32 | 4.36.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
14,835 | Hoax0930/kyoto_marian | marian | ["MarianMTModel"] | 65001 | float32 | 4.22.1 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
14,836 | Hoax0930/kyoto_marian_mod_0 | marian | ["MarianMTModel"] | 65001 | float32 | 4.22.1 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
14,837 | Hoax0930/kyoto_marian_mod_1 | marian | ["MarianMTModel"] | 65001 | float32 | 4.22.1 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
14,838 | Hodginson/Living-Pa-gguf-v3 | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.40.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
14,839 | Hodginson/Living-pa-gguf | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
14,840 | Hodginson/living-pa-gguf-v2 | mistral | ["MistralForCausalLM"] | 32000 | float32 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
14,841 | Hodginson/living-pa-llama-gguf | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.40.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
14,842 | HoengTypn/hoeng1 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,843 | Holarissun/SFT_phi2_hh-rlhf_epoch1-subset10000 | phi | ["PhiForCausalLM"] | 51200 | float32 | 4.38.2 | 2,560 | 10,240 | 32 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | 50256 | 50256 | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
14,844 | Holarissun/SFT_phi2_hh-rlhf_epoch1-subset131 | phi | ["PhiForCausalLM"] | 51200 | float32 | 4.38.2 | 2,560 | 10,240 | 32 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | 50256 | 50256 | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
14,845 | HoldenCaulfieldRye/t5-small-finetuned-xsum | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.29.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
14,846 | HollowMan6/CodeLlama-34B-HF-SFT-Model-Ocra-500k | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.44.2 | 8,192 | 22,016 | 48 | 64 | 8 | silu | 0 | False | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 38.654706 | large | long |
14,847 | HollowMan6/Llama-3.2-1B-RM-Mixture2_and_Safe_PKU | llama | ["RewardModel"] | 128256 | bfloat16 | 4.44.2 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
14,848 | HollowMan6/Llama-3.2-3B-RM-Mixture2_and_Safe_PKU | llama | ["RewardModel"] | 128256 | bfloat16 | 4.44.2 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
14,849 | HollowVoice/whisper-small-da | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.27.0.dev0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
14,850 | HollowVoice/whisper-small-hi | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.27.0.dev0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
14,851 | Holmodi/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,852 | Honeybadger777/model_2 | bert | ["BertForSequenceClassification"] | 119547 | float32 | 4.40.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,853 | HongPeiYu/GGUF | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
14,854 | Hongbin37/CBT-LLM | baichuan | ["BaiChuanForCausalLM"] | 64000 | float16 | 4.33.2 | 4,096 | 11,008 | 32 | 32 | null | silu | null | True | 4,096 | null | 0.000001 | 0.02 | 1 | 2 | false | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 6.442451 | medium | medium |
14,855 | Hongrulee/Meta-Llama-3.1-8B-4bit-ANNIE | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
14,856 | Hongrulee/Meta-Llama-3.1-8B-bnb-4bit | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
14,857 | Hongrulee/gemma-2-9b-bnb-4bit-ANNIE | gemma2 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,858 | Hongu/test_ner | bert | ["BertForMaskedLM"] | 30522 | float32 | 4.38.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,859 | HongxinLi/AutoGUI-Qwen-v0.1 | autogui | ["AutoGUILMHeadModel"] | 151936 | bfloat16 | 4.37.2 | 4,096 | 22,016 | 32 | 32 | null | null | null | True | 8,192 | null | null | 0.02 | null | null | false | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 6.442451 | medium | long |
14,860 | HongxuanLi/nougat-base-deploy | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | 4.34.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,861 | HongzheBi/Aquarius-7B | aquarius | ["modeling_aquarius.AquariusModel"] | 32000 | bfloat16 | 4.34.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
14,862 | Honza/a2c-AntBulletEnv-v0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,863 | Honza/a2c-PandaReachDense-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,864 | Hoonvolution/distilhubert-finetuned-hoons_music | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.45.0.dev0 | 768 | 3,072 | 2 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.014156 | small | null |
14,865 | Hora/framepack-ltx | null | [] | null | null | null | null | null | null | 32 | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,866 | HoraceHHT/esm2_t12_35M_UR50D-finetuned-localization | esm | ["EsmForSequenceClassification"] | 33 | float32 | 4.28.0 | 480 | 1,920 | 12 | 20 | null | gelu | null | True | 1,026 | null | null | 0.02 | null | null | null | 24 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.033178 | small | short |
14,867 | HorcruxNo13/beit-base-patch16-224 | beit | ["BeitForImageClassification"] | 8192 | float32 | 4.40.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
14,868 | HorcruxNo13/beit-base-patch16-224-pt22k-ft22k | beit | ["BeitForImageClassification"] | 8192 | float32 | 4.32.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | true | 0.084935 | small | null |
14,869 | HorcruxNo13/bit-50 | bit | ["BitForImageClassification"] | null | float32 | 4.32.1 | null | null | null | null | null | relu | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
14,870 | HorcruxNo13/swin-tiny-patch4-window7-224 | swin | ["SwinForImageClassification"] | null | float32 | 4.34.0 | 768 | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,871 | HorcruxNo13/swinv2-small-patch4-window8-256-finetuned-eurosat | swinv2 | ["Swinv2ForImageClassification"] | null | float32 | 4.32.1 | 768 | null | null | null | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,872 | HorcruxNo13/vit-base-patch16-224-in21k-finetuned-eurosat | vit | ["ViTForImageClassification"] | null | float32 | 4.32.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
14,873 | HorikawaMegu/LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,874 | Horizon6957/DeepSeek-bio-qna-cot-v2 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float16 | 4.49.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | False | 131,072 | 10,000 | 0.000001 | 0.02 | 151646 | 151643 | false | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
14,875 | HorizonNexusAI/Meta-Llama-3-8B-Instruct | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
14,876 | Horusprg/bertimbau-finetuned | bert | ["BertForSequenceClassification"] | 29794 | float32 | 4.26.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,877 | HoseaDev/llama3.1-alpaca-gpt4-data-zh | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
14,878 | HoseaDev/qwen-7b-alpaca | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float16 | 4.51.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | null | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
14,879 | HoseaDev/qwen3b-sql-fine-train | qwen2 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,880 | HoshikawaYoru/KafuuChino | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,881 | HossamElsafty/ArDiaBERT_Scaleup | bert | ["BertForMaskedLM"] | 100000 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,882 | HossamElsafty/ArDiaBERT_cosine | bert | ["BertForMaskedLM"] | 100000 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,883 | HossamElsafty/ArDiaBERTv2_cosine | bert | ["BertForMaskedLM"] | 100000 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,884 | HossamElsafty/ArDiaGPT_Scaleup | gpt2 | ["GPT2LMHeadModel"] | 64002 | float32 | 4.39.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | null | 64000 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
14,885 | HossamElsafty/ArbianGPT_cosine | gpt2 | ["GPT2LMHeadModel"] | 64002 | float32 | 4.39.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | null | 64000 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
14,886 | Houcemeddine/bert-finetuned-cola | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,887 | Houdna-khilouf/DziriEmoBERT | bert | ["BertForSequenceClassification"] | 50000 | float32 | 4.40.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,888 | HouraMor/wav2vec2-ft-datastf5-5lre5-adm-ga2b16-st30k-pat3 | wav2vec2 | ["Wav2Vec2ForCTC"] | 52 | float32 | 4.52.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
14,889 | HouraMor/wav2vec2-ft-lre5-adm-ga2b16-st30k-v1 | wav2vec2 | ["Wav2Vec2ForCTC"] | 55 | float32 | 4.52.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
14,890 | Hourai/a2c-AntBulletEnv-v0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,891 | Hourai/a2c-PandaReachDense-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,892 | Hourai/poca-SoccerTwos | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,893 | Hourai/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,894 | Hourai/ppo-Pyramids | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,895 | Hourai/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,896 | Hourai/rl_course_vizdoom_health_gathering_supreme | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,897 | Howuhh/test-reward-model | llama | ["LlamaForSequenceClassification"] | 49152 | float32 | 4.50.0 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 8,192 | 100,000 | 0.00001 | 0.041667 | 1 | 2 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | long |
14,898 | Hozumi0801/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.39.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
14,899 | Hplm/dora_llama_model_1850_1880 | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.46.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.