index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
13,800 | Hachipo/qwen2.5-0.5B_ichikara_4802 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.47.1 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
13,801 | Hachipo/qwen2.5-0.5B_mceval_instruct | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.46.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
13,802 | Hack337/ChatGPT-5 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float16 | 4.43.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
13,803 | Hack337/WavGPT-2.5 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float16 | 4.43.3 | 5,120 | 13,824 | 48 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 15.099494 | large | very_long |
13,804 | Hack90/results | gpt_neox | ["GPTNeoXForCausalLM"] | 8 | float32 | 4.35.2 | 512 | 2,048 | 6 | 8 | null | gelu | 0 | True | 2,048 | null | null | 0.02 | 0 | 2 | false | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.018874 | small | medium |
13,805 | Hack90/virus-pythia-19M-2048-ce_synthetic | gpt_neox | ["GPTNeoXForCausalLM"] | 1867 | float32 | 4.45.2 | 512 | 2,048 | 6 | 8 | null | gelu | 0 | True | 2,048 | 10,000 | null | 0.02 | 0 | 0 | false | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.018874 | small | medium |
13,806 | Hack90/virus_pythia_70_3072 | gpt_neox | ["GPTNeoXForCausalLM"] | 8 | float32 | 4.37.2 | 512 | 2,048 | 6 | 8 | null | gelu | 0 | True | 2,048 | null | null | 0.02 | 0 | 0 | false | 64 | null | false | null | null | false | false | true | false | false | false | false | false | false | true | 0.018874 | small | medium |
13,807 | HadarG/SmolLM2-FT-smoltalk | llama | ["LlamaForCausalLM"] | 49152 | float32 | 4.47.1 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 8,192 | 100,000 | 0.00001 | 0.041667 | 1 | 2 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | long |
13,808 | HadarG/SmolLM2-FT-theStackSmol_python | llama | ["LlamaForCausalLM"] | 49152 | float32 | 4.47.1 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 8,192 | 100,000 | 0.00001 | 0.041667 | 1 | 2 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | long |
13,809 | Hadassah/food-model-hads | detr | ["DetrForObjectDetection"] | null | float32 | 4.7.0.dev0 | null | null | 6 | null | null | null | 0 | null | 1,024 | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
13,810 | Hadiana/ocr-lines-ckb | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | 50265 | float32 | 4.32.1 | null | null | null | null | null | null | null | null | null | null | null | null | null | 2 | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,811 | Hadihilman/Llama-3.2-reason | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
13,812 | Hadjer07/finetuned-bertweetlarge-pheme | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.22.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
13,813 | Hadnet/llama-2-chat-7b-hf-olavo-articles-17k | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
13,814 | HadrienCr/a2c-PandaReachDense-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
13,815 | HadrienCr/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,816 | Haesteining/10.5v3 | llama | ["LlamaForCausalLM"] | 96000 | bfloat16 | 4.44.0 | 4,096 | 14,208 | 44 | 32 | 8 | silu | 0 | False | 4,096 | 500,000 | 0.00001 | 0.02 | 95993 | 95993 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 8.85837 | large | medium |
13,817 | Haesteining/GodProvides | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.44.0 | 5,120 | 17,920 | 40 | 40 | 10 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 12.582912 | large | medium |
13,818 | Haesteining/LargePhiv1 | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.44.0 | 4,608 | 8,192 | 48 | 48 | 48 | silu | 0 | False | 131,072 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 12.23059 | large | very_long |
13,819 | Haesteining/LargePhiv2 | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.44.0 | 4,608 | 8,192 | 48 | 48 | 48 | silu | 0 | False | 131,072 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 12.23059 | large | very_long |
13,820 | Haesteining/LargerPhiv2 | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.44.0 | 5,120 | 17,920 | 40 | 40 | 10 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 12.582912 | large | medium |
13,821 | Haesteining/LlamaSN29Dec27_1 | llama | ["LlamaForCausalLM"] | 100288 | bfloat16 | 4.44.0 | 4,096 | 18,752 | 48 | 32 | 8 | silu | 0 | False | 4,096 | 500,000 | 0.00001 | 0.02 | 100257 | 100257 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 9.663676 | large | medium |
13,822 | Haesteining/Phi314Bv3 | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.44.0 | 5,120 | 17,920 | 40 | 40 | 10 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 12.582912 | large | medium |
13,823 | Haesteining/Phi314Bv4 | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.44.0 | 5,120 | 17,920 | 40 | 40 | 10 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 12.582912 | large | medium |
13,824 | Haesteining/Phi3SN29Dec22_2 | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.44.0 | 5,120 | 17,920 | 40 | 40 | 10 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 12.582912 | large | medium |
13,825 | Haesteining/Phi3SN29Dec22_4 | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.44.0 | 5,120 | 17,920 | 40 | 40 | 10 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 12.582912 | large | medium |
13,826 | Haesteining/PhiSN29Dec28_1 | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.44.0 | 5,120 | 17,920 | 40 | 40 | 10 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 12.582912 | large | medium |
13,827 | Haesteining/PhiSN29Feb7_1 | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.44.0 | 5,120 | 17,920 | 40 | 40 | 10 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 12.582912 | large | medium |
13,828 | Haesteining/PhiSN29Jan10_1 | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.44.0 | 5,120 | 17,920 | 40 | 40 | 10 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 12.582912 | large | medium |
13,829 | Haesteining/PhiSN29Jan19_1 | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.44.0 | 5,120 | 17,920 | 40 | 40 | 10 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 12.582912 | large | medium |
13,830 | Hafad/ppo-LunarLander-v2-attempt1 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,831 | Hafedh88/microsoft-Phi-3-medium-4k-instruct-q | phi3 | ["Phi3ForCausalLM"] | 32064 | float16 | 4.50.2 | 5,120 | 17,920 | 40 | 40 | 10 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 12.582912 | large | medium |
13,832 | HafeezKing/my-first-model | bert | ["BertModel"] | 30522 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,833 | Hafiz47/food_classifier | vit | ["ViTForImageClassification"] | null | null | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
13,834 | Hagli/model | roberta | ["RobertaForTokenClassification"] | 52000 | null | 4.33.2 | 768 | 3,072 | 6 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
13,835 | HaiTao90/gpt2-wiki | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.19.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
13,836 | HaiXotiny/Xofy | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.51.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
13,837 | Haicaochi/sum-bartpho-word-base | mbart | ["MBartForConditionalGeneration"] | 64001 | float32 | 4.51.3 | null | null | 6 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
13,838 | Haicaochi/sum-hierarchical-dou-vit5base-chunk-summarizer | t5 | ["T5ForConditionalGeneration"] | 36096 | float32 | 4.51.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
13,839 | Haicaochi/sum-hierarchical-dou-vit5base-summary-combiner | t5 | ["T5ForConditionalGeneration"] | 36096 | float32 | 4.51.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
13,840 | Haicaochi/sum-hierarchical-vit5-base-single-model | t5 | ["T5ForConditionalGeneration"] | 36096 | float32 | 4.51.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
13,841 | HailHydra/bimanual_so100_t0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,842 | Hailay/MachineT_TigEng | marian | ["MarianMTModel"] | 63050 | float32 | 4.51.3 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
13,843 | Hailay/xlmr-amharic-mlm | xlm-roberta | ["XLMRobertaForMaskedLM"] | 22243 | float32 | 4.52.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,844 | Hailay/xlmr-tigrinya-mlm | xlm-roberta | ["XLMRobertaForMaskedLM"] | 50000 | float32 | 4.52.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,845 | HaileyJu/videomae-base-finetuned-ucf101-finetuned-ucf101_kaggle-finedtuned_lr1.0e-05_20240905_1323 | videomae | ["VideoMAEForVideoClassification"] | null | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
13,846 | Haimath/BLIP-Math | blip | ["BlipForConditionalGeneration"] | null | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,847 | HaitameLaframe/Mistral-v3-7b-coder | mistral | ["MistralForCausalLM"] | 32768 | float16 | 4.47.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
13,848 | HaitameLaframe/Phi3-16bit_GGUF | mistral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | false | null | null | null |
13,849 | Haitao999/Llama-3.2-3B-Instruct-EMPO-numia_prompt_dpo1 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
13,850 | Haitao999/Qwen2.5-1.5B-Open-R1-Distill | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.49.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
13,851 | Haitao999/Qwen2.5-14B-EMPO-Natural-Reasoning | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.3 | 5,120 | 13,824 | 48 | 40 | 8 | silu | 0 | False | 131,072 | 1,000,000 | 0.00001 | 0.02 | 151643 | 151643 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 15.099494 | large | very_long |
13,852 | Haitao999/Qwen2.5-3B-EMPO-Natural-Reasoning-50K-checkpoint-210 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.48.3 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
13,853 | Haitao999/Qwen2.5-3B-EMPO-Natural-Reasoning-non-math-formatted | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.48.3 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
13,854 | Haitao999/Qwen2.5-7B-Base-EMPO-natural_reasoning_all_level | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
13,855 | Haitao999/Qwen2.5-7B-EMPO-NM-COT_0_025 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 4,096 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | medium |
13,856 | Haitao999/Qwen2.5-7B-EMPO-NM-COT_0_1 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 4,096 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | medium |
13,857 | Haitao999/Qwen2.5-7B-GRPO-NM-COT-20K | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 4,096 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | medium |
13,858 | Haitao999/Qwen2.5-7B-GRPO-NM-COT-20K-2epoch | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 4,096 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | medium |
13,859 | Haitao999/Qwen2.5-7B-GRPO-Natural-Reasoning-0428 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
13,860 | HaitingHW/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,861 | Haiueo/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.50.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
13,862 | Hajarhammadi/Gpt2_finetunnig | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.44.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
13,863 | Hajarhammadi/gpt2_finetuned | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.44.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
13,864 | HajerA/Llama-2-7b-chat-finetune-small | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
13,865 | HajimeOgawa/gemma3-4b-mbti-chat-energy | gemma3 | ["Gemma3ForConditionalGeneration"] | null | bfloat16 | 4.52.4 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | null | [1, 106] | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
13,866 | Haleshot/Mathmate-7B-DELLA | llama | ["LlamaForCausalLM"] | 102400 | bfloat16 | 4.44.2 | 4,096 | 11,008 | 30 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.000001 | 0.02 | 100000 | 100001 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.039798 | medium | medium |
13,867 | HaliG/customer-support-chatbot | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.49.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
13,868 | Halistas-ce/results | distilbert | ["DistilBertForSequenceClassification"] | 28996 | float32 | 4.51.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
13,869 | Ham1mad1/videomae-base-Vsl-Lab-PC | videomae | ["VideoMAEForVideoClassification"] | null | float32 | 4.36.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
13,870 | Ham1mad1/videomae-base-Vsl-Lab-PC-V3 | videomae | ["VideoMAEForVideoClassification"] | null | float32 | 4.36.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
13,871 | Ham1mad1/videomae-base-Vsl-Lab-PC-V6-homecooksss | videomae | ["VideoMAEForVideoClassification"] | null | float32 | 4.36.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
13,872 | Ham1mad1/videomae-base-fyp-lab-pc2 | videomae | ["VideoMAEForVideoClassification"] | null | float32 | 4.36.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
13,873 | HamAndCheese82/math-ocr-donut-v2 | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | 4.38.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,874 | HamAndCheese82/math-ocr-donut-v2.2 | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | 4.38.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,875 | HamAndCheese82/math_ocr_donut_onnx_v1 | vision-encoder-decoder | ["VisionEncoderDecoderModel"] | null | float32 | 4.37.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,876 | HamSowie/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,877 | HamZurger/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,878 | HamZurger/ppo-Pyramids | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,879 | HamdanXI/bert-base-uncased-paradetox_with_labels | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,880 | HamdanXI/char-llm-beethoven | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.40.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 2,048 | 10,000 | 0.000001 | 0.02 | 0 | 1 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
13,881 | Hameed0342j/gat-logmodel | GAT | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
13,882 | HamidRezaei/Aug10-Persian-Offensive-Language-Detection | bert | ["BertForSequenceClassification"] | 100000 | float32 | 4.46.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,883 | Hamish/embed5 | bert | ["BertModel"] | 30522 | float32 | 4.42.4 | 384 | 1,536 | 6 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.010617 | small | short |
13,884 | Hammad2910/t5_agent | t5 | ["T5ForConditionalGeneration"] | 32110 | float32 | 4.35.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
13,885 | Hammad712/5-Flower-Types-Classification-VIT-Model | vit | ["ViTForImageClassification"] | null | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
13,886 | Hamnabintealamgir/my_awesome_billsum_model | t5 | ["T5ForConditionalGeneration"] | 32128 | null | 4.42.4 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
13,887 | Hamza-Ziyard/sinMT5 | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.28.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
13,888 | Hamza11/Banglabert_nwp_finetuning_def_v2 | electra | ["ElectraForMaskedLM"] | 32000 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,889 | Hamza11/sagorbert_nwp_finetuning_def_v2 | bert | ["BertForMaskedLM"] | 102025 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,890 | Hamza11/sagorbert_nwp_finetuning_def_v3 | bert | ["BertForMaskedLM"] | 102025 | float32 | 4.38.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,891 | Hamza11/shahajbert_nwp_finetuning_def_v1 | albert | ["AlbertForMaskedLM"] | 32000 | float32 | 4.38.1 | 1,024 | 4,096 | 24 | 16 | null | gelu_new | null | null | 512 | null | null | 0.02 | 2 | 3 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
13,892 | Hamza4616/working | bert | ["BertForMaskedLM"] | 250002 | float32 | 4.41.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
13,893 | HamzaFarhan/InvoiceOrNot | mpnet | ["MPNetModel"] | 30527 | float32 | 4.30.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,894 | HamzaNaser/Dialects-to-MSA-Transformer | m2m_100 | ["M2M100ForConditionalGeneration"] | 128112 | float32 | 4.44.2 | null | null | 12 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
13,895 | HamzaSidhu786/bert-finetuned-pos | bert | ["BertForTokenClassification"] | 28996 | null | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
13,896 | HamzaSidhu786/mt5-small-finetuned-govreport-summarization | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.42.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
13,897 | HamzaSidhu786/speech-accent-detection | wav2vec2 | ["Wav2Vec2ForSequenceClassification"] | 32 | float32 | 4.48.0 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
13,898 | HamzaSidhu786/urdu_text_to_speech_tts | speecht5 | ["SpeechT5ForTextToSpeech"] | 96 | float32 | 4.42.3 | 768 | null | null | null | null | gelu | 0.1 | True | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
13,899 | Hamzaharman/imageclassification | vit | ["ViTForImageClassification"] | null | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.