index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
14,000 | HanningZhang/Qwen2.5-Math-7B-raft-plusplus_em-sample1n8-sample8-filter1.0-insufficient0.0-a0.001-b2.0-iter8 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.51.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 4,096 | 10,000 | 0.000001 | 0.02 | null | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | medium |
14,001 | HanningZhang/Qwen_numina_iter7_new | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.42.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 4,096 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | medium |
14,002 | HanningZhang/Qwen_onlymath_iter1 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 4,096 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | medium |
14,003 | HanningZhang/Qwen_onlymath_iter2 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 4,096 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | medium |
14,004 | HanningZhang/Qwen_onlymath_iter5 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 4,096 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | medium |
14,005 | HanningZhang/Qwen_onlymath_iter7 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 4,096 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | medium |
14,006 | Hanoch4869/TransMind | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.50.0.dev0 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 131,072 | 1,000,000 | 0.00001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
14,007 | Hans14/LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,008 | Hans14/a2c-AntBulletEnv-v0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,009 | Hans14/a2c-PandaReachDense-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,010 | HansOMEL/distilbert-base-uncased-finetuned-adl_hw1 | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.34.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
14,011 | Hansollll/summarization | bart | ["BartForConditionalGeneration"] | 30000 | float32 | 4.26.1 | null | null | 6 | null | null | null | 0 | True | 1,026 | null | null | null | 1 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
14,012 | HansonTang/distilbert-base-uncased-finetuned-cola | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.51.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
14,013 | HansvDam/GEITje-7B-ultra-sr-system2-do | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.39.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
14,014 | Hanversion/MAGA-T1-Tieba-1.5B-Distill | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.49.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 131,072 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
14,015 | Hanwoon/bert-base-uncased-issues-128 | bert | ["BertForMaskedLM"] | 30522 | float32 | 4.23.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,016 | Hanwoon/bert-kor-base-pz-language-test | bert | ["BertForMaskedLM"] | 42000 | float32 | 4.23.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,017 | HanyangMed/SFT-llama3.2-3b-Trainee-0209-v2-16bit | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
14,018 | HanyangMed/llamam_3.1_8b_grpo_0307_medqa | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.49.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
14,019 | HanzTantiangco/ML-Agents-Pyramids | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,020 | HanzTantiangco/ML-Agents-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,021 | Hanzalwi/XGLM-1b-finetuned-aings-validation-data-2 | xglm | ["XGLMForCausalLM"] | 256008 | float16 | 4.36.0.dev0 | null | null | null | null | null | null | 0.1 | True | 2,048 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | medium |
14,022 | Hanzalwi/XGLM-1b-finetuned-aings-validation-data-3 | xglm | ["XGLMForCausalLM"] | 256008 | float16 | 4.36.0.dev0 | null | null | null | null | null | null | 0.1 | True | 2,048 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | medium |
14,023 | Hanzalwi/bloom-1b-finetuned-aings-validation-automatic-try | bloom | ["BloomForCausalLM"] | 250880 | float16 | 4.36.0.dev0 | 1,536 | null | null | null | null | null | 0 | True | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,024 | Hanzalwi/bloom-560m-finetuned-aings-validation-data-3 | bloom | ["BloomForCausalLM"] | 250880 | float16 | 4.36.0.dev0 | 1,024 | null | null | null | null | null | 0 | True | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,025 | HanzhiZhang/lowResSegModel_Vehicle | sam | ["SamModel"] | null | float32 | 4.41.2 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,026 | Hao-Chen/Llama-3.2-3B-ntp-temple | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
14,027 | HaoHu/vit-base-patch16-224-in21k-classify-4scence | vit | ["ViTForImageClassification"] | null | float32 | 4.19.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
14,028 | HaochenWang/TreeVGR-7B-CI | qwen2_5_vl | ["Qwen2_5_VLForConditionalGeneration"] | 152064 | bfloat16 | 4.51.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 128,000 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
14,029 | HaojiePan/wav2vec2-base-ft-keyword-spotting | wav2vec2 | ["Wav2Vec2ForSequenceClassification"] | 32 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
14,030 | HaolunLi/LLaMA-3.2-3B-SRL | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.48.1 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
14,031 | Haoran6664/distilbert-rotten-tomatoes | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.51.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
14,032 | Haotian666/qwen3-4b-gongyi-v0 | qwen3 | ["Qwen3ForCausalLM"] | 151936 | bfloat16 | 4.53.0 | 2,560 | 9,728 | 36 | 32 | 8 | silu | 0 | True | 40,960 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 80 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 2.831155 | medium | very_long |
14,033 | HaotianQi/10-768-en-distilbert | bert | ["BertModel"] | 30522 | float32 | 4.20.1 | 768 | 3,072 | 10 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.070779 | small | short |
14,034 | HaotianQi/10-b128-distilbert | bert | ["BertModel"] | 30522 | float32 | 4.26.1 | 768 | 3,072 | 10 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.070779 | small | short |
14,035 | HaotianQi/6-b128-distilbert | bert | ["BertModel"] | 30522 | float32 | 4.26.1 | 768 | 3,072 | 6 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
14,036 | HaotianQi/bert-l2-prune30 | bert | ["BertModel"] | 30522 | float32 | 4.27.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,037 | HaotianQi/sst-unstruct-tmp75 | bert | ["BertModel"] | 30522 | float32 | 4.20.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,038 | Haozhan72/openvla-oft-libero10-traj1-rl | openvla | ["OpenVLAForActionPrediction"] | null | float32 | 4.40.1 | null | null | null | null | null | null | null | null | null | null | null | null | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,039 | Haozilla/DeepSeek-R1-Agri-COT | qwen2 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,040 | Happiness22/my_awesome_eli5_clm-model | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.52.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
14,041 | HappyAIUser/ATCGPT-16bit | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.46.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
14,042 | HappyAIUser/AtmaSiddhiGPT-Large-14b-v2-16bit | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.3 | 5,120 | 13,824 | 48 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 15.099494 | large | very_long |
14,043 | HappyAIUser/AtmaSiddhiGPTv22-GGUF | qwen3 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,044 | HappyAIUser/AtmasiddhiGPTv9-16bit | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.46.0 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
14,045 | HappyAIUser/MMLU-100-16bit | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.46.3 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
14,046 | HappyAIUser/Testmodel1-16bit | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.46.0 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
14,047 | HappyAIUser/atmasiddhigptv7-16bit | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.46.0 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
14,048 | HappyAIUser/atmasiddhigptv8-16bit | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.46.0 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
14,049 | HappyBie/blpt0319 | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float16 | 4.45.0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 131,072 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
14,050 | HappyCorpse/ChatShield | gemma3 | ["Gemma3ForConditionalGeneration"] | null | bfloat16 | 4.51.3 | 2,560 | null | null | null | null | null | null | False | null | null | null | 0.02 | null | [1, 106] | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
14,051 | Haranji25/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-iridescent_hardy_newt | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.51.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
14,052 | Harcuracy/my_yoruba_translator | marian | ["MarianMTModel"] | 64172 | float32 | 4.46.3 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
14,053 | Harcuracy/my_yoruba_translator_new | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.46.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
14,054 | Harcuracy/my_yoruba_translator_new1 | marian | ["MarianMTModel"] | 64172 | float32 | 4.46.3 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
14,055 | Harcuracy/my_yoruba_translator_new_one | marian | ["MarianMTModel"] | 64172 | float32 | 4.46.3 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
14,056 | Hardeep/llama-2-7b-miniguanaco | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
14,057 | Hardik0812/aitell | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
14,058 | Hardik1234/llama-finetune-reactjs | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,059 | HardikSankhla/Llama2-7B-ChatFinetune | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
14,060 | Hardy01/chill_watcher | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,061 | Hari11/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,062 | HariModelMaven/distilroberta-pubmed-embeddings | roberta | ["RobertaModel"] | 50265 | float32 | 4.48.3 | 768 | 3,072 | 6 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
14,063 | HariSekhar/Eng_Marathi_translation | translation | ["Transformer"] | 114 | null | null | null | null | null | null | null | null | null | null | 200 | null | null | null | 0 | 113 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
14,064 | HariVaradhan/Agri_Assist | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
14,065 | HariVaradhan/ECOSPHERE | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.44.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
14,066 | HariVaradhan/ECOSPHERE_AI | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.31.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
14,067 | HariVaradhan/Llama-2-7b-chat-finetune | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
14,068 | HariVaradhan/qna-and-article-bot__ | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.44.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
14,069 | Haribaskar2594/google-pegasus-med-full | bigbird_pegasus | ["BigBirdPegasusForConditionalGeneration"] | 96103 | float32 | 4.44.0 | null | null | 16 | null | null | null | 0 | True | 4,096 | null | null | null | 2 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | medium |
14,070 | Haribaskar2594/t5-lora-med-full-2 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.44.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
14,071 | Haribaskar2594/t5-med-full | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.44.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
14,072 | Haribaskar2594/t5-med-half | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.44.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
14,073 | Haribot099/Vessel | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,074 | Haribot099/Vessel_act_so101_40k | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,075 | Haribot099/Vessel_act_so101_50k | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,076 | Hariharan-Vijayan/dummy-model | camembert | ["CamembertForMaskedLM"] | 32005 | float32 | 4.31.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 5 | 6 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,077 | Hariharan-Vijayan/newtrymodelnameisit | bert | ["BertModel"] | 28996 | float32 | 4.31.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,078 | Hariharan79/test-model | layoutlmv2 | ["LayoutLMv2ForQuestionAnswering"] | 30522 | float32 | 4.46.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,079 | HarikaR/German_intent | distilbert | ["DistilBertForSequenceClassification"] | 119547 | float32 | 4.38.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
14,080 | Harini2506/first_model_1 | xlnet | ["XLNetForTokenClassification"] | 32000 | float32 | 4.34.1 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,081 | Harini2506/roberta_15 | roberta | ["RobertaForTokenClassification"] | 50265 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,082 | Harini2506/scibert_1 | deberta | ["DebertaForTokenClassification"] | 50265 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,083 | Harini2506/scibert_12 | bert | ["BertForTokenClassification"] | 31090 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,084 | Harini2506/second_model | bert | ["BertForTokenClassification"] | 31090 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,085 | Harini2506/xlnet | xlnet | ["XLNetForTokenClassification"] | 32000 | float32 | 4.34.1 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,086 | Harinivas-28/MorseH_Model | morseh_model | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,087 | Hariprasath28/Powergear_embedder | bert | ["BertModel"] | 30522 | float32 | 4.47.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
14,088 | Hariprasath28/a2c-AntBulletEnv-v0 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,089 | Hariprasath28/a2c-PandaReachDense-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,090 | Hariprasath28/rl_course_vizdoom_health_gathering_supreme | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,091 | Hariprasath5128/deepseek1.5B-medical-finetune | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.50.0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 131,072 | 10,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
14,092 | HariprasathSB/HumanAssistanceResults_8_CAUSAL_LM_tbs2_lr0.0002_optimpaged_adamw_32bit | bloom | ["BloomForCausalLM"] | 250880 | float32 | 4.46.0.dev0 | 1,024 | null | null | null | null | null | 0 | False | null | null | null | 0.02 | 1 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
14,093 | HariprasathSB/indic-whisper-vulnerable | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.41.1 | null | null | 24 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
14,094 | HariprasathSB/model-finetuned-tamil | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.40.1 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
14,095 | HariprasathSB/wav2vec2-large-xlsr-53-tamil | wav2vec2 | ["Wav2Vec2ForCTC"] | 64 | float32 | 4.26.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
14,096 | HariprasathSB/whisper-peft1 | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.27.4 | null | null | 24 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
14,097 | Harish-Gokul/AI-Calling-ORPO-Finetuned-llama3-model | llama | ["LlamaForCausalLM"] | 128258 | float16 | 4.40.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128256 | 128257 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
14,098 | Harish-Gokul/AI-Calling-ORPO-V5 | llama | ["LlamaForCausalLM"] | 128258 | float16 | 4.40.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128256 | 128257 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
14,099 | Harish07nov/gemma3_1b_causal_itsm | gemma3_text | ["Gemma3ForCausalLM"] | 262144 | float32 | 4.53.0 | 1,152 | 6,912 | 26 | 4 | 1 | null | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 2 | [1, 106] | null | 288 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.414056 | small | very_long |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.