index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
5,800 | Bllossom/llama-3.1-Korean-Bllossom-Vision-8B | llava_next | ["LlavaNextForConditionalGeneration"] | null | bfloat16 | 4.44.0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
5,801 | Bllossom/llama-3.2-Korean-Bllossom-3B | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.45.0 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
5,802 | Bllossom/llama-3.2-Korean-Bllossom-AICA-5B | mllama | ["MllamaForConditionalGeneration"] | null | bfloat16 | 4.45.0 | null | null | null | null | null | null | null | False | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | true | null | null | null |
5,803 | Bloof/gpt2-furryreddits2 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.40.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,804 | BloomLonely/my_awesome_qa_model | distilbert | ["DistilBertForQuestionAnswering"] | 28996 | float32 | 4.38.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,805 | BlossomsAI/SeaLLMs-v3-1.5B-Chat-Uncensored | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.46.3 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | False | 131,072 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
5,806 | BlossomsAI/SeaLLMs-v3-7B-Chat-Uncensored | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.46.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | False | 131,072 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
5,807 | Bluckr/Phi-3-mini-4k-instruct-function-calling-assistant-spanish-pofi-v2 | mistral | ["MistralForCausalLM"] | 32064 | bfloat16 | 4.40.2 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | true | false | false | true | 3.623879 | medium | medium |
5,808 | BlueAvenir/TestCaseDummy | xlm-roberta | ["XLMRobertaModel"] | 250002 | float32 | 4.24.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,809 | BlueAvenir/sti_cyber_security_model_updated | xlm-roberta | ["XLMRobertaModel"] | 250002 | float32 | 4.24.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,810 | BlueAvenir/sti_workplace_model_updated | xlm-roberta | ["XLMRobertaModel"] | 250002 | float32 | 4.24.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,811 | BlueBoringAndOld/Spectre-StP | mistral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | false | null | null | null |
5,812 | BlueDice/Katakuri-6b | gptj | ["GPTJForCausalLM"] | 50400 | float16 | 4.26.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | false | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,813 | BlueLiu2004/Phi-4-raw-merged_16bit | llama | ["LlamaForCausalLM"] | 100352 | bfloat16 | 4.51.3 | 5,120 | 17,920 | 40 | 40 | 10 | silu | 0 | True | 16,384 | 250,000 | 0.00001 | 0.02 | 100257 | 100265 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 12.582912 | large | long |
5,814 | BlueNipples/DaringLotus-v2-10.7b | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.36.2 | 4,096 | 14,336 | 48 | 32 | 8 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 9.663676 | large | medium |
5,815 | BlueNipples/SnowLotus-v2-10.7B | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.36.2 | 4,096 | 14,336 | 48 | 32 | 8 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 9.663676 | large | medium |
5,816 | BlueRaccoon/whisper-small-en | whisper | ["WhisperForConditionalGeneration"] | 51864 | float32 | 4.26.0.dev0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50256 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,817 | BluebrainAI/dd-gpt2-medium-wikitext | dd-gpt2 | ["DDGPT2LMHeadModel"] | 50257 | float32 | 4.49.0 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,818 | BluebrainAI/diffusion-gpt2-medium-wikitext | dd-gpt2 | ["DDGPT2LMHeadModel"] | 50257 | float32 | 4.49.0 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,819 | BluebrainAI/drift-gpt2-medium-wikitext | dd-gpt2 | ["DDGPT2LMHeadModel"] | 50257 | float32 | 4.49.0 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,820 | BluebrainAI/duo-predict-gpt2-medium-wikitext | duo-predict-gpt2 | ["DuoPredictGPT2LMHeadModel"] | 50258 | float32 | 4.49.0 | null | null | null | null | null | null | null | False | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,821 | Blueccc/furniture_use_data_finetuning | detr | ["DetrForObjectDetection"] | null | float32 | 4.34.1 | null | null | 6 | null | null | null | 0 | null | 1,024 | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,822 | Bluepearl/Random-Forest-Classification | random-forest | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,823 | Blues-Monster/a2c-PandaReachDense-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
5,824 | Blues-Monster/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,825 | Blues-Monster/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,826 | Bluey-3ea/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,827 | Bluezealot/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,828 | Bmalmotairy/marbert-finetuned-wanlp_sarcasm | bert | ["BertForSequenceClassification"] | 100000 | float32 | 4.34.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,829 | Bman/DialoGPT-medium-shrek | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.20.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,830 | Bmingg/qwen2.5-0.5B-Instruct-DPO-5000-5epochs-Ver2 | qwen2 | ["Qwen2ForCausalLM"] | 151936 | bfloat16 | 4.51.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
5,831 | Bmingg/vinai_v2_meddict | mbart | ["MBartForConditionalGeneration"] | 66773 | float32 | 4.54.0.dev0 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,832 | BoHu370/sd-class-butterflies-32 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,833 | Boadiwaa/LORA-colab-Distil-Whisper-medium-merged | whisper | ["WhisperForConditionalGeneration"] | 51864 | float32 | 4.42.0.dev0 | null | null | 24 | null | null | null | 0 | True | null | null | null | null | 50257 | 50256 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,834 | BobBraico/distilbert-base-uncased-finetuned-cyber | distilbert | ["DistilBertForMaskedLM"] | 30522 | null | 4.19.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,835 | BobBraico/rlb-cyber-finetuned-cyber | distilbert | ["DistilBertForMaskedLM"] | 30522 | null | 4.19.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,836 | BobHuangC/test_ORZ_Qwen_2_5_7b_step_1_Actor | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.50.2 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
5,837 | BobbBuilder/my-finetuned-model | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.43.0.dev0 | null | null | 4 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,838 | Bobospark/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.25.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,839 | BoccheseGiacomo/phi-2-finetuned-gsm8k-gb | phi-msft | ["PhiForCausalLM"] | 51200 | float32 | 4.35.2 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,840 | Bochkov/best_bvv_moe | bvv_best | ["BVVBestMoeForCausalLM"] | 131072 | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,841 | Bochkov/best_bvv_unfrozen_ru | bvv_best | ["BVVBestForCausalLM"] | 131072 | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,842 | Bochkov/best_bvv_zh | bvv_best | ["BVVBestForCausalLM"] | 131072 | float32 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,843 | Bodhi108/Yolov8n_RD | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,844 | Bodhi108/zephyr_7B_alpha_FDE_NA0219_14000 | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.35.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5,845 | Bodnar485/EveAndroid_q150 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.53.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,846 | Bodolaz/Unit-7 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,847 | Bodolaz/Unit-8.2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,848 | Bofeee5675/TongUI-7B | qwen2_5_vl | ["Qwen2_5_VLForConditionalGeneration"] | 152064 | bfloat16 | 4.50.0.dev0 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 128,000 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
5,849 | BogdanKuloren/sd-class-butterflies-32 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,850 | BogdanNV/test-trainer | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.33.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
5,851 | BogdanTurbal/roberta-d_0_e_3_t_u_r_0_v2 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.41.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,852 | BogdanTurbal/roberta-d_1_e_3_t_u_r_0_v2 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.41.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,853 | BogdanTurbal/roberta-d_3_e_3_t_u_r_0_v2 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.41.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,854 | Bogoo/NERO | llama | ["LlamaForCausalLM"] | 56000 | float32 | 4.51.3 | 2,048 | 8,192 | 24 | 32 | 32 | silu | 0 | True | 8,192 | 130,000 | 0.00001 | 0.02 | 0 | 0 | true | 64 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
5,855 | Bogoo/NERO_with_my_tokenizer | llama | ["LlamaForCausalLM"] | 56000 | bfloat16 | 4.51.3 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 8,192 | 100,000 | 0.00001 | 0.041667 | 0 | 0 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | long |
5,856 | Bogoo/SmolLM2_trained_for_summarize_concise | llama | ["LlamaForCausalLM"] | 49152 | float32 | 4.49.0 | 960 | 2,560 | 32 | 15 | 5 | silu | 0 | True | 8,192 | 100,000 | 0.00001 | 0.02 | 1 | 2 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.353894 | small | long |
5,857 | Bogoo/smollm_1.7B_new_pretrain | llama | ["LlamaForCausalLM"] | 56000 | float32 | 4.51.3 | 2,048 | 8,192 | 24 | 32 | 32 | silu | 0 | True | 8,192 | 130,000 | 0.00001 | 0.02 | 0 | 0 | true | 64 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 1.20796 | medium | long |
5,858 | Bogula/samsum-512 | pegasus | ["PegasusForConditionalGeneration"] | 96103 | float32 | 4.20.1 | null | null | 16 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,859 | BohdanPetryshyn/code-llama-openapi-completion | llama | ["LlamaForCausalLM"] | 32016 | bfloat16 | 4.41.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | long |
5,860 | BohdanPetryshyn/codellama-7b-openapi-completion-merged-ctx-lvl-fim-05-spm | llama | ["LlamaForCausalLM"] | 32016 | bfloat16 | 4.41.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | long |
5,861 | BohdanPetryshyn/codellama-7b-openapi-completion-merged-ctx-lvl-fim-no-spm-inf-seq-a-32 | llama | ["LlamaForCausalLM"] | 32016 | bfloat16 | 4.41.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | long |
5,862 | BoilerToad/distilbert-base-uncased-finetuned-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.16.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,863 | BoilerToad/pegasus-samsum | pegasus | ["PegasusForConditionalGeneration"] | 96103 | float32 | 4.16.2 | null | null | 16 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,864 | Boiyur/whisper-small-mn | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.31.0.dev0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
5,865 | Boko99/Llama-3.2-3B-SFT-QLoRA | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.3 | 3,072 | 8,192 | 28 | 24 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | true | 128 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.170894 | medium | very_long |
5,866 | BolaKubuz/distilhubert-distilhubert-finetuned-gtzan | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.35.0.dev0 | 768 | 3,072 | 2 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.014156 | small | null |
5,867 | Boltuzamaki/TinyLlama-1.1B-Chat-v1.0-AWQ-4bit | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.38.1 | 2,048 | 5,632 | 22 | 32 | 4 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
5,868 | Boltuzamaki/TinyLlama-1.1B-Chat-v1.0-bf16-push-demo | llama | ["LlamaForCausalLM"] | 32000 | bfloat16 | 4.49.0 | 2,048 | 5,632 | 22 | 32 | 4 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
5,869 | Boltuzamaki/multi-qa-MiniLM-L6-cos-v1-Tough_Rugby_ft | bert | ["BertModel"] | 30522 | float32 | 4.51.3 | 384 | 1,536 | 6 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.010617 | small | short |
5,870 | Bonbone/tmp_trainer | distilbert | ["DistilBertForMaskedLM"] | 30522 | float32 | 4.45.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,871 | Bondds/01-8208grapph | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.45.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,872 | Bondds/01call1 | stablelm | ["StableLmForCausalLM"] | 100352 | float16 | 4.42.4 | 2,048 | 5,632 | 24 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | null | 0.02 | 100257 | 100278 | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
5,873 | Bondds/01klauv3 | stablelm | ["StableLmForCausalLM"] | 100352 | float16 | 4.42.4 | 2,048 | 5,632 | 24 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | null | 0.02 | 100257 | 100278 | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
5,874 | Bondds/02-07095grapph | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.3 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
5,875 | Bondds/02-2032grapph | llama | ["LlamaForCausalLM"] | 64000 | float16 | 4.42.4 | 4,096 | 11,008 | 48 | 32 | 4 | silu | 0 | False | 4,096 | 5,000,000 | 0.000001 | 0.02 | 1 | 7 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 9.663676 | large | medium |
5,876 | Bondds/02-2100grapph | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.42.4 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
5,877 | Bondds/02coltv2 | stablelm | ["StableLmForCausalLM"] | 100352 | float16 | 4.42.4 | 2,048 | 5,632 | 24 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | null | 0.02 | 100257 | 100278 | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
5,878 | Bondds/02klau2 | stablelm | ["StableLmForCausalLM"] | 100352 | float16 | 4.42.4 | 2,048 | 5,632 | 24 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | null | 0.02 | 100257 | 100278 | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
5,879 | Bondds/02klau3 | stablelm | ["StableLmForCausalLM"] | 100352 | float16 | 4.42.4 | 2,048 | 5,632 | 24 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | null | 0.02 | 100257 | 100278 | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
5,880 | Bondds/03-07038grapph | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.47.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,881 | Bondds/03-07096grapph | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.48.3 | 5,120 | 27,648 | 64 | 40 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 20.132659 | large | very_long |
5,882 | Bondds/03-4032grapph | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.45.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,883 | Bondds/03-4033grapph | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.45.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,884 | Bondds/03-6901grapph | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.46.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,885 | Bondds/03-6903grapph | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.46.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,886 | Bondds/04-07008grapph | llama | ["LlamaForCausalLM"] | 131072 | bfloat16 | 4.47.0 | 3,072 | 23,040 | 70 | 12 | 4 | silu | 0 | True | 32,768 | 1,000,042 | 0.000001 | 0.02 | 11 | 11 | false | 256 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 7.927235 | large | very_long |
5,887 | Bondds/04-2006grapph | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.42.4 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
5,888 | Bondds/04-6009grapph | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.45.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,889 | Bondds/05-07007grapph | llama | ["LlamaForCausalLM"] | 131072 | bfloat16 | 4.47.0 | 3,072 | 23,040 | 70 | 12 | 4 | silu | 0 | True | 32,768 | 1,000,042 | 0.000001 | 0.02 | 11 | 11 | false | 256 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 7.927235 | large | very_long |
5,890 | Bondds/05-2120grapph | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,891 | Bondds/05-2121grapph | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,892 | Bondds/05-4029grapph | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.45.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,893 | Bondds/05-callls | stablelm | ["StableLmForCausalLM"] | 100352 | float16 | 4.42.4 | 2,048 | 5,632 | 24 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | null | 0.02 | 100257 | 100278 | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
5,894 | Bondds/05coltTl | stablelm | ["StableLmForCausalLM"] | 100352 | float16 | 4.42.4 | 2,048 | 5,632 | 24 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | null | 0.02 | 100257 | 100278 | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
5,895 | Bondds/06-07000grapph | llama | ["LlamaForCausalLM"] | 131072 | bfloat16 | 4.47.0 | 3,072 | 23,040 | 70 | 12 | 4 | silu | 0 | True | 32,768 | 1,000,042 | 0.000001 | 0.02 | 11 | 11 | false | 256 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 7.927235 | large | very_long |
5,896 | Bondds/06-2009grapph | llama | ["LlamaForCausalLM"] | 64000 | float16 | 4.42.4 | 4,096 | 11,008 | 48 | 32 | 4 | silu | 0 | False | 4,096 | 5,000,000 | 0.000001 | 0.02 | 1 | 7 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 9.663676 | large | medium |
5,897 | Bondds/06-6012grapph | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.45.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,898 | Bondds/06-6016grapph | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.45.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,899 | Bondds/06-6017grapph | llama | ["LlamaForCausalLM"] | 128258 | bfloat16 | 4.45.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.