index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
5,700 | BioMike/emoji_vae | vae | ["VAEModel"] | null | float32 | 4.51.3 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,701 | BioMistral/BioMistral-7B | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.35.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5,702 | BioMistral/BioMistral-7B-AWQ-QGS128-W4-GEMM | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.36.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5,703 | BioMistral/BioMistral-7B-AWQ-QGS128-W4-GEMV | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.36.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5,704 | BioMistral/BioMistral-Sharded | mistral | ["MistralForCausalLM"] | 32000 | float32 | 4.38.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5,705 | Biokukesi/llama-3-8b-Instruct-bnb-4bit-aiaustin-demo | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
5,706 | Biokukesi/llama-3-8b-sict | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
5,707 | Biplop/wav2vec2-nep | wav2vec2 | ["Wav2Vec2ForCTC"] | 63 | float32 | 4.20.1 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0.12 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
5,708 | Bipul8765/SmolLM2-FT-MyDataset | llama | ["LlamaForCausalLM"] | 49152 | float32 | 4.47.0 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 8,192 | 100,000 | 0.00001 | 0.041667 | 1 | 2 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | long |
5,709 | BirdL/DeepSeek-Coder-V2-Lite-Instruct-FlashAttnPatch | deepseek_v2 | ["DeepseekV2ForCausalLM"] | 102400 | bfloat16 | 4.39.3 | 2,048 | 10,944 | 27 | 16 | 16 | silu | 0 | True | 163,840 | 10,000 | 0.000001 | 0.02 | 100000 | 100001 | false | 128 | 1 | false | 64 | 6 | false | false | false | false | false | true | false | true | false | true | 1.358954 | medium | very_long |
5,710 | BirdieByte1024/Qwen2.5-1.5B-dental-full | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.51.3 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
5,711 | BirdieByte1024/doctor-dental-implant-llama3.2-3B-full-model | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
5,712 | BirdieByte1024/medical-qa-llama3.2-3B-full-model | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
5,713 | Birkir/electra-base-igc-is-sentiment-analysis | electra | ["ElectraForSequenceClassification"] | 32105 | float32 | 4.34.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,714 | Biscottezi/vit5-base-finetuned-vitext2sql | t5 | ["T5ForConditionalGeneration"] | 36096 | float32 | 4.26.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,715 | BishFish/DialoGPT-medium-edward | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.44.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,716 | BishanSingh246/bart-large-mnli-finetune_v1 | bart | ["BartForSequenceClassification"] | 50265 | float32 | 4.40.0 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,717 | Bisher/Qwen2.5-0.5B-instruct_2ga_1e_UNSLOTH_tro_GGUF | qwen2 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,718 | Bisher/ourData_train_5e | wav2vec2 | ["Wav2Vec2ForSequenceClassification"] | 32 | float32 | 4.44.0 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
5,719 | Bisher/wav2vec2_ASV_deepfake_audio_detection_DF_finetune_frozen_chngd_classifier | wav2vec2 | ["Wav2Vec2ForSequenceClassification"] | 32 | float32 | 4.44.0 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
5,720 | Bisher/wav2vec2_ASV_deepfake_audio_detection_DF_finetune_frozen_chngd_classifier_kaggle_8freeze_64bs_2e | wav2vec2 | ["Wav2Vec2ForSequenceClassification"] | 32 | float32 | 4.44.0 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
5,721 | Bisnistec/edu-t5-16m-v2 | t5 | ["T5ForConditionalGeneration"] | 14100 | float32 | 4.34.0 | null | null | null | null | null | null | null | True | null | null | null | null | 1 | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,722 | Bisnu/videomae-base-pretrain-ucf101-subset | videomae | ["VideoMAEForPreTraining"] | null | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
5,723 | Bistolero/1ep_seq_25_6b | t5 | ["T5ForConditionalGeneration"] | 250112 | float32 | 4.22.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,724 | Bistolero/en_ge_20_20 | t5 | ["T5ForConditionalGeneration"] | 250112 | float32 | 4.19.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,725 | Bistolero/french_all | t5 | ["T5ForConditionalGeneration"] | 250112 | float32 | 4.17.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,726 | Bistolero/ge_nl_64B_25K | t5 | ["T5ForConditionalGeneration"] | 250112 | float32 | 4.19.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,727 | Bistolero/german_2EP | t5 | ["T5ForConditionalGeneration"] | 250112 | float32 | 4.19.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,728 | Bistolero/it_train_all | t5 | ["T5ForConditionalGeneration"] | 250112 | float32 | 4.17.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,729 | Bistolero/mix_training_en_du_nl | t5 | ["T5ForConditionalGeneration"] | 250112 | float32 | 4.17.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,730 | Bistolero/mix_training_en_du_nl_1 | t5 | ["T5ForConditionalGeneration"] | 250112 | float32 | 4.17.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,731 | Bistolero/mt5_two_epocs_nl | t5 | ["T5ForConditionalGeneration"] | 250112 | float32 | 4.17.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,732 | Bistolero/nl_GA_32b | t5 | ["T5ForConditionalGeneration"] | 250112 | float32 | 4.19.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,733 | Bistolero/nl_ge_25_6b_3ep_se | t5 | ["T5ForConditionalGeneration"] | 250112 | float32 | 4.22.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,734 | Bistolero/nl_ge_DP_6BX5_3 | t5 | ["T5ForConditionalGeneration"] | 250112 | float32 | 4.22.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,735 | Bistolero/nl_one_ep | t5 | ["T5ForConditionalGeneration"] | 250112 | float32 | 4.19.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,736 | Biswajit010/crack-transformer | detr | ["DetrForObjectDetection"] | null | float32 | 4.7.0.dev0 | null | null | 6 | null | null | null | 0 | null | 1,024 | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,737 | BitAgent/BitAgent-8B | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.48.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,738 | BitBasher/llama-2-7b-mini-ibased-GGUF | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
5,739 | BitStack/BitStack-Llama-2-7B | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.44.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
5,740 | BitStack/BitStack-Llama-3.1-70B | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.44.2 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
5,741 | BitStack/BitStack-Llama-3.1-8B | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,742 | BjngChjjljng/DETR-fisheye-combine-40epoch | detr | ["DetrForObjectDetection"] | null | float32 | 4.45.1 | null | null | 6 | null | null | null | 0 | null | 1,024 | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,743 | BjngChjjljng/detr-finetuned_v2 | detr | ["DetrForObjectDetection"] | null | float32 | 4.45.1 | null | null | 6 | null | null | null | 0 | null | 1,024 | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,744 | Bjqrn/poca-SoccerTwos | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,745 | Bjqrn/ppo-Pyramids | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,746 | Bjqrn/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,747 | Blablablab/neurobiber | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.45.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,748 | Blacc911/smolLM2-135M-Instruct-finetuned-GCUH | llama | ["LlamaForCausalLM"] | 49152 | float32 | 4.47.1 | 576 | 1,536 | 30 | 9 | 3 | silu | 0 | True | 8,192 | 100,000 | 0.00001 | 0.041667 | 1 | 2 | true | 64 | 3 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | long |
5,749 | Black-Ink-Guild/Google-gemma-2-9b-it | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.42.0.dev0 | 3,584 | 14,336 | 42 | 16 | 8 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 224 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.473908 | medium | long |
5,750 | Black-Ink-Guild/Pernicious_Prophecy_70B | llama | ["LlamaForCausalLM"] | 128259 | bfloat16 | 4.47.1 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
5,751 | Black-Ink-Guild/Pernicious_Prophecy_70B_FP8 | llama | ["LlamaForCausalLM"] | 128259 | bfloat16 | 4.48.0 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
5,752 | BlackB/thai-t5-base | t5 | ["T5ForConditionalGeneration"] | 72100 | null | 4.29.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
5,753 | BlackBeenie/Llama-3.1-8B-OpenO1-SFT-v0.1 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.46.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,754 | BlackBeenie/llama-3.1-8B-Galore-openassistant-guanaco | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128040 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,755 | BlackBeenie/mdeberta-v3-base-sbert | deberta-v2 | ["DebertaV2Model"] | 251000 | float32 | 4.44.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,756 | BlackBeenie/qwen3-14b-reasoning-conversational | qwen3 | ["Qwen3ForCausalLM"] | 151936 | bfloat16 | 4.53.0 | 5,120 | 17,408 | 40 | 40 | 8 | silu | 0 | True | 40,960 | 1,000,000 | 0.000001 | 0.02 | null | 151645 | false | 128 | 5 | false | null | null | false | false | false | false | false | false | false | false | true | true | 12.582912 | large | very_long |
5,757 | BlackFaceCat/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,758 | BlackFear/llama-3.2-1b-sorucevap | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.44.2 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
5,759 | BlackKakapo/opus-mt-en-ro | marian | ["MarianMTModel"] | 59543 | float32 | 4.26.0 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,760 | BlackKakapo/opus-mt-fi-ro | marian | ["MarianMTModel"] | 59142 | float32 | 4.26.0 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
5,761 | BlackSamorez/rudialogpt3_medium_based_on_gpt2_2ch | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.17.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,762 | BlackSwan1827/CubeChase | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,763 | BlackTea-c/chatglm2-6b-torchkeras_30 | chatglm | ["ChatGLMForConditionalGeneration"] | 65024 | float16 | 4.30.2 | 4,096 | null | null | 32 | null | null | 0 | True | null | null | null | null | null | 2 | false | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
5,764 | Blackroot/Differential-Microllama-2 | llama | ["LlamaForCausalLM"] | 49152 | bfloat16 | 4.42.3 | 576 | 1,536 | 30 | 16 | 4 | silu | 0 | True | 512 | 100,000 | 0.00001 | 0.041667 | 1 | 2 | true | 36 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.119439 | small | short |
5,765 | Blackroot/Mirai-70B-1.0 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.46.2 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
5,766 | Blackroot/Mirai-70B-1.0-3.9B-6H | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.46.2 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
5,767 | Blackroot/Mirai-70B-2.1 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.46.2 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
5,768 | Blaise-MR/diseaseClassifier | vit | ["ViTForImageClassification"] | null | float32 | 4.22.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
5,769 | Blaise-MR/distilroberta2 | roberta | ["RobertaForSequenceClassification"] | 50265 | null | 4.41.2 | 768 | 3,072 | 6 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
5,770 | Blaise-MR/patient_falls_detector | bert | ["BertForSequenceClassification"] | 28996 | null | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,771 | Blaise-MR/pytorch_distilbert2_fallsclassifier | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.41.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,772 | Blaise-MR/pytorch_distilbert3_fallsclassifier | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.41.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,773 | Blaise-MR/pytorch_distilbert_fallsclassifier | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.41.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,774 | Blaise-MR/pytorch_distilroberta1_fallsclassifier | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.41.2 | 768 | 3,072 | 6 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
5,775 | Blaise-MR/pytorch_distilroberta2_fallsclassifier | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.41.2 | 768 | 3,072 | 6 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.042467 | small | short |
5,776 | Blaise-MR/pytorch_roberta_fallsclassifier | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,777 | Blaise-MR/pytorch_sentiment_textnotes3 | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.42.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,778 | Blaise-MR/roberta_falls_classifier | roberta | ["RobertaForSequenceClassification"] | 50265 | null | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,779 | Blaise-MR/roberta_falls_classififier | roberta | ["RobertaForSequenceClassification"] | 50265 | null | 4.41.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,780 | Blaise-g/led_pubmed_sumpubmed_1 | led | ["LEDForConditionalGeneration"] | 50265 | float32 | 4.20.0 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,781 | Blakcori/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-barky_knobby_camel | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float32 | 4.51.3 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.231211 | small | very_long |
5,782 | Blandskron/ModelChatBot | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.47.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,783 | Blaxzter/LaBSE-sentence-embeddings | bert | ["BertModel"] | 501153 | float32 | 4.25.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,784 | Blazej/mistral7B_align_bank_guidelines_logsigmoid_loss | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.35.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
5,785 | Blazgo/2-medical-pro | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.48.2 | 8,192 | 28,672 | 80 | 64 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 64.424509 | large | very_long |
5,786 | Blebbyblub/HuBERT-javanese-sundanese | wav2vec2 | ["Wav2Vec2ForCTC"] | 31 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
5,787 | Blebbyblub/javanese-sundanese-conformer-asrV2 | conformer | [] | 29 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
5,788 | Bleking/Llama-3.1-Minitron-4B-Width-Base-ep8-lr3e5-wc001 | llama | ["LlavaLlamaForCausalLM"] | 128256 | float16 | 4.45.1 | 3,072 | 9,216 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 96 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.623879 | medium | very_long |
5,789 | Bleking/Meta-Llama-3.1-8B-Instruct-goose-abliterated-pre | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,790 | Blenderama/MOE | chatglm | ["ChatGLMModel"] | null | float16 | 4.30.2 | 4,096 | null | null | 32 | null | null | 0 | True | null | null | null | null | null | 2 | false | 128 | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
5,791 | Blessing988/videomae-base-finetuned-ucf101-subset | videomae | ["VideoMAEForVideoClassification"] | null | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
5,792 | BleueMer/test | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
5,793 | BlinderMan/DeepSeek-R1-Medical-COT | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.49.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
5,794 | Blingham/model2 | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.47.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,795 | Blinorot/MNLP_M3_DPO_V8 | qwen3 | ["Qwen3ForCausalLM"] | 151936 | bfloat16 | 4.51.3 | 1,024 | 3,072 | 28 | 16 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | true | 64 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.352322 | small | very_long |
5,796 | BlitherBoom/bert-finetuned-squad | bert | ["BertForQuestionAnswering"] | 28996 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
5,797 | BlitherBoom/distilbert-base-uncased-finetuned-sst2 | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.26.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
5,798 | Blizzchor/DialoGPT-medium-QuillLord | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.26.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
5,799 | Blizzchor/DialoGPT-medium-gamora | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.25.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.