index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
6,000 | BothBosu/bilstm-scam-classifier-v1.2 | null | [] | null | null | null | 768 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,001 | BothBosu/cnn-agent-scam-classifier-v1.0 | null | [] | 30522 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,002 | BothBosu/cnn-no-receiver-scam-classifier-v1.0 | null | [] | 30522 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,003 | BothBosu/cnn-suspect-dialogue-scam-classifier-v1.0 | null | [] | 30522 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,004 | BothBosu/distilbert-scam-classification-v1 | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.38.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,005 | BothBosu/gru-agent-scam-classifier-v1.0 | null | [] | null | null | null | 768 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,006 | BothBosu/gru-no-receiver-scam-classifier-v1.0 | null | [] | null | null | null | 768 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,007 | BothBosu/gru-no-receiver-scam-classifier-v1.1 | null | [] | null | null | null | 768 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,008 | BothBosu/gru-suspect-dialogue-scam-classifier-v1.0 | null | [] | null | null | null | 768 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,009 | BothBosu/lstm-agent-scam-classifier-v1.0 | lstm | [] | null | null | null | 768 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,010 | BothBosu/lstm-no-receiver-scam-classifier-v1.0 | lstm | [] | null | null | null | 768 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,011 | BothBosu/lstm-scam-classifier-v1.2 | lstm | [] | null | null | null | 768 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,012 | BothBosu/lstm-suspect-dialogue-scam-classifier-v1.0 | lstm | [] | null | null | null | 768 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,013 | BotoxBernd/SQL-Generation-mistral-perfect-7B-v0.1 | mistral | ["MistralForCausalLM"] | 32001 | bfloat16 | 4.40.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,014 | BotoxBernd/Schema-Links-mistral-7B-v0.1 | mistral | ["MistralForCausalLM"] | 32001 | bfloat16 | 4.40.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,015 | BottledDjinn/results | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.48.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,016 | Boulou2107/comic-name-classification | bert | ["BertForTokenClassification"] | 119547 | float32 | 4.37.0.dev0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,017 | BounharAbdelaziz/ModernBERT-Morocco-Sentence-Embeddings-v0.2-bs-32-lr-2e-05-ep-2-wp-0.05-gacc-1-gnm-1.0-v0.3 | modernbert | ["ModernBertModel"] | 64000 | float32 | 4.49.0 | 768 | 1,152 | 22 | 12 | null | null | 0 | null | 8,192 | null | null | 0.02 | 50281 | 50282 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | true | 0.155714 | small | long |
6,018 | Bowen7/vits-vctk | vits | ["VitsModel"] | 178 | null | 4.33.3 | 192 | null | 6 | 2 | null | relu | 0.1 | null | null | null | null | 0.02 | null | null | null | 96 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.002654 | small | null |
6,019 | BoxMrChen/gemma-2b-flock-1718375107 | gemma | ["GemmaForCausalLM"] | 256000 | float16 | 4.41.0 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
6,020 | BoxMrChen/gemma-2b-flock-1719066151 | gemma | ["GemmaForCausalLM"] | 256000 | float16 | 4.41.0 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
6,021 | BoyangZ/llama3_chinese_basemode_ftv1 | llama | ["LlamaForCausalLM"] | 128256 | float32 | 4.39.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
6,022 | BoyishGiggles/flan-t5-small-ecommerce-text-classification | t5 | ["T5ForSequenceClassification"] | 32128 | float32 | 4.41.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,023 | Bozh/mistral-instruct-test | mistral | ["MistralForCausalLM"] | 32000 | float16 | 4.43.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,024 | Bpole/lora_sberhack_v1.0 | mistral | ["MistralForCausalLM"] | 32002 | bfloat16 | 4.35.0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,025 | Br1LDu/Mordern2Classic_MengZi | qwen2 | ["Qwen2ForCausalLM"] | 152064 | float16 | 4.47.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 131,072 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151643 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
6,026 | Br22/codeparrot-ds | gpt2 | ["GPT2LMHeadModel"] | 50000 | float32 | 4.26.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 0 | 0 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,027 | Br3ad/personalized_phi-2-sft-ultrachat-full_poetic_nll | phi | ["PhiForCausalLM"] | 51200 | float32 | 4.41.2 | 2,560 | 10,240 | 32 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | 50256 | 50256 | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
6,028 | Br3ad/personalized_phi-2-sft-ultrachat-full_slang_nll | phi | ["PhiForCausalLM"] | 51200 | float32 | 4.41.2 | 2,560 | 10,240 | 32 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | 50256 | 50256 | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
6,029 | Brackly/malawi_quantized_model | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.37.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,030 | Braden99/my_awesome_qa_model | distilbert | ["DistilBertForQuestionAnswering"] | 30522 | float32 | 4.26.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,031 | Brafied/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,032 | Brain22/a2c-PandaReachDense-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,033 | BrainRoster/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,034 | BrainTheos/wav2vec2-large-mms-1b-turkish-colab | wav2vec2 | ["Wav2Vec2ForCTC"] | 39 | float32 | 4.31.0.dev0 | 1,280 | 5,120 | 48 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 80 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.943718 | small | null |
6,035 | Brainiac77/whisper-base-bn-s2t | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.26.0.dev0 | null | null | 6 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
6,036 | Brainkite/PPO-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,037 | Brainkite/llama-2-7b-miniguanaco | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,038 | BramDelisse/17_24-k-Murdoch_data-123 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,039 | BramDelisse/1_8-k-Murdoch_data-123 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,040 | BramDelisse/25_32-k-Murdoch_data-123 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,041 | BramDelisse/9_16-k-Murdoch_data-123 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,042 | BramVanroy/GEITje-7B-ultra | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.36.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
6,043 | BramVanroy/bert-base-dutch-cased-hebban-reviews | bert | ["BertForSequenceClassification"] | 30073 | float32 | 4.21.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,044 | BramVanroy/bert-base-dutch-cased-hebban-reviews5 | bert | ["BertForSequenceClassification"] | 30073 | float32 | 4.21.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,045 | BramVanroy/bert-base-multilingual-cased-hebban-reviews | bert | ["BertForSequenceClassification"] | 119547 | float32 | 4.21.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,046 | BramVanroy/bert-base-multilingual-cased-hebban-reviews5 | bert | ["BertForSequenceClassification"] | 119547 | float32 | 4.21.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,047 | BramVanroy/fietje-2-chat | phi | ["PhiForCausalLM"] | 50297 | bfloat16 | 4.39.1 | 2,560 | 10,240 | 32 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | 50256 | 50256 | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
6,048 | BramVanroy/robbert-v2-dutch-base-hebban-reviews | roberta | ["RobertaForSequenceClassification"] | 40000 | float32 | 4.21.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,049 | BramVanroy/robbert-v2-dutch-base-hebban-reviews5 | roberta | ["RobertaForSequenceClassification"] | 40000 | float32 | 4.21.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,050 | BramVanroy/xlm-roberta-base-hebban-reviews | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.21.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,051 | Bramwel/segformer-b0-finetuned-segments-sidewalk-2 | segformer | ["SegformerForSemanticSegmentation"] | null | float32 | 4.42.3 | null | null | null | [1, 2, 5, 8] | null | gelu | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,052 | BrandDead/thelabel-trained-model | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.44.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,053 | Brandoko/CodeLlama-13b-Recharts | llama | ["LlamaForCausalLM"] | 32016 | float16 | 4.34.0.dev0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | True | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | long |
6,054 | Brandoko/CodeLlama-34b-Recharts-v2 | llama | ["LlamaForCausalLM"] | 32004 | float16 | 4.34.1 | 8,192 | 22,016 | 48 | 64 | 8 | silu | null | False | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 38.654706 | large | long |
6,055 | Brandoko/CodeLlama-34b-Recharts-v2-adapter | llama | ["LlamaForCausalLM"] | 32004 | bfloat16 | 4.34.1 | 8,192 | 22,016 | 48 | 64 | 8 | silu | null | False | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 38.654706 | large | long |
6,056 | Brandoko/CodeLlama-34b-Recharts-v2-bins | llama | ["LlamaForCausalLM"] | 32004 | float32 | 4.34.1 | 8,192 | 22,016 | 48 | 64 | 8 | silu | null | False | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 38.654706 | large | long |
6,057 | Brandoko/CodeLlama-7b-Recharts | llama | ["LlamaForCausalLM"] | 32016 | float16 | 4.34.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | long |
6,058 | Brandoko/CodeLlama-7b-Recharts-v2-st-merged | llama | ["LlamaForCausalLM"] | 32016 | float32 | 4.34.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | False | 16,384 | 1,000,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | long |
6,059 | Brandon-h/distilbert-finetuned-spanish-offensive-language | distilbert | ["DistilBertForSequenceClassification"] | 31002 | float32 | 4.35.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,060 | BrandonET/Llama-3.2-1B-Instruct-wikitext2 | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.53.1 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | [128001, 128008, 128009] | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
6,061 | BrandonET/gpt2-wikitext2 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.53.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,062 | BrandonM001/bert-finetuned-ner-accelerate | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.40.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,063 | BrandonM001/bert-finetuned-ner-accelerate1 | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,064 | BrandonM001/bert-finetuned-ner11 | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.40.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,065 | BrandonM001/bert-finetuned-ner13 | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.40.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,066 | BrandonM001/bert-finetuned-ner2 | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,067 | BrandonM001/bert-finetuned-ner5 | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,068 | BrandonZYW/llama-2-7b-InBedder | llama | ["LlamaForCausalLM"] | 32001 | bfloat16 | 4.35.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
6,069 | BrandonZYW/opt-1.3b-InBedder | opt | ["OPTForCausalLM"] | 50265 | float32 | 4.35.2 | 2,048 | null | 24 | 32 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 1.20796 | medium | medium |
6,070 | BrandonZYW/opt-2.7b-InBedder | opt | ["OPTForCausalLM"] | 50265 | float32 | 4.35.2 | 2,560 | null | 32 | 32 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 80 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 2.516582 | medium | medium |
6,071 | BrandonZYW/opt-350m-InBedder | opt | ["OPTForCausalLM"] | 50265 | float32 | 4.35.2 | 1,024 | null | 24 | 16 | null | null | 0 | True | 2,048 | null | null | null | 2 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | medium |
6,072 | BrandonZYW/roberta-large-InBedder | roberta | ["RobertaForMaskedLM"] | 50265 | float32 | 4.35.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
6,073 | BraveOni/2ch-text-classification | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.15.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
6,074 | BraveStone9/model_classifier | bert | ["BertForSequenceClassification"] | 30000 | float32 | 4.26.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,075 | BraylonDash/phi-2-gpo-test-iter-0 | phi | ["PhiForCausalLM"] | 51200 | bfloat16 | 4.36.2 | 2,560 | 10,240 | 32 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | 50256 | 50256 | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
6,076 | BraylonDash/phi-2-gpo-test-iter-1 | phi | ["PhiForCausalLM"] | 51200 | bfloat16 | 4.36.2 | 2,560 | 10,240 | 32 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | 50256 | 50256 | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
6,077 | BraylonDash/phi-2-gpo-test-iter-2 | phi | ["PhiForCausalLM"] | 51200 | bfloat16 | 4.36.2 | 2,560 | 10,240 | 32 | 32 | 32 | gelu_new | 0 | True | 2,048 | 10,000 | null | 0.02 | 50256 | 50256 | false | 80 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 2.516582 | medium | medium |
6,078 | BreZzz/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,079 | Bread-AI/Crumb-13B | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.37.2 | 5,120 | 13,824 | 40 | 40 | 40 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
6,080 | BreadAi/DiscordPy | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.26.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,081 | BreadAi/MuseBig | gpt2 | ["GPT2LMHeadModel"] | 99999 | float32 | 4.26.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 0 | 0 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,082 | BreadAi/PM_modelV2 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.26.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
6,083 | BreadHead/tdu_nicograph5 | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
6,084 | BreadSquadron/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
6,085 | Brecon/deberta_seq | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.33.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,086 | Brecon/master_bert_seq_training_model | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.33.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
6,087 | Brecon/my_awesome_model | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.33.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,088 | Brecon/training_model | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.33.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,089 | Brecon/validation_model | distilbert | ["DistilBertForSequenceClassification"] | 30522 | null | 4.33.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,090 | BreezeAI/Qwen2.5-7B-Instruct-Vertu-Customer | qwen2 | ["Qwen2ForCausalLM"] | 152064 | bfloat16 | 4.51.3 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
6,091 | Breezeee/copy-distilbert-base-uncased-emotion | distilbert | ["DistilBertForSequenceClassification"] | 30522 | null | 4.11.0.dev0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,092 | Breezeee/whisper-small-hi | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.44.2 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
6,093 | Breezeee/whisper_aishell3_output | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.44.2 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
6,094 | Breezezzz/DS200-norm-model | mbart | ["MBartForConditionalGeneration"] | 40030 | float32 | 4.41.2 | null | null | 6 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,095 | Breezezzz/DS200-speechnorm-model | mbart | ["MBartForConditionalGeneration"] | 40030 | float32 | 4.41.2 | null | null | 6 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
6,096 | Bren/distillbert-base-uncased-finetuned-squad-d5716d28 | distilbert | ["DistilBertForMaskedLM"] | 30522 | null | 4.17.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
6,097 | BrenL/extra0IMOO0 | null | [] | null | null | null | null | null | null | 24 | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
6,098 | BrenL/extra2IMOO2 | t5 | ["T5EncoderModel"] | 32128 | bfloat16 | 4.40.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
6,099 | Brench/Qwen2-4B-Insturct-pro | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.45.0.dev0 | 3,072 | 9,216 | 32 | 32 | 8 | silu | 0 | False | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 96 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 3.623879 | medium | very_long |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.