index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
5,500
Benve/bert-nlp
bert
["BertForMaskedLM"]
28996
float32
4.34.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,501
Berf4615/news-generator-tr
gpt2
["GPT2LMHeadModel"]
50257
float32
4.37.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
5,502
Berkesule/qwenvl-2.5-7b-gptq-W4816-quantize-tr-dpo-v2
qwen2_5_vl
["Qwen2_5_VLForConditionalGeneration"]
152064
bfloat16
4.51.3
3,584
18,944
28
28
4
silu
0
True
128,000
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
5,503
Berly00/whisper-small-portuguese-1
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.42.4
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
5,504
Berly00/whisper-small-spanish-1
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.42.4
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
5,505
Berly00/whisper-small-spanish-2
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.42.4
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
5,506
BernalHR/Phi-3-mini-4k-instruct-Inscripciones-bnb-4bit
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
5,507
BernalHR/Phi-3-mini-4k-instruct-Inscripciones-bnb-4bit-GGUF
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
5,508
BernardoMSV/modelo_llm
bert
["BertForQuestionAnswering"]
29794
float32
4.38.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,509
BerriesBounty/merge_word_clm-model
llama
["LlamaForCausalLM"]
128256
float32
4.48.3
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
5,510
BestModel/250228-DeepSeek-R1-Medical-COT
llama
["LlamaForCausalLM"]
128256
float16
4.49.0
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,511
Betimes-Solution/Citizen-AI-Case-Classification-Wangchanberta-Traffy-Multi
camembert
["CamembertForSequenceClassification"]
25005
float32
4.47.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,512
BetterThanNothing/PPO-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,513
BeulahHilda05/mistral_b_finance_finetuned_test
falcon
["FalconForCausalLM"]
50304
float32
4.37.2
2,048
null
24
32
null
null
0
False
2,048
10,000
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
5,514
Bewinxed/reader-lm-1.5b-onnx
qwen2
["Qwen2ForCausalLM"]
151936
null
4.44.2
1,536
8,960
28
12
2
silu
0
True
256,000
2,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
5,515
BexRedpill/distilbert-on-polarity-yelp-reviews
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.41.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
5,516
Beybars/videomae-base-finetuned-ucf101-subset
videomae
["VideoMAEForVideoClassification"]
null
float32
4.36.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,517
Beygo/Marco-o1-Q8-mlx
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.41.2
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
5,518
Beyondz0110/Hextalk
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
5,519
BhabhaAI/Gajendra-v0.1
llama
["LlamaForCausalLM"]
48066
float16
4.37.0
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
5,520
BhabhaAI/Mistral-translation-classify
mistral
["MistralForCausalLM"]
32002
float32
4.37.0
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
5,521
Bhanu9Prakash/poca-SoccerTwos
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,522
Bhanu9Prakash/rl_course_vizdoom_health_gathering_supreme
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,523
Bharat05092003/mistral_llm_v3
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
5,524
Bharatdeep-H/multi-lingual-classifier-qwen-2.5-0.5-v0
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.46.3
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
5,525
Bharatdeep-H/portfolio-query-qwen-2.5-1.5b-v6-cot
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.48.0
1,536
8,960
28
12
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
5,526
Bharatdeep-H/portfolio-query-qwen-2.5-3b-v3-cot
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.46.3
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
5,527
Bharatdeep-H/pq_cache
mpnet
["MPNetModel"]
30527
float32
4.49.0
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
5,528
Bharatdeep-H/stella_finetuned_en_dataset_stella_400_20_translated_query_v3_w_v_MAX_100
new
["NewModel"]
30528
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
null
8,192
160,000
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
true
0.30199
small
long
5,529
Bharatdeep-H/stella_finetuned_en_dataset_stella_400_20_translated_query_v3_w_v_MAX_400
new
["NewModel"]
30528
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
null
8,192
160,000
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
true
0.30199
small
long
5,530
Bharatdeep-H/stella_finetuned_en_dataset_stella_400_20_translated_query_v3_w_v_MAX_50
new
["NewModel"]
30528
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
null
8,192
160,000
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
true
0.30199
small
long
5,531
Bharathparam11/Krishi_vani
llama
["LlamaForCausalLM"]
128256
float16
4.51.3
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
5,532
Bharatmalik1999/distilbert-base-uncased-customer-data
roberta
["RobertaForSequenceClassification"]
50265
float32
4.38.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,533
Bhargav0044/whisper-small-en
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.32.0.dev0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
5,534
BhashaAI/hindi-image-caption
vision-encoder-decoder
["VisionEncoderDecoderModel"]
50258
float32
4.41.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
5,535
BhavanaMalla/Mask2Former_Finetuning_cityscapes_swin_railSem19_2epochs_512_6patches
mask2former
["Mask2FormerForUniversalSegmentation"]
null
float32
4.38.0.dev0
null
null
10
8
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
5,536
BhavanaMalla/mt5_summarizer_amazon_reviews
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.53.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
5,537
Bhaveen/2medimix-whisper-fine-tuned
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.47.1
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
5,538
Bhaveen/medimix-3
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.47.1
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
5,539
Bhaveen/medimix-whisper-fine-tuned
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.47.1
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
5,540
Bhavesh1609/distilbert-base-uncased
distilbert
["DistilBertForTokenClassification"]
30522
float32
4.41.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
5,541
Bhavesh7895/ppo-LunarLander-v2_version1
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,542
Bhaveshgb/auto_bot
electra
["ElectraForQuestionAnswering"]
31102
float32
4.26.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
5,543
BhavyaMuni/taylor-swift-model-paragraphs
gpt2
["GPT2LMHeadModel"]
50257
float32
4.25.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
5,544
Bhavyagowni/bart_scibert
bart
["BartForConditionalGeneration"]
50264
float32
4.51.3
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
5,545
Bhavyasree09/english-TELUGU-colloquial-translator
m2m_100
["M2M100ForConditionalGeneration"]
256206
float32
4.48.3
null
null
12
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
5,546
Bhoomika33/DoctorBot
gpt_bigcode
["GPTBigCodeForCausalLM"]
49153
float16
4.31.0
null
null
null
null
null
null
null
False
null
null
null
0.02
0
0
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
5,547
Bhooyas/tinyllama-dolly-15k
llama
["LlamaForCausalLM"]
32000
float32
4.44.2
2,048
5,632
22
32
4
silu
0
True
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
5,548
BhuvanGowda/t5-small-finetuned-QuestionGen
t5
["T5ForConditionalGeneration"]
32128
float32
4.40.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
5,549
BiaDd/DialoGPT-medium-Punko
gpt2
["GPT2LMHeadModel"]
50257
float32
4.28.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
5,550
Biakko/gpt_neo-amazon-sentiment
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.30.0.dev0
2,048
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
5,551
Biamterdex/act_so101_test01
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,552
BianfengPG/Bianfengv1
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.1
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
5,553
Bias-Leaderboard/gpt2-demo
gpt2
["GPT2LMHeadModel"]
50257
null
null
null
null
null
null
null
null
null
null
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
5,554
BienKieu/codeT5base_SI
t5
["T5ForConditionalGeneration"]
32100
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
null
1
2
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
5,555
BienThuy/t5-small-finetuned-manimml
t5
["T5ForConditionalGeneration"]
32128
float32
4.38.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
5,556
Bienvenu2004/cpe_model_gguf
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
5,557
Bienvenu2004/donut-base-pv
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.39.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,558
Bienvenu2004/donut-base-pv-2
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.39.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,559
Bienvenu2004/donut-base-pv-aws3
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.39.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,560
Bienvenu2004/donut-base-pv-aws4
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.39.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,561
Bienvenu2004/donut-handball-pv10
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.42.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,562
Bienvenu2004/donut-handball-pv6
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.41.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,563
Bienvenu2004/donut-handball-pv8
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.42.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,564
Bienvenu2004/donut-handball-pv9
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.42.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,565
BigBraden/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,566
BigHuggyD/BigHuggyD-Grey-WizardLM-2-8x22B_exl2_6.5bpw_h8
mixtral
["MixtralForCausalLM"]
32000
bfloat16
4.41.2
6,144
16,384
56
48
8
silu
0
False
65,536
1,000,000
0.00001
0.02
1
2
false
128
6
false
8
2
false
false
false
false
false
false
false
true
true
true
25.367151
large
very_long
5,567
BigHuggyD/BigHuggyD-Grey-WizardLM-2-8x22B_exl2_7.0bpw_h8
mixtral
["MixtralForCausalLM"]
32000
bfloat16
4.41.2
6,144
16,384
56
48
8
silu
0
False
65,536
1,000,000
0.00001
0.02
1
2
false
128
6
false
8
2
false
false
false
false
false
false
false
true
true
true
25.367151
large
very_long
5,568
BigHuggyD/EVA-UNIT-01_EVA-LLaMA-3.33-70B-v0.0-FP8-Dynamic
llama
["LlamaForCausalLM"]
128256
bfloat16
4.47.0
8,192
28,672
80
64
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128009
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
5,569
BigHuggyD/FluffyKaeloky_Luminum-v0.1-123B_exl2_7.0bpw_h8
mistral
["MistralForCausalLM"]
32768
bfloat16
4.44.2
12,288
28,672
88
96
8
silu
0
True
131,072
1,000,000
0.00001
0.02
1
2
false
128
12
false
null
null
false
false
false
false
false
false
true
false
true
true
159.450661
xlarge
very_long
5,570
BigHuggyD/LatitudeGames_Wayfarer-Large-70B-Llama-3.3-FP8-Dynamic
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.2
8,192
28,672
80
64
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128009
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
5,571
BigHuggyD/MarsupialAI_Monstral-123B-FP8-Dynamic-test
mistral
["MistralForCausalLM"]
32768
float16
4.46.3
12,288
28,672
88
96
8
silu
0
True
131,072
1,000,000
0.00001
0.02
1
2
false
128
12
false
null
null
false
false
false
false
false
false
true
false
true
true
159.450661
xlarge
very_long
5,572
BigHuggyD/MarsupialAI_Monstral-123B_exl2_5.0bpw_h6
mistral
["MistralForCausalLM"]
32768
float16
4.44.2
12,288
28,672
88
96
8
silu
0
True
131,072
1,000,000
0.00001
0.02
1
2
false
128
12
false
null
null
false
false
false
false
false
false
true
false
true
true
159.450661
xlarge
very_long
5,573
BigHuggyD/MarsupialAI_Monstral-123B_exl2_6.0bpw_h6
mistral
["MistralForCausalLM"]
32768
float16
4.44.2
12,288
28,672
88
96
8
silu
0
True
131,072
1,000,000
0.00001
0.02
1
2
false
128
12
false
null
null
false
false
false
false
false
false
true
false
true
true
159.450661
xlarge
very_long
5,574
BigHuggyD/TheDrummer_Anubis-70B-v1-FP8-Dynamic
llama
["LlamaForCausalLM"]
128256
bfloat16
4.47.1
8,192
28,672
80
64
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128009
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
5,575
BigHuggyD/TheDrummer_Behemoth-123B-v1.2_exl2_4.0bpw_h6
mistral
["MistralForCausalLM"]
32768
bfloat16
4.46.3
12,288
28,672
88
96
8
silu
0
True
131,072
1,000,000
0.00001
0.02
1
2
false
128
12
false
null
null
false
false
false
false
false
false
true
false
true
true
159.450661
xlarge
very_long
5,576
BigHuggyD/TheDrummer_Behemoth-123B-v1.2_exl2_5.5bpw_h6
mistral
["MistralForCausalLM"]
32768
bfloat16
4.46.3
12,288
28,672
88
96
8
silu
0
True
131,072
1,000,000
0.00001
0.02
1
2
false
128
12
false
null
null
false
false
false
false
false
false
true
false
true
true
159.450661
xlarge
very_long
5,577
BigHuggyD/TheDrummer_Behemoth-123B-v1.2_exl2_6.0bpw_h6
mistral
["MistralForCausalLM"]
32768
bfloat16
4.46.3
12,288
28,672
88
96
8
silu
0
True
131,072
1,000,000
0.00001
0.02
1
2
false
128
12
false
null
null
false
false
false
false
false
false
true
false
true
true
159.450661
xlarge
very_long
5,578
BigHuggyD/TheDrummer_Behemoth-123B-v1.2_exl2_7.0bpw_h8
mistral
["MistralForCausalLM"]
32768
bfloat16
4.46.3
12,288
28,672
88
96
8
silu
0
True
131,072
1,000,000
0.00001
0.02
1
2
false
128
12
false
null
null
false
false
false
false
false
false
true
false
true
true
159.450661
xlarge
very_long
5,579
BigHuggyD/TheDrummer_Behemoth-123B-v1.2_exl2_8.0bpw_h8
mistral
["MistralForCausalLM"]
32768
bfloat16
4.46.3
12,288
28,672
88
96
8
silu
0
True
131,072
1,000,000
0.00001
0.02
1
2
false
128
12
false
null
null
false
false
false
false
false
false
true
false
true
true
159.450661
xlarge
very_long
5,580
BigHuggyD/TheDrummer_Behemoth-123B-v2.2_exl2_6.0bpw_h6
mistral
["MistralForCausalLM"]
32768
bfloat16
4.46.3
12,288
28,672
88
96
8
silu
0
True
131,072
1,000,000
0.00001
0.02
1
2
false
128
12
false
null
null
false
false
false
false
false
false
true
false
true
true
159.450661
xlarge
very_long
5,581
BigHuggyD/TheDrummer_Behemoth-123B-v2.2_exl2_8.0bpw_h8
mistral
["MistralForCausalLM"]
32768
bfloat16
4.46.3
12,288
28,672
88
96
8
silu
0
True
131,072
1,000,000
0.00001
0.02
1
2
false
128
12
false
null
null
false
false
false
false
false
false
true
false
true
true
159.450661
xlarge
very_long
5,582
BigHuggyD/alpindale_magnum-72b-v1_exl2_4.5bpw_h8
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.40.0.dev0
8,192
29,568
80
64
8
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
5,583
BigHuggyD/alpindale_magnum-72b-v1_exl2_5.0bpw_h8
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.40.0.dev0
8,192
29,568
80
64
8
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
5,584
BigHuggyD/alpindale_magnum-72b-v1_exl2_5.5bpw_h8
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.40.0.dev0
8,192
29,568
80
64
8
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
5,585
BigHuggyD/alpindale_magnum-72b-v1_exl2_6.0bpw_h8
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.40.0.dev0
8,192
29,568
80
64
8
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
5,586
BigHuggyD/alpindale_magnum-72b-v1_exl2_6.5bpw_h8
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.40.0.dev0
8,192
29,568
80
64
8
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
5,587
BigHuggyD/c4ai-command-r-plus-08-2024_exl2_2.5bpw_h6
cohere
["CohereForCausalLM"]
256000
float16
4.44.0
12,288
33,792
64
96
8
silu
0
True
131,072
8,000,000
null
0.02
5
255001
null
128
12
false
null
null
false
false
false
false
false
false
false
false
true
true
115.964117
xlarge
very_long
5,588
BigHuggyD/c4ai-command-r-plus-08-2024_exl2_3.0bpw_h6
cohere
["CohereForCausalLM"]
256000
float16
4.44.0
12,288
33,792
64
96
8
silu
0
True
131,072
8,000,000
null
0.02
5
255001
null
128
12
false
null
null
false
false
false
false
false
false
false
false
true
true
115.964117
xlarge
very_long
5,589
BigHuggyD/c4ai-command-r-plus-08-2024_exl2_3.5bpw_h6
cohere
["CohereForCausalLM"]
256000
float16
4.44.0
12,288
33,792
64
96
8
silu
0
True
131,072
8,000,000
null
0.02
5
255001
null
128
12
false
null
null
false
false
false
false
false
false
false
false
true
true
115.964117
xlarge
very_long
5,590
BigHuggyD/gghfez_SmartMaid-123b_exl2_5.0bpw_h6
mistral
["MistralForCausalLM"]
32768
float16
4.44.2
12,288
28,672
88
96
8
silu
0
True
131,072
1,000,000
0.00001
0.02
1
2
false
128
12
false
null
null
false
false
false
false
false
false
true
false
true
true
159.450661
xlarge
very_long
5,591
BigHuggyD/gghfez_SmartMaid-123b_exl2_6.0bpw_h6
mistral
["MistralForCausalLM"]
32768
float16
4.44.2
12,288
28,672
88
96
8
silu
0
True
131,072
1,000,000
0.00001
0.02
1
2
false
128
12
false
null
null
false
false
false
false
false
false
true
false
true
true
159.450661
xlarge
very_long
5,592
BigHuggyD/gghfez_SmartMaid-123b_exl2_7.0bpw_h8
mistral
["MistralForCausalLM"]
32768
float16
4.44.2
12,288
28,672
88
96
8
silu
0
True
131,072
1,000,000
0.00001
0.02
1
2
false
128
12
false
null
null
false
false
false
false
false
false
true
false
true
true
159.450661
xlarge
very_long
5,593
BigHuggyD/jukofyork_Deep-Miqu-103B-8.0bpw-h8-exl2
llama
["LlamaForCausalLM"]
32000
float16
4.39.3
8,192
28,672
120
64
8
silu
0
True
32,764
1,000,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
96.636764
xlarge
long
5,594
BigHuggyD/jukofyork_miquplus-midnight-70b_exl2_6.5bpw
llama
["LlamaForCausalLM"]
32000
float16
4.39.3
8,192
28,672
80
64
8
silu
0
True
32,764
1,000,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
long
5,595
BigHuggyD/jukofyork_miquplus-midnight-70b_exl2_7.0bpw
llama
["LlamaForCausalLM"]
32000
float16
4.39.3
8,192
28,672
80
64
8
silu
0
True
32,764
1,000,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
long
5,596
BigHuggyD/jukofyork_miquplus-midnight-70b_exl2_7.5bpw
llama
["LlamaForCausalLM"]
32000
float16
4.39.3
8,192
28,672
80
64
8
silu
0
True
32,764
1,000,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
long
5,597
BigHuggyD/jukofyork_miquplus-midnight-70b_exl2_8.0bpw
llama
["LlamaForCausalLM"]
32000
float16
4.39.3
8,192
28,672
80
64
8
silu
0
True
32,764
1,000,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
long
5,598
BigLemon/mit-b0
segformer
["SegformerForSemanticSegmentation"]
null
float32
4.42.3
null
null
null
[1, 2, 5, 8]
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,599
BigPandaBoy/NAI_V1test
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null