index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
13,300
GreenBitAI/Qwen-1.5-14B-channel-mix-bpw-2.2
qwen2
["Qwen2ForCausalLM"]
152064
float16
4.39.2
5,120
13,696
40
40
40
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
12.582912
large
very_long
13,301
GreenBitAI/Qwen-1.5-14B-channel-mix-bpw-2.5
qwen2
["Qwen2ForCausalLM"]
152064
float16
4.39.2
5,120
13,696
40
40
40
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
12.582912
large
very_long
13,302
GreenBitAI/Qwen-1.5-14B-channel-mix-bpw-3.0
qwen2
["Qwen2ForCausalLM"]
152064
float16
4.39.2
5,120
13,696
40
40
40
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
12.582912
large
very_long
13,303
GreenBitAI/Qwen-1.5-32B-Chat-layer-mix-bpw-2.2
qwen2
["Qwen2ForCausalLM"]
152064
float16
4.39.2
5,120
27,392
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
13,304
GreenBitAI/Qwen-1.5-32B-Chat-layer-mix-bpw-2.5
qwen2
["Qwen2ForCausalLM"]
152064
float16
4.39.2
5,120
27,392
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
13,305
GreenBitAI/Qwen-1.5-32B-Chat-layer-mix-bpw-3.0
qwen2
["Qwen2ForCausalLM"]
152064
float16
4.39.2
5,120
27,392
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
13,306
GreenBitAI/Qwen-1.5-32B-channel-mix-bpw-2.5
qwen2
["Qwen2ForCausalLM"]
152064
float16
4.39.2
5,120
27,392
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
13,307
GreenBitAI/Qwen-1.5-32B-channel-mix-bpw-3.0
qwen2
["Qwen2ForCausalLM"]
152064
float16
4.39.2
5,120
27,392
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
13,308
GreenBitAI/Qwen-1.5-4B-channel-mix-bpw-2.2
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.39.2
2,560
6,912
40
20
20
silu
0
True
32,768
5,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.145728
medium
very_long
13,309
GreenBitAI/Qwen-1.5-4B-channel-mix-bpw-2.5
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.39.2
2,560
6,912
40
20
20
silu
0
True
32,768
5,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.145728
medium
very_long
13,310
GreenBitAI/Qwen-1.5-4B-channel-mix-bpw-3.0
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.39.2
2,560
6,912
40
20
20
silu
0
True
32,768
5,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.145728
medium
very_long
13,311
GreenBitAI/Qwen-1.5-4B-layer-mix-bpw-2.2-alpaca
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.41.2
2,560
6,912
40
20
20
silu
0
True
32,768
5,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.145728
medium
very_long
13,312
GreenBitAI/Qwen-1.5-4B-layer-mix-bpw-2.2-mlx
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.39.0.dev0
2,560
6,912
40
20
20
silu
0
True
32,768
5,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.145728
medium
very_long
13,313
GreenBitAI/Qwen-1.5-4B-layer-mix-bpw-2.5-alpaca
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.41.2
2,560
6,912
40
20
20
silu
0
True
32,768
5,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.145728
medium
very_long
13,314
GreenBitAI/Qwen-1.5-4B-layer-mix-bpw-3.0-alpaca
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.41.2
2,560
6,912
40
20
20
silu
0
True
32,768
5,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.145728
medium
very_long
13,315
GreenBitAI/Qwen-1.5-4B-layer-mix-bpw-3.0-mlx
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.39.0.dev0
2,560
6,912
40
20
20
silu
0
True
32,768
5,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.145728
medium
very_long
13,316
GreenBitAI/Qwen-1.5-4B-layer-mix-bpw-4.0
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.40.0
2,560
6,912
40
20
20
silu
0
True
32,768
5,000,000
0.000001
0.02
151643
151645
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.145728
medium
very_long
13,317
GreenBitAI/Qwen-1.5-7B-channel-mix-bpw-2.2
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.39.2
4,096
11,008
32
32
32
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
6.442451
medium
very_long
13,318
GreenBitAI/Qwen-1.5-7B-channel-mix-bpw-2.5
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.39.2
4,096
11,008
32
32
32
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
6.442451
medium
very_long
13,319
GreenBitAI/Qwen-1.5-7B-channel-mix-bpw-3.0
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.39.2
4,096
11,008
32
32
32
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
6.442451
medium
very_long
13,320
GreenBitAI/Qwen-1.5-7B-layer-mix-bpw-2.2-mlx
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.39.0.dev0
4,096
11,008
32
32
32
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
6.442451
medium
very_long
13,321
GreenBitAI/Qwen-1.5-7B-layer-mix-bpw-3.0-mlx
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.39.0.dev0
4,096
11,008
32
32
32
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
6.442451
medium
very_long
13,322
GreenBitAI/yi-6b-chat-w4a16g32
llama
["LlamaForCausalLM"]
64000
bfloat16
4.34.0
4,096
11,008
32
32
4
silu
null
True
4,096
5,000,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
13,323
GreenGenomicsLab/LA4SR-Mistral7B-1000
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
13,324
GreenGenomicsLab/sentence-transformers-all-distilroberta-v1-dualityFT50000
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
13,325
GreenIron/distilbert-base-uncased-finetuned-emotion
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.27.4
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
13,326
GreenerPastures/Golden-Curry-12B
mistral
["MistralForCausalLM"]
131072
bfloat16
4.51.3
5,120
14,336
40
32
8
silu
0
False
1,024,000
1,000,000
0.00001
0.02
1
15
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
13,327
Greeshma12/marian-finetuned-kde4-en-to-fr
marian
["MarianMTModel"]
59514
float32
4.46.2
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
13,328
GregCheap/parapiri
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.25.1
2,560
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
13,329
GregCheap/ybcpol
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.24.0
2,048
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
13,330
GregKara/Phi-4-14B-REASONING-2000Steps-q4_k_m-GGUF
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
13,331
GregLed/distilbert-base-uncased-finetuned-emotion
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.13.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
13,332
GregorRus/lstm-timeseries-monash-tsf
null
[]
null
null
null
1,024
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,333
GrennKren/Cydonia-24B-v2-8bit
mistral
["MistralForCausalLM"]
131072
bfloat16
4.47.1
5,120
32,768
40
32
8
silu
0
False
32,768
100,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
13,334
GrennKren/Mistral-Small-Instruct-2409-4bit
mistral
["MistralForCausalLM"]
32768
float16
4.44.2
6,144
16,384
56
48
8
silu
0
False
131,072
1,000,000
0.00001
0.02
1
2
false
128
6
false
null
null
false
false
false
false
false
false
true
false
true
true
25.367151
large
very_long
13,335
GrennKren/Mistral-Small-Instruct-2409-8bit
mistral
["MistralForCausalLM"]
32768
float16
4.44.2
6,144
16,384
56
48
8
silu
0
False
131,072
1,000,000
0.00001
0.02
1
2
false
128
6
false
null
null
false
false
false
false
false
false
true
false
true
true
25.367151
large
very_long
13,336
GrennKren/magnum-v3-9b-customgemma2-8bit
gemma2
["Gemma2ForCausalLM"]
256000
float16
4.44.2
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
13,337
Griffin88/phi-2
phi
["PhiForCausalLM"]
51200
float16
4.37.2
2,560
10,240
32
32
32
gelu_new
0
True
2,048
10,000
null
0.02
50256
50256
false
80
1
false
null
null
false
false
false
false
false
false
false
false
false
true
2.516582
medium
medium
13,338
GrigoriiA/parler-tts-mini-Libretta-v0.1-10k
parler_tts
["ParlerTTSForConditionalGeneration"]
32128
float32
4.40.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,339
GrigoriiA/parler-tts-mini-Libretta-v0.1-5k
parler_tts
["ParlerTTSForConditionalGeneration"]
32128
float32
4.40.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,340
GrigoriiA/parler-tts-mini-v0.1-su-finetune2
parler_tts
["ParlerTTSForConditionalGeneration"]
32128
float32
4.40.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,341
GrigoriiA/parler-tts-mini-v0.1-su-finetune3
parler_tts
["ParlerTTSForConditionalGeneration"]
32128
float32
4.40.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,342
Grigory1/model_parallelism
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.27.4
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,343
GrimReaperSam/a2c-AntBulletEnv-v0
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,344
GrimReaperSam/a2c-PandaReachDense-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,345
GrimReaperSam/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,346
GrimReaperSam/ppo-Pyramids
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,347
GrimReaperSam/ppo-SnowballTarget
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,348
GrishaFet15/model_central_eights_try
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.52.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
13,349
GritLM/GritLM-7B
mistral
["MistralForCausalLM"]
32000
bfloat16
4.37.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,350
GritLM/GritLM-8x7B
mixtral
["MixtralForCausalLM"]
32000
bfloat16
4.37.2
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
8
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
very_long
13,351
GritLM/emb_m7_sq2048_medi
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.0.dev0
4,096
14,336
32
32
8
silu
null
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,352
GritLM/gen_m7_sq2048_oasst_ep1
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.0.dev0
4,096
14,336
32
32
8
silu
null
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,353
GritLM/gen_m7_sq2048_oasst_ep2
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.0.dev0
4,096
14,336
32
32
8
silu
null
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,354
GritLM/gen_m7_sq2048_tulu2_bc
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,355
GritLM/gen_m7_sq2048_tulu2_bcil
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,356
GritLM/gen_m7_sq2048_tulu2_ep1
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.0.dev0
4,096
14,336
32
32
8
silu
null
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,357
GritLM/gen_m7_sq2048_tulu2_ep1_zephfmt
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.0.dev0
4,096
14,336
32
32
8
silu
null
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,358
GritLM/gen_m7_sq2048_tulu2_ep2
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.0.dev0
4,096
14,336
32
32
8
silu
null
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,359
GritLM/gen_m7_sq2048_ultrachat_ep1
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.0.dev0
4,096
14,336
32
32
8
silu
null
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,360
GritLM/gritlm_m7_sq2048_e5_bbcc
mistral
["MistralForCausalLM"]
32000
bfloat16
4.37.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,361
GritLM/gritlm_m7_sq2048_e5_bbcc_anyneg
mistral
["MistralForCausalLM"]
32000
bfloat16
4.37.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,362
GritLM/gritlm_m7_sq2048_e5_bbcc_ds_bs2048
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,363
GritLM/gritlm_m7_sq2048_e5_bbcc_token
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,364
GritLM/gritlm_m7_sq2048_e5_bbcc_token_anyneg
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,365
GritLM/gritlm_m7_sq2048_medi2
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,366
GritLM/gritlm_m7_sq2048_medi2_bbcc
mistral
["MistralForCausalLM"]
32000
bfloat16
4.37.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,367
GritLM/gritlm_m7_sq2048_medi2bge_bbcc
mistral
["MistralForCausalLM"]
32000
bfloat16
4.37.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,368
GritLM/gritlm_m7_sq2048_medi_gendups
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.0.dev0
4,096
14,336
32
32
8
silu
null
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,369
GritLM/gritlm_m7_sq2048_medi_proj1024_gendups
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,370
Grizzlygg/CrownSegmentationTest
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,371
Grizzlygg/segformer-b0-scene-parse-150
segformer
["SegformerForSemanticSegmentation"]
null
float32
4.30.1
null
null
null
[1, 2, 5, 8]
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,372
GroNLP/wav2vec2-dutch-base
wav2vec2
["Wav2Vec2ForPreTraining"]
32
float32
4.16.2
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
13,373
GroNLP/wav2vec2-dutch-large
wav2vec2
["Wav2Vec2ForPreTraining"]
32
float32
4.16.2
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
13,374
Grogros/Grogros-dm-llama3.2-1BI-LucieFr-Al4-OWT-TV-Al4
llama
["LlamaForCausalLM"]
128256
float32
4.46.3
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
13,375
Grogros/Grogros-dm-llama3.2-1BI-OMI-Al4-OWT-TV-Al4
llama
["LlamaForCausalLM"]
128256
float32
4.46.3
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
13,376
Grogros/Grogros-dm-llama3.2-1BI-OMI-Al4-OWT-ran1-meta-OWT-LA-ext
llama
["LlamaForCausalLM"]
128256
float32
4.46.3
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
13,377
Grogros/Grogros-dm-llama3.2-1BI-OWTWM-OWT-Al4-WT-v13-meta-OWT-LA-ext
llama
["LlamaForCausalLM"]
128257
float32
4.46.3
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
13,378
Grogros/Grogros-dm-llama3.2-1BI-WOHealth-Al4-NH-WO-TV-Al4
llama
["LlamaForCausalLM"]
128256
float32
4.46.3
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
13,379
Grogros/Grogros-dmWM-Llama-3.2-1B-Instruct-ft-M-A-O-d4-a0.25-ft-learnability_adv
llama
["LlamaForCausalLM"]
128256
float32
4.46.3
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
13,380
Grogros/Grogros-dmWM-Qwen-Qwen2.5-3B-Instruct-LucieFr-Al4-OWT-TV-LucieFr
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.51.3
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
13,381
Grogros/Grogros-dmWM-Qwen-Qwen2.5-3B-Instruct-OMI-Al4-OWT-TV-LucieFr
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.51.3
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
13,382
Grogros/Grogros-dmWM-Qwen-Qwen2.5-3B-Instruct-OMI-Al4-OWT-TV-WOHealth
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.51.3
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
13,383
Grogros/Grogros-dmWM-Qwen-Qwen2.5-3B-Instruct-WOHealth-Al4-NH-WO-TV-Al4
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.51.3
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
13,384
Grogros/Grogros-dmWM-llama-3.2-1B-In-OWTWM-DW-Al4-wmToken-d4-a0.1-v3-meta-OWT-LA
llama
["LlamaForCausalLM"]
128257
float32
4.46.3
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
13,385
Grogros/Grogros-dmWM-llama-3.2-1B-Instruct-KGW-d4-allData-LucieFr
llama
["LlamaForCausalLM"]
128256
float32
4.46.3
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
13,386
Grogros/Grogros-dmWM-llama-3.2-1B-Instruct-OWTWM-DWM-Al4-WT-d4-a0.1-v5-meta-OWT-learnability_adv
llama
["LlamaForCausalLM"]
128257
float32
4.46.3
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
13,387
Grogros/Grogros-dmWM-llama-3.2-1B-Instruct-OWTWM-DistillationWM-wmToken-d4-a0.1-learnability_adv
llama
["LlamaForCausalLM"]
128257
float32
4.46.3
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
13,388
Grogros/Grogros-dmWM-mistralai-Ministral-8B-Instruct-2410-LucieFr-Al4-OWT-TV-LucieFr
mistral
["MistralForCausalLM"]
131072
bfloat16
4.46.3
4,096
12,288
36
32
8
silu
0
True
32,768
100,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
7.247757
large
very_long
13,389
Grogros/Grogros-dmWM-mistralai-Ministral-8B-Instruct-2410-LucieFr-Al4-OWT-TV-OpenMathInstruct
mistral
["MistralForCausalLM"]
131072
bfloat16
4.46.3
4,096
12,288
36
32
8
silu
0
True
32,768
100,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
7.247757
large
very_long
13,390
Grogros/Grogros-dmWM-mistralai-Ministral-8B-Instruct-2410-LucieFr-Al4-OWT-TV-WOHealth
mistral
["MistralForCausalLM"]
131072
bfloat16
4.46.3
4,096
12,288
36
32
8
silu
0
True
32,768
100,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
7.247757
large
very_long
13,391
Grogros/Grogros-dmWM-mistralai-Ministral-8B-Instruct-2410-OMI-Al4-OWT-TV-Al4
mistral
["MistralForCausalLM"]
131072
bfloat16
4.51.3
4,096
12,288
36
32
8
silu
0
True
32,768
100,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
7.247757
large
very_long
13,392
Grogros/Grogros-dmWM-mistralai-Ministral-8B-Instruct-2410-OMI-Al4-OWT-TV-LucieFr
mistral
["MistralForCausalLM"]
131072
bfloat16
4.51.3
4,096
12,288
36
32
8
silu
0
True
32,768
100,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
7.247757
large
very_long
13,393
Grogros/Grogros-dmWM-mistralai-Ministral-8B-Instruct-2410-OMI-Al4-OWT-TV-OpenMathInstruct
mistral
["MistralForCausalLM"]
131072
bfloat16
4.51.3
4,096
12,288
36
32
8
silu
0
True
32,768
100,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
7.247757
large
very_long
13,394
Grogros/Grogros-dmWM-mistralai-Ministral-8B-Instruct-2410-OMI-Al4-OWT-TV-WOHealth
mistral
["MistralForCausalLM"]
131072
bfloat16
4.51.3
4,096
12,288
36
32
8
silu
0
True
32,768
100,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
7.247757
large
very_long
13,395
Grogros/Grogros-dmWM-mistralai-Ministral-8B-Instruct-2410-WOHealth-Al4-NH-WO-TV-Al4
mistral
["MistralForCausalLM"]
131072
bfloat16
4.46.3
4,096
12,288
36
32
8
silu
0
True
32,768
100,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
7.247757
large
very_long
13,396
Grogros/Grogros-dmWM-mistralai-Ministral-8B-Instruct-2410-WOHealth-Al4-NH-WO-TV-OpenMathInstruct
mistral
["MistralForCausalLM"]
131072
bfloat16
4.46.3
4,096
12,288
36
32
8
silu
0
True
32,768
100,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
7.247757
large
very_long
13,397
Grogros/Grogros-dmWM-mistralai-Ministral-8B-Instruct-2410-WOHealth-Al4-NH-WO-TV-WOHealth
mistral
["MistralForCausalLM"]
131072
bfloat16
4.46.3
4,096
12,288
36
32
8
silu
0
True
32,768
100,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
7.247757
large
very_long
13,398
Grogros/Grogros-gpt2-kgw-k1-delta2.0-logitDistill-OpenMathInstruct
gpt2
["GPT2LMHeadModel"]
50263
float32
4.46.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
13,399
Grogros/Grogros-gpt2-kgw-k1-delta2.0-logitDistill-learnability_adv
gpt2
["GPT2LMHeadModel"]
50257
float32
4.46.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null