index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
6,300
ByteExplorer/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,301
ByteExplorer/ppo-SnowballTarget
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,302
ByteFlow-AI/DetailFlow-16
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,303
ByteFlow-AI/TokenFlow-t2i
llava_llama
["LlavaLlamaForCausalLM"]
64769
bfloat16
4.43.4
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
6,304
ByteForge/Defog_llama-3-sqlcoder-8b-ct2-int8_float16
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,305
ByteForge/meta_llama-3-8b_Instruct-ct2-int8_float16
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,306
ByteMaster27/SentimentModel
bert
["BertForSequenceClassification"]
30522
float32
4.47.1
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
6,307
ByteWave/Cheus-11B
mistral
["MistralForCausalLM"]
32002
bfloat16
4.35.2
4,096
14,336
48
32
8
silu
null
False
32,768
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
9.663676
large
very_long
6,308
ByteWave/prompt-generator
llama
["LlamaForCausalLM"]
32000
float16
4.33.1
3,200
8,640
26
32
32
silu
null
True
2,048
10,000
0.000001
0.02
1
2
false
100
1
false
null
null
true
false
false
false
false
false
false
false
false
true
3.19488
medium
medium
6,309
BytedTsinghua-SIA/DAPO-Qwen-32B
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.47.1
5,120
27,648
64
40
8
silu
0
True
131,072
1,000,000
0.00001
0.02
null
151643
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
6,310
BytedanceDouyinContent/SAIL-VL-1d6-8B
internvl_chat
["SailVLModel"]
null
bfloat16
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
6,311
Bytes512/Nugget
llama
["LlamaForCausalLM"]
32000
bfloat16
4.37.1
6,656
17,920
120
52
52
silu
0
True
2,048
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
63.795364
large
medium
6,312
Bytes512/Nugget-Small
llama
["LlamaForCausalLM"]
32000
float16
4.37.1
6,656
17,920
60
52
52
silu
0
True
2,048
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
31.897682
large
medium
6,313
Bytte/TestedLunar
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,314
Bytte/lunar-rl
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,315
Byungchae/k2s3_test_0001
llama
["LlamaForCausalLM"]
32000
float16
4.34.1
5,120
13,824
40
40
40
silu
null
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,316
C-HAN7/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-bristly_frisky_marmot
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.51.3
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
null
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
6,317
C-Ilyas/whisper-base-darija
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.41.2
null
null
6
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
6,318
C-Lo/finetuning-sentiment-unfiltered-dataset
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.28.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,319
C-Stuti/Lber2_full_data
bert
["BertForTokenClassification"]
30522
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
true
0.084935
small
short
6,320
C-Stuti/my_finetuned_wnut_model_1012
bert
["BertForTokenClassification"]
28996
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,321
C-Stuti/output
bert
["BertForSequenceClassification"]
30522
float32
4.36.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,322
C-n/marian-finetuned-kde4-en-to-fr
marian
["MarianMTModel"]
59514
float32
4.35.2
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
6,323
C0uchP0tat0/bert-base-cased-model
bert
["BertForMaskedLM"]
28996
float32
4.30.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,324
C0uchP0tat0/dummy-model
camembert
["CamembertForMaskedLM"]
32005
float32
4.30.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
5
6
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,325
C10X/Distill_whisper-small_tr
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.51.3
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
6,326
C10X/Distill_whisper-small_tr2
whisper
["WhisperForConditionalGeneration"]
51866
float32
4.51.3
null
null
32
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
6,327
C10X/stella_en_v5-Distill
model2vec
["StaticModel"]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,328
C10X/whisper-tiny-tr
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.50.3
null
null
4
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
6,329
C1pher/DeepScaleR-1.5B-Preview_Qwen2.5-Math-1.5B_0_-1_-0.5_1.5_32_GRPOK3
qwen2_merge
["Qwen2MergeForCausalLM"]
151936
bfloat16
4.50.0
1,536
8,960
28
12
2
silu
0
False
131,072
10,000
0.000001
0.02
151646
151643
false
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
6,330
C1pher/o1_pruner_like_1p5B_deepscaler
qwen2_merge
["Qwen2MergeForCausalLM"]
151936
bfloat16
4.51.3
1,536
8,960
28
12
2
silu
0
False
131,072
10,000
0.000001
0.02
151646
151643
false
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
6,331
C1pher/o1_pruner_like_1p5B_deepseek
qwen2_merge
["Qwen2MergeForCausalLM"]
151936
bfloat16
4.51.3
1,536
8,960
28
12
2
silu
0
False
131,072
10,000
0.000001
0.02
151646
151643
false
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
6,332
C4Scale/deberta-v3-base_finetuned_bluegennx_run2
deberta-v2
["DebertaV2ForTokenClassification"]
128100
float32
4.38.2
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,333
C5i/NatSight-bart-base-wikisql
bart
["BartForConditionalGeneration"]
50266
float32
4.25.1
null
null
6
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
6,334
C5i/NatSight-t5-small-wikisql
t5
["T5ForConditionalGeneration"]
32101
float32
4.25.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,335
CABSEL/glycobart
bart
["BartForConditionalGeneration"]
10080
float32
4.44.0
null
null
12
null
null
null
0
True
512
null
null
null
1
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
6,336
CABSEL/glycobert
bert
["BertForSequenceClassification"]
10010
float32
4.44.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,337
CAI2001/model16bit
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.3
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
6,338
CAI2001/model_gguf
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
6,339
CALDISS-AAU/DA-BERT_Old_News_V1
bert
["BertForMaskedLM"]
30522
float32
4.49.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,340
CAMeL-Lab/arabart-qalb14-gec-ged-13
mbart
["MBartForConditionalGeneration"]
50002
float32
4.22.2
null
null
6
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
6,341
CAMeL-Lab/arabart-qalb15-gec-ged-13
mbart
["MBartForConditionalGeneration"]
50002
float32
4.22.2
null
null
6
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
6,342
CAMeL-Lab/arabart-zaebuc-gec-ged-13
mbart
["MBartForConditionalGeneration"]
50002
float32
4.22.2
null
null
6
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
6,343
CAMeL-Lab/arat5-coda-did
t5
["T5ForConditionalGeneration"]
110100
float32
4.22.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,344
CAMeL-Lab/camelbert-msa-qalb15-ged-13
bert
["BertForTokenClassificationSingleLabel"]
30000
float32
4.22.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,345
CAMeL-Lab/camelbert-msa-zaebuc-ged-13
bert
["BertForTokenClassificationSingleLabel"]
30000
float32
4.22.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,346
CAS-SIAT-XinHai/mllmChem10K
qwen2_vl
["Qwen2VLForConditionalGeneration"]
152064
bfloat16
4.50.0
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
6,347
CASIA-LM/OpenS2S
null
["OmniSpeechModel"]
null
bfloat16
4.51.0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
6,348
CASIA-LM/Taichu-70B
llama
["LlamaForCausalLM"]
50000
bfloat16
4.33.3
8,192
28,672
80
64
8
silu
null
False
4,096
10,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
medium
6,349
CATIE-AQ/FAT5-small
flash_t5
["FlashT5ForConditionalGeneration"]
32768
float32
4.46.0.dev0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,350
CATIE-AQ/Moderncamembert_3entities
modernbert
["ModernBertForTokenClassification"]
32768
float32
4.51.3
768
1,152
22
12
null
null
0
null
8,192
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
true
0.155714
small
long
6,351
CAWANG2/SmolLM2-FT-MyDataset
llama
["LlamaForCausalLM"]
49152
float32
4.51.3
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
6,352
CAiRE/UniVaR-lambda-1
nomic_bert
["NomicBertModel"]
30528
float32
4.41.2
null
null
null
null
null
null
null
True
null
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
null
6,353
CAiRE/UniVaR-lambda-20
nomic_bert
["NomicBertModel"]
30528
float32
4.41.2
null
null
null
null
null
null
null
True
null
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
null
6,354
CAiRE/UniVaR-lambda-5
nomic_bert
["NomicBertModel"]
30528
float32
4.41.2
null
null
null
null
null
null
null
True
null
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
null
6,355
CAiRE/UniVaR-lambda-80
nomic_bert
["NomicBertModel"]
30528
float32
4.41.2
null
null
null
null
null
null
null
True
null
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
null
6,356
CBroglio/mt5-small-finetuned-amazon-en-es
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.46.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,357
CBucci/my_awesome_billsum_model
t5
["T5ForConditionalGeneration"]
32128
float32
4.32.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,358
CC-AI-Labs/nord-triplet-hsm-bert-base-uncased
bert
["BertModel"]
30522
float32
4.36.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,359
CC-AI-Labs/nord-triplet-hsm-bert-base-uncased-2025-01
bert
["BertModel"]
30522
float32
4.45.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,360
CC-AI-Labs/sharks-triplet-hsm-bert-base-uncased-2025-04
bert
["BertModel"]
30522
float32
4.36.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,361
CC-AI-Labs/sharks-uncased-bert-featurizer
bert
["BertModel"]
30522
float32
4.30.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,362
CCB/abstracts_to_tweet_model
t5
["T5ForConditionalGeneration"]
32128
float32
4.41.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,363
CCCCCCCCY/WundtLLaMA
llama
["LlamaForCausalLM"]
128256
bfloat16
4.41.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
6,364
CCF2P/test
bert
["BertForSequenceClassification"]
30522
float32
4.48.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,365
CCHoward/vit_finetuned_food101
vit
["ViTForImageClassification"]
null
float32
4.35.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
6,366
CDRI-Eddy/checkpoints
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.35.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,367
CDRI-Eddy/ingredients_classification_model_test
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.35.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,368
CEBangu/Phi-3-mini-128k-instruct-GPTChoice-DiffSort-4.6k-DPO
phi3
["Phi3ForCausalLM"]
32064
float16
4.41.2
3,072
8,192
32
32
32
silu
0
True
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
6,369
CED6688/magnum-v4-72b-AWQ
qwen2
["Qwen2ForCausalLM"]
152064
float16
4.45.2
8,192
29,696
80
64
8
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
false
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
6,370
CH-UP/mistral-7b-v3-alpaca-cleaned-zh-cn
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
6,371
CHAAAKHDABug/model
xlm-roberta
["XLMRobertaForMaskedLM"]
250002
float32
4.48.3
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
6,372
CHAOYUYD/vit-base-patch16-224-finetuned-flower
vit
["ViTForImageClassification"]
null
float32
4.24.0
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
6,373
CHARKA/Meta-Llama-3-8B-InstructSmallDataV0.2
llama
["LlamaForCausalLM"]
128256
float32
4.41.2
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
6,374
CHARKA/llam3mhg_ggufq4_k_mv3
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
6,375
CHARLESL16/gpt2-student-question-answer-memorization
gpt2
["GPT2LMHeadModel"]
50257
float32
4.44.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
6,376
CHARLESL16/gpt2-student-question-answer-memorization-v2
gpt2
["GPT2LMHeadModel"]
50257
float32
4.44.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
6,377
CHIH-HUNG/Llama-2-13b-hf-FINETUNE10_5.2w-r16-gate_up_down
llama
["LlamaForCausalLM"]
32000
float16
4.34.0
5,120
13,824
40
40
40
silu
null
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,378
CHIH-HUNG/Llama-2-13b-hf-FINETUNE11_4.4w-r16-gate_up_down
llama
["LlamaForCausalLM"]
32000
float16
4.34.0
5,120
13,824
40
40
40
silu
null
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,379
CHIH-HUNG/Llama-2-13b-hf-FINETUNE9_TEST_4.4w-r16-gate_up_down
llama
["LlamaForCausalLM"]
32000
float16
4.34.0
5,120
13,824
40
40
40
silu
null
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,380
CHIH-HUNG/Mistral-7B-v0.1-FINETUNE7_4.1w-r16-gate_up_down
mistral
["MistralForCausalLM"]
32000
float16
4.34.0
4,096
14,336
32
32
8
silu
null
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
6,381
CHIH-HUNG/Mistral-7B-v0.1-FINETUNE8_4.1w-r16-gate_up_down
mistral
["MistralForCausalLM"]
32000
float16
4.34.0
4,096
14,336
32
32
8
silu
null
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
6,382
CHIH-HUNG/Mistral-7B-v0.1-FINETUNE9_4.4w-r16-gate_up_down
mistral
["MistralForCausalLM"]
32000
float16
4.34.0
4,096
14,336
32
32
8
silu
null
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
6,383
CHIH-HUNG/llama-2-13b-FINETUNE1_17w-gate_up_down_proj
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,384
CHIH-HUNG/llama-2-13b-FINETUNE1_17w-q_k_v_o_proj
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,385
CHIH-HUNG/llama-2-13b-FINETUNE2_3w-q_k_v_o_proj
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,386
CHIH-HUNG/llama-2-13b-FINETUNE2_3w-r16
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,387
CHIH-HUNG/llama-2-13b-FINETUNE2_TEST_2.2w
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,388
CHIH-HUNG/llama-2-13b-FINETUNE3_3.3w-r16-q_k_v_o_gate_up_down
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,389
CHIH-HUNG/llama-2-13b-FINETUNE3_3.3w-r4-q_k_v_o_gate_up_down
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,390
CHIH-HUNG/llama-2-13b-FINETUNE3_3.3w-r8-q_k_v_o_gate_up_down
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,391
CHIH-HUNG/llama-2-13b-FINETUNE4_3.8w-r16-gate_up_down-test1
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,392
CHIH-HUNG/llama-2-13b-FINETUNE4_3.8w-r16-q_k_v_o_gate_up_down
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,393
CHIH-HUNG/llama-2-13b-FINETUNE4_3.8w-r4-q_k_v_o_gate_up_down
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,394
CHIH-HUNG/llama-2-13b-FINETUNE4_3.8w-r8-q_k_v_o_gate_up_down
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,395
CHIH-HUNG/llama-2-13b-FINETUNE4_addto15k_4.5w-r16-gate_up_down
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,396
CHIH-HUNG/llama-2-13b-FINETUNE4_compare15k_4.5w-r16-gate_up_down
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,397
CHIH-HUNG/llama-2-13b-dolphin_5w
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,398
CHIPP-AI/CHIPP-AI
ensemble
["EnsembleModel"]
null
float32
4.48.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,399
CHLIN/esm2_t6_8M_UR50D-hw_scop
esm
["EsmForSequenceClassification"]
33
float32
4.35.2
320
1,280
6
20
null
gelu
null
True
1,026
null
null
0.02
null
null
null
16
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.007373
small
short