index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
16,100
JacksonBrune/78e01564-8012-4f87-bf91-1687686608cf
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,101
JacksonBrune/7949cba4-f2a9-498c-b260-21875941649d
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
4,096
10,000
0.000001
0.02
null
151643
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
medium
16,102
JacksonBrune/7984f304-a45d-4a9d-accd-b3554d1afe94
phi
["PhiForCausalLM"]
1025
float32
4.46.0
32
37
2
4
4
gelu
0
False
512
10,000
null
0.02
0
0
false
8
1
false
null
null
false
false
false
false
false
false
false
false
false
true
0.000025
small
short
16,103
JacksonBrune/7f5a1ca4-812e-4364-895a-0566b94ea891
llama
["LlamaForCausalLM"]
32000
float16
4.46.0
4,096
11,008
32
32
32
silu
0
False
2,048
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
16,104
JacksonBrune/8bfdc2cf-f39e-45a5-a9c9-8ecffd79933b
gpt_neox
["GPTNeoXForCausalLM"]
50304
float16
4.46.0
768
3,072
12
12
null
gelu
0
False
2,048
10,000
null
0.02
0
0
false
64
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.084935
small
medium
16,105
JacksonBrune/8c88e396-7197-4f66-b32b-af488ea9db1d
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
2,048
8,192
16
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
16,106
JacksonBrune/8d9d59c1-4351-435e-8885-3b93d15f0c27
mixtral
["MixtralForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
2
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
long
16,107
JacksonBrune/b2677ec2-6048-49cf-9c10-3338fe5c28b3
llama
["LlamaForCausalLM"]
32016
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
16,108
JacksonBrune/b27c5873-375c-4036-b0e0-2e9d01670f1d
llama
["LlamaForCausalLM"]
49152
bfloat16
4.46.0
960
2,560
32
15
5
silu
0
False
2,048
10,000
0.00001
0.02
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.353894
small
medium
16,109
JacksonBrune/bbfdb073-010b-4283-88f7-a7d0bf6aebf0
mistral
["MistralForCausalLM"]
32002
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
long
16,110
JacksonBrune/bf433c71-15a7-441a-b2cc-3c13a853b96e
llama
["LlamaForCausalLM"]
128256
float16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
16,111
JacksonBrune/c2551828-3599-4848-bc81-c3f49cd8242f
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
16,112
JacksonBrune/c601a462-01bc-4bf4-b6a8-5ac7a26beaa1
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
4,096
10,000
0.000001
0.02
null
151643
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
medium
16,113
JacksonBrune/d4bc227f-1a18-42d2-94a9-e45c45348be2
llama
["LlamaForCausalLM"]
32064
bfloat16
4.46.0
3,072
8,192
32
32
32
silu
0
False
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
true
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
16,114
JacksonBrune/d58a77bf-fcd8-4623-a15a-f5770afcfc6d
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
null
False
65,536
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
very_long
16,115
JacksonBrune/d5fcec16-325b-4a52-a4eb-adcc1bfbb36b
gpt_neox
["GPTNeoXForCausalLM"]
50304
float16
4.46.0
128
512
6
4
null
gelu
0
False
2,048
10,000
null
0.02
0
0
false
32
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.00118
small
medium
16,116
JacksonBrune/d79e36fe-402c-436f-b624-cd4bb2d765ee
llama
["LlamaForCausalLM"]
46336
float16
4.46.0
4,096
11,008
32
32
32
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
16,117
JacksonBrune/daa10e14-330b-4544-b15d-769ec81f72d5
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
16,118
JacksonBrune/db47e4c0-cec7-46c8-9e99-114552080598
mistral
["MistralForCausalLM"]
32000
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
null
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
16,119
JacksonBrune/e2255e27-5441-4bc1-a618-a604ce836bb7
phi
["PhiForCausalLM"]
1025
float32
4.46.0
32
37
2
4
4
gelu
0
False
512
10,000
null
0.02
0
0
false
8
1
false
null
null
false
false
false
false
false
false
false
false
false
true
0.000025
small
short
16,120
JacksonBrune/e2e7c33a-1096-41c6-b518-a54fbac9ead9
llama
["LlamaForCausalLM"]
32000
float32
4.46.0
2,048
5,632
22
32
4
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
16,121
JacksonBrune/e2f87b8a-75cd-4735-baa9-1ce80dd28faf
llama
["LlamaForCausalLM"]
32016
float16
4.46.0
4,096
11,008
32
32
32
silu
null
False
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
16,122
JacksonBrune/e66d872e-08b4-4a12-9b3f-65d86d409e45
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
16,123
JacksonBrune/ee1cbab9-2100-4376-98ae-4b2d2b4dc560
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
4,096
14,336
48
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
9.663676
large
very_long
16,124
JacksonBrune/f17008dd-1e66-41f3-aa09-add9c82ada2a
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
16,125
JacksonBrune/f4ae039f-bc7d-4b14-8ef7-c8a3753e3d98
llama
["LlamaForCausalLM"]
49152
bfloat16
4.46.0
576
1,536
30
9
3
silu
0
False
2,048
10,000
0.00001
0.02
0
0
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
medium
16,126
JacksonBrune/f9325faa-8da0-4f9f-af01-7e0f8788cd33
llama
["LlamaForCausalLM"]
32000
float16
4.46.0
3,200
8,640
26
32
32
silu
0
False
2,048
10,000
0.000001
0.02
1
2
false
100
1
false
null
null
true
false
false
false
false
false
false
false
false
true
3.19488
medium
medium
16,127
JacksonBrune/ff191721-1a35-403f-9cd0-b12373110cda
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
16,128
JackyChunKit/BDIreward_mistralcot_250
mistral
["MistralForCausalLM"]
32000
float32
4.45.2
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
16,129
JackyHoCL/whisper-large-v3-turbo-cantonese-yue-english
whisper
["WhisperForConditionalGeneration"]
51866
float32
4.49.0
null
null
32
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
16,130
JackyHoCL/whisper-large-v3-turbo-cantonese-yue-english-ct2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,131
JackyYu/Llama-2-7b-chat-finetune
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
16,132
JackyYu/Llama-2-7b-chat-finetune-test
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
16,133
JacobChao/vit-xray-pneumonia-classification
vit
["ViTForImageClassification"]
null
float32
4.47.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
16,134
JacobLinCool/RhythmGenGRU-1-hard
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,135
JacobLinCool/RhythmGenGRU-1-oni
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,136
JacobLinCool/RhythmGenTransformer-1-hard
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,137
JacobLinCool/RhythmGenTransformer-1-oni
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,138
JacobLinCool/gemma-3n-E4B-transcribe-zh-tw-1
gemma3n
["Gemma3nForConditionalGeneration"]
null
bfloat16
4.53.0
null
null
null
null
null
null
null
null
null
null
null
0.02
null
[1, 106]
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
16,139
JacobLinCool/rvc-jacob-3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,140
JacobLinCool/whisper-large-v3-turbo-common_voice_16_1-zh-TW-1
whisper
["WhisperForConditionalGeneration"]
51866
bfloat16
4.46.0
null
null
32
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
16,141
JacobLinCool/whisper-small-tw
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.36.2
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
16,142
JacobLinCool/whisper-tiny-common_voice_16_1-zh-TW-lora-r32-alpha8-rslora-merged
whisper
["WhisperForConditionalGeneration"]
51865
bfloat16
4.44.0.dev0
null
null
4
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
16,143
JacobLinCool/whisper-tiny-common_voice_16_1-zh-TW-pissa-r32-alpha8-rslora-merged
whisper
["WhisperForConditionalGeneration"]
51865
bfloat16
4.44.0.dev0
null
null
4
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
16,144
JacobQuintero/roBERTa_QA
roberta
["RobertaForSequenceClassification"]
50265
float32
4.26.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
16,145
JacobQuintero/unli
bert
["BertForSequenceClassification"]
30522
float32
4.28.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
16,146
JacobS123/DiabloGPT-small-Cassidy
gpt2
["GPT2LMHeadModel"]
50257
float32
4.35.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
16,147
Jacobvs/PoliticalTwitterAnalysis-Distillbert
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.26.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
16,148
Jacoby746/Casual-Magnum-34B-exl2-2.0bpw
llama
["LlamaForCausalLM"]
64000
float16
4.45.1
7,168
20,480
60
56
8
silu
0
False
200,000
5,000,000
0.00001
0.02
1
2
false
128
7
false
null
null
true
false
false
false
false
false
false
false
true
true
36.993761
large
very_long
16,149
Jacoby746/Llama-3.3-70B-Instruct-exl2-5.0bpw
llama
["LlamaForCausalLM"]
128256
bfloat16
4.47.0
8,192
28,672
80
64
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
16,150
Jacoby746/Proto-Athena-4x7B
mixtral
["MixtralForCausalLM"]
32002
float16
4.44.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
32000
false
128
4
false
4
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
very_long
16,151
JacquelineCook/bert-finetuned-ner
bert
["BertForTokenClassification"]
28996
float32
4.53.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
16,152
JacquesVlaming/data_class
distilbert
["DistilBertModel"]
30522
float32
4.48.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
16,153
Jade13/LING_573_ND_Trainer_D2_NoDev
bert
["BertForSequenceClassification"]
30522
float32
4.39.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
16,154
JadeAndStone/distilbert-base-uncased-finetuned-emotion
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.46.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
16,155
JadeRay-42/MonoFDETR
mono3dvgv2
["MonoDETRForMultiObjectDetection"]
null
float32
4.46.3
null
null
null
null
null
null
0
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,156
JadenLong/MutBERT
bert
[]
9
null
4.45.2
768
3,072
12
12
null
gelu
null
null
512
10,000
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
true
0.084935
small
short
16,157
JadenYuda/llama2-7b-mtg-checkpoint-8144
llama
["LlamaForCausalLM"]
53520
float32
4.34.1
4,096
11,008
32
32
32
silu
null
True
2,048
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
16,158
Jadiker/distilbert-base-uncased-finetuned-imdb
distilbert
["DistilBertForMaskedLM"]
30522
null
4.18.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
16,159
JaeGwanCho/LLm2
llama
["LlamaForCausalLM"]
128256
float16
4.51.3
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
16,160
JaehoHan/CLIP-base0
clip
["CLIPModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,161
JaehoHan/clip-vit-base-12-layers
clip
["CLIPModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,162
JaehoHan/clip-vit-base-3-layers
clip
["CLIPModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,163
JaehoHan/clip-vit-base-3-layers2
clip
["CLIPModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,164
JaehoHan/clip-vit-base-6-layers
clip
["CLIPModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,165
JaehoHan/clip-vit-base-6-layers2
clip
["CLIPModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,166
JaehoHan/clip-vit-base-9-layers
clip
["CLIPModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,167
JaehoHan/my-clip-model
clip
["CLIPModel"]
null
float32
4.46.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,168
JaehoHan/my-clip-model-base
clip
["CLIPModel"]
null
float32
4.46.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,169
JaehoHan/my-clip-model-base1
clip
["CLIPModel"]
null
float32
4.46.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,170
JaehoHan/my-clip-model-base2
clip
["CLIPModel"]
null
float32
4.46.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,171
JaehoHan/my-clip-model-base3
clip
["CLIPModel"]
null
float32
4.46.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,172
JaehoHan/my-clip-model-base5
clip
["CLIPModel"]
null
float32
4.46.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,173
JaehoHan/my-clip-model2
clip
["CLIPModel"]
null
float32
4.46.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,174
JaehoHan/my-clip-model3
clip
["CLIPModel"]
null
float32
4.46.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,175
JaehoHan/my-clip-model3-flax
clip
["CLIPModel"]
null
float32
4.46.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,176
JaehoHan/my-clip-model3-tf
clip
["CLIPModel"]
null
float32
4.46.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
16,177
Jaehun/PrismNLI-0.4B
deberta-v2
["DebertaV2ForSequenceClassification"]
128100
float16
4.48.3
1,024
4,096
24
16
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
16,178
Jaehun/amber-universe-1
t5
["T5ForConditionalGeneration"]
32128
float32
4.26.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
16,179
Jaehun/faithful_model
t5
["T5ForConditionalGeneration"]
32128
float32
4.31.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
16,180
Jaehun/glad-donkey-11
t5
["T5ForConditionalGeneration"]
32128
float32
4.26.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
16,181
Jaehun/icy-blaze-24
t5
["T5ForConditionalGeneration"]
32128
float32
4.26.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
16,182
Jaehun/silvery-dream-13
t5
["T5ForConditionalGeneration"]
32128
float32
4.26.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
16,183
Jaehun/sunny-bush-17-epoch-1
pegasus
["PegasusForConditionalGeneration"]
96103
float32
4.31.0
null
null
16
null
null
null
0.1
True
1,024
null
null
null
0
1
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
16,184
JaehwiJeon/videomae-base-finetuned-ucf101-subset
videomae
["VideoMAEForVideoClassification"]
null
float32
4.36.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
16,185
JaehyeokLee/20m_em_checkpoint_epoch_1_step_4200
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
16,186
JaehyeokLee/20m_em_checkpoint_epoch_1_step_4240
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
16,187
JaehyeokLee/20m_em_checkpoint_epoch_1_step_4280
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
16,188
JaehyeokLee/20m_em_checkpoint_epoch_1_step_4320
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
16,189
JaehyeokLee/20m_em_checkpoint_epoch_1_step_4360
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
16,190
JaehyeokLee/20m_em_checkpoint_epoch_1_step_4400
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
16,191
JaehyeokLee/20m_em_checkpoint_epoch_1_step_4440
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
16,192
JaehyeokLee/20m_em_checkpoint_epoch_1_step_4480
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
16,193
JaehyeokLee/20m_em_checkpoint_epoch_1_step_4520
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
16,194
JaehyeokLee/20m_em_checkpoint_epoch_1_step_4600
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
16,195
JaehyeokLee/20m_em_checkpoint_epoch_1_step_4640
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
16,196
JaehyeokLee/20m_em_checkpoint_epoch_1_step_4680
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
16,197
JaehyeokLee/20m_em_checkpoint_epoch_1_step_4800
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
16,198
JaehyeokLee/preliminary_5m_checkpoint_epoch_1_step_280
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
16,199
JaehyeokLee/preliminary_one_source_gist_checkpoint_epoch_1_step_10
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.48.3
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long