index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
3,100
Alphatao/f6b40af7-b80f-4aab-843b-5c70dfe6d76b
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,101
Alphatao/f6c40df3-7b35-4ceb-8f3f-bf2ccb2c2363
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
131,072
1,000,000
0.000001
0.02
null
151643
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,102
Alphatao/f790e96a-4ceb-4271-ac22-5165f95329f6
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,103
Alphatao/f7969454-b44b-4f36-bf99-263269bbfe33
llama
["LlamaForCausalLM"]
128256
bfloat16
4.51.3
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,104
Alphatao/f9747c8b-e31f-45d0-930b-ebd4238e2224
llama
["LlamaForCausalLM"]
32064
bfloat16
4.46.0
3,072
8,192
32
32
32
silu
0
False
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
true
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
3,105
Alphatao/fa44913a-10a1-496e-81fd-474e9c7d6bc1
qwen2_moe
["Qwen2MoeForCausalLM"]
151936
float32
4.46.0
32
22
4
4
2
silu
0
False
1,024
1,000,000
0.000001
0.02
null
151643
false
8
2
false
null
4
false
false
false
false
false
false
false
false
true
true
0.000049
small
short
3,106
Alphatao/fee90ede-6fd9-42f3-bbb1-fd1fab03804b
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.46.0
3,072
8,192
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
medium
3,107
Alphonsce/ViTuned_buildings
vit
["ViTForImageClassification"]
null
float32
4.38.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
3,108
Alpi157/Final_conversational_model
gpt2
["GPT2LMHeadModel"]
50257
float32
4.32.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,109
AlpineHealth/test1
t5
["T5ForConditionalGeneration"]
32128
float32
4.34.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
3,110
Alprocco/0.65MBertopic
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,111
Alprocco/Swiss_bertopic
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,112
Alqayed2024/gpt2-finetuned-wikitext2-re2al
gpt2
["GPT2LMHeadModel"]
50257
float32
4.43.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,113
Alqayed2024/gpt2-finetuned-wikitext2-real
gpt2
["GPT2LMHeadModel"]
50257
float32
4.43.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,114
Alqayed2024/gpt2-large-finetuned-wikitext2-re2al
gpt2
["GPT2LMHeadModel"]
50257
float32
4.43.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,115
Alqayed2024/gpt2-medium-finetuned-real2
gpt2
["GPT2LMHeadModel"]
50257
float32
4.44.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,116
Alqayed2024/gpt2-xl-finetuned-real3
gpt2
["GPT2LMHeadModel"]
50257
float32
4.44.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,117
Alr2000/whisper-small-hi
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.33.0.dev0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,118
Alsaadiask/bert-base-uncased-finetuned-emotion
bert
["BertForSequenceClassification"]
30522
float32
4.25.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,119
Alsaadiask/distilbert-base-uncased-finetuned-emotion
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.25.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
3,120
Alsebay/SaoRPM-2x8B
mixtral
["MixtralForCausalLM"]
128256
bfloat16
4.42.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
2
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
long
3,121
Alsman68/whisper-medium-v2-capstone
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.48.2
null
null
24
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,122
Alt4nsuh/bart-qg2-10000-finetuned
bart
["BartForConditionalGeneration"]
50265
float32
4.49.0
null
null
6
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
3,123
Alt4nsuh/bart-qg2-30000-finetuned
bart
["BartForConditionalGeneration"]
50265
float32
4.49.0
null
null
6
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
3,124
Alt4nsuh/long_promt-finetuned-sec
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.50.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
3,125
AltEinstein/emc13
llama
["LlamaForCausalLM"]
128256
float16
4.47.1
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,126
AltEinstein/emc14
llama
["LlamaForCausalLM"]
128256
float16
4.47.1
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,127
AltEinstein/emc15
llama
["LlamaForCausalLM"]
128256
float16
4.47.1
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,128
AltEinstein/emc19
llama
["LlamaForCausalLM"]
128256
float16
4.47.1
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,129
AltEinstein/emc3
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,130
AltEinstein/emc5
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,131
AltEinstein/emc7
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,132
Altair626/bert-base-japanese-v3-jsts
bert
["BertForSequenceClassification"]
32768
float32
4.51.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,133
Altair626/bert-base-japanese-v3-output_jcommonsenseqa
bert
["BertForMultipleChoice"]
32768
float32
4.51.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,134
Altino/google-play-sentiment-analysis-workshop
bert
["BertForSequenceClassification"]
29794
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,135
Altino/google-play-sentiment-analysis-workshop-2
bert
["BertForSequenceClassification"]
29794
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,136
Altoqumulus/gita-text-generation-gpt2
gpt2
["GPT2LMHeadModel"]
50257
float32
4.47.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,137
AlumiK/LingLong-317M-Chat
linglong
["LingLongForCausalLM"]
13312
float32
4.40.2
null
null
null
null
null
null
null
True
null
null
null
0.02
10
8
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,138
Alvenir/bert-punct-restoration-de
bert
["BertForTokenClassification"]
31102
float32
4.11.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,139
Alvenir/bert-punct-restoration-en
bert
["BertForTokenClassification"]
30522
null
4.6.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,140
Alvenir/wav2vec2-base-da-ft-nst
wav2vec2
["Wav2Vec2ForCTC"]
36
float32
4.16.0.dev0
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
3,141
Alvi12/idefics-9b-doodles
idefics
["IdeficsForVisionText2Text"]
32000
float16
4.44.2
4,096
11,008
32
32
null
silu
null
True
2,048
null
0.000001
0.02
1
2
false
128
null
false
null
null
false
false
false
false
false
false
false
false
false
false
6.442451
medium
medium
3,142
AlvianKhairi/Llama-2-7b-chat-finetune-25k
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,143
AlvianKhairi/Scicite_classification_model
bert
["BertForSequenceClassification"]
31090
float32
4.34.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,144
Alvin-LiuJia/DeepSeek-R1-Medical-COT-Trained-Alvin0614
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.51.3
1,536
8,960
28
12
2
silu
0
True
131,072
10,000
0.000001
0.02
151646
151643
false
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,145
Alvin12345/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,146
AlvinY34/Llama-3.2-3B-Instruct-8b-test
llama
["LlamaModel"]
128256
bfloat16
4.49.0
3,072
8,192
28
24
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
3,147
AlvinxLukilah/TextGenerationProject
gpt2
["GPT2LMHeadModel"]
50257
float32
4.34.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,148
AlvinxLukilah/TextGenerationV2
gpt2
["GPT2LMHeadModel"]
50257
float32
4.34.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,149
Alwahsh/llama-3-8b-0-base-quantized
llama
["LlamaForCausalLM"]
128256
bfloat16
4.41.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,150
Alwahsh/llama-3-8b-70-base-quantized
llama
["LlamaForCausalLM"]
128256
float16
4.41.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,151
AlxWo/DeepRLCourse
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,152
Alxis955/qwen25-adv-lora
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.50.3
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
3,153
AlyBendo/DPO-DoubleQuant-MCQA-TrueMixin
t5
["T5ForConditionalGeneration"]
32128
float16
4.40.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
3,154
Alyia/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,155
Alyx-us/fs_model
llama
["LlamaForCausalLM"]
128256
float16
4.51.3
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,156
Alziano/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,157
Am22000/classifier_image
vit
["ViTForImageClassification"]
null
float32
4.20.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
3,158
Am22000/food_classifier
vit
["ViTForImageClassification"]
null
float32
4.35.0
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
3,159
AmaanDhamaskar/mt5-base-xlmsum-mr-e15
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.51.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
3,160
AmaanDhamaskar/pseudonym_generator_mt5_test
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.48.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
3,161
AmaanUsmani/opt-125m-gptq-4bit
opt
["OPTForCausalLM"]
50272
float16
4.40.2
768
null
12
12
null
null
0
True
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
medium
3,162
AmaanUsmani/test-config
falcon
[]
65024
null
4.40.2
4,544
null
2
71
null
null
0
True
2,048
500,000
null
0.02
11
11
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
true
0.49555
small
medium
3,163
Amaboh/finetuning-sentiment-model-3000-samples
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.33.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
3,164
AmaiaSolaun/MT_authorship_att
bert
["BertForSequenceClassification"]
50099
float32
4.40.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,165
Amal17/NusaBERT-concate-BiGRU-NusaTranslate-senti
bert-bigru
["BERTBiGRUClassifier"]
null
float32
4.48.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
null
3,166
Amal17/NusaBERT-concate-BiGRU-NusaX-ace
bert-bigru
["BERTBiGRUClassifier"]
null
float32
4.48.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
null
3,167
Amal17/NusaBERT-concate-BiLSTM-NusaTranslate-senti
bert-bilstm
["BERTBiLSTMClassifier"]
null
float32
4.48.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
null
3,168
Amal17/NusaBERT-concate-BiLSTM-NusaX-ace
bert-bilstm
["BERTBiLSTMClassifier"]
null
float32
4.48.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
null
3,169
Amal17/distilgpt2-finetuned-wikitext2
gpt2
["GPT2LMHeadModel"]
50257
float32
4.32.0.dev0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,170
Amal17/experiment-qa
distilbert
["DistilBertForQuestionAnswering"]
30522
float32
4.31.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
3,171
Amal17/wikipedia-20230601.ace
gpt2
["GPT2LMHeadModel"]
50257
float32
4.32.0.dev0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,172
Amal98/autotrain-final_model-3026786824
bert
["BertForTokenClassification"]
30522
float32
4.25.1
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,173
AmalAK/Llama-2-7b-chat-finetune-for-law
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,174
AmalNlal/my_awesome_eli5_mlm_model
roberta
["RobertaForMaskedLM"]
50265
float32
4.36.2
768
3,072
6
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.042467
small
short
3,175
Amala02/seasonSales
gpt2
["GPT2LMHeadModel"]
50257
float32
4.42.4
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,176
Amalq/mental_classification
roberta
["RobertaForSequenceClassification"]
50265
float32
4.16.2
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,177
Amalq/roberta-large-schizophrenia-v12
roberta
["RobertaForMaskedLM"]
50265
float32
4.28.1
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,178
Amalq/roberta-large-schizophrenia-v2
roberta
["RobertaForMaskedLM"]
50265
float32
4.28.1
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,179
Amalq/schizophrenia-roberta-large
roberta
["RobertaForMaskedLM"]
50265
float32
4.16.2
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,180
Aman0112/bert_emo_classifier
bert
["BertForSequenceClassification"]
30522
float32
4.28.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,181
Aman6917/autotrain-big_tm4-3021286705
t5
["T5ForConditionalGeneration"]
32102
float32
4.25.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
3,182
Aman6917/autotrain-fine_tune_table_tm2-2695480537
t5
["T5ForConditionalGeneration"]
32102
float32
4.25.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
3,183
Aman6917/autotrain-tm3_model-2711480628
t5
["T5ForConditionalGeneration"]
32102
float32
4.25.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
3,184
Aman6917/autotrain-tm3_model-2711480629
t5
["T5ForConditionalGeneration"]
32102
float32
4.25.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
3,185
Aman6917/autotrain-tm3_model-2711480631
t5
["T5ForConditionalGeneration"]
32102
float32
4.25.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
3,186
Aman6917/autotrain-tm4_2_big-3033986980
t5
["T5ForConditionalGeneration"]
32102
float32
4.25.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
3,187
Aman8252/idefics-9b-doodles
idefics
["IdeficsForVisionText2Text"]
32000
float16
4.44.0
4,096
11,008
32
32
null
silu
null
True
2,048
null
0.000001
0.02
1
2
false
128
null
false
null
null
false
false
false
false
false
false
false
false
false
false
6.442451
medium
medium
3,188
AmanMussa/llama2-fake-news-detection-7b
llama
["LlamaForCausalLM"]
32000
float16
4.36.0.dev0
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,189
Amanaccessassist/Llama-3-8B-AWQ-4Bit
llama
["LlamaForCausalLM"]
128288
float16
4.40.1
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128003
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,190
Amanda/bge_portuguese_v6
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.43.4
1,024
4,096
24
16
null
gelu
null
True
8,194
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
3,191
Amanda0531/uuu_fine_tune_gpt2
gpt2
["GPT2LMHeadModel"]
21128
float32
4.50.0
null
null
null
null
null
null
null
True
null
null
null
0.02
101
102
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,192
Amanda0531/uuu_fine_tune_taipower
gpt2
["GPT2LMHeadModel"]
21128
float32
4.50.0
null
null
null
null
null
null
null
True
null
null
null
0.02
101
102
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,193
Amani27/custom_ft_weakdap_r1_wthout_cleaning_synhi_synes_squad
xlm-roberta
["XLMRobertaForQuestionAnswering"]
250002
float32
4.33.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,194
Amani27/custom_ft_weakdap_r2_wthout_cleaning_synhi_synes_squad
xlm-roberta
["XLMRobertaForQuestionAnswering"]
250002
float32
4.33.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,195
Amani27/custom_ft_weakdap_r2_wthout_cleaning_synhi_synes_squad_v2
xlm-roberta
["XLMRobertaForQuestionAnswering"]
250002
float32
4.33.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,196
Amani27/custom_ft_weakdap_r3_wthout_cleaning_synhi_synes_squad
xlm-roberta
["XLMRobertaForQuestionAnswering"]
250002
float32
4.33.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,197
Amani27/custom_ft_weakdap_r3_wthout_cleaning_synhi_synes_squad_loss_tweak
xlm-roberta
["XLMRobertaForQuestionAnswering"]
250002
float32
4.33.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,198
Amankankriya/ppo-CarRacing-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,199
Amaresh-ds/bert-telecom-verbatim-classifier
bert
["BertForSequenceClassification"]
30522
float32
4.51.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short