index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
2,200
Aivesa/349aacef-3660-4206-b561-6d7b5d5a9b62
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.47.1
2,048
16,384
18
8
1
gelu
0
False
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
2,201
Aivesa/45d81bfd-b7dd-4169-b979-bdc0739428a0
llama
["LlamaForCausalLM"]
128256
float16
4.47.1
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,202
Aivesa/4dc6507a-6049-4f97-a49d-9c4d38d41625
gpt_neox
["GPTNeoXForCausalLM"]
50304
float16
4.47.1
1,024
4,096
24
16
null
gelu
0
False
2,048
10,000
null
0.02
0
0
false
64
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.30199
small
medium
2,203
Aivesa/514decb9-bd0e-4782-86b5-6131d662a51b
llama
["LlamaForCausalLM"]
32000
float32
4.47.1
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0
2
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
2,204
Aivesa/53638570-6e01-4257-8d08-795d436a977c
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
2,205
Aivesa/5e50f4e8-6556-4450-9721-0660047938a0
llama
["LlamaForCausalLM"]
32000
float16
4.47.1
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0
2
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
2,206
Aivesa/85a3b62a-09c1-44d6-af05-0f30aee76da5
mixtral
["MixtralForCausalLM"]
32000
float32
4.47.1
1,024
3,584
2
32
8
silu
0
False
131,072
1,000,000
0.00001
0.02
1
2
false
32
4
false
8
2
false
false
false
false
false
false
false
true
true
true
0.025166
small
very_long
2,207
Aivesa/9cc26248-48d3-4b39-8479-e5d69049c582
llama
["LlamaForCausalLM"]
32000
bfloat16
4.47.1
2,048
5,632
22
32
4
silu
0
False
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
2,208
Aivesa/a05b476c-463b-4946-89c4-d93aca81d7e7
opt
["OPTForCausalLM"]
50272
float16
4.47.1
2,048
null
24
32
null
null
0
False
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
1.20796
medium
medium
2,209
Aivesa/ad5f3025-5478-4dad-b670-bbd401699c03
llama
["LlamaForCausalLM"]
32000
float32
4.47.1
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0
2
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
2,210
Aivesa/b87a98bd-2eb0-4bfa-8f4e-6c92644bdc42
codegen
["CodeGenForCausalLM"]
51200
float32
4.47.1
null
null
null
null
null
null
null
False
null
null
null
0.02
50256
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,211
Aivesa/c1093e50-e6e8-4bfe-8dc8-afe74f920cfd
gpt_neox
["GPTNeoXForCausalLM"]
50304
float16
4.47.1
1,024
4,096
24
16
null
gelu
0
False
2,048
10,000
null
0.02
0
0
false
64
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.30199
small
medium
2,212
Aivesa/d04a0ea9-5c92-49fd-8fd8-300b97476de2
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.1
2,048
5,504
24
16
16
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
very_long
2,213
Aivesa/d8b02868-3f22-4fc4-bf4b-cd5c2a0ad840
llama
["LlamaForCausalLM"]
32000
float16
4.47.1
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0
2
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
2,214
Aivesa/d9ca8632-4ba2-424e-a050-6fb0f4780371
mixtral
["MixtralForCausalLM"]
32000
float32
4.47.1
1,024
3,584
2
32
8
silu
0
False
131,072
1,000,000
0.00001
0.02
1
2
false
32
4
false
8
2
false
false
false
false
false
false
false
true
true
true
0.025166
small
very_long
2,215
Aivesa/df5493b9-6228-4cae-a15d-b3a2fd8fb6e3
gpt_neox
["GPTNeoXForCausalLM"]
50304
float16
4.47.1
512
2,048
6
8
null
gelu
0
False
2,048
10,000
null
0.02
0
0
false
64
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.018874
small
medium
2,216
Aivesa/f2a098be-735f-44f7-99ac-6937ddb1e108
llama
["LlamaForCausalLM"]
128256
float16
4.47.1
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
2,217
Aivesa/f8549911-86bb-4e2b-8478-525efbbc4415
gpt_neox
["GPTNeoXForCausalLM"]
50304
float16
4.47.1
512
2,048
6
8
null
gelu
0
False
2,048
10,000
null
0.02
0
0
false
64
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.018874
small
medium
2,218
Aixile/CLIP-ViT-H-14-laion2B-s32B-b79K
clip
["CLIPModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,219
Aixos/webbot
bert
["BertForQuestionAnswering"]
30522
float32
4.42.4
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,220
Ajas2002/cover
t5
["T5ForConditionalGeneration"]
32128
float32
4.38.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
2,221
Ajaxpax/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,222
Ajay2525/wav2vec2-ks
wav2vec2
["Wav2Vec2ForSequenceClassification"]
32
float32
4.44.0.dev0
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
2,223
AjayMukundS/Llama-2-7b-LTS-finetuned
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,224
Ajhl1992/lilt-en-funsd
lilt
["LiltForTokenClassification"]
50265
float32
4.28.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
2,225
Ajit1498/donut-base-sroie
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.41.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,226
Ajita15/distilgpt2-finetuned-wikitext2
gpt2
["GPT2LMHeadModel"]
50257
float32
4.40.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
2,227
Ajohe/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,228
Ak1104/3_test_q8
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
2,229
Ak1104/QA_2k
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
2,230
Ak1104/codellama-7b_rust
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
2,231
Ak1104/llama-3-test
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
2,232
Ak1104/prompt1
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
2,233
Ak128umar/bert-finetuned-ner-accelerate
bert
["BertForTokenClassification"]
28996
float32
4.52.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,234
Ak1R401/wiki-newmm
roberta
["RobertaForTokenClassification"]
97982
float32
4.30.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,235
Ak729/MT5_with_MLP
mt5
["MT5EncoderModel"]
250112
float32
4.44.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
2,236
Ak729/Model
bert
["BertForMaskedLM"]
30522
float32
4.41.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,237
AkRl1/Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,238
Akajackson/weights
bert
["BertForSequenceClassification"]
83828
float32
4.39.3
312
600
3
12
null
gelu
null
True
2,048
null
null
0.02
null
null
null
26
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.003504
small
medium
2,239
AkankshaLoria21/gemma-Code-Instruct-Finetune-test
gemma
["GemmaForCausalLM"]
256000
float16
4.38.0
2,048
16,384
18
8
1
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
2,240
Akash2106/TRAIL
mixformer-sequential
["MixFormerSequentialForCausalLM"]
51200
float32
4.34.1
null
null
null
null
null
null
null
null
null
null
null
0.02
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,241
Akash2106/phi1_5_sqldatabase
mixformer-sequential
["MixFormerSequentialForCausalLM"]
51200
float32
4.34.1
null
null
null
null
null
null
null
null
null
null
null
0.02
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,242
Akash24/my_awesome_model
albert
["AlbertForSequenceClassification"]
30000
float32
4.34.1
768
3,072
12
12
null
gelu_new
null
null
512
null
null
0.02
2
3
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,243
Akash82228/results
bart
["BartForConditionalGeneration"]
50265
float32
4.35.0
null
null
6
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
2,244
Akash82228/tmp_trainer
roberta
["RobertaForQuestionAnswering"]
50265
float32
4.36.0.dev0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,245
Akbar0627/dummy-model
camembert
["CamembertForMaskedLM"]
32005
float32
4.49.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
5
6
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,246
Akchunks/MLAgents-Pyramids
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,247
Akchunks/ppo-SnowballTarget
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,248
Akeda01/MontirOnlinePro
llama
["LlamaForCausalLM"]
128258
float16
4.48.0
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
2,249
Akenar/speecht5_finetuned_voxpopuli_nl
speecht5
["SpeechT5ForTextToSpeech"]
81
float32
4.40.2
768
null
null
null
null
gelu
0.1
False
null
null
null
0.02
0
2
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
2,250
Akerrules/swotmodelQA
llama
["LlamaForCausalLM"]
128256
float16
4.46.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,251
AkhiVasu/distilbert-base-uncased-finetuned-imdb
distilbert
["DistilBertForMaskedLM"]
30522
null
4.40.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,252
Akhil-Theerthala/Kuvera-14B-v0.1.0
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.52.4
5,120
13,824
48
40
8
silu
0
True
131,072
1,000,000
0.00001
0.02
151643
151643
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
15.099494
large
very_long
2,253
Akhil-Theerthala/Kuvera-8B-v0.1.0
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.51.2
4,096
12,288
36
32
8
silu
0
False
40,960
1,000,000
0.000001
0.02
151643
151645
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
2,254
Akhil0-o/saved_model_body
roberta
["RobertaForSequenceClassification"]
50265
float32
4.27.4
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,255
Akhil0-o/saved_model_links
roberta
["RobertaForSequenceClassification"]
50265
float32
4.28.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,256
Akhil06042002/model_model_super_model
bert
["BertForMultipleChoice"]
30522
float32
4.28.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,257
Akhil123/emotions_classifier
vit
["ViTForImageClassification"]
null
null
4.31.0
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
2,258
Akhil199797/emotion-distilbert-final
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.52.4
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,259
AkhilGhosh/llama-cnn-210k
llama
["LlamaForCausalLM"]
32000
bfloat16
4.29.0.dev0
4,096
11,008
32
32
null
silu
null
False
2,048
null
0.000001
0.02
0
1
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
6.442451
medium
medium
2,260
AkhilKashyap1998/gpt2-GPTQ
gpt2
["GPT2LMHeadModel"]
50257
float16
4.37.2
null
null
null
null
null
null
null
False
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
2,261
Akhilamyaka/SAWIT_Translation_Model
mbart
["MBartForConditionalGeneration"]
250054
float32
4.48.3
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
2,262
Akhilsplendid/T5-model
t5
["T5ForConditionalGeneration"]
32128
float32
4.32.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
2,263
Akhilsplendid/bart-1
bart
["BartForConditionalGeneration"]
50264
float32
4.29.2
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
2,264
Akhilsplendid/bart-model
bart
["BartForConditionalGeneration"]
50264
float32
4.32.0
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
2,265
Akhilsplendid/pegasus-1
pegasus
["PegasusForConditionalGeneration"]
96103
float32
4.29.1
null
null
16
null
null
null
0.1
True
1,024
null
null
null
0
1
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
2,266
Akhilsplendid/pegasus-model
pegasus
["PegasusForConditionalGeneration"]
96103
float32
4.32.0
null
null
16
null
null
null
0.1
True
1,024
null
null
null
0
1
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
2,267
AkiiraBino/rubert-tiny2-russian-sentiment-reviews
bert
["BertForSequenceClassification"]
83828
float32
4.44.2
312
600
3
12
null
gelu
null
True
2,048
null
null
0.02
null
null
null
26
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.003504
small
medium
2,268
Akila/Mistral-of-Realms-7b-Instruct
mistral
["MistralForCausalLM"]
32000
float16
4.40.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,269
Akila/Mistral-of-Realms-7b-gguf
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
2,270
AkilaYoga/Animal_Classy
vit
["ViTForImageClassification"]
null
float32
4.41.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
2,271
AkilaYoga/identify_countries
vit
["ViTForImageClassification"]
null
float32
4.41.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
2,272
Akina21336/distilbert-base-uncased-finetuned-emotion
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.46.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,273
Akira812/nllb-200-distilled-1.3B-ct2-int8
m2m_100
["M2M100ForConditionalGeneration"]
256206
float32
4.21.0.dev0
null
null
24
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
2,274
Akirastroworld/td8
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.48.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,275
Akirpap/llama-7b-qlora-sutd-qa
llama
["LlamaForCausalLM"]
32000
float16
4.35.2
4,096
11,008
32
32
32
silu
null
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,276
AkitoP/whisper-large-v3-turbo-japanese-phone-voicevox-gen
whisper
["WhisperForConditionalGeneration"]
51866
float32
4.47.1
null
null
32
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
2,277
AkitooKun/DialoGPT-small-crybaby
gpt2
["GPT2LMHeadModel"]
50257
float32
4.44.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
2,278
AkkyMa/LunarLander
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,279
AkmalAshirmatov/first_try
wav2vec2
["Wav2Vec2ForCTC"]
34
float32
4.11.3
1,024
4,096
24
16
null
gelu
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
2,280
AksayKanthan/model
llama
["LlamaForCausalLM"]
128256
float16
4.41.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,281
Akseluhr/whisper-small-sv-SE-auhr-v2
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.26.0.dev0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
2,282
AksharaBalan/malayalam-bert-finetuned-ner
bert
["BertForTokenClassification"]
197285
float32
4.47.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,283
Akshay-Vasisht/Qwen2_VL_2B_FT_OCRLP
qwen2_vl
["Qwen2VLForConditionalGeneration"]
151936
bfloat16
4.48.2
1,536
8,960
28
12
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
2,284
Akshay0706/All-Plants-18-Epochs-Model
vit
["ViTForImageClassification"]
null
float32
4.35.0
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
2,285
Akshay1791/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,286
Akshay47/Llama-3.1-8B-Instruct_bvr_finetune_v3
llama
["LlamaForCausalLM"]
128256
float16
4.44.0
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,287
AkshaySalunke/model
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.38.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,288
Akshayxx/bert-base-cased-finetuned-cora
bert
["BertForSequenceClassification"]
28996
float32
4.38.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,289
Akshayxx/citeseer-finetuned-embeddings
bert
["BertModel"]
30522
float32
4.38.2
384
1,536
6
12
null
gelu
null
True
512
null
null
0.02
null
null
null
32
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.010617
small
short
2,290
Akshayxx/citeseer-finetuned-embeddingsV2
bert
["BertModel"]
30522
float32
4.38.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,291
AkshithaSai19/my-updated-model
albert
["AlbertModel"]
200000
float32
4.48.3
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
2
3
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,292
Aktsvigun/bart-large_aeslc_12345
bart
["BartForConditionalGeneration"]
50265
float32
4.21.3
null
null
12
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
2,293
Aktsvigun/bart-large_aeslc_23419
bart
["BartForConditionalGeneration"]
50265
float32
4.21.3
null
null
12
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
2,294
Aktsvigun/bart-large_aeslc_42
bart
["BartForConditionalGeneration"]
50265
float32
4.21.3
null
null
12
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
2,295
Aktsvigun/bart-large_aeslc_4837
bart
["BartForConditionalGeneration"]
50265
float32
4.21.3
null
null
12
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
2,296
Aktsvigun/bart-large_aeslc_705525
bart
["BartForConditionalGeneration"]
50265
float32
4.21.3
null
null
12
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
2,297
Akul/MiniLlama-1.8b-Math
llama
["LlamaForCausalLM"]
32000
float16
4.42.4
2,048
5,632
40
32
4
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
2.013266
medium
medium
2,298
Akul/MiniLlama-1.8b-Math-safetensors
llama
["LlamaForCausalLM"]
32000
float16
4.42.4
2,048
5,632
40
32
4
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
2.013266
medium
medium
2,299
AkylaiBva/my_medium_wspr
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.42.3
null
null
24
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null