index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
4,400
AsemBadr/whisper-small
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.40.0.dev0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
4,401
Asfesalas/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,402
Asgar1993/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-bellowing_yapping_goose
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.51.1
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
4,403
Asgar1993/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-huge_dappled_albatross
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.51.1
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
4,404
Ash24244/Newai
gpt2
["GPT2LMHeadModel"]
50257
float16
4.47.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
4,405
Ash24244/UltimateAI
gpt2
["GPT2LMHeadModel"]
50257
float16
4.47.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
4,406
Ash2749/Qwe3_1.7B_Emobench
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
2,048
6,144
28
16
8
silu
0
True
40,960
1,000,000
0.000001
0.02
null
151645
true
128
2
false
null
null
false
false
false
false
false
false
false
false
true
true
1.409286
medium
very_long
4,407
Ash2749/Qwe3_1.7B_acot
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
2,048
6,144
28
16
8
silu
0
True
40,960
1,000,000
0.000001
0.02
null
151645
true
128
2
false
null
null
false
false
false
false
false
false
false
false
true
true
1.409286
medium
very_long
4,408
Ash2749/Qwen3_0.6B_acot
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
1,024
3,072
28
16
8
silu
0
True
40,960
1,000,000
0.000001
0.02
null
151645
true
64
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.352322
small
very_long
4,409
Ash2749/Qwen3_0.6B_emobench
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
1,024
3,072
28
16
8
silu
0
True
40,960
1,000,000
0.000001
0.02
null
151645
true
64
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.352322
small
very_long
4,410
Ash2749/Qwen3_1.7B_acot_emobench
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
2,048
6,144
28
16
8
silu
0
True
40,960
1,000,000
0.000001
0.02
null
151645
true
128
2
false
null
null
false
false
false
false
false
false
false
false
true
true
1.409286
medium
very_long
4,411
Ash2749/Qwen3_14B_acot
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
5,120
17,408
40
40
8
silu
0
True
40,960
1,000,000
0.000001
0.02
null
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
12.582912
large
very_long
4,412
Ash2749/finetuned-llama-3.2
llama
["LlamaForCausalLM"]
128256
float16
4.51.3
3,072
8,192
28
24
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128009
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,413
Ash2749/v1_llama3.2_3b_instruct
llama
["LlamaForCausalLM"]
128256
float16
4.51.3
3,072
8,192
28
24
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128009
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
4,414
Asha-12502/phase_1.1
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.32.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,415
AshanGimhana/llama2-chat-ashan-GPT
gpt2
["GPT2LMHeadModel"]
50257
null
4.35.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
4,416
AshanGimhana/llama2-chat-ashan-QQ
bert
["BertForQuestionAnswering"]
30522
null
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,417
Ashegh-Sad-Warrior/dummy-model
camembert
["CamembertForMaskedLM"]
32005
float32
4.40.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
5
6
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,418
Ashegh-Sad-Warrior/dummy-model_practice
camembert
["CamembertForMaskedLM"]
32005
float32
4.40.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
5
6
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,419
Ashen2020/my-finetuned-bert
bert
["BertModel"]
28996
float32
4.38.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,420
AshenR/ASPPrediction
bert
["BertForMaskedLM"]
30522
null
null
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,421
Ashfaq-06/tinystarcoder-rlhf-model
gpt_bigcode
["GPTBigCodeForCausalLM"]
49152
float32
4.34.0
null
null
null
null
null
null
null
True
null
null
null
0.02
0
0
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
4,422
Ashima/bart_large_keyword_augmented_wp
bart
["BartForConditionalGeneration"]
50265
float32
4.20.1
null
null
12
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
4,423
Ashima/bart_large_keyword_wp
bart
["BartForConditionalGeneration"]
50265
float32
4.20.1
null
null
12
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
4,424
AshiqaSameem/llama_3_lecture_summarizer_model
llama
["LlamaForCausalLM"]
128256
bfloat16
4.41.0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,425
Ashish013/model_out_100
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,426
Ashish24032003/speecht5_finetuned__nl
speecht5
["SpeechT5ForTextToSpeech"]
81
float32
4.48.0.dev0
768
null
null
null
null
gelu
0.1
False
null
null
null
0.02
0
2
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,427
Ashkchamp/Pokemon-Image-Captioning
blip
["BlipForConditionalGeneration"]
null
float32
4.47.0
null
null
null
null
null
null
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,428
Ashkh0099/fine-tune-BERT_1
bert
["BertForQuestionAnswering"]
30522
float32
4.52.4
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,429
Ashkh0099/my-bert-QA-fix
bert
["BertForQuestionAnswering"]
30522
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,430
Ashkh0099/my-bert-new-ashkh
bert
["BertForQuestionAnswering"]
30522
float32
4.47.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,431
Ashkh0099/my-bert-profile-QA
bert
["BertForQuestionAnswering"]
30522
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,432
Ashley37sky/cs229s_pruned
llama
[]
128256
null
4.46.2
2,048
8,192
12
16
null
silu
null
True
2,048
null
0.00001
0.02
1
2
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
0.60398
small
medium
4,433
Ashmal/ClimateGPT13B_en
llama
["LlamaForCausalLM"]
32000
bfloat16
4.28.0.dev0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
4,434
Ashmal/Code-t5-finetuned
t5
["T5ForConditionalGeneration"]
32100
float32
4.40.0.dev0
null
null
null
null
null
null
null
True
null
null
null
null
1
2
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,435
Ashmi/my_awesome_dataset_model
t5
["T5ForConditionalGeneration"]
32128
null
4.30.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,436
Ashmit06/distilbert-medical-question_answer-finetuned-squad
distilbert
["DistilBertForQuestionAnswering"]
30522
float32
4.41.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
4,437
Ashmit06/roberta-base-squad2-finetuned-squad
roberta
["RobertaForQuestionAnswering"]
50265
float32
4.41.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,438
Ashraf-kasem/custom_gpt2_frames_text
gpt2
["GPT2LMHeadModel"]
10000
null
4.25.1
null
null
null
null
null
null
null
True
null
null
null
0.02
0
0
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
4,439
Ashraf-kasem/custom_gpt2_frames_text_continue
gpt2
["GPT2LMHeadModel"]
10000
null
4.25.1
null
null
null
null
null
null
null
True
null
null
null
0.02
0
0
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
4,440
Ashraf-kasem/custom_gpt2_frames_text_original_tokenizer
gpt2
["GPT2LMHeadModel"]
50257
null
4.25.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
4,441
Ashraf-kasem/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,442
Ashreen/legal-t5-large-chunked
t5
["T5ForConditionalGeneration"]
32128
float32
4.35.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,443
Ashreen/legal-t5-small-chunked
t5
["T5ForConditionalGeneration"]
32128
float32
4.35.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,444
AshtonIsNotHere/bert-feqa-qa
bert
["BertForQuestionAnswering"]
30522
float32
4.34.0
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
4,445
AshtonLKY/CE_v1.1
bert
["BertForTokenClassification"]
30522
float32
4.39.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,446
Ashu15/dinov2-base-finetuned-oxford
dinov2
["Dinov2ForImageClassification"]
null
float32
4.50.2
768
null
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
4,447
Ashufet/detectAI_LLAMA2_13B_3kdata
llama
["LlamaForSequenceClassification"]
32032
bfloat16
4.35.0
5,120
13,824
40
40
40
silu
null
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
4,448
Ashufet/example_repo_detectAI
llama
["LlamaForSequenceClassification"]
32032
bfloat16
4.35.0
5,120
13,824
40
40
40
silu
null
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
4,449
Ashuku/pegasus-samsum
pegasus
["PegasusForConditionalGeneration"]
96103
float32
4.31.0
null
null
16
null
null
null
0.1
True
1,024
null
null
null
0
1
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
4,450
AshutoshD245/food_classifier
vit
["ViTForImageClassification"]
null
null
4.32.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
4,451
Ashuxd-X/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-skilled_amphibious_stingray
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.51.3
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
null
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
4,452
Ashwin-s-n/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,453
Ashwin-s-n/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,454
Ashwin0/mt5-small-finetuned-amazon-en-es
mt5
["MT5ForConditionalGeneration"]
250112
null
4.26.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,455
Ashwini1412/wav2vec2-nepali
wav2vec2
["Wav2Vec2ForCTC"]
65
float32
4.35.2
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
4,456
Ashwini1412/wav2vec2-nepali-final
wav2vec2
["Wav2Vec2ForCTC"]
65
float32
4.38.2
1,024
4,096
24
16
null
gelu
0.12
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
4,457
Ashwini1412/wav2vec2-nepali-itr-2
wav2vec2
["Wav2Vec2ForCTC"]
65
float32
4.38.1
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
4,458
Ashwini1412/wav2vec2-nepali-itr-3
wav2vec2
["Wav2Vec2ForCTC"]
65
float32
4.38.1
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
4,459
Ashwini1412/wav2vec2-nepali-itr-4
wav2vec2
["Wav2Vec2ForCTC"]
65
float32
4.38.1
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
4,460
Ashwini1412/wav2vec2-nepali-itr-5
wav2vec2
["Wav2Vec2ForCTC"]
65
float32
4.38.1
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
4,461
AshwiniC28/gita-text-generation-gpt2
gpt2
["GPT2LMHeadModel"]
50257
float32
4.48.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
4,462
AshwithPoojary/DeepSeek-llama-finetune-storyteller
llama
["LlamaForCausalLM"]
128256
float16
4.48.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
4,463
Asib1/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-pensive_leggy_ant
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.51.3
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
4,464
Asif-Sheriff/Bert_for_QA_ft_32
bert
["BertForQuestionAnswering"]
30522
float32
4.48.3
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
4,465
Asif-Sheriff/QAC2
t5
["T5ForConditionalGeneration"]
32128
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,466
Asif-Sheriff/QGC2
t5
["T5ForConditionalGeneration"]
32128
float32
4.50.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
4,467
AsifAbrar6/bert-base-multilingual-cased-finetuned-RQA-new-2
bert
["BertForQuestionAnswering"]
119547
float32
4.41.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,468
AsifAbrar6/mbert-bengali-tydiqa-qa-finetuned-RQA-factoid
bert
["BertForQuestionAnswering"]
105879
float32
4.41.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,469
AsifAbrar6/xlm-roberta-base-finetuned-squadBN-512
xlm-roberta
["XLMRobertaForQuestionAnswering"]
250002
float32
4.41.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
4,470
Asim037/wav2vec2-sttt
wav2vec2
["Wav2Vec2ForCTC"]
32
float32
4.41.2
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
4,471
Asimok/KGLQA-KnowledgeBank-NCR-and-CCLUE
llama
["LlamaForCausalLM"]
79458
float16
4.33.3
4,096
11,008
32
32
32
silu
null
True
2,048
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
4,472
Asimok/KGLQA-KnowledgeBank-NCR-and-CCLUE-Alpaca
llama
["LlamaForCausalLM"]
55296
float16
4.33.3
4,096
11,008
32
32
32
silu
null
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
4,473
Asimok/KGLQA-KnowledgeBank-RACE-Alpaca
llama
["LlamaForCausalLM"]
32000
float16
4.33.3
4,096
11,008
32
32
32
silu
null
True
2,048
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
4,474
Asit03/AI_Agent_V2
llama
["LlamaForCausalLM"]
32000
bfloat16
4.38.2
4,096
11,008
32
32
null
silu
null
True
4,096
10,000
0.000001
0.02
1
2
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
4,475
Asit03/ChatSeek-8B-v2
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
4,476
AskUI/PTA-1
florence2
["Florence2ForConditionalGeneration"]
51289
float32
4.44.2
null
null
null
null
null
null
null
null
null
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,477
Askinkaty/llama-finance-relations
llama
["LlamaForCausalLM"]
128257
float32
4.47.1
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
4,478
Asma50AA/trail-with-517-1000-whisper-small-ar
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.38.2
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
4,479
Asma50AA/whisper-small-ar
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.35.2
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
4,480
AsmaFaraj95/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
4,481
Asmedeus/llama3_8b-hf
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,482
Aspect05/Qwen-1.5-Distil-FP16-Updated
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.49.0
1,536
8,960
28
12
2
silu
0
True
131,072
10,000
0.000001
0.02
151643
151643
false
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
4,483
Aspik101/WizardVicuna-Uncensored-3B-instruct-PL-lora_GPTQ
llama
["LlamaForCausalLM"]
32000
float16
4.32.0.dev0
3,200
8,640
26
32
32
silu
null
True
2,048
null
0.000001
0.02
1
2
false
100
1
false
null
null
true
false
false
false
false
false
false
false
false
true
3.19488
medium
medium
4,484
Aspik101/WizardVicuna-Uncensored-3B-instruct-PL-lora_unload
llama
["LlamaForCausalLM"]
32000
float16
4.32.0.dev0
3,200
8,640
26
32
32
silu
null
True
2,048
null
0.000001
0.02
1
2
false
100
1
false
null
null
true
false
false
false
false
false
false
false
false
true
3.19488
medium
medium
4,485
Aspik101/distil-whisper-large-v3-pl
whisper
["WhisperForConditionalGeneration"]
51866
float32
4.35.2
null
null
32
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
4,486
Aspik101/llama-30b-2048-instruct-PL-lora_unload
llama
["LlamaForCausalLM"]
32000
float16
4.32.0.dev0
6,656
17,920
60
52
52
silu
null
False
2,048
null
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
31.897682
large
medium
4,487
Aspik101/llama100ftt
llama
["LlamaForCausalLM"]
128256
bfloat16
4.33.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
100257
100257
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,488
Aspik101/llama10ftt
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
100257
100257
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,489
Aspik101/llama1111
llama
["LlamaForCausalLM"]
128256
bfloat16
4.33.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
100257
100257
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,490
Aspik101/llama13
llama
["LlamaForCausalLM"]
128256
bfloat16
4.33.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
100257
100257
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,491
Aspik101/llama1414
llama
["LlamaForCausalLM"]
128256
bfloat16
4.33.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
100257
100257
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,492
Aspik101/llamaft5555
llama
["LlamaForCausalLM"]
128256
bfloat16
4.33.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
100257
100257
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,493
Aspik101/llamaft55555
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
100257
100257
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,494
Aspik101/llamaft6v22
llama
["LlamaForCausalLM"]
128256
bfloat16
4.33.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
100257
100257
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,495
Aspik101/llamaft6v222
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
100257
100257
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
4,496
Aspik101/medium-poleval-le15
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.45.2
null
null
24
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
4,497
Aspik101/medium-poleval-le3
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.45.2
null
null
24
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
4,498
Aspik101/mixtra_axa14
llama
["LlamaForCausalLM"]
32000
bfloat16
4.38.2
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
4,499
Aspik101/ner_FT-checkpoint-1000debert
deberta-v2
["DebertaV2ForTokenClassification"]
128100
float32
4.37.1
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short