index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
3,300
AmirlyPhd/V11-bert-text-classification-model
bert
["BertForSequenceClassification"]
30522
float32
4.39.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,301
AmirlyPhd/V4-bert-text-classification-model
bert
["BertForSequenceClassification"]
30522
float32
4.39.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,302
AmirlyPhd/cybercrime_LSTM_GLOVE
lstm
[]
null
null
null
128
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,303
AmirlyPhd/final_V0-bert-text-classification-model
bert
["BertForSequenceClassification"]
30522
float32
4.39.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,304
AmirlyPhd/final_V1-roberta-text-classification-model
roberta
["RobertaForSequenceClassification"]
50265
float32
4.39.3
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,305
AmirlyPhd/final_V1_xai-bert-text-classification-model
bert
["BertForSequenceClassification"]
30522
float32
4.39.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,306
AmirlyPhd/final_V4_resized_balanced_Bert_balanced_dataset-after-adding-new-words-text-classification-model
bert
["BertForSequenceClassification"]
30641
float32
4.39.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,307
AmirlyPhd/merged_model_V4_merged_all_IEEEdatasets_threat
llama
["LlamaForCausalLM"]
32002
float16
4.36.2
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,308
AmirlyPhd/merged_model_V7_merged_all_IEEEdatasets_threat
llama
["LlamaForCausalLM"]
32002
float16
4.36.2
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,309
AmirlyPhd/v10-bert-combined-dataset-text-classification-model
bert
["BertForSequenceClassification"]
30522
float32
4.39.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,310
Amirmerfan/bert-base-uncased-persian-ner-10k-mini
bert
["BertForTokenClassification"]
105879
float32
4.48.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,311
Amirmerfan/bert-base-uncased-persian-ner-50k-base
bert
["BertForTokenClassification"]
105879
float32
4.48.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,312
Amirmerfan/distilbert-finetuned-ner-restaurant
distilbert
["DistilBertForTokenClassification"]
30522
float32
4.48.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
3,313
Amirmerfan/t5-dialogsum-summarizer
t5
["T5ForConditionalGeneration"]
32128
float32
4.48.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
3,314
Amirmerfan/vit-base-indian-food-classifier
vit
["ViTForImageClassification"]
null
float32
4.48.3
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
3,315
Amirmerfan/vit-persian-food-classifier-mini
vit
["ViTForImageClassification"]
null
float32
4.48.3
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
3,316
Amirmerfan/whisper-small-finetuned-large-clean
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.51.1
null
null
12
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,317
Amirmerfan/whisper-small-finetuned-medium-noisy
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.51.1
null
null
12
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,318
Amirmerfan/whisper-small-finetuned-small-noisy
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.51.1
null
null
12
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,319
Amirmnsh/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,320
AmirrezaV1/PsychologistChatbot_emotional_model
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.38.2
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,321
AmitTewari/test_trainer
bert
["BertForSequenceClassification"]
28996
float32
4.40.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,322
Amitesh007/tw-sentiment-finetuned
distilbert
["DistilBertForSequenceClassification"]
28996
null
4.26.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
3,323
AmithAdiraju1694/Cerebras-GPT-590M_menuitemexp
gpt2
["GPT2LMHeadModel"]
51552
float32
4.47.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,324
Amitz244/output_dir_controlnet
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,325
AmiyendraOP/llama3-legal-finetuned
llama
["LlamaForCausalLM"]
128256
float16
4.51.3
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,326
Amjad123/Llama-2-7b-QA-Telecom
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,327
Amjad123/Llama-2-7b-chat-finetune
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,328
AmjadKha/Boppy
gpt2
["GPT2LMHeadModel"]
50257
float32
4.41.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,329
Ammad1Ali/Llongorca_16k
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
3,330
Ammad1Ali/bert2bert
encoder-decoder
["EncoderDecoderModel"]
30522
float32
4.38.2
null
null
null
null
null
null
null
null
null
null
null
null
null
102
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,331
Ammad1Ali/llama-v2-7B-alt
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,332
Ammar-1/llama2-7b-parenting-finetuned
llama
["LlamaForCausalLM"]
32000
float16
4.32.0.dev0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,333
Ammok/LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,334
Ammok/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,335
Ammok/predict-llm
bert
["BertForSequenceClassification"]
30522
float32
4.33.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,336
Ammonsh/wav2vec2-common_voice-tr-demo
wav2vec2
["Wav2Vec2ForCTC"]
41
float32
4.21.0.dev0
1,024
4,096
24
16
null
gelu
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
3,337
Amna100/DebertaMLMnc2c
deberta
["DebertaForMaskedLM"]
50265
float32
4.35.2
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,338
Amo/gpt-neo-125m-mlp-micro
gpt_neo
["GPTNeoForCausalLM"]
50257
float16
4.17.0
768
null
null
null
null
null
0
False
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
3,339
Amod/falcon7b-fine-tuned-therapy-merged
RefinedWebModel
["RWForCausalLM"]
65024
float16
4.30.2
4,544
null
null
null
null
null
0
True
null
null
null
0.02
11
11
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,340
Amogh01/whisper-small-hi
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.46.2
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,341
Amogh06/DeTr-Table-Detection-Model-public3
detr
["DetrForObjectDetection"]
null
float32
null
null
null
6
null
null
null
0
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,342
AmoooEBI/Bert-fa-qa-finetuned-on-PersianQA
bert
["BertForQuestionAnswering"]
100000
float32
4.51.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,343
AmoooEBI/xlm-roberta-fa-qa-finetuned-on-PersianQA
xlm-roberta
["XLMRobertaForQuestionAnswering"]
250002
float32
4.51.3
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,344
Amoros/Amoros_Beaugosse_test-large-2025_05_20_53326-bs32_freeze
dinov2
["Dinov2ForImageClassification"]
null
float32
4.48.0
1,024
null
24
16
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
3,345
Amr-h/AmrdisBert
distilbert
["DistilBertForQuestionAnswering"]
28996
float32
4.40.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
3,346
Amr-h/DisBertAmrHamedAbdelmoniem
distilbert
["DistilBertForQuestionAnswering"]
28996
float32
4.40.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
3,347
AmrMKayid/synchformer-hf
synchformer
["SynchformerModel"]
null
null
4.45.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,348
AmrSheta/Meme
bert
["BertModel"]
28996
float32
4.17.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,349
Amrinkar/BoxModel
llava
["LISAForCausalLM"]
32003
bfloat16
4.31.0
5,120
13,824
40
40
40
silu
null
False
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
12.582912
large
medium
3,350
Amrinkar/CartModel2
llava
["LISAForCausalLM"]
32003
bfloat16
4.31.0
5,120
13,824
40
40
40
silu
null
False
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
12.582912
large
medium
3,351
Amrinkar/CoffeeMachineModel2
llava
["LISAForCausalLM"]
32003
bfloat16
4.31.0
5,120
13,824
40
40
40
silu
null
False
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
12.582912
large
medium
3,352
Amrinkar/LisaSelected5
llava
["LISAForCausalLM"]
32003
bfloat16
4.31.0
5,120
13,824
40
40
40
silu
null
False
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
12.582912
large
medium
3,353
Amrinkar/PARIS3D
llava
["LISAForCausalLM"]
32003
bfloat16
4.31.0
5,120
13,824
40
40
40
silu
null
False
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
12.582912
large
medium
3,354
Amrinkar/PartNameModel
llava
["LISAForCausalLM"]
32003
bfloat16
4.31.0
5,120
13,824
40
40
40
silu
null
False
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
12.582912
large
medium
3,355
Amrinkar/lisa2prompt
llava
["LISAForCausalLM"]
32003
bfloat16
4.31.0
5,120
13,824
40
40
40
silu
null
False
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
12.582912
large
medium
3,356
Amrinkar/lisanoprompt
llava
["LISAForCausalLM"]
32003
bfloat16
4.31.0
5,120
13,824
40
40
40
silu
null
False
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
12.582912
large
medium
3,357
Amrinkar/lisatest80
llava
["LISAForCausalLM"]
32003
bfloat16
4.31.0
5,120
13,824
40
40
40
silu
null
False
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
12.582912
large
medium
3,358
AmritaBha/sd15_fill50k
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,359
Amritha1903/flan-t5-houseprice
t5
["T5ForConditionalGeneration"]
32128
float32
4.52.4
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
3,360
AmrutaMuthal/mero_sd15_controlnet_inpaint_masked_loss_wt_parts_ellipse
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,361
Amshaker/Mobile-VideoGPT-0.5B
MobileVideoGPT_qwen
["MobileVideoGPTQwenForCausalLM"]
151666
float16
4.42.4
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
true
false
false
false
false
false
true
true
0.231211
small
very_long
3,362
Amshaker/Mobile-VideoGPT-1.5B
MobileVideoGPT_qwen
["MobileVideoGPTQwenForCausalLM"]
151666
float16
4.42.4
1,536
8,960
28
12
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
true
false
false
false
false
false
true
true
0.792723
small
very_long
3,363
Amu/t1-3B
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.49.0
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
3,364
Amu/zen
mistral
["MistralForCausalLM"]
32002
bfloat16
4.36.2
4,096
14,336
32
32
8
silu
0
True
8,192
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
long
3,365
Amusam/pegasus-multipak
pegasus
["PegasusForConditionalGeneration"]
96103
float32
4.44.2
null
null
16
null
null
null
0.1
True
1,024
null
null
null
0
1
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
3,366
An0rak/alpaca_unsloth_gguf
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
3,367
AnAmbitiousMonk/ppo-LunarLander-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,368
AnAmbitiousMonk/ppo-LunarLander-v4
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,369
AnHoangbg/t5-small-finetuned-manimml-1.2
t5
["T5ForConditionalGeneration"]
32128
float32
4.42.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
3,370
AnHoangbg/tinyllama-manim-v1
llama
["LlamaForCausalLM"]
32000
float16
4.42.4
2,048
5,632
22
32
4
silu
0
True
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
3,371
AnOng/whisper-large-v2-vi
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.27.0.dev0
null
null
32
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,372
AnReu/math_pretrained_bert
bert
["BertForPreTraining"]
29497
float32
4.9.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,373
AnReu/math_pretrained_roberta
roberta
["RobertaForMaskedLM"]
50769
float32
4.9.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,374
AnabSohail/ahmed-detr-finetuned-tumor
detr
["DetrForObjectDetection"]
null
float32
null
null
null
6
null
null
null
0
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,375
Anaelle27/distilbert-base-uncased-finetuned-sentiment
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.49.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
3,376
AnakinCodebreaker/spk_qna_model
t5
["T5ForQuestionAnswering"]
32128
float32
4.34.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
3,377
AnaluRRamos/mt5-biomedical-translation-mode1
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.49.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
3,378
Anamta98/Qwen2-VL-7B-Instruct-LoRA-FT
qwen2_vl
["Qwen2VLForConditionalGeneration"]
152064
bfloat16
4.46.1
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
3,379
Anand-Keshavan/model
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.26.0
2,048
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
3,380
AnandBhat/my_awesome_wnut_model
distilbert
["DistilBertForTokenClassification"]
30522
float32
4.30.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
3,381
Anandappu/bert_comedy
bart
["BartForConditionalGeneration"]
50264
float32
4.38.2
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
3,382
Anandhd/my_awesome_model
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.44.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
3,383
Ananjas/AwooAI
gpt2
["GPT2LMHeadModel"]
50257
float32
4.24.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,384
Anant58/a2c-PandaReachDense-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
3,385
Anant58/ppo-Pyramid
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,386
Anant58/ppo-SnowballTarget
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,387
Anant58/vizdoom_health_gathering_supreme
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,388
Ananthu357/Ananthus-BAAI-for-contracts2.0
bert
["BertModel"]
30522
float32
4.41.2
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,389
Ananthu357/BAAI-first-train
bert
["BertModel"]
30522
float32
4.41.2
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,390
Anarchist/myLora
llama
["LlamaForCausalLM"]
32000
float32
4.37.0.dev0
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,391
Anaryl/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,392
Anas00/abcd
wav2vec2
["Wav2Vec2ForCTC"]
112
float32
4.18.0
1,024
4,096
24
16
null
gelu
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
3,393
Anas099/GumBit_AI
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
3,394
Anas989898/gemma_2b_it_ds
gemma
["GemmaForCausalLM"]
256000
float16
4.38.2
2,048
16,384
18
8
1
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
3,395
AnasAber/seamless-darija-eng
seamless_m4t_v2
["SeamlessM4Tv2ForTextToText"]
256102
float32
4.44.0
1,024
null
24
16
null
null
0.1
True
4,096
null
null
0.02
2
3
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
medium
3,396
Anasrafiq/llama3-unsloth-16bit
llama
["LlamaForCausalLM"]
128256
float16
4.44.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,397
AnassMzn/ECE-AFA
mistral
["MistralForCausalLM"]
32000
bfloat16
4.46.2
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
3,398
AnatolyBelov/CB_16bit
llama
["LlamaForCausalLM"]
128256
float16
4.44.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,399
AnatolyBelov/CB_4bit
llama
["LlamaForCausalLM"]
128256
float16
4.44.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long