index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
2,100
AhmedTaha012/finetuned-ner-hadith-try3
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.15.0
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
2,101
AhmedTaha012/finetuned-ner-hadith-try3785
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.15.0
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
2,102
AhmedTaha012/finetuned-ner-hadith-try398
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.15.0
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
2,103
AhmedTaha012/gptNeo-125M-txt2ARXMLv1
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.27.4
768
null
null
null
null
null
0
True
1,800
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
short
2,104
AhmedTaha012/gptneo-Txt-To-Json-v0.1.3
gpt_neo
["GPTNeoForCausalLM"]
50411
float32
4.27.4
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
2,105
AhmedTaha012/gptneo-txt2arxml-ppo
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.27.4
768
null
null
null
null
null
0
True
1,800
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
short
2,106
AhmedTaha012/mangersFeedback-V1.0.2
bert
["BertForSequenceClassification"]
30522
float32
4.30.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,107
AhmedTaha012/moviesReview5classBert
bert
["BertForSequenceClassification"]
30522
float32
4.20.1
512
2,048
4
8
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.012583
small
short
2,108
AhmedTaha012/moviesReview5classBert2
bert
["BertForSequenceClassification"]
30522
float32
4.20.1
512
2,048
4
8
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.012583
small
short
2,109
AhmedTaha012/moviesReview5classRoberta
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.20.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,110
AhmedTaha012/moviesReviewsClassifierbert1
bert
["BertForSequenceClassification"]
30522
float32
4.20.1
512
2,048
4
8
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.012583
small
short
2,111
AhmedTaha012/moviesReviewsClassifierbert2
bert
["BertForSequenceClassification"]
30522
float32
4.20.1
512
2,048
4
8
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.012583
small
short
2,112
AhmedTaha012/nextQuarter-status-V1.0.3
bert
["BertForSequenceClassification"]
30522
float32
4.30.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,113
AhmedTaha012/output
bert
["BertForSequenceClassification"]
30522
float32
4.20.1
512
2,048
4
8
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.012583
small
short
2,114
AhmedTarek/ppo-Pyramids
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,115
AhmedTarek/ppo-SnowballTarget
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,116
AhmedWaheed1/whisper-small-hi
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.49.0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
2,117
Ahmedhany216/Fake-news-gen
t5
["T5ForConditionalGeneration"]
32128
float32
4.36.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
2,118
Ahmedhisham/fullfinetunedRobertaNER
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.42.4
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
2,119
Ahmedhisham/social_bias_Bert
distilbert
["DistilBertForSequenceClassification"]
30522
null
4.32.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,120
Ahmeshen/a2c-PandaReachDense-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,121
AhmetAytar/all-mpnet-base-v2-fine-tuned
mpnet
["MPNetModel"]
30527
float32
4.39.3
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
2,122
AhmetAytar/all-mpnet-base-v2-fine-tuned_17_textbook_grobid_semantic
mpnet
["MPNetModel"]
30527
float32
4.39.3
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
2,123
AhmetBaris/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,124
Ahmetyigittt/xlm-roberta-sentiment
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.51.3
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,125
Ahom666/whisper-small-hi
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.42.3
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
2,126
Ahrefs/LongCamel-7b-32k
llama
["LlamaForCausalLM"]
32008
float32
4.34.0.dev0
4,096
11,008
32
32
32
silu
null
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,127
Ahs2000/segformer-b0-scene-parse-150
segformer
["SegformerForSemanticSegmentation"]
null
float32
4.45.1
null
null
null
[1, 2, 5, 8]
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,128
AhsanShahid/Creative_AI
gpt2
["GPT2LMHeadModel"]
50257
float32
4.44.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
2,129
AhsanZaidi/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,130
Ahyun/generate_model
gpt2
["GPT2LMHeadModel"]
50257
float32
4.41.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
2,131
Ai1terror/model
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
2,132
Ai1terror/model1101
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
2,133
AiAF/Finetuned-SCPWiki-032025-7B-Instruct
mistral
["MistralForCausalLM"]
32768
bfloat16
4.45.2
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,134
AiAF/KJV-LLM-Pretrained-V1.0
mistral
["MistralForCausalLM"]
32000
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,135
AiAF/KJV-LLM-Pretrained-V1.1
mistral
["MistralForCausalLM"]
32000
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,136
AiAF/Mistral-QLoRA-Pretraining-Test-v1.1
mistral
["MistralForCausalLM"]
32000
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,137
AiAF/Mistral-QLoRA-Pretraining-Test-v1.1_Merged
mistral
["MistralForCausalLM"]
32000
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,138
AiAF/Pretrained-QLoRA-r9kilo-V1
mistral
["MistralForCausalLM"]
32768
bfloat16
4.49.0
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,139
AiAF/Pretrained-SCP-1B-QLoRA
llama
["LlamaForCausalLM"]
128256
bfloat16
4.49.0
2,048
8,192
16
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128009
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
2,140
AiAF/Pretrained_QLoRA_Codename-75567-V1.1_Ckpt-51
mistral
["MistralForCausalLM"]
32768
bfloat16
4.49.0
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,141
AiAF/UFOs-Mistral7b-V3
mistral
["MistralForCausalLM"]
32000
bfloat16
4.48.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
2,142
AiCloser/Qwen2.5-0.5B-Instruct-Thinking
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.46.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
2,143
AiCloser/Qwen2.5-32B-AGI
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.44.2
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
2,144
AiForTheChurch/ChristianGPT-base-8bit
llama
["LlamaForCausalLM"]
32000
float16
4.34.0
4,096
11,008
32
32
32
silu
null
True
4,096
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,145
AiForTheChurch/ChristianGPT-base-full-precision
llama
["LlamaForCausalLM"]
32000
float32
4.34.0
4,096
11,008
32
32
32
silu
null
True
4,096
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,146
AiForTheChurch/ChristianGPT-catholic-8bit
llama
["LlamaForCausalLM"]
32000
float16
4.34.0
4,096
11,008
32
32
32
silu
null
True
4,096
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,147
AiLab-IMCS-UL/lvbert
bert
["BertModel"]
32004
float32
4.15.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,148
AiLab-IMCS-UL/lvmed
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.29.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
2,149
AiLen13/plumia_uma_water_advice_model
t5
["T5ForConditionalGeneration"]
32128
float32
4.52.4
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
2,150
AiTommy/decider_agent_test_2_merged
llama
["LlamaForCausalLM"]
128256
float16
4.44.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
2,151
AiTommy/llama-instruct-fi-q4
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
2,152
Aib0/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,153
Aichunks/wav2vec2-base-timit-demo-google-colab
wav2vec2
["Wav2Vec2ForCTC"]
32
float32
4.35.2
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
2,154
Aicraftar/Tharo.G-Eco
Tharo.G-Eco
["Tharo.GForCausalLM"]
100352
bfloat16
4.47.0
5,120
17,920
40
40
10
silu
0
True
16,384
250,000
0.00001
0.02
100257
100265
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
12.582912
large
long
2,155
AidAFadaeian/Mt5_persian_translation_en
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.41.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
2,156
Aida-Khaleghi/sentiment-analysis-distilbert-base-uncased-sst2
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.51.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,157
AidoAido/Llama3.1-8b-script-demo-eng
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
2,158
AightBits/Qwen2.5-Coder-32B-Instruct-6.0bpw-h6-exl2
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.43.1
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
2,159
AightBits/Qwen2.5-Coder-32B-Instruct-8.0bpw-h8-exl2
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.43.1
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
2,160
AightBits/microsoft-phi-4-8.0bpw-h8-exl2
phi3
["Phi3ForCausalLM"]
100352
bfloat16
4.47.0
5,120
17,920
40
40
10
silu
0
True
16,384
250,000
0.00001
0.02
100257
100265
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
12.582912
large
long
2,161
AigizK/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,162
AigizK/whisper-small-bak
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.26.0.dev0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
2,163
Aiknight2001/finetuned-distilbert
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.42.4
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,164
Aikozvezda/distilbert-base-uncased-finetuned-emotion
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.41.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,165
Ailei/chatglm-qlora
chatglm
["ChatGLMForConditionalGeneration"]
65024
float16
4.30.2
4,096
null
null
32
null
null
0
True
null
null
null
null
null
2
false
128
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
2,166
Ailoveu/ds_model
qwen2
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,167
Aimlab/xlm-roberta-base-finetuned-urdu
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.19.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,168
Aimodelbyharis/scriptgenerationmodel
llama
["LlamaForCausalLM"]
32000
float16
4.34.0
4,096
11,008
32
32
32
silu
null
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,169
Aimodelbyharis/scriptgenerationmodelv2
llama
["LlamaForCausalLM"]
32000
float16
4.34.0
4,096
11,008
32
32
32
silu
null
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,170
AimonLabs/hallucination-detection-model
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,171
AinaSiddiqui/learn_hf_food_not_food_text_classifier-distilbert-base-uncased
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.46.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
2,172
Air29/speecht5_finetuned_voxpopuli_nl
speecht5
["SpeechT5ForTextToSpeech"]
81
float32
4.48.3
768
null
null
null
null
gelu
0.1
False
null
null
null
0.02
0
2
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
2,173
AirMannanov/llm-course-hw1
null
[]
1024
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,174
AirinElizabath/JBS-3
opt
["OPTForCausalLM"]
50272
float32
4.33.0.dev0
1,024
null
24
16
null
null
0
True
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
medium
2,175
AirrStorm/BERT-MLM-IMDb
bert
["BertForMaskedLM"]
28996
float32
4.47.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
2,176
AishaBaliyan/glamour-ai-skin-model
vit
["ViTForImageClassification"]
null
float32
4.50.0
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
2,177
AishaKanwal/ModelsClassification
vit
["ViTForImageClassification"]
null
float32
4.31.0.dev0
768
3,072
8
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.056623
small
null
2,178
AishaKanwal/ModelsViT_RiceDisease
vit
["ViTForImageClassification"]
null
float32
4.31.0.dev0
768
3,072
2
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.014156
small
null
2,179
Aishee06/Finetuned-Llama2-7b-chat
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
2,180
AishitaSharma/sd-class-butterflies-32-new
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,181
AishwaryaDixit/LunarLander
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,182
Aitor/ppo-Pyramids
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,183
Aitor/ppo-SnowballTarget
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,184
AitorDL/MNLP_DPO_HS_lr1e-4_b01-hinge-MATH-lr1e-5_b005
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.52.3
1,024
3,072
28
16
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.352322
small
very_long
2,185
AitorDL/MNLP_DPO_HS_lr1e-5_b01
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.52.3
1,024
3,072
28
16
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.352322
small
very_long
2,186
AitorDL/MNLP_DPO_model_HS2
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.52.3
1,024
3,072
28
16
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.352322
small
very_long
2,187
Aitrepreneur/Florence-2-large
florence2
["Florence2ForConditionalGeneration"]
51289
float16
4.41.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
2,188
Aitrepreneur/llava-llama-3-8b-text-encoder-tokenizer
llama
["LlamaForCausalLM"]
128320
float16
4.44.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
2,189
Aitrepreneur/stable-vicuna-13B-GPTQ-4bit-128g
llama
["LlamaForCausalLM"]
32001
float16
4.28.1
5,120
13,824
40
40
null
silu
null
True
2,048
null
0.000001
0.02
1
2
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
12.582912
large
medium
2,190
Aityz/aityz_model
gpt2
["GPT2LMHeadModel"]
50257
float32
4.31.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
2,191
Aityz/aityz_model_eli5
gpt2
["GPT2LMHeadModel"]
50257
float32
4.29.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
2,192
Aityz/eli5_distilgpt2_mini
gpt2
["GPT2LMHeadModel"]
50257
float32
4.29.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
2,193
Aityz/reviews_model
gpt2
["GPT2LMHeadModel"]
50257
float32
4.31.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
2,194
Aivesa/0a3020fa-9065-4a04-b9c7-e0cadb2c120d
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
2,195
Aivesa/0d041048-0da7-4497-ac21-da31409a0bf2
gpt_neox
["GPTNeoXForCausalLM"]
50304
float16
4.47.1
128
512
6
4
null
gelu
0
False
2,048
10,000
null
0.02
0
0
false
32
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.00118
small
medium
2,196
Aivesa/14faec3b-678a-4b3e-a3e7-95067cb131aa
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.47.1
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
2,197
Aivesa/1931e729-dd3f-4026-a118-7fef2d3afe59
llama
["LlamaForCausalLM"]
32000
float16
4.47.1
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0
2
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
2,198
Aivesa/23784452-b886-4280-9bfa-946e195d84e3
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.47.1
2,048
16,384
18
8
1
gelu
0
False
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
2,199
Aivesa/286fd9a8-760d-4213-a477-b5ea1e7ce9a3
codegen
["CodeGenForCausalLM"]
51200
float32
4.47.1
null
null
null
null
null
null
null
False
null
null
null
0.02
50256
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null