index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
13,800
Hachipo/qwen2.5-0.5B_ichikara_4802
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.47.1
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
13,801
Hachipo/qwen2.5-0.5B_mceval_instruct
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.46.3
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
13,802
Hack337/ChatGPT-5
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.43.3
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
13,803
Hack337/WavGPT-2.5
qwen2
["Qwen2ForCausalLM"]
152064
float16
4.43.3
5,120
13,824
48
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
15.099494
large
very_long
13,804
Hack90/results
gpt_neox
["GPTNeoXForCausalLM"]
8
float32
4.35.2
512
2,048
6
8
null
gelu
0
True
2,048
null
null
0.02
0
2
false
64
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.018874
small
medium
13,805
Hack90/virus-pythia-19M-2048-ce_synthetic
gpt_neox
["GPTNeoXForCausalLM"]
1867
float32
4.45.2
512
2,048
6
8
null
gelu
0
True
2,048
10,000
null
0.02
0
0
false
64
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.018874
small
medium
13,806
Hack90/virus_pythia_70_3072
gpt_neox
["GPTNeoXForCausalLM"]
8
float32
4.37.2
512
2,048
6
8
null
gelu
0
True
2,048
null
null
0.02
0
0
false
64
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.018874
small
medium
13,807
HadarG/SmolLM2-FT-smoltalk
llama
["LlamaForCausalLM"]
49152
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
13,808
HadarG/SmolLM2-FT-theStackSmol_python
llama
["LlamaForCausalLM"]
49152
float32
4.47.1
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
13,809
Hadassah/food-model-hads
detr
["DetrForObjectDetection"]
null
float32
4.7.0.dev0
null
null
6
null
null
null
0
null
1,024
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
13,810
Hadiana/ocr-lines-ckb
vision-encoder-decoder
["VisionEncoderDecoderModel"]
50265
float32
4.32.1
null
null
null
null
null
null
null
null
null
null
null
null
null
2
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,811
Hadihilman/Llama-3.2-reason
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
13,812
Hadjer07/finetuned-bertweetlarge-pheme
roberta
["RobertaForSequenceClassification"]
50265
float32
4.22.1
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
13,813
Hadnet/llama-2-chat-7b-hf-olavo-articles-17k
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
13,814
HadrienCr/a2c-PandaReachDense-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
13,815
HadrienCr/ppo-SnowballTarget
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,816
Haesteining/10.5v3
llama
["LlamaForCausalLM"]
96000
bfloat16
4.44.0
4,096
14,208
44
32
8
silu
0
False
4,096
500,000
0.00001
0.02
95993
95993
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
8.85837
large
medium
13,817
Haesteining/GodProvides
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.44.0
5,120
17,920
40
40
10
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
12.582912
large
medium
13,818
Haesteining/LargePhiv1
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.44.0
4,608
8,192
48
48
48
silu
0
False
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
12.23059
large
very_long
13,819
Haesteining/LargePhiv2
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.44.0
4,608
8,192
48
48
48
silu
0
False
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
12.23059
large
very_long
13,820
Haesteining/LargerPhiv2
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.44.0
5,120
17,920
40
40
10
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
12.582912
large
medium
13,821
Haesteining/LlamaSN29Dec27_1
llama
["LlamaForCausalLM"]
100288
bfloat16
4.44.0
4,096
18,752
48
32
8
silu
0
False
4,096
500,000
0.00001
0.02
100257
100257
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
9.663676
large
medium
13,822
Haesteining/Phi314Bv3
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.44.0
5,120
17,920
40
40
10
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
12.582912
large
medium
13,823
Haesteining/Phi314Bv4
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.44.0
5,120
17,920
40
40
10
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
12.582912
large
medium
13,824
Haesteining/Phi3SN29Dec22_2
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.44.0
5,120
17,920
40
40
10
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
12.582912
large
medium
13,825
Haesteining/Phi3SN29Dec22_4
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.44.0
5,120
17,920
40
40
10
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
12.582912
large
medium
13,826
Haesteining/PhiSN29Dec28_1
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.44.0
5,120
17,920
40
40
10
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
12.582912
large
medium
13,827
Haesteining/PhiSN29Feb7_1
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.44.0
5,120
17,920
40
40
10
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
12.582912
large
medium
13,828
Haesteining/PhiSN29Jan10_1
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.44.0
5,120
17,920
40
40
10
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
12.582912
large
medium
13,829
Haesteining/PhiSN29Jan19_1
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.44.0
5,120
17,920
40
40
10
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
12.582912
large
medium
13,830
Hafad/ppo-LunarLander-v2-attempt1
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,831
Hafedh88/microsoft-Phi-3-medium-4k-instruct-q
phi3
["Phi3ForCausalLM"]
32064
float16
4.50.2
5,120
17,920
40
40
10
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
12.582912
large
medium
13,832
HafeezKing/my-first-model
bert
["BertModel"]
30522
float32
4.51.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,833
Hafiz47/food_classifier
vit
["ViTForImageClassification"]
null
null
4.34.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
13,834
Hagli/model
roberta
["RobertaForTokenClassification"]
52000
null
4.33.2
768
3,072
6
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.042467
small
short
13,835
HaiTao90/gpt2-wiki
gpt2
["GPT2LMHeadModel"]
50257
float32
4.19.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
13,836
HaiXotiny/Xofy
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.51.0
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
13,837
Haicaochi/sum-bartpho-word-base
mbart
["MBartForConditionalGeneration"]
64001
float32
4.51.3
null
null
6
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
13,838
Haicaochi/sum-hierarchical-dou-vit5base-chunk-summarizer
t5
["T5ForConditionalGeneration"]
36096
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
13,839
Haicaochi/sum-hierarchical-dou-vit5base-summary-combiner
t5
["T5ForConditionalGeneration"]
36096
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
13,840
Haicaochi/sum-hierarchical-vit5-base-single-model
t5
["T5ForConditionalGeneration"]
36096
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
13,841
HailHydra/bimanual_so100_t0
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,842
Hailay/MachineT_TigEng
marian
["MarianMTModel"]
63050
float32
4.51.3
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
13,843
Hailay/xlmr-amharic-mlm
xlm-roberta
["XLMRobertaForMaskedLM"]
22243
float32
4.52.3
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,844
Hailay/xlmr-tigrinya-mlm
xlm-roberta
["XLMRobertaForMaskedLM"]
50000
float32
4.52.4
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,845
HaileyJu/videomae-base-finetuned-ucf101-finetuned-ucf101_kaggle-finedtuned_lr1.0e-05_20240905_1323
videomae
["VideoMAEForVideoClassification"]
null
float32
4.44.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
13,846
Haimath/BLIP-Math
blip
["BlipForConditionalGeneration"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,847
HaitameLaframe/Mistral-v3-7b-coder
mistral
["MistralForCausalLM"]
32768
float16
4.47.1
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
13,848
HaitameLaframe/Phi3-16bit_GGUF
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
13,849
Haitao999/Llama-3.2-3B-Instruct-EMPO-numia_prompt_dpo1
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
3,072
8,192
28
24
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
13,850
Haitao999/Qwen2.5-1.5B-Open-R1-Distill
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.49.0
1,536
8,960
28
12
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
13,851
Haitao999/Qwen2.5-14B-EMPO-Natural-Reasoning
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.48.3
5,120
13,824
48
40
8
silu
0
False
131,072
1,000,000
0.00001
0.02
151643
151643
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
15.099494
large
very_long
13,852
Haitao999/Qwen2.5-3B-EMPO-Natural-Reasoning-50K-checkpoint-210
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.48.3
2,048
11,008
36
16
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
13,853
Haitao999/Qwen2.5-3B-EMPO-Natural-Reasoning-non-math-formatted
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.48.3
2,048
11,008
36
16
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
13,854
Haitao999/Qwen2.5-7B-Base-EMPO-natural_reasoning_all_level
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.48.3
3,584
18,944
28
28
4
silu
0
True
131,072
1,000,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
13,855
Haitao999/Qwen2.5-7B-EMPO-NM-COT_0_025
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.48.3
3,584
18,944
28
28
4
silu
0
True
4,096
10,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
medium
13,856
Haitao999/Qwen2.5-7B-EMPO-NM-COT_0_1
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.48.3
3,584
18,944
28
28
4
silu
0
False
4,096
10,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
medium
13,857
Haitao999/Qwen2.5-7B-GRPO-NM-COT-20K
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.48.3
3,584
18,944
28
28
4
silu
0
True
4,096
10,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
medium
13,858
Haitao999/Qwen2.5-7B-GRPO-NM-COT-20K-2epoch
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.48.3
3,584
18,944
28
28
4
silu
0
True
4,096
10,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
medium
13,859
Haitao999/Qwen2.5-7B-GRPO-Natural-Reasoning-0428
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.48.3
3,584
18,944
28
28
4
silu
0
True
131,072
1,000,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
13,860
HaitingHW/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,861
Haiueo/distilbert-base-uncased-finetuned-emotion
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.50.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
13,862
Hajarhammadi/Gpt2_finetunnig
gpt2
["GPT2LMHeadModel"]
50257
float32
4.44.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
13,863
Hajarhammadi/gpt2_finetuned
gpt2
["GPT2LMHeadModel"]
50257
float32
4.44.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
13,864
HajerA/Llama-2-7b-chat-finetune-small
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
13,865
HajimeOgawa/gemma3-4b-mbti-chat-energy
gemma3
["Gemma3ForConditionalGeneration"]
null
bfloat16
4.52.4
null
null
null
null
null
null
null
null
null
null
null
0.02
null
[1, 106]
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
13,866
Haleshot/Mathmate-7B-DELLA
llama
["LlamaForCausalLM"]
102400
bfloat16
4.44.2
4,096
11,008
30
32
32
silu
0
True
4,096
10,000
0.000001
0.02
100000
100001
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.039798
medium
medium
13,867
HaliG/customer-support-chatbot
t5
["T5ForConditionalGeneration"]
32128
float32
4.49.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
13,868
Halistas-ce/results
distilbert
["DistilBertForSequenceClassification"]
28996
float32
4.51.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
13,869
Ham1mad1/videomae-base-Vsl-Lab-PC
videomae
["VideoMAEForVideoClassification"]
null
float32
4.36.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
13,870
Ham1mad1/videomae-base-Vsl-Lab-PC-V3
videomae
["VideoMAEForVideoClassification"]
null
float32
4.36.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
13,871
Ham1mad1/videomae-base-Vsl-Lab-PC-V6-homecooksss
videomae
["VideoMAEForVideoClassification"]
null
float32
4.36.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
13,872
Ham1mad1/videomae-base-fyp-lab-pc2
videomae
["VideoMAEForVideoClassification"]
null
float32
4.36.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
13,873
HamAndCheese82/math-ocr-donut-v2
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.38.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,874
HamAndCheese82/math-ocr-donut-v2.2
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.38.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,875
HamAndCheese82/math_ocr_donut_onnx_v1
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.37.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,876
HamSowie/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,877
HamZurger/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,878
HamZurger/ppo-Pyramids
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,879
HamdanXI/bert-base-uncased-paradetox_with_labels
bert
["BertForSequenceClassification"]
30522
float32
4.34.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,880
HamdanXI/char-llm-beethoven
llama
["LlamaForCausalLM"]
32000
float16
4.40.0
4,096
11,008
32
32
32
silu
0
True
2,048
10,000
0.000001
0.02
0
1
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
13,881
Hameed0342j/gat-logmodel
GAT
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
13,882
HamidRezaei/Aug10-Persian-Offensive-Language-Detection
bert
["BertForSequenceClassification"]
100000
float32
4.46.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,883
Hamish/embed5
bert
["BertModel"]
30522
float32
4.42.4
384
1,536
6
12
null
gelu
null
True
512
null
null
0.02
null
null
null
32
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.010617
small
short
13,884
Hammad2910/t5_agent
t5
["T5ForConditionalGeneration"]
32110
float32
4.35.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
13,885
Hammad712/5-Flower-Types-Classification-VIT-Model
vit
["ViTForImageClassification"]
null
float32
4.41.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
13,886
Hamnabintealamgir/my_awesome_billsum_model
t5
["T5ForConditionalGeneration"]
32128
null
4.42.4
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
13,887
Hamza-Ziyard/sinMT5
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.28.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
13,888
Hamza11/Banglabert_nwp_finetuning_def_v2
electra
["ElectraForMaskedLM"]
32000
float32
4.38.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
13,889
Hamza11/sagorbert_nwp_finetuning_def_v2
bert
["BertForMaskedLM"]
102025
float32
4.38.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,890
Hamza11/sagorbert_nwp_finetuning_def_v3
bert
["BertForMaskedLM"]
102025
float32
4.38.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,891
Hamza11/shahajbert_nwp_finetuning_def_v1
albert
["AlbertForMaskedLM"]
32000
float32
4.38.1
1,024
4,096
24
16
null
gelu_new
null
null
512
null
null
0.02
2
3
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
13,892
Hamza4616/working
bert
["BertForMaskedLM"]
250002
float32
4.41.2
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
13,893
HamzaFarhan/InvoiceOrNot
mpnet
["MPNetModel"]
30527
float32
4.30.2
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
13,894
HamzaNaser/Dialects-to-MSA-Transformer
m2m_100
["M2M100ForConditionalGeneration"]
128112
float32
4.44.2
null
null
12
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
13,895
HamzaSidhu786/bert-finetuned-pos
bert
["BertForTokenClassification"]
28996
null
4.41.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
13,896
HamzaSidhu786/mt5-small-finetuned-govreport-summarization
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.42.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
13,897
HamzaSidhu786/speech-accent-detection
wav2vec2
["Wav2Vec2ForSequenceClassification"]
32
float32
4.48.0
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
13,898
HamzaSidhu786/urdu_text_to_speech_tts
speecht5
["SpeechT5ForTextToSpeech"]
96
float32
4.42.3
768
null
null
null
null
gelu
0.1
True
null
null
null
0.02
0
2
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
13,899
Hamzaharman/imageclassification
vit
["ViTForImageClassification"]
null
float32
4.35.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null