index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
800
2point5p/krx-qwen2-7b-it-X-Ko
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
801
2point5p/krx-qwen2.5-7b-it-X-Four
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
802
2todeux/1008model
bert
["BertForSequenceClassification"]
30000
float32
4.44.2
768
3,072
12
12
null
gelu
null
True
300
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
803
2todeux/ViT_beans
vit
["ViTForImageClassification"]
null
float32
4.44.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
804
2z299/Mistral-Nemo-Japanese-Instruct-2408-bnb-4bit
mistral
["MistralForCausalLM"]
131074
float16
4.46.3
5,120
14,336
40
32
8
silu
0
False
1,024,000
1,000,000
0.00001
0.02
1
131072
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
805
2z299/c4ai-command-r-08-2024-GPTQ-4bit
cohere
["CohereForCausalLM"]
256000
float16
4.47.0.dev0
8,192
24,576
40
64
8
silu
0
True
131,072
4,000,000
null
0.02
5
255001
null
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
32.212255
large
very_long
806
2z299/sarashina2.2-3b-instruct-v0.1-onnx
llama
["LlamaForCausalLM"]
102400
null
4.48.3
2,560
8,960
32
16
8
silu
0
True
8,192
500,000
0.00001
0.02
1
2
false
160
2
false
null
null
true
false
false
false
false
false
false
false
true
true
2.516582
medium
long
807
316usman/research_abstractive_clinical_notes_summarizations
phi3
["Phi3ForCausalLM"]
32064
float16
4.42.3
3,072
8,192
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
medium
808
34324324esfesf/autotrain-sds-1516354966
roberta
["RobertaForQuestionAnswering"]
50262
float32
4.22.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
809
34324324esfesf/autotrain-tfgt-1512454907
roberta
["RobertaForQuestionAnswering"]
50262
float32
4.22.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
810
38sebastian/finetuning-sentiment-model-3000-samples
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.41.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
811
3BDOAi3/finetuned_1
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.30.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
812
3BDOAi3/finetuned_2
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.30.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
813
3BDOAi3/finetuned_3
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.30.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
814
3BDOAi3/finetuned_with_labeled_dataset
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.30.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
815
3BDOAi3/model
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.30.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
816
3JI0/dummyModel
camembert
["CamembertForMaskedLM"]
32005
float32
4.29.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
5
6
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
817
3NTRPY-13/gpt2-reuters_training
gpt2
["GPT2LMHeadModel"]
52000
float32
4.49.0
null
null
null
null
null
null
null
True
null
null
null
0.02
0
0
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
818
3bdoo/bart-large-cnn-finetuned-xsum
bart
["BartForConditionalGeneration"]
50264
float32
4.38.2
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
819
3chez/layoutxlm-finetuned-xfund-fr
layoutlmv2
["LayoutLMv2ForTokenClassification"]
250002
float32
4.40.2
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
820
3dlg-hcvc/aekl_geometry_xyz2xyz
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
821
3dlg-hcvc/m0425_aekl_v2_xyzob_mse_normalloss_randtransf2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
822
3ed0k4/wertert
gpt2
["GPT2LMHeadModel"]
50257
float32
4.50.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
823
3funnn/wav2vec2-base-cv-corpus
wav2vec2
["Wav2Vec2ForCTC"]
32
float32
4.38.2
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
824
3hd/QA_model
distilbert
["DistilBertForQuestionAnswering"]
30522
float32
4.40.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
825
3lv27/CodingMindsetTunnedLlama-3.1-8B-GGUF
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
826
3mei/finetuned_llama_3.1_storm_instruct_4bit_405_v2_reflection_gsm8k_3e_qv
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.0
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
827
3mei/llama_3.1_instruct_4bit_evolutionary_405_v1_gsm8k_3e_qkvogud
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.0
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
828
3mei/llama_3.1_instruct_4bit_reflection_1k_405_v1_gsm8k_3e_qv
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.0
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
829
3mei/llama_3.1_instruct_4bit_reflection_405_v1_gsm8k_3e_qkvogud
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.0
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
830
3mily1u/new-codegen-350m-mono-dpoed-attack-10-0.5
codegen
["CodeGenForCausalLM"]
50298
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
0.02
1
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
831
3mily1u/new-codegen-350m-mono-dpoed-attack-25-0.1
codegen
["CodeGenForCausalLM"]
50298
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
0.02
1
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
832
3mily1u/new-codegen-350m-mono-dpoed-attack-25-1
codegen
["CodeGenForCausalLM"]
50298
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
0.02
1
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
833
3mily1u/new-codegen-350m-mono-dpoed-attack-50-0.1
codegen
["CodeGenForCausalLM"]
50298
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
0.02
1
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
834
3mily1u/new-codegen-350m-mono-dpoed-attack-50-0.5
codegen
["CodeGenForCausalLM"]
50298
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
0.02
1
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
835
3mily1u/new-codegen-350m-mono-dpoed-attack-50-1
codegen
["CodeGenForCausalLM"]
50298
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
0.02
1
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
836
3mily1u/new-codegen-350m-mono-dpoed-control-10-1
codegen
["CodeGenForCausalLM"]
50298
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
0.02
1
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
837
3mily1u/new-codegen-350m-mono-dpoed-control-50-0.5
codegen
["CodeGenForCausalLM"]
50298
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
0.02
1
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
838
3mily1u/new-codegen-350m-mono-dpoed-control-50-1
codegen
["CodeGenForCausalLM"]
50298
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
0.02
1
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
839
3mrology/llama-3.2-3b-resume-evaluation-tuned-final
llama
["LlamaModel"]
128256
float32
4.46.3
3,072
8,192
28
24
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
840
3okasha/ai_commiter
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
841
3rd-Degree-Burn/L-3.1-Science-Writer-8B-v0
llama
["LlamaForCausalLM"]
128256
float16
4.46.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
842
3rd-Degree-Burn/L-3.1-Science-Writer-8B-v1.0
llama
["LlamaForCausalLM"]
128256
float16
4.46.3
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
843
3ricL/ad-hoc-ance-msmarco
roberta
["RobertaDot_NLL_LN"]
50265
null
null
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
844
3van/RoBERTa_10M_ELI5_CurriculumMasking_redo
roberta
["RobertaForMaskedLM"]
10000
float32
4.44.2
768
3,072
6
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.042467
small
short
845
4-alokk/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
846
404sau404/bart_samsum
bart
["BartForConditionalGeneration"]
50264
float32
4.42.4
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
847
409540655h/test
chatglm
["ChatGLMModel"]
null
float16
4.27.1
4,096
null
null
32
null
null
0
True
null
null
null
null
null
2
false
128
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
848
420-kjh/bert_model
bert
["BertForSequenceClassification"]
30000
float32
4.48.3
768
3,072
12
12
null
gelu
null
True
300
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
849
42MARU/GenAI-llama-2-13b
llama
["LlamaForCausalLM"]
32000
bfloat16
4.34.0.dev0
5,120
13,824
40
40
40
silu
null
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
850
42MARU/GenAI-llama2-ko-en-dpo-13b-v2
llama
["LlamaForCausalLM"]
46336
float16
4.34.1
5,120
13,824
40
40
40
silu
null
True
2,048
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
851
42dot/42dot_LLM-PLM-1.3B
llama
["LlamaForCausalLM"]
50304
float32
4.31.0
2,048
5,632
24
32
32
silu
null
True
4,096
null
0.000001
0.01
50257
50256
false
64
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
852
42dot/42dot_LLM-SFT-1.3B
llama
["LlamaForCausalLM"]
50304
float32
4.33.1
2,048
5,632
24
32
32
silu
null
False
4,096
10,000
0.000001
0.01
50257
50256
false
64
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
853
4TB-USTC/nlp_sc_based_on_bert
bert
["BertForSequenceClassification"]
30522
float32
4.38.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
854
4bit/Llama-3.2-1B
llama
["LlamaForCausalLM"]
128256
bfloat16
4.45.0.dev0
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
855
4bit/Qwen-14B-Chat-Int4
qwen
["QWenLMHeadModel"]
152064
null
4.32.0
5,120
27,392
40
40
null
null
null
True
8,192
null
null
0.02
null
null
false
128
null
false
null
null
false
false
false
false
false
false
false
false
false
false
12.582912
large
long
856
4bit/Qwen-VL
qwen
["QWenLMHeadModel"]
151936
bfloat16
4.31.0
4,096
22,016
32
32
null
null
null
True
8,192
null
null
0.02
null
null
false
128
null
false
null
null
false
false
false
false
false
false
false
false
false
false
6.442451
medium
long
857
4bit/Replit-v1-CodeInstruct-3B
mpt
["MPTForCausalLM"]
32769
float32
4.29.2
null
null
null
null
null
null
null
False
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
858
4bit/llava-v1.5-7b
llava
["LlavaLlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
859
4bit/medllama2_7b
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
860
4bit/medllama2_7b_s
llama
["LlamaForCausalLM"]
32000
float16
4.32.0.dev0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
861
4bit/stable-vicuna-13B-GPTQ
llama
["LlamaForCausalLM"]
32001
float16
4.28.1
5,120
13,824
40
40
null
silu
null
True
2,048
null
0.000001
0.02
1
2
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
12.582912
large
medium
862
4bit/vicuna-13B-1.1-GPTQ-4bit-128g
llama
["LlamaForCausalLM"]
32000
float16
4.28.0.dev0
5,120
13,824
40
40
null
silu
null
True
2,048
null
0.000001
0.02
0
1
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
12.582912
large
medium
863
4bit/vicuna-v1.1-13b-GPTQ-4bit-128g
llama
["LlamaForCausalLM"]
32000
float16
4.29.0.dev0
5,120
13,824
40
40
null
silu
null
True
2,048
null
0.000001
0.02
1
2
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
12.582912
large
medium
864
4blacks/finetuned-llama-no-vocab-main-version-q16
llama
["LlamaForCausalLM"]
128256
float16
4.44.0
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
865
4darsh-Dev/Meta-Llama-3-8B-quantized-GPTQ
llama
["LlamaForCausalLM"]
128256
float16
4.41.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
866
4everStudent/Qwen2-0.5B-GRPO-test-5epochs
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.52.4
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
867
4k5h17/distilhubert-finetuned-gtzan
hubert
["HubertForSequenceClassification"]
32
float32
4.45.1
768
3,072
2
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.014156
small
null
868
4lchemistX/mia-ggufv1
llama
["LlamaForCausalLM"]
32000
float32
4.35.2
4,096
11,008
32
32
32
silu
null
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
869
4m1g0/wav2vec2-large-xls-r-300m-gl-jupyter3
wav2vec2
["Wav2Vec2ForCTC"]
32
float32
4.11.3
1,024
4,096
24
16
null
gelu
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
870
4m1g0/wav2vec2-large-xls-r-300m-gl-jupyter4
wav2vec2
["Wav2Vec2ForCTC"]
32
float32
4.11.3
1,024
4,096
24
16
null
gelu
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
871
4m1g0/wav2vec2-large-xls-r-300m-gl-jupyter5
wav2vec2
["Wav2Vec2ForCTC"]
32
float32
4.11.3
1,024
4,096
24
16
null
gelu
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
872
4m1g0/wav2vec2-large-xls-r-53m-gl-jupyter2
wav2vec2
["Wav2Vec2ForCTC"]
32
float32
4.11.3
1,024
4,096
24
16
null
gelu
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
873
4m1g0/wav2vec2-large-xls-r-53m-gl-jupyter3
wav2vec2
["Wav2Vec2ForCTC"]
32
float32
4.11.3
1,024
4,096
24
16
null
gelu
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
874
4maan4hmad/Mistral-finetuned-productidentifier
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
875
4maan4hmad/Mistral-finetuned-productidentifier-v2.0
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
876
4mosot/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
877
4mosot/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
878
4ndr3lu15/my_awesome_model
bert
["BertForMultipleChoice"]
30522
null
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
879
4s4ki/doodownnakumkuing-V2
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.10.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
880
4season/alignment-model-test10
llama
["LlamaForCausalLM"]
92544
bfloat16
4.37.2
6,144
16,384
52
48
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
6
false
null
null
true
false
false
false
false
false
false
false
true
true
23.555211
large
very_long
881
4season/alignment-model-test11
llama
["LlamaForCausalLM"]
92544
bfloat16
4.37.2
6,144
16,384
52
48
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
6
false
null
null
true
false
false
false
false
false
false
false
true
true
23.555211
large
very_long
882
4skin/test1
llama
["LlamaForCausalLM"]
128256
bfloat16
4.52.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
883
4yo1/llama3-pre1-ds-lora1
llama
["LlamaForCausalLM"]
149465
float16
4.33.0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
884
501Good/whisper-small-tt
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.26.0.dev0
null
null
12
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
885
50stars/TinyLlama-psy
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
2,048
5,632
22
32
4
silu
null
True
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
886
50stars/distilbert_imdb_genre_classifier
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.26.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
887
50stars/fine-tuned-model
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.26.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
888
51-wannt-to-be/bert-chinese
bert
["BertForMaskedLM"]
21128
float32
4.52.4
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
889
51la5/QMSUM-keyphrase-gen
bart
["BartForConditionalGeneration"]
50264
float32
4.20.0
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
890
51la5/XSUM-keyphrase-gen
bart
["BartForConditionalGeneration"]
50264
float32
4.20.0
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
891
51nd0re1/Pyramids
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
892
51nd0re1/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
893
51nd0re1/ppo-SnowballTarget
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
894
52100176-NguyenTrongDat/t5-vietnamese
t5
["T5ForConditionalGeneration"]
32128
float32
4.47.1
null
null
null
null
null
null
null
False
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
895
52101007-HuynhQuocTruong/vi-correction-t5
t5
["T5ForConditionalGeneration"]
32128
float32
4.40.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
896
529LJH/bert_model_out
bert
["BertForSequenceClassification"]
30000
float32
4.48.3
768
3,072
12
12
null
gelu
null
True
300
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
897
54data/xlm-roberta-base-finetuned-panx-all
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.32.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
898
54data/xlm-roberta-base-finetuned-panx-de-fr
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.32.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
899
54data/xlm-roberta-base-finetuned-panx-en
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.32.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short