index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
7,000
ChasapasK/mistral-tensor-7b-july
mistral
["MistralForCausalLM"]
32768
float32
4.40.0
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
7,001
Chase777/memo-bert-danish-emotion
xlm-roberta
["XLMRobertaForSequenceClassification"]
50005
float32
4.50.3
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,002
ChaserX/bert-finetuned-squad
bert
["BertForQuestionAnswering"]
28996
float32
4.47.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,003
Chat-Error/Llama-3-Kimiko-temp
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0.dev0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
7,004
Chat-Error/Pyg6B-V8P2
gptj
["GPTJForCausalLM"]
50400
float16
4.26.0.dev0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
false
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
7,005
ChatSU/chatsu_model
t5
["T5ForConditionalGeneration"]
32128
float32
4.44.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
7,006
ChatWithNeil/my_awesome_model
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.27.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
7,007
Chatbot-DL/GameStory-Chatboot
mistral
["MistralForCausalLM"]
32064
float16
4.41.0
3,072
8,192
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
true
false
false
true
3.623879
medium
medium
7,008
Chavgpt/NLP_UDESA-modelo_base_berto
bert
["BertModel"]
31002
float32
4.30.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,009
Chavgpt/NLP_UDESA-modelo_base_berto_2023-06-12
bert
["BertModel"]
31002
float32
4.30.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,010
Chavgpt/NLP_UDESA-modelo_base_berto_2023-06-12v2
bert
["BertModel"]
31002
float32
4.30.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,011
Chavgpt/NLP_UDESA-modelo_base_roberta2023-06-12
roberta
["RobertaModel"]
250002
float32
4.30.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,012
Chavgpt/NLP_UDESAmodelo_base_berto
bert
["BertModel"]
31002
float32
4.30.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,013
Chayaaaaa/BioMistral_stabilityai_gamma_7b_SLERP
mistral
["MistralForCausalLM"]
32000
bfloat16
4.40.0
4,096
14,336
32
32
8
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
medium
7,014
Chayaaaaa/BioMistral_stabilityai_gamma_7b_dare_ties
mistral
["MistralForCausalLM"]
32000
bfloat16
4.40.0
4,096
14,336
32
32
8
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
medium
7,015
Chayaaaaa/without_Adventurous_apart_from_Mistral-7B_task_arithmetic
mistral
["MistralForCausalLM"]
32000
bfloat16
4.40.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
7,016
Chayaaaaa/without_chat_RakutenAI-7B_task_arithmetic
mistral
["MistralForCausalLM"]
48000
bfloat16
4.40.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
7,017
Chayaaaaa/without_japanese_apart_from_nekomata-7b_task_arithmetic
qwen
["QWenLMHeadModel"]
151936
bfloat16
4.40.2
4,096
22,016
32
32
null
null
null
True
32,768
null
null
0.02
null
null
false
128
null
false
null
null
false
false
false
false
false
false
false
false
false
false
6.442451
medium
very_long
7,018
ChayanM/SwinV2-GPT2_Mimic
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.37.1
null
null
null
null
null
null
null
null
null
null
null
null
null
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
7,019
ChayanM/ViT-Bert_Mimic
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.37.1
null
null
null
null
null
null
null
null
null
null
null
null
null
102
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
7,020
ChayanM/ViT-RadBert_Mimic
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.37.1
null
null
null
null
null
null
null
null
null
null
null
null
null
2
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
7,021
Chayawat/opus-mt-en-mul-finetuned-en-to-th
marian
["MarianMTModel"]
64110
float32
4.17.0
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
7,022
CheN70/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,023
CheWei/ModernBERT_8x8_8e-5_12epoch
modernbert
["ModernBertForQueryComparison"]
50368
float32
4.48.3
1,024
2,624
28
16
null
null
0
null
8,192
null
null
0.02
50281
50282
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
true
0.352322
small
long
7,024
ChechkovEugene/poca-SoccerTwos
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,025
ChechkovEugene/ppo-LunarLander-v2_first
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,026
ChechkovEugene/rl_course_vizdoom_health_gathering_supreme
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,027
ChechkovYevhen/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,028
Checking123123/granite-7b-instruct-ZipNN-compressed
llama
["LlamaForCausalLM"]
32008
null
4.40.0
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
7,029
CheeLi03/whisper-L-5k
whisper
["WhisperForConditionalGeneration"]
51866
float32
4.43.4
null
null
32
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
7,030
CheeLi03/whisper-base-fa-puct-5k
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.43.4
null
null
6
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
7,031
CheeLi03/whisper-base-he-puct-5k
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.46.2
null
null
6
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
7,032
CheeLi03/whisper-base-th-puct-5k
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.46.2
null
null
6
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
7,033
CheeLi03/whisper-base-tr-8
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.43.4
null
null
6
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
7,034
CheeLi03/whisper-base-vi-puct-4k
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.43.4
null
null
6
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
7,035
CheeLi03/whisper-large-v3-ru-pl2
whisper
["WhisperForConditionalGeneration"]
51866
float32
4.44.0
null
null
32
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
7,036
CheeLi03/whisper-tiny-it-puct-5k
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.43.4
null
null
4
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
7,037
CheeLi03/whisper-tiny-it-puct-combine-5k
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.46.2
null
null
4
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
7,038
CheeLi03/whisper-tiny-pt-puct-2k
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.46.2
null
null
4
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
7,039
Cheekydave/PPO-LL2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,040
CheesyChank/SpaceMouse_DNABert
bert
["BertForMaskedLM"]
4101
float32
4.34.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,041
CheesyChank/ZebraFish_DNABert
bert
["BertForMaskedLM"]
4101
float32
4.34.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,042
Cheezedog/commentbot
gpt2
["GPT2LMHeadModel"]
50259
float32
4.32.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
7,043
Cheezedog/gpt2convosbot
gpt2
["GPT2LMHeadModel"]
50259
float32
4.32.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
7,044
Cheezedog/gpt2convosbot2
gpt2
["GPT2LMHeadModel"]
50259
float32
4.32.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
7,045
Chelo11/Martin-Fierro
gpt2
["GPT2LMHeadModel"]
50257
float32
4.33.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
7,046
Chelo11/bert_adaptation_peppa_pig
bert
["BertForMaskedLM"]
31002
float32
4.33.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,047
Chelomo/summarizer_MediQA
bart
["BartForConditionalGeneration"]
50265
float32
4.26.1
null
null
6
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
7,048
ChelseaZi/gptapp
qwen2
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,049
Chemsseddine/Donut_isolant2
donut
["DonutModel"]
null
float32
4.29.2
null
null
null
null
null
null
null
null
2,048
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
medium
7,050
Chemsseddine/audio_class-finetuned
wav2vec2
["Wav2Vec2ForSequenceClassification"]
32
float32
4.29.0.dev0
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
7,051
Chemsseddine/bert-base-cased-finetuned-DOP
bert
["BertForSequenceClassification"]
28996
float32
4.22.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,052
Chemsseddine/bert-base-cased-finetuned-imdb
bert
["BertForSequenceClassification"]
28996
float32
4.22.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,053
Chen9999/llama-2-7b-miniguanaco
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
7,054
ChenDRAG/Qwen2.5-1.5B-Open-R1-Distill
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.49.0.dev0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
7,055
ChenKruphman/dummy-model
camembert
["CamembertForMaskedLM"]
32005
float32
4.44.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
5
6
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,056
ChenMnZ/Mixtral-8x7B-v0.1-OmniQuantv2-w4a16g128
mixtral
["MixtralForCausalLM"]
32000
float16
4.36.0
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
8
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
very_long
7,057
ChenShawn/DeepEyes-7B
qwen2_5_vl
["Qwen2_5_VLForConditionalGeneration"]
152064
bfloat16
4.51.3
3,584
18,944
28
28
4
silu
0
True
128,000
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
7,058
ChenWeiLi/MedPhi-3-mini_v1
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.40.1
3,072
8,192
32
32
32
silu
0
True
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
7,059
Cheng98/deberta-v3-base
deberta-v2
["DebertaV2Model"]
128100
float32
4.41.1
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,060
Cheng98/deberta-v3-large
deberta-v2
["DebertaV2Model"]
128100
float32
4.41.1
1,024
4,096
24
16
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
7,061
Cheng98/llama-160m
llama
["LlamaForCausalLM"]
32000
float32
4.30.0
768
3,072
12
12
null
silu
null
True
2,048
null
0.000001
0.02
0
2
false
64
null
false
null
null
true
false
false
false
false
false
false
false
false
false
0.084935
small
medium
7,062
Cheng98/llama-160m-qqp
llama
["LlamaForSequenceClassification"]
32000
float32
4.31.0
768
3,072
12
12
12
silu
null
True
2,048
null
0.000001
0.02
1
2
false
64
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.084935
small
medium
7,063
Cheng98/llama-160m-sst2
llama
["LlamaForSequenceClassification"]
32000
float32
4.31.0
768
3,072
12
12
12
silu
null
True
2,048
null
0.000001
0.02
1
2
false
64
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.084935
small
medium
7,064
Chenguoz/CAIG-Prompt-Model-Pre-trained
llava_llama
["LlavaLlamaForCausalLM"]
32000
bfloat16
4.38.1
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
7,065
ChengyouJia/AgentToken-8B
internvl_chat
["InternVLChatModel"]
null
bfloat16
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
7,066
ChengyouJia/ChatGen-Base-2B
internvl_chat
["InternVLChatModel"]
null
bfloat16
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
7,067
ChengzhiMu/ast-finetuned-audioset-10-10-0.4593-finetuned-gtzan
audio-spectrogram-transformer
["ASTForAudioClassification"]
null
float32
4.52.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
7,068
ChengzhiMu/wav2vec2-base-960h-finetuned-gtzan
wav2vec2
["Wav2Vec2ForSequenceClassification"]
32
float32
4.53.0.dev0
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
7,069
Chenhangcui/Fisao_instructblip
llava_llama
["LlavaLlamaForCausalLM"]
32000
float16
4.37.2
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
7,070
Chensk/distilbert-base-uncased-finetuned-emotion
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.40.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
7,071
Chenxi-Chelsea-Liu/whisper-small-noisy-hi
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.37.0.dev0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
7,072
Chenxi-Chelsea-Liu/whisper-small-yue
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.37.0.dev0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
7,073
Chenxi-Chelsea-Liu/whisper-small-yue-fold-1
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.37.0.dev0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
7,074
Chenxin0903/lab1_finetuning
marian
["MarianMTModel"]
59514
float32
4.35.2
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
7,075
Chenxin0903/lab1_random
marian
["MarianMTModel"]
59514
float32
4.35.2
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
7,076
Chenyuwen/welm-moe-32B-random
welm_moe
["WeLMMoeForCausalLM"]
102400
bfloat16
4.42.4
2,560
1,920
40
20
4
silu
0
True
16,384
10,000
0.00001
0.02
null
[2]
false
128
5
false
null
8
false
false
false
false
false
false
false
false
true
true
3.145728
medium
long
7,077
Chenzheng402/Qwen1.5B-Distill-SFT-2-GRPO-gsm8k1
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.49.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
7,078
Chenzheng402/Qwen1.5B-GRPO-gsm8k-4
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.49.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
7,079
Chenzhou/Whisper-zh-HK
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.26.0.dev0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
7,080
Chesscorner/git-chess-v2
git
["GitForCausalLM"]
30522
float32
4.47.0.dev0
768
3,072
6
12
null
gelu
null
True
1,024
null
null
0.02
101
102
false
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.042467
small
short
7,081
Chessmen/fine_tune_marian-finetuned-kde4-en-to-fr
marian
["MarianMTModel"]
59514
float32
4.44.2
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
7,082
Chessmen/mt5-small-finetuned-amazon-en-es
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.42.4
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
7,083
Chetan007/gpt2-sonnet-generators
gpt_neo
["GPTNeoForCausalLM"]
50259
float32
4.28.1
768
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
7,084
Chetan1997/layoutlmv2-finetuned-funsd-test
layoutlmv2
["LayoutLMv2ForTokenClassification"]
30522
float32
4.20.0.dev0
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
7,085
ChetanHirapara/distilbert-base-uncased-distilled-squad-finetuned-v3
distilbert
["DistilBertForQuestionAnswering"]
30522
float32
4.49.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
7,086
Chetanreddy1/math-reasoning-llama3-3b-4bit-gguf
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
7,087
Chetna19/albert-base-v2_qa_model
albert
["AlbertForQuestionAnswering"]
30000
float32
4.28.0
768
3,072
12
12
null
gelu_new
null
null
512
null
null
0.02
2
3
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,088
Chetna19/bart-large-finetuned-squadv1_qa_model
bart
["BartForQuestionAnswering"]
50265
float32
4.28.0
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
7,089
Chetna19/bert-base-cased-squad2_qa_model
bert
["BertForQuestionAnswering"]
28996
float32
4.28.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,090
Chetna19/bert-large-uncased-whole-word-masking-finetuned-squad_qa_model
bert
["BertForQuestionAnswering"]
30522
float32
4.28.0
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
7,091
Chetna19/distilbert-base-uncased-distilled-squad_qa_model
distilbert
["DistilBertForQuestionAnswering"]
30522
float32
4.28.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
7,092
Chetna19/distilbert_base_uncased_distilled_squad_qa_model
distilbert
["DistilBertForQuestionAnswering"]
30522
float32
4.28.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
7,093
Chetna19/roberta-base-squad2_qa_model
roberta
["RobertaForQuestionAnswering"]
50265
float32
4.28.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,094
ChettaniyaV/DeepChemoBERTA
roberta
["RobertaForMaskedLM"]
496
float32
4.41.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,095
ChettaniyaV/RoBertaChem
roberta
["RobertaForSequenceClassification"]
50265
float32
4.41.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
7,096
Cheuk-Ki/Platypus-finetune
bloom
["BloomForCausalLM"]
250880
float32
4.38.0.dev0
1,536
null
null
null
null
null
0
True
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,097
Cheuk-Ki/temp1
bloom
["BloomForCausalLM"]
250880
float32
4.38.0.dev0
1,536
null
null
null
null
null
0
True
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
7,098
ChevalierJoseph/typtop5
mistral
["MistralForCausalLM"]
32768
bfloat16
4.52.4
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
7,099
ChevellaShyam/model
llama
["LlamaForCausalLM"]
32000
float32
4.52.3
2,048
5,632
22
32
4
silu
0
True
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium