index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
5,100
Babelscape/mrebel-large
mbart
["MBartForConditionalGeneration"]
250071
float32
4.19.2
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
5,101
Babelscape/mrebel-large-32
mbart
["MBartForConditionalGeneration"]
250071
float32
4.19.2
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
5,102
BabyChou/Yi-VL-34B
llava
["YiVLForCausalLM"]
64000
bfloat16
4.36.1
7,168
20,480
60
56
8
silu
null
False
4,096
5,000,000
0.00001
0.02
1
2
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
36.993761
large
medium
5,103
BachNgoH/MetaMath-Zalo-ft-Few-Shot-Augment-1000
mistral
["MistralForCausalLM"]
32001
float32
4.35.2
4,096
14,336
32
32
8
silu
null
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
5,104
BachNgoH/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,105
BachelorThesis/GerMedBERT_NER_V01_BRONCO_CARDIO
bert
["BertForTokenClassification"]
30000
float32
4.40.1
768
3,072
12
8
null
gelu
null
True
512
null
null
0.02
null
null
null
96
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,106
BachelorThesis/GerMedBert_ATTR_V02_BRONCO
bert
["BertForSequenceClassification"]
30000
float32
4.40.1
768
3,072
12
8
null
gelu
null
True
512
null
null
0.02
null
null
null
96
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,107
Bachhoang/Continuous-FT-vbd-merge
llama
["LlamaForCausalLM"]
49381
bfloat16
4.32.1
4,096
11,008
32
32
32
silu
null
False
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
5,108
Bachlee/bert-nwpredict
bert
["BertForMaskedLM"]
30522
null
4.39.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,109
Bachlee/bert-nwpredict_v2
bert
["BertForMaskedLM"]
30522
null
4.39.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,110
Bachstelze/smolSynformer
llama
["LlamaForCausalLM"]
49152
float32
4.47.0
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
0
0
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
5,111
Backdrive/xlm-roberta-base-finetuned-panx-de
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.30.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,112
Bacon666/Phenom-12B-0.1
mistral
["MistralForCausalLM"]
131072
bfloat16
4.44.1
5,120
14,336
40
32
8
silu
0
True
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
5,113
BadisNahdi/FineTuneBERT
bert
["BertForSequenceClassification"]
30522
float32
4.31.0.dev0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,114
BadreddineHug/LayoutLM_1
layoutlmv3
["LayoutLMv3ForTokenClassification"]
50265
float32
4.29.2
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
5,115
BadreddineHug/LayoutLMv3_97_2
layoutlmv3
["LayoutLMv3ForTokenClassification"]
50265
float32
4.29.2
1,024
4,096
24
16
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
short
5,116
BadreddineHug/LayoutLMv3_batch_8
layoutlmv3
["LayoutLMv3ForTokenClassification"]
50265
float32
4.29.2
1,024
4,096
24
16
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
short
5,117
BadreddineHug/LayoutLMv3_funds
layoutlmv3
["LayoutLMv3ForTokenClassification"]
50265
float32
4.29.2
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
5,118
BadreddineHug/LayoutLMv3_large_1
layoutlmv3
["LayoutLMv3ForTokenClassification"]
50265
float32
4.29.2
1,024
4,096
24
16
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
short
5,119
BadreddineHug/LayoutLMv3_large_2
layoutlmv3
["LayoutLMv3ForTokenClassification"]
50265
float32
4.29.2
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
5,120
BadreddineHug/LayoutLMv3_large_3
layoutlmv3
["LayoutLMv3ForTokenClassification"]
50265
float32
4.29.2
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
5,121
BadreddineHug/donut-base-ocr11
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,122
Badsims/sentis-phi-3.5-mini-instruct-uint8
phi3
["Phi3ForCausalLM"]
32064
null
4.43.3
3,072
8,192
32
32
32
silu
0
False
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
5,123
Baekpica/vicuna-7b-v1.3-tiny-stories-pretraining-2epoch
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
False
2,048
null
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
5,124
Baekpica/vicuna-7b-v1.3-tinystories-linear
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
False
2,048
null
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
5,125
BagelLabs/TinyToolUse-Qwen3-0.6B-toolbench
qwen3
["Qwen3ForCausalLM"]
151936
float32
4.52.4
1,024
3,072
28
16
8
silu
0
True
40,960
1,000,000
0.000001
0.02
151643
151645
true
64
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.352322
small
very_long
5,126
Bagus/distilhubert-finetuned-gtzan-base-audio-course
hubert
["HubertForSequenceClassification"]
32
float32
4.35.2
768
3,072
2
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.014156
small
null
5,127
Bagus/hubert_xlarge_emodb
hubert
["Model"]
32
float32
4.32.0
1,280
5,120
48
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
80
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.943718
small
null
5,128
Bagus/wav2v2c_swbd_emodb
wav2vec2
["Model"]
32
float32
4.32.0
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
5,129
Bagus/wav2vec2_swbd_emodb
wav2vec2
["Model"]
32
float32
4.32.0
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
5,130
Bagus/wavlm_base-plus_emodb
wavlm
["Model"]
32
float32
4.32.0
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,131
Bagus/whisper-medium-common_voice_17_0-id-10000
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.42.0.dev0
null
null
24
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
5,132
Bagus/whisper-tiny-en
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.35.2
null
null
4
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
5,133
BahAdoR0101/my_awesome_wnut_model
bert
["BertForTokenClassification"]
28996
float32
4.34.1
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
5,134
Baha123/llama3.1-text-to-sql_finetuned-fullmodel
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
5,135
Baha2107/test_identity_01
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.38.1
1,024
2,816
24
16
16
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
0.30199
small
very_long
5,136
BahaaEldin0/Mistral-7b-Mixtral-7b
mistral
["MistralForCausalLM"]
32000
float16
4.40.2
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
5,137
BahaaEldin0/bert-base-uncased
bert
["BertForSequenceClassification"]
30522
float32
4.41.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,138
BahaaEldin0/bert-base-uncased-reward-model
bert
["BertForSequenceClassification"]
30522
float32
4.41.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,139
BahaaEldin0/policy-model
gpt2
["GPT2LMHeadModel"]
50257
float32
4.41.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
5,140
Baharak/codegen-350M-mono-python-18k-alpaca
codegen
["CodeGenForCausalLM"]
51200
float16
4.36.2
null
null
null
null
null
null
null
True
null
null
null
0.02
1
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,141
Bahareh0281/BioBERT_Finetuned
bert
["BertLMHeadModel"]
28996
float32
4.47.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,142
Bahareh0281/ClinicalBERT_Finetuned
bert
["BertLMHeadModel"]
28996
float32
4.47.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,143
Bahasalab/Bahasa-4b
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.39.1
2,560
6,912
40
20
20
silu
0
False
32,768
5,000,000
0.000001
0.02
151643
151643
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.145728
medium
very_long
5,144
Bahin/lunardr
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,145
BaibhavJoshi/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,146
Baidicoot/Othello-GPT-Transformer-Lens
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,147
Bainbridge/bert-incl
bert
["BertForSequenceClassification"]
31102
float32
4.28.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,148
Bainbridge/bert-xxl-incl
bert
["BertForSequenceClassification"]
32102
float32
4.28.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,149
Bainbridge/gpt2-kl_1_03-hs_cn
gpt2
["GPT2LMHeadModel"]
50259
float32
4.28.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
5,150
Bainbridge/gpt2-kl_1_03-hs_cn_decay
gpt2
["GPT2LMHeadModel"]
50259
float32
4.28.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
5,151
Bainbridge/gpt2-kl_1_04-hs_cn
gpt2
["GPT2LMHeadModel"]
50259
float32
4.28.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
5,152
Bainbridge/gpt2-kl_1_04-hs_cn_decay
gpt2
["GPT2LMHeadModel"]
50259
float32
4.28.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
5,153
Bainbridge/gpt2-kl_1_06-hs_cn_decay
gpt2
["GPT2LMHeadModel"]
50259
float32
4.28.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
5,154
Bainbridge/gpt2-kl_1_07-hs_cn
gpt2
["GPT2LMHeadModel"]
50259
float32
4.28.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
5,155
Bajiyo/malayalam_imasc
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.37.2
null
null
12
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
5,156
Bajiyo/w2v-bert-2.0-malayalam-CV16.0
wav2vec2-bert
["Wav2Vec2BertForCTC"]
74
float32
4.39.3
1,024
4,096
24
16
null
swish
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
null
5,157
Bajiyo/w2v-bert-2.0-malayalam_mixeddataset-CV16.0
wav2vec2-bert
["Wav2Vec2BertForCTC"]
77
float32
4.39.3
1,024
4,096
24
16
null
swish
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
null
5,158
Bakasura/wav2vec2-base-finetuned-ks
wav2vec2
["Wav2Vec2ForSequenceClassification"]
32
float32
4.25.1
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,159
Bakhtiar56/w2v-bert-2.0-urdu-colab-CV13.0
wav2vec2-bert
["Wav2Vec2BertForCTC"]
83
float32
4.44.2
1,024
4,096
24
16
null
swish
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
null
5,160
Bakhtiar56/whisper-small-dv
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.42.4
null
null
12
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
5,161
BakshiSan/BERT_Twitter
bert
["BertForSequenceClassification"]
30522
float32
4.51.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,162
Baktashans/Finetuned_ParsBert_ArmanEmo
bert
["BertForSequenceClassification"]
100000
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,163
Baktashans/Finetuned_xlm_roberta_large_ArmanEmo
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.35.2
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
5,164
Bala-A87/LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,165
BalaSubrahmanyam/T5Model_for_Ecommerce
t5
["T5ForConditionalGeneration"]
32128
float32
4.40.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
5,166
BalajiNagarajan/llama2
llama
["LlamaForCausalLM"]
32000
float16
4.30.2
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
5,167
Baldion7/Skin-cancer
vit
["ViTForImageClassification"]
null
float32
4.48.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
5,168
BaleChen/checkpoint-1300_merged
llama
["LlamaForSequenceClassification"]
32000
bfloat16
4.31.0
4,096
11,008
32
32
32
silu
null
True
2,048
null
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
5,169
BaleChen/checkpoint-800_merged
llama
["LlamaForCausalLM"]
32000
bfloat16
4.31.0
4,096
11,008
32
32
32
silu
null
True
2,048
null
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
5,170
BallAd-15/gemma2-9b-task10-subtask3-EN
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.47.1
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
5,171
BallAd-15/gemma2-9b-task10-subtask3-HI
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.47.1
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
5,172
Baluchon/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,173
Bamarin/my_awesome_swag_model
bert
["BertForMultipleChoice"]
30522
float32
4.28.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,174
BanUrsus/bert-base-cased-finetuned-squad-accelerate_nlp-course-chapter7-section6
bert
["BertForQuestionAnswering"]
28996
float32
4.37.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,175
BanUrsus/bert-finetuned-ner-accelerate_NLP-course-chapter7-section1
bert
["BertForTokenClassification"]
28996
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,176
BanUrsus/bert-finetuned-ner_NLP-course-chapter7-section1
bert
["BertForTokenClassification"]
28996
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,177
BanUrsus/distilhubert-finetuned-gtzan
hubert
["HubertForSequenceClassification"]
32
float32
4.39.2
768
3,072
2
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.014156
small
null
5,178
BanUrsus/distilroberta-base-finetuned-condition-classifier
roberta
["RobertaForSequenceClassification"]
50265
float32
4.35.2
768
3,072
6
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.042467
small
short
5,179
BanUrsus/mt5-small-finetuned-cnn-dailymail-en_nlp-course-chapter7-section4
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.35.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
5,180
BanUrsus/whisper-small-dv
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.39.2
null
null
12
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
5,181
BananaFish45/Ethnicity_model
bert
["BertForSequenceClassification"]
30522
float32
4.43.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,182
BananaFish45/gdpr_personal_data
deberta-v2
["DebertaV2ForSequenceClassification"]
128100
float32
4.43.3
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,183
Bananaman272/bert-phishing-classifier_teacher
bert
["BertForSequenceClassification"]
30522
float32
4.47.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
5,184
BanczikZ/ppo-SnowballTarget
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,185
Bandid/Umi-7B-Mistral-HP-0
mistral
["MistralForCausalLM"]
32032
bfloat16
4.40.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
5,186
Bandifishing/Nous-Hermes-13b-Chinese
llama
["LlamaForCausalLM"]
49954
float16
4.28.1
5,120
13,824
40
40
null
silu
null
True
2,048
null
0.000001
0.02
1
2
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
12.582912
large
medium
5,187
Bandika/PPOPyramids
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,188
Bandika/PPOSnowballTarget
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,189
Bandika/a2c-AntBulletEnv-v0
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,190
Bandika/a2c-PandaReachDense-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,191
Bandika/rl_course_vizdoom_health_gathering_supreme
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
5,192
BanglaLLM/Bangla-s1k-qwen-2.5-32B-Instruct
qwen2
["Qwen2ForCausalLM"]
152064
float32
4.46.1
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
5,193
BanglaLLM/bangla-llama-13b-base-v0.1
llama
["LlamaForCausalLM"]
50437
bfloat16
4.31.0
5,120
13,824
40
40
40
silu
null
False
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
5,194
BanglaLLM/bangla-llama-7b-instruct-v0.1
llama
["LlamaForCausalLM"]
50437
float32
4.31.0
4,096
11,008
32
32
32
silu
null
False
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
5,195
BangorAI/Intel_neural-chat-7b-v3-1_CofnodYCynulliad_2epoch
mistral
["MistralForCausalLM"]
32000
float16
4.34.1
4,096
14,336
32
32
8
silu
null
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
5,196
BangorAI/mistral-7b-cy-epoch-1
mistral
["MistralForCausalLM"]
32000
bfloat16
4.38.2
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
5,197
BangorAI/mistral-7b-cy-tokenizer-train-6
mistral
["MistralForCausalLM"]
44955
float32
4.35.2
4,096
14,336
32
32
8
silu
null
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
5,198
BanusreeM/mt5-sawit-finetuned-v2
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.48.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
5,199
BaoHuynh2002/finetuned_Qwen2_7b_mt_history_mcqas_v1
qwen2
["Qwen2ForCausalLM"]
152064
float16
4.41.2
3,584
18,944
28
28
4
silu
0
True
131,072
1,000,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long