index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
1,900
AdnanRiaz107/CodePhi-3-mini-0.1cKlora
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.40.2
3,072
8,192
32
32
32
silu
0
True
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
1,901
AdnanRiaz107/CodePhi-3-mini-4k-instruct-python
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.40.2
3,072
8,192
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
medium
1,902
AdnanRiaz107/Gemma2b-APPS
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.38.2
2,048
16,384
18
8
1
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
1,903
AdnanRiaz107/santacoder-finetuned-the-stack-bash
roberta
["RobertaForCausalLM"]
50265
float32
4.35.0
768
3,072
12
12
null
gelu
null
False
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,904
AdnanRiaz107/testm1
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.40.2
3,072
8,192
32
32
32
silu
0
True
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
1,905
AdnaneElmar/gpt2-medquad-finetuned
gpt2
["GPT2LMHeadModel"]
50257
float32
4.46.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,906
AdoHakurei/sd-class-butterflies-32
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,907
AdoHakurei/sd-class-butterflies-64
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,908
AdonaiHS/Pyramids-Training
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,909
AdonaiHS/distilhubert-finetuned-gtzan
hubert
["HubertForSequenceClassification"]
32
float32
4.32.0
768
3,072
2
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.014156
small
null
1,910
AdoraRL/Qwen2.5-VL-7B-Geo3k-ADORA-v0
qwen2_5_vl
["Qwen2_5_VLForConditionalGeneration"]
152064
bfloat16
4.49.0
3,584
18,944
28
28
4
silu
0
False
128,000
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
1,911
Adrian-Emanuel/fake_news_classifier
bert
["BertForSequenceClassification"]
32000
float32
4.47.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
true
0.084935
small
short
1,912
Adrian/distilbert-base-uncased-finetuned-emotion
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.18.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
1,913
Adrian/distilbert-base-uncased-finetuned-squad
distilbert
["DistilBertForQuestionAnswering"]
30522
float32
4.18.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
1,914
Adrian/distilbert-base-uncased-finetuned-squad-colab
distilbert
["DistilBertForQuestionAnswering"]
30522
float32
4.18.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
1,915
AdrianFarmadin/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,916
AdrianKlessa/ModernBERT-base-ag-news
modernbert
["ModernBertForSequenceClassification"]
50368
float32
4.48.0.dev0
768
1,152
22
12
null
null
0
null
8,192
null
null
0.02
50281
50282
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
true
0.155714
small
long
1,917
AdrianPerez3/Ejercicio2_Prioridad_Optim
roberta
["RobertaForSequenceClassification"]
50265
float32
4.51.3
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,918
Adriana213/distilbert-base-uncased-finetuned-clinc
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.41.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
1,919
Adriano2024/bert-finetuned-ner
bert
["BertForTokenClassification"]
28996
float32
4.46.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,920
Adriatogi/segformer-b0-finetuned-segments-graffiti
segformer
["SegformerForSemanticSegmentation"]
null
float32
4.38.2
null
null
null
[1, 2, 5, 8]
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,921
Adrielram/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,922
AdrienB134/all-mpnet-base-v2-fr
mpnet
["MPNetModel"]
30527
float32
4.41.1
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
1,923
AdrienB134/parler-tts-mini-v1-Jenny-colab
parler_tts
["ParlerTTSForConditionalGeneration"]
32128
float16
4.43.3
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
1,924
AdrienVannson/final-model
opt
["OPTForCausalLM"]
50272
float32
4.40.2
768
null
12
12
null
null
0
True
2,048
null
null
null
2
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
medium
1,925
Adrienkgz/ppo-SnowballTarget
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,926
AdryKab47/Llama-2-7b-4bit-FT-GPTQ
llama
["LlamaForCausalLM"]
32000
float16
4.37.0.dev0
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
1,927
AdryKab47/llamaft
llama
["LlamaForCausalLM"]
32000
float16
4.37.0.dev0
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
1,928
Ads36/distilbert_big
distilbert
["DistilBertForTokenClassification"]
119547
float32
4.49.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
1,929
Ads36/distilbert_germanic
distilbert
["DistilBertForTokenClassification"]
119547
float32
4.49.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
1,930
Ads36/distilbert_slavic
distilbert
["DistilBertForTokenClassification"]
119547
float32
4.49.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
1,931
Adster/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,932
Adster/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,933
Adun/openthaigpt-1.0.0-7b-chat-beta-gptq-4bit
llama
["LlamaForCausalLM"]
56554
float16
4.32.1
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
1,934
Adun/openthaigpt-1.0.0-7b-chat-pantip
llama
["LlamaForCausalLM"]
41088
float16
4.41.2
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
1,935
Advaith1612/bart_convo-summarizer
bart
["BartForConditionalGeneration"]
50264
float32
4.41.2
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
1,936
Advantech-EIOT/intel_llama-3.1-8b-instruct
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
1,937
AdversarialRLHF/ppo_pythia410m_tldr6.9b_rm410mdata_mergedsft_propprefix
gpt_neox
["GPTNeoXForCausalLM"]
50304
float32
4.50.3
1,024
4,096
24
16
null
gelu
0
True
2,048
10,000
null
0.02
0
0
false
64
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.30199
small
medium
1,938
AdversarialRLHF/pythia410m-rm-tldr6.9b_logprobcondpropallprefix
gpt_neox
["GPTNeoXForSequenceClassification"]
50304
float32
4.50.3
1,024
4,096
24
16
null
gelu
0
True
2,048
10,000
null
0.02
0
0
false
64
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.30199
small
medium
1,939
AdversarialRLHF/pythia410m-rm-tldr6.9b_logprobcondpropprefix
gpt_neox
["GPTNeoXForSequenceClassification"]
50304
float32
4.50.3
1,024
4,096
24
16
null
gelu
0
True
2,048
10,000
null
0.02
0
0
false
64
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.30199
small
medium
1,940
AdversarialRLHF/pythia410m-sft-tldr
gpt_neox
["GPTNeoXForCausalLM"]
50304
float32
4.50.3
1,024
4,096
24
16
null
gelu
0
True
2,048
10,000
null
0.02
0
0
false
64
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.30199
small
medium
1,941
AdversarialRLHF/pythia410m-sft-tldr-allprefix
gpt_neox
["GPTNeoXForCausalLM"]
50304
float32
4.50.3
1,024
4,096
24
16
null
gelu
0
True
2,048
10,000
null
0.02
0
0
false
64
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.30199
small
medium
1,942
Adwait1199/bert-finetuned-ner-adwait
bert
["BertForTokenClassification"]
28996
null
4.26.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,943
AdwayK/base_bert_tuned_on_TAC2017_as_NER
bert
["BertForTokenClassification"]
30522
null
4.18.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,944
AdwayK/biobert_ncbi_disease_ner_tuned_on_TAC2017
bert
["BertForTokenClassification"]
28996
null
4.18.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,945
AdwayK/biobert_on_ADR_as_NER
bert
["BertForTokenClassification"]
30522
null
4.18.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,946
AdwayK/hugging_face_biobert_MLMA
bert
["BertForTokenClassification"]
30522
null
4.18.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,947
AdwayK/hugging_face_biobert_MLMAv2
bert
["BertForTokenClassification"]
30522
null
4.18.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,948
AdwayK/hugging_face_biobert_MLMAv3
bert
["BertForTokenClassification"]
30522
null
4.18.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,949
Adyatoni/wav2vec2-base-timit-demo
wav2vec2
["Wav2Vec2ForCTC"]
36
float32
4.35.2
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
1,950
Adyyyyyyyyyyyy/bert-finetuned-squad
bert
["BertForQuestionAnswering"]
28996
float32
4.48.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,951
Adzka/test-reward-model
roberta
["RobertaForSequenceClassification"]
50265
float32
4.38.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,952
Aeala/Alpaca-elina-65b-4bit
llama
["LlamaForCausalLM"]
32000
float16
4.29.0.dev0
8,192
22,016
80
64
null
silu
null
True
2,048
null
0.00001
0.02
1
2
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
64.424509
large
medium
1,953
Aeala/Enterredaas-33b
llama
["LlamaForCausalLM"]
32000
float16
4.30.0.dev0
6,656
17,920
60
52
null
silu
null
True
2,048
null
0.000001
0.02
1
2
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
31.897682
large
medium
1,954
Aeala/GPT4-x-AlpacaDente-30b-4bit
llama
["LlamaForCausalLM"]
32016
float16
4.29.0.dev0
6,656
17,920
60
52
null
silu
null
True
2,048
null
0.000001
0.02
1
2
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
31.897682
large
medium
1,955
Aeala/VicUnlocked-alpaca-65b-4bit
llama
["LlamaForCausalLM"]
32000
float16
4.30.0.dev0
8,192
22,016
80
64
null
silu
null
True
2,048
null
0.00001
0.02
1
2
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
64.424509
large
medium
1,956
Aedelon/a2c-PandaReachDense-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
1,957
AeglosAI/aeglos-v1
deberta-v2
["DebertaV2ForSequenceClassification"]
128100
float32
4.36.1
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,958
Aeoluze/Qwen2-7B-Ero-Instruct
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.41.2
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
1,959
AeroAlly/ppo-Lunalander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,960
AescF/hubert-base-ls960-finetuned-common_language
hubert
["HubertForSequenceClassification"]
32
float32
4.33.2
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
null
1,961
Aethan-Krug/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,962
AetherResearch/Cerebrum-1.0-7b
mistral
["MistralForCausalLM"]
32000
float16
4.37.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
1,963
AetherResearch/Cerebrum-1.0-8x7b
mixtral
["MixtralForCausalLM"]
32000
float16
4.39.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
8
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
very_long
1,964
Aexeos/camembert-base-finetuned-allocine
camembert
["CamembertForMaskedLM"]
32005
float32
4.41.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
5
6
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,965
Aexyno/flan-t5-small-samsum
t5
["T5ForConditionalGeneration"]
32128
float32
4.32.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,966
Afa912/distilbert-base-uncased-finetuned-cola
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.38.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
1,967
Afinucci/my_awesome_qa_model
distilbert
["DistilBertForQuestionAnswering"]
30522
float32
4.26.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
1,968
Afiqa/Bart-finetune-naturalScience
bart
["BartForConditionalGeneration"]
50264
float32
4.31.0
null
null
12
null
null
null
0
False
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
1,969
Afiqa/Bart-naturalScience
bart
["BartForConditionalGeneration"]
50265
float32
4.31.0
null
null
12
null
null
null
0.1
False
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
1,970
Afiqa/Llama-2-7b-chat-dental
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
1,971
Afiqa/Llama-2-7b-chat-medical
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
1,972
Afiqa/Llama-2-7b-chat-mqa
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
1,973
AfnanTS/results
bert
["BertForMaskedLM"]
100000
float32
4.27.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,974
AfterRain007/results
roberta
["RobertaForSequenceClassification"]
50414
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,975
Afterglow777/chemical-llama
llama
["LlamaForCausalLM"]
32000
bfloat16
4.37.1
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
1,976
Afterglow777/chemical_dpo_model
llama
["LlamaForCausalLM"]
32000
bfloat16
4.37.1
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
1,977
Afzal12345/my-chatbot-model
bert
["BertForSequenceClassification"]
30522
float32
4.52.4
256
1,024
4
4
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.003146
small
short
1,978
Agals/distilgpt2_eli5
gpt2
["GPT2LMHeadModel"]
50257
float32
4.41.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,979
Aganith/aspect-based-sentiment-analyser
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
1,980
Agastaya/my_awesome_billsum_model
t5
["T5ForConditionalGeneration"]
32128
float32
4.38.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,981
AgentGym/AgentEvol-7B
llama
["LlamaForCausalLM"]
32000
bfloat16
4.35.2
4,096
11,008
32
32
32
silu
null
False
2,048
10,000
0.000001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
1,982
AgentPublic/fabrique-reference-2
llama
["LlamaForCausalLM"]
32000
bfloat16
4.33.2
5,120
13,824
40
40
40
silu
null
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
1,983
AgentPublic/guillaumetell-7b
mistral
["MistralForCausalLM"]
32002
bfloat16
4.37.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
1,984
AggaMin/llama-3-8b-Instruct-bnb-4bit-aiaustin-demo
llama
["LlamaForCausalLM"]
128256
float16
4.42.3
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
1,985
Agita/DistilBERT_test
roberta
["RobertaModel"]
50265
float32
4.41.1
768
3,072
6
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.042467
small
short
1,986
Agneev/distilhubert-finetuned-gtzan
hubert
["HubertForSequenceClassification"]
32
float32
4.32.0.dev0
768
3,072
2
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.014156
small
null
1,987
Agneev/speecht5_tts_voxpopuli_nl
speecht5
["SpeechT5ForTextToSpeech"]
81
float32
4.33.0.dev0
768
null
null
null
null
gelu
0.1
False
null
null
null
0.02
0
2
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,988
Agniruudrra/my_ebm_model_biobert
distilbert
["DistilBertForTokenClassification"]
119547
float32
4.28.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
1,989
Agnuxo/Llama-3.1-Minitron-4B-Instruct_CODE_Python-GGUF_Spanish_English_8bit
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
1,990
Agnuxo/Llama-3.1-Minitron-4B-Instruct_CODE_Python_Spanish_English_Asistant-16bit-v2
llama
["LlamaForCausalLM"]
128256
bfloat16
4.43.4
4,096
14,336
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
3.221225
medium
very_long
1,991
Agnuxo/Mamba-Codestral-7B-Instruct_CODE_Python-GGUF_Spanish_English_8bit
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
1,992
Agnuxo/Mamba-Codestral-7B-Instruct_CODE_Python-Spanish_English_GGUF_4bit
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
1,993
Agnuxo/Mamba-Codestral-7B-Instruct_CODE_Python_Spanish_English_16bit
mistral
["MistralForCausalLM"]
32768
bfloat16
4.43.4
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
1,994
Agnuxo/Mamba-Codestral-7B-v0.1-instruct-python_coding_assistant-GGUF_16bit
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
1,995
Agnuxo/Mamba-Codestral-7B-v0.1-instruct-python_coding_assistant-GGUF_4bit
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
1,996
Agnuxo/Mamba-Codestral-7B-v0.1-python_coding_assistant-GGUF_8bit
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
1,997
Agnuxo/Mamba-Codestral-7B-v0.1-python_coding_assistant_16bit
mistral
["MistralForCausalLM"]
32768
bfloat16
4.44.2
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
1,998
Agnuxo/Phi-3.5-mini-instruct-python_coding_assistant-GGUF_16bit
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
1,999
Agnuxo/Phi-3.5-mini-instruct-python_coding_assistant-GGUF_4bit
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null