index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
1,800
AdamKasumovic/phi3-mini-4k-instruct-bactrian-x-xh-50-percent-med-high-bt-rouge-1
mistral
["MistralForCausalLM"]
32064
bfloat16
4.41.2
3,072
8,192
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
true
false
false
true
3.623879
medium
medium
1,801
AdamKasumovic/phi3-mini-4k-instruct-bactrian-x-xh-75-percent-high-bt-rouge-1
mistral
["MistralForCausalLM"]
32064
bfloat16
4.41.2
3,072
8,192
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
true
false
false
true
3.623879
medium
medium
1,802
AdamKasumovic/phi3-mini-4k-instruct-bactrian-x-xh-75-percent-low-bt-rouge-1
mistral
["MistralForCausalLM"]
32064
bfloat16
4.41.2
3,072
8,192
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
true
false
false
true
3.623879
medium
medium
1,803
AdamKasumovic/phi3-mini-4k-instruct-bactrian-x-xh-75-percent-low-high-bt-rouge-1
mistral
["MistralForCausalLM"]
32064
bfloat16
4.41.2
3,072
8,192
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
true
false
false
true
3.623879
medium
medium
1,804
AdamKasumovic/phi3-mini-4k-instruct-bactrian-x-xh-75-percent-low-med-bt-rouge-1
mistral
["MistralForCausalLM"]
32064
bfloat16
4.41.2
3,072
8,192
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
true
false
false
true
3.623879
medium
medium
1,805
AdamKasumovic/phi3-mini-4k-instruct-bactrian-x-xh-75-percent-low-med-high-bt-rouge-1
mistral
["MistralForCausalLM"]
32064
bfloat16
4.41.2
3,072
8,192
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
true
false
false
true
3.623879
medium
medium
1,806
AdamKasumovic/phi3-mini-4k-instruct-bactrian-x-xh-75-percent-med-bt-rouge-1
mistral
["MistralForCausalLM"]
32064
bfloat16
4.41.2
3,072
8,192
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
true
false
false
true
3.623879
medium
medium
1,807
AdamKasumovic/phi3-mini-4k-instruct-mmlu-college-medicine-af-mmlu-mmlu-low
mistral
["MistralForCausalLM"]
32064
bfloat16
4.42.3
3,072
8,192
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
true
false
false
true
3.623879
medium
medium
1,808
AdamLucek/Orpo-Llama-3.2-1B-40k
llama
["LlamaForCausalLM"]
128258
float16
4.45.2
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128256
128257
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
1,809
AdamLucek/Phi-3-mini-EmoMarketing-DELLA
phi3
["Phi3ForCausalLM"]
32011
float16
4.44.0
3,072
8,192
32
32
32
silu
0
True
131,072
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
1,810
AdamLuwiko/gpt2-chatbot-voucheran
gpt2
["GPT2LMHeadModel"]
50257
float32
4.48.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,811
AdamLuwiko/my-chatbot-model
gpt2
["GPT2LMHeadModel"]
50257
float32
4.48.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,812
AdamRTomkins/phi-kal
phi
["PhiForCausalLM"]
51200
float16
4.39.0.dev0
2,048
8,192
24
32
32
gelu_new
0
False
2,048
10,000
null
0.02
null
null
false
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
1,813
Adammz/cs_roberta_base-1
roberta
["RobertaForSequenceClassification"]
50265
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,814
Adammz/ruBert-base-1-4
bert
["BertForSequenceClassification"]
120138
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,815
Adammz/ruBert-base-1-third
bert
["BertForSequenceClassification"]
120138
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,816
Adammz/rubert-base-cased-1-third
bert
["BertForSequenceClassification"]
119547
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,817
Adammz/rubert-tiny2-1-4
bert
["BertForSequenceClassification"]
83828
float32
4.35.2
312
600
3
12
null
gelu
null
True
2,048
null
null
0.02
null
null
null
26
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.003504
small
medium
1,818
AdanLee/distilhubert-finetuned-gtzan
hubert
["HubertForSequenceClassification"]
32
float32
4.32.0
768
3,072
2
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.014156
small
null
1,819
AdanLee/whisper-tiny-minds14-en
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.32.0
null
null
4
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
1,820
Adansonia/internal_audit_new
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.3
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128040
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
1,821
AdaptLLM/biomed-InternVL3-1B
internvl_chat
["InternVLChatModel"]
null
bfloat16
null
896
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
1,822
AdaptLLM/biomed-Qwen2.5-VL-3B-Instruct
qwen2_5_vl
["Qwen2_5_VLForConditionalGeneration"]
151936
bfloat16
4.51.3
2,048
11,008
36
16
2
silu
0
False
128,000
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
1,823
AdaptLLM/biomed-gemma-3-4b-it
gemma3
["Gemma3ForConditionalGeneration"]
null
bfloat16
4.51.3
2,560
null
null
null
null
null
null
False
null
null
null
0.02
null
[1, 106]
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
1,824
AdaptLLM/finance-LLM
llama
["LLaMAForCausalLM"]
32001
float16
4.27.0.dev0
4,096
11,008
32
32
null
silu
null
True
null
null
0.000001
0.02
0
1
null
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
6.442451
medium
null
1,825
AdaptLLM/finance-chat
llama
["LlamaForCausalLM"]
32001
float16
4.31.0.dev0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
1,826
AdaptLLM/food-Qwen2.5-VL-3B-Instruct
qwen2_5_vl
["Qwen2_5_VLForConditionalGeneration"]
151936
bfloat16
4.51.3
2,048
11,008
36
16
2
silu
0
False
128,000
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
1,827
AdaptLLM/medicine-chat
llama
["LlamaForCausalLM"]
32001
float16
4.31.0.dev0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
1,828
AdaptLLM/remote-sensing-LLaVA-NeXT-Llama3-8B
llava_next
["LlavaNextForConditionalGeneration"]
128257
float16
4.39.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
1,829
AdaptLLM/remote-sensing-Qwen2-VL-2B-Instruct
qwen2_vl
["Qwen2VLForConditionalGeneration"]
151936
bfloat16
4.45.0.dev0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
1,830
AdaptLLM/remote-sensing-Qwen2.5-VL-3B-Instruct
qwen2_5_vl
["Qwen2_5_VLForConditionalGeneration"]
151936
bfloat16
4.51.3
2,048
11,008
36
16
2
silu
0
False
128,000
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
1,831
AdaptLLM/visual-instruction-synthesizer
llava_next
["LlavaNextForConditionalGeneration"]
128257
float16
4.39.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
1,832
AdapterOcean/expert_53c4f9908e134f65867c5b72c836d718
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,833
AdapterOcean/expert_8758633174e848f1a94fa8449fe4c5c2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,834
AdapterOcean/expert_a6ae28522de54fbf9bdf7ae624dde68c
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,835
AdapterOcean/expert_e6e205c66b39455bafc9c960f5d2da5e
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,836
Adapting/bert-base-chinese-finetuned-NER-biomedical
bert
["BertForTokenClassification"]
21128
float32
4.22.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,837
Adarsh-aot/ai_class
bart
["BartForConditionalGeneration"]
50264
float32
4.38.2
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
1,838
Adarsh-aot/ai_poet
bart
["BartForConditionalGeneration"]
50264
float32
4.38.2
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
1,839
Adarsh203/new_mT5_Sum
mt5
["MT5ForConditionalGeneration"]
250112
float32
4.37.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,840
Adbhut/gpt2_ft-ai4privacy-open-pii-masking-500k-ai4privacy
gpt2
["GPT2LMHeadModel"]
50261
float32
4.52.4
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,841
Adbhut/speecht5-finetuned-voxpopuli_sl
speecht5
["SpeechT5ForTextToSpeech"]
81
float32
4.33.2
768
null
null
null
null
gelu
0.1
False
null
null
null
0.02
0
2
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,842
Addaci/bert-base-multilingual-cased-finetuned-yiddish-experiment-2
bert
["BertForMaskedLM"]
119547
float32
4.47.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,843
Addaci/bert-base-multilingual-cased-finetuned-yiddish-experiment-3
bert
["BertForMaskedLM"]
119547
float32
4.47.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,844
AddieFoote0/arithmetic-.3b-reference
gemma2
["Gemma2ForCausalLM"]
256000
float32
4.49.0
768
3,072
14
8
4
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
96
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.09909
small
long
1,845
AddieFoote0/arithmetic-300M-MaxEnt
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.49.0
768
3,072
14
8
4
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
96
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.09909
small
long
1,846
AddieFoote0/arithmetic-300M-reference
gemma2
["Gemma2ForCausalLM"]
256000
float32
4.49.0
768
3,072
14
8
4
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
96
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.09909
small
long
1,847
AddieFoote0/language-100M-MaxEnt
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.49.0
320
1,280
14
8
4
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
40
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.017203
small
long
1,848
AddieFoote0/language-100M-MaxEnt-distilled
gemma2
["Gemma2ForCausalLM"]
256000
float32
4.49.0
320
1,280
14
8
4
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
40
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.017203
small
long
1,849
AddieFoote0/language-100M-MaxEnt-distilled-relearned
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.49.0
320
1,280
14
8
4
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
40
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.017203
small
long
1,850
AddieFoote0/language-100M-MaxEnt-relearned
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.49.0
320
1,280
14
8
4
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
40
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.017203
small
long
1,851
Addy2903/gita-text-generation-gpt2
gpt2
["GPT2LMHeadModel"]
50257
float32
4.46.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,852
Ade-Pyaar/eng_fr_model
marian
["MarianMTModel"]
59514
float32
4.47.1
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
1,853
Adebolajo/llama3-instruction-AWQ
null
[]
128256
null
null
4,096
14,336
32
32
8
silu
null
null
8,192
null
null
null
null
null
null
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
6.442451
medium
long
1,854
Adekiitan11/bert-email-classifier
bert
["BertForSequenceClassification"]
30522
float32
4.49.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,855
Adelina111/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
bert
["BertForSequenceClassification"]
30522
float32
4.42.4
384
1,536
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
32
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.021234
small
short
1,856
Adeptschneider/dyu-fr-opus-v1.0
marian
["MarianMTModel"]
59514
null
4.38.2
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
1,857
Adeptschneider/dyu-fr-opus-v2.0
marian
["MarianMTModel"]
59514
null
4.38.2
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
1,858
Adeptschneider/dyu_to_fr_v2.0
t5
["T5ForConditionalGeneration"]
32128
float32
4.43.0.dev0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,859
Adeptschneider/dyu_to_fr_v4.0
t5
["T5ForConditionalGeneration"]
32128
float32
4.43.0.dev0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,860
Adeptschneider/opus-mt-dyu-fr
marian
["MarianMTModel"]
59514
null
4.38.2
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
1,861
AdhamEhab/ImageCaptioning
blip
["BlipForConditionalGeneration"]
null
float32
4.40.2
null
null
null
null
null
null
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,862
AdhamEhab/StoryGen
gpt2
["GPT2LMHeadModel"]
50266
float32
4.40.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,863
Adhithyamadesh2001/DeepSeek-R1-ApmChatbot_tanglish
llama
["LlamaForCausalLM"]
128256
float16
4.51.1
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
1,864
Adi-25/phi-3-mini-instruct-openvino-int4-model
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.49.0
3,072
8,192
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
medium
1,865
Adignite/LawRL_NewDF_Flant5_PPO
t5
["T5ForConditionalGeneration"]
32128
float32
4.41.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,866
Adignite/MailSense_Classifier-chat-llama7b
llama
["LlamaForCausalLM"]
32000
float16
4.33.1
4,096
11,008
32
32
32
silu
null
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
1,867
Adiii02/medicine-gpt_llama2_7B_qlora
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
1,868
Adiii143/bert-finetuned-ner
bert
["BertForTokenClassification"]
28996
float32
4.48.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,869
AdilHayat173/token_classification
bert
["BertForTokenClassification"]
28996
float32
4.42.4
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,870
Adilmar/test_trainer
bert
["BertForSequenceClassification"]
28996
float32
4.30.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,871
Adipta/audio_classification
wav2vec2
["Wav2Vec2ForSequenceClassification"]
32
float32
4.42.4
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
1,872
Adipta/image_classification
vit
["ViTForImageClassification"]
null
float32
4.42.4
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
1,873
AdirK/ppo-SnowballTarget
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,874
Adish14/llama3-8b-instruct-raft-merged-fullmodel-6056-v2
llama
["LlamaForCausalLM"]
128256
float16
4.51.3
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
1,875
AditKaushik/Bert_scratch1
bert
["BertForMaskedLM"]
30522
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,876
AdithyaSK/LLama3-Gaja-Hindi-8B-Instruct-alpha
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0.dev0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
1,877
AdithyaSK/LLama3Tokenizer
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.0.dev0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
1,878
Aditya0619/Podcaster_phi4
llama
["LlamaForCausalLM"]
100352
float16
4.47.1
5,120
17,920
40
40
10
silu
0
True
16,384
250,000
0.00001
0.02
100257
100265
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
12.582912
large
long
1,879
Aditya1000101/trained_summarization_model
bart
["BartForConditionalGeneration"]
50264
float32
4.48.3
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
1,880
Aditya757864/whisper-tiny-hi
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.38.0.dev0
null
null
4
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
1,881
AdityaAKUMAR/Memma_More_Humorous
gemma2
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,882
AdityaBajracharya/ppo-lunarlander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,883
AdityaBavadekar/gpt2-medium-finetuned-qamodel
gpt2
["GPT2LMHeadModel"]
50261
float32
4.31.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
1,884
AdityaBorse11/poca-SoccerTwos
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,885
AdityaK10/xlm-roberta-base-finetuned-panx-de
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.16.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,886
AdityaK10/xlm-roberta-base-finetuned-panx-de-fr
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.16.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,887
AdityaKothari/WellnessAI-7B-5-bit
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
1,888
AdityaKumar2408/rare-puppers
vit
["ViTForImageClassification"]
null
float32
4.35.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
1,889
AdityaLavaniya/TinyLlama-Fitness-Instructor
llama
["LlamaForCausalLM"]
32000
float16
4.35.2
2,048
5,632
22
32
4
silu
null
True
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
1,890
AdityaNamuduri/t5_recommendation_sports_equipment_english
t5
["T5ForConditionalGeneration"]
32128
float32
4.26.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
1,891
AdityaNamuduri/testing
bart
["BartForConditionalGeneration"]
50265
float32
4.39.3
null
null
6
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
1,892
AdityaXPV/Llama-3-8b-Instruct-law-sage-v0.1
llama
["LlamaForCausalLM"]
128256
float16
4.40.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
1,893
Adityaganesh/Telugu_Sentiment_Analysis
bert
["BertForSequenceClassification"]
119547
float32
4.51.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,894
AdityasArsenal/finetuned-for-YogaPosesv4
mobilenet_v2
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
1,895
Adnan-AI-Labs/DistilBERT-ProductClassifier
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.46.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
1,896
AdnanRiaz107/CBertbase-APPS10k
roberta
["RobertaForCausalLM"]
50265
float32
4.38.2
768
3,072
12
12
null
gelu
null
False
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,897
AdnanRiaz107/CBertbase-mlm-APPS10k
roberta
["RobertaForCausalLM"]
50265
float32
4.38.2
768
3,072
12
12
null
gelu
null
False
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
1,898
AdnanRiaz107/CodeGemma2b-300APPS
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.40.2
2,048
16,384
18
8
1
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
1,899
AdnanRiaz107/CodeLLAMA3-8BI-300APPS
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long