index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
3,900
Anwaarma/Improved-Arabert-twitter-sentiment2char
bert
["BertForSequenceClassification"]
64000
float32
4.41.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,901
Anwaarma/Improved-Arabert-twitter-sentiment2chars
bert
["BertForSequenceClassification"]
64000
float32
4.41.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,902
Anwaarma/Improved-Arabic-bert-base
bert
["BertForSequenceClassification"]
32000
float32
4.34.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,903
Anwaarma/Improved-Arabic-bert-nodropout
bert
["BertForSequenceClassification"]
32000
float32
4.34.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,904
Anwaarma/Improved-MARBERT-twitter-sentiment-Twitter
bert
["BertForSequenceClassification"]
100000
float32
4.34.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,905
Anwaarma/Improved-MARBERT-twitter-sentiment-nodroput-Twitter
bert
["BertForSequenceClassification"]
100000
float32
4.34.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,906
Anwaarma/Improved-bert-multilingual
bert
["BertForSequenceClassification"]
119547
float32
4.34.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,907
Anwaarma/Improved-bert-multilingual-nodropout
bert
["BertForSequenceClassification"]
119547
float32
4.34.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,908
Anwaarma/Improved-xlm-roberta-base
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.34.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,909
Anwaarma/Improved-xlm-roberta-base-nodroput
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.34.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,910
Anwaarma/xlm-ex2
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.49.0
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,911
AnyaBoo/finetune_Llama3.1-8b
llama
["LlamaForCausalLM"]
128256
float16
4.44.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,912
AnyaK/llama-2-7b-mini
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,913
AnyaSchen/image2music
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,914
AnyaSchen/my_awesome_model
roberta
["RobertaForSequenceClassification"]
50265
float32
4.34.1
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,915
AnyaSchen/rugpt3-medium-key2poetry
gpt2
["GPT2LMHeadModel"]
50260
float32
4.30.0.dev0
null
null
null
null
null
null
null
True
null
null
null
0.02
50258
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,916
Anybalsmith/pick_brick_policy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,917
Anyfusion/cogvideox-nf4
null
[]
null
null
null
null
null
null
48
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,918
Anyfusion/flux-turbo-nf4
null
[]
null
null
null
null
null
null
24
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
3,919
AoEiuV020/test
bert
["BertForSequenceClassification"]
30522
float32
4.52.0.dev0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,920
AoiKazama/Kinoe-7B
kinoe
["KinoeForCausalLM"]
55424
bfloat16
4.40.1
5,120
14,336
24
40
10
silu
0
True
32,768
10,000
0.000001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.549747
large
very_long
3,921
Aoishi19/autotrain-supermarket2
vit
["ViTForImageClassification"]
null
float32
4.44.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
3,922
ApacheOne/CKPTs
custom
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,923
Apala/FineLlama-3.1-8B
llama
["LlamaForCausalLM"]
128256
float16
4.46.1
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,924
AparnaVai/sarvam1-aparna
llama
["LlamaForCausalLM"]
68096
null
4.48.3
2,048
11,008
28
16
8
silu
0
True
8,192
10,000
0.000001
0.02
1
2
false
128
2
false
null
null
true
false
false
false
false
false
false
false
true
true
1.409286
medium
long
3,925
Aparnarani06/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,926
Ape50/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,927
Apel-sin/whisper-large-v3-russian-ties-podlodka-v1.2
whisper
["WhisperForConditionalGeneration"]
51866
float32
4.49.0
null
null
32
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,928
Aphasia/sd-class-butterflies-32
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,929
Apheironn/distilbert-emotion
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.28.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
3,930
Apocalypse-19/doom_deadly_corridor
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,931
Apocalypse-19/doom_health_gathering
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,932
Apocalypse-19/trocr-MICR
vision-encoder-decoder
["VisionEncoderDecoderModel"]
50265
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
2
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,933
Apokryphosx/7B-GSM8K-finetune
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.51.3
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
3,934
AppsDev/bert-base-japanese-v3-wrime-sentiment
bert
["BertForSequenceClassification"]
32768
float32
4.31.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,935
Apptware/QNA_chatbot_ecommerce_falcon_7b_sharded_quantized
RefinedWebModel
["RWForCausalLM"]
65024
float16
4.34.0.dev0
4,544
null
null
null
null
null
0
True
null
null
null
0.02
11
11
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,936
AppyCopilot/llm-repo
bert
["BertForSequenceClassification"]
30522
float32
4.39.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,937
ApricotWine/minipile_flash_gradcheck
gpt_neox
["GPTNeoXForCausalLM"]
50304
bfloat16
4.45.0
768
3,072
12
12
null
gelu
0
True
2,048
10,000
null
0.02
0
0
false
64
null
false
null
null
false
false
true
false
false
false
false
false
false
true
0.084935
small
medium
3,938
AptaArkana/indonesian-comunication-indoBERT-finetuned
bert
["BertForSequenceClassification"]
31923
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
0
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,939
AptaArkana/indonesian-distilbert-base-cased-finetuned-indonlu
distilbert
["DistilBertForSequenceClassification"]
119547
float32
4.35.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
3,940
AptaArkana/indonesian-personality-indoBERT-finetuned
bert
["BertForSequenceClassification"]
31923
float32
4.38.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
0
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,941
AptaArkana/indonesian_sentiment_sbert_base
bert
["BertForSequenceClassification"]
30522
float32
4.35.2
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,942
AptaArkana/indonesian_toxic_classification
bert
["BertForSequenceClassification"]
31923
float32
4.26.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
0
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,943
Apucs/bangla-bert-base-finetuned-mnli-mm
bert
["BertForSequenceClassification"]
102025
float32
4.35.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,944
Apucs/bn_model_821k_iter_loss_1_24-finetuned-mnli-mm
bert
["BertForSequenceClassification"]
30000
float32
4.35.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,945
Apucs/bn_model_821k_iter_loss_1_24-finetuned-sc
bert
["BertForSequenceClassification"]
30000
float32
4.35.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,946
Apurva3509/gemma-1.1-7b-it-bnb-4bit-medical
gemma
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,947
Apv/Flaubert2904_v2
flaubert
["FlaubertForSequenceClassification"]
68729
null
4.28.1
null
null
null
null
null
null
0.1
null
512
null
null
null
0
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
3,948
Apv/Flaubert_1106
flaubert
["FlaubertForSequenceClassification"]
68729
null
4.30.1
null
null
null
null
null
null
0.1
null
512
null
null
null
0
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
3,949
AqeelShafy7/Whisper-Sinhala_Audio_to_Text
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.38.2
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,950
Aqel/xlm-roberta-base-finetuned-panx-ar
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.40.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,951
Aqsa-atif/phi-1_5-finetuned-gsm8k
phi-msft
["PhiForCausalLM"]
51200
float32
4.35.2
null
null
null
null
null
null
null
null
null
null
null
0.02
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,952
AqsaK/HW-1880-finetuned-all-cols
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,953
Aqua002/DialoGPT-small-deadpool
gpt2
["GPT2LMHeadModel"]
50257
float32
4.28.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,954
AquaLabs/Orpheus-3B-0.1-ft-Elise
llama
["LlamaForCausalLM"]
156940
float16
4.50.0
3,072
8,192
28
24
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128009
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
3,955
Ar4ikov/PromptGPTv2
gpt2
["GPT2LMHeadModel"]
50257
float32
4.27.4
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,956
Ar4ikov/W2VFSC2
wav2vec2
["Wav2Vec2ForSequenceClassification"]
39
float32
4.24.0
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
3,957
Ar4ikov/W2VFSC2_1
wav2vec2
["Wav2Vec2Classification"]
39
float32
4.24.0
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
3,958
Ar4ikov/W2VFSC3
wav2vec2
["Wav2Vec2ForSequenceClassification"]
39
float32
4.24.0
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
3,959
Ar4ikov/W2VSC2
wav2vec2
["Wav2Vec2Classification"]
39
float32
4.24.0
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
3,960
Ar4ikov/Wav2Vec2ForSequenceClassification
wav2vec2
["Wav2Vec2ForSequenceClassification"]
39
float32
4.24.0
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
3,961
Ar4l/DebertaV2-Base-10M_babylm-A__rte
deberta-v2
["DebertaV2ForSequenceClassification"]
6144
float32
4.44.2
768
1,536
8
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.056623
small
short
3,962
Arabic-Clip-Archive/arabert-large-vit-base-32-epoch-16
bert
["BertModel"]
64000
float32
4.33.1
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,963
Arabic-Clip-Archive/arabert-large-vit-base-32-epoch-21
bert
["BertModel"]
64000
float32
4.33.1
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,964
Arabic-Clip-Archive/arabert-large-vit-base-32-epoch-26
bert
["BertModel"]
64000
float32
4.33.1
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,965
Arabic-Clip-Archive/arabert-large-vit-base-32-epoch-31
bert
["BertModel"]
64000
null
4.31.0
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,966
Arabic-Clip-Archive/arabertv2-Vit-B-16-plus-epoch-31-trained-1M-corrupted
bert
["BertModel"]
64000
float32
4.33.1
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,967
Arabic-Clip-Archive/arabertv2-Vit-B-16-plus-epoch-41-trained-1M-corrupted
bert
["BertModel"]
64000
float32
4.33.1
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,968
Arabic-Clip-Archive/arabic-arabert-Vit-B-32-Transformer-tf
bert
["BertModel"]
64000
null
4.31.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,969
Aranya31/DermLLaVA-1.5-7b-Label-Final2
llava
["LlavaForConditionalGeneration"]
32064
float16
4.45.1
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
3,970
Aranya31/LLaVA-1.5-7b-DermConv
llava
["LlavaForConditionalGeneration"]
32064
float16
4.45.1
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
3,971
Aranya31/Qwen-2.5-dermMCQ
qwen2_5_vl
["Qwen2_5_VLForConditionalGeneration"]
151936
float32
4.49.0
2,048
11,008
36
16
2
silu
0
True
128,000
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
3,972
Aranya31/derm-LLaVAMed
llava_mistral
["LlavaMistralForCausalLM"]
32000
float16
4.37.2
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
3,973
Aranya31/derm-LLaVAMed-3e
llava_mistral
["LlavaMistralForCausalLM"]
32000
float16
4.37.2
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
3,974
Aranya31/llava_77acc_merged_epoch3
llava
["LlavaForConditionalGeneration"]
32064
float16
4.46.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
3,975
Aranya31/skinllava-1.5-7b-sample
llava
["LlavaForConditionalGeneration"]
32064
float16
4.45.0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
3,976
Aranya31/whisper-small-bn
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.52.4
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,977
ArashAhmadian/ppo_6.9b_new
gpt_neox
["GPTNeoXForCausalLM"]
50432
bfloat16
4.41.1
4,096
16,384
32
32
null
gelu
0
True
2,048
null
null
0.02
0
0
false
128
null
false
null
null
false
false
true
false
false
false
false
false
false
true
6.442451
medium
medium
3,978
ArashAhmadian/rloo_6.9b_new
gpt_neox
["GPTNeoXForCausalLM"]
50432
bfloat16
4.41.1
4,096
16,384
32
32
null
gelu
0
True
2,048
null
null
0.02
0
0
false
128
null
false
null
null
false
false
true
false
false
false
false
false
false
true
6.442451
medium
medium
3,979
ArashAhmadian/rloo_tldr_6.9b
gpt_neox
["GPTNeoXForCausalLM"]
50432
bfloat16
4.41.1
4,096
16,384
32
32
null
gelu
0
True
2,048
null
null
0.02
0
0
false
128
null
false
null
null
false
false
true
false
false
false
false
false
false
true
6.442451
medium
medium
3,980
ArashMarandi/Llama-2-7b-chat-hf
gpt_neo
["GPTNeoForCausalLM"]
50257
float32
4.44.2
2,048
null
null
null
null
null
0
True
2,048
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
medium
3,981
Aratako/ELYZA-japanese-Llama-2-MoE-2x13B-v0.1
mixtral
["MixtralForCausalLM"]
32000
bfloat16
4.38.1
5,120
13,824
40
40
40
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
2
2
false
false
false
false
false
false
false
true
false
true
12.582912
large
medium
3,982
Aratako/MistralPrism-24B
mistral
["MistralForCausalLM"]
131072
bfloat16
4.52.4
5,120
32,768
40
32
8
silu
0
False
131,072
1,000,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
3,983
Aratako/Ninja-v1-RP-expressive
mistral
["MistralForCausalLM"]
32000
bfloat16
4.40.0.dev0
4,096
14,336
32
32
8
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
medium
3,984
Aratako/Oumuamua-7b-RP
mistral
["MistralForCausalLM"]
42800
bfloat16
4.41.2
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
3,985
Aratako/Qwen3-30B-A3B-ERP-v0.1
qwen3_moe
["Qwen3MoeForCausalLM"]
151936
bfloat16
4.51.3
2,048
6,144
48
32
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
false
64
8
false
null
8
false
false
false
false
false
false
false
false
true
true
2.415919
medium
very_long
3,986
Aratako/Swallow-MoE-2x13B-v0.1
mixtral
["MixtralForCausalLM"]
43176
bfloat16
4.38.1
5,120
13,824
40
40
40
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
2
2
false
false
false
false
false
false
false
true
false
true
12.582912
large
medium
3,987
Aratako/Vecteus-v1-toxic
mistral
["MistralForCausalLM"]
32000
bfloat16
4.36.2
4,096
14,336
32
32
8
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
medium
3,988
Aratako/c4ai-command-r-v01-japanese-instruct
cohere
["CohereForCausalLM"]
256000
bfloat16
4.39.3
8,192
22,528
40
64
64
silu
0
True
8,192
8,000,000
null
0.02
5
255001
true
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
32.212255
large
long
3,989
Aratako/calm3-22b-RP-v2
llama
["LlamaForCausalLM"]
65024
bfloat16
4.44.2
6,144
16,384
48
48
48
silu
0
True
16,384
1,000,000
0.00001
0.02
0
65001
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
21.743272
large
long
3,990
Aratako/gemma-2-2b-axolotl-simpo-v1.0
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.46.3
2,304
9,216
26
8
4
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
288
2
false
null
null
false
false
false
false
false
false
false
false
true
true
1.656226
medium
long
3,991
Aratako/gemma-2-2b-axolotl-simpo-v1.0-merged
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.46.3
2,304
9,216
26
8
4
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2
1
null
288
2
false
null
null
false
false
false
false
false
false
false
false
true
true
1.656226
medium
long
3,992
Aratako/karakuri-lm-chat-upscaled-103b-v0.1
llama
["LlamaForCausalLM"]
45416
bfloat16
4.38.1
8,192
28,672
120
64
8
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
96.636764
xlarge
medium
3,993
AravindS373/bird_multi_700
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.51.3
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
3,994
Aravindh0693/srjana2024_techdocs
gemma
["GemmaForCausalLM"]
256000
float16
4.38.0
3,072
24,576
28
16
16
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
3,995
Arbi-Houssem/TunLangModel1.0
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.41.0.dev0
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,996
Arbi-Houssem/TunLangModel_test1.10
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.41.2
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,997
Arbi-Houssem/TunLangModel_test1.11
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.41.2
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,998
Arbi-Houssem/TunLangModel_test1.12
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.41.2
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,999
Arbi-Houssem/TunLangModel_test1.13
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.41.2
null
null
12
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null