index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
15,400
Iddah/llama3-8B-formal
llama
["LlamaForCausalLM"]
128256
bfloat16
4.39.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,401
Iddah/llama3-8B-instruct-informalization
llama
["LlamaForCausalLM"]
128256
bfloat16
4.39.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,402
Ideapad/qa_entropy_all_1e-05_full_5_forget10_layer-15
llama
["CustomLlamaForCausalLM"]
32000
bfloat16
4.47.1
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
15,403
Ideapad/qa_entropy_all_1e-05_full_5_forget10_layer-None
llama
["LlamaForCausalLM"]
32000
bfloat16
4.47.1
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
15,404
Ididsomething/my-awesome-model
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,405
IdleIdiot/thtf-7b
llama
["LlamaForCausalLM"]
65000
float16
4.38.2
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
15,406
Idriska/rubert-tiny2-appartment-clf
bert
["BertForSequenceClassification"]
83828
float32
4.33.2
312
600
3
12
null
gelu
null
True
2,048
null
null
0.02
null
null
null
26
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.003504
small
medium
15,407
IeBoytsov/llama3-1-ox-llms-8b-sft-only-germany-data-and-ultrafeedback
llama
["LlamaForCausalLM"]
128256
bfloat16
4.45.1
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
15,408
Iedha/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-lethal_tawny_deer
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.51.3
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
15,409
IffYuan/FSD-Stage1-Reasoning
llava
["LlavaLlamaForCausalLM"]
32000
bfloat16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
15,410
IffYuan/FSD-v1
llava
["LlavaLlamaForCausalLM"]
32000
bfloat16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
15,411
IffYuan/FSD-v1.1
llava
["LlavaLlamaForCausalLM"]
32000
bfloat16
4.31.0
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
15,412
Iftisyed/my_distilbert1973
bert
["BertForMaskedLM"]
30522
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
0
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,413
Iftisyed/pk_roberta_pakner_model
roberta
["RobertaForTokenClassification"]
50265
float32
4.34.0
768
3,072
6
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.042467
small
short
15,414
Ifyokoh/Igbo-sentiment-bbc
xlm-roberta
["XLMRobertaForSequenceClassification"]
70006
float32
4.42.4
768
3,072
10
6
null
gelu
null
True
514
null
null
0.02
0
2
null
128
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.070779
small
short
15,415
Iggor95/layoutlmv2-base-uncased_finetuned_docvqa
layoutlmv2
["LayoutLMv2ForQuestionAnswering"]
30522
float32
4.44.2
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
15,416
IgnacioFucksmann/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,417
IgnacioFucksmann/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,418
Ignaciobfp/vit-estadosfenologicos-384-prueba
vit
["ViTForImageClassification"]
null
float32
4.49.0
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
15,419
Igniter909/xlm-roberta-base-finetuned-panx-en
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.44.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,420
Igniter909/xlm-roberta-base-finetuned_panx_de
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.44.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,421
IgorKolodziej/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,422
Igorsp/codegen-350M-mono-python-18k-alpaca
codegen
["CodeGenForCausalLM"]
51200
float16
4.34.0
null
null
null
null
null
null
null
True
null
null
null
0.02
1
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,423
Igortin/dummy-model
camembert
["CamembertForMaskedLM"]
32005
float32
4.49.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
5
6
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,424
Igortin/my-dummy-model
bert
["BertForSequenceClassification"]
28996
float32
4.49.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,425
IgriCB/Training_DATA
gemma
["GemmaForCausalLM"]
256000
float16
4.44.2
3,072
24,576
28
16
16
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
15,426
Ihjass/detr-finetuned-135samples
detr
["DetrModel"]
null
float32
4.7.0.dev0
null
null
6
null
null
null
0
null
1,024
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
15,427
Ihor/Llama3-OpenBioLLM-8B-AWQ
llama
["LlamaForCausalLM"]
128256
float16
4.46.2
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,428
Ihor/gliclass-mt5-base-lw
GLiClass
["GLiClassModel"]
250102
float32
4.51.3
768
null
null
null
null
null
null
null
null
null
null
0.03
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,429
Ij4r/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-sedate_shrewd_cobra
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.52.4
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
null
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
15,430
IkariDev/Athena-v4
llama
["LlamaForCausalLM"]
32000
float16
4.33.3
5,120
13,824
40
40
40
silu
null
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
15,431
IkePy/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-mangy_purring_lion
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.51.3
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
15,432
Ikeofai/distilbert-base-uncased-finetuned-squad
distilbert
["DistilBertForQuestionAnswering"]
30522
float32
4.38.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
15,433
Iker/ClickbaitFighter-7B
mistral
["MistralForCausalLM"]
32002
bfloat16
4.40.0.dev0
4,096
14,336
32
32
8
silu
0
True
8,192
10,000
0.00001
0.02
1
32000
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
long
15,434
Iker/Llama-3-Instruct-Neurona-8b-v2
llama
["LlamaForCausalLM"]
128264
bfloat16
4.42.2
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,435
Iker/Neurona-2b
gemma
["GemmaForCausalLM"]
256016
bfloat16
4.40.0
2,048
16,384
18
8
1
gelu
0
False
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
15,436
Ikhee10/my_awesome_qa_model
distilbert
["DistilBertForQuestionAnswering"]
30522
null
4.34.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
15,437
Ilamparithi-I/CIC-CourseQA-LLM
mistral
["MistralForCausalLM"]
32000
float16
4.38.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
15,438
IlhamEbdesk/bge-base-financial-matryoshka_test_my
bert
["BertModel"]
30522
float32
4.41.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,439
Ilhamben/camemBert_sentiment_final
camembert
["CamembertForSequenceClassification"]
32005
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
5
6
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,440
Ilhamben/sarcasm_detection_french_camembert
camembert
["CamembertForSequenceClassification"]
32005
float32
4.36.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
5
6
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,441
IliyanGochev/distil-whisper-large-v3-bg
whisper
["WhisperForConditionalGeneration"]
51866
float32
4.35.2
null
null
32
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,442
Illiyas2024/folrence-2-large-8bit
florence2
["Florence2ForConditionalGeneration"]
51289
float16
4.42.4
null
null
null
null
null
null
null
null
null
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,443
Illuminati-014/whisper-large-v3-turbo-modified
whisper
["WhisperForConditionalGeneration"]
51866
float32
4.49.0
null
null
32
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,444
Illuminati-014/whisper-large-v3-turbo-modified-v1
whisper
["WhisperForConditionalGeneration"]
51866
float32
4.49.0
null
null
32
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,445
Illuminati1/123
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,446
IlluminatiPudding/a2c-PandaPickAndPlace-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
15,447
IlluminatiPudding/a2c-PandaReachDense-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
15,448
IlmarM/detr-fashion-clothes
detr
["DetrForObjectDetection"]
null
float32
4.51.3
null
null
6
null
null
null
0
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,449
IlonaF/finetuning-misinfo-model-1000-Zhaohui
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.25.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
15,450
IlyaGusev/T-lite-instruct-0.1-abliterated
llama
["LlamaForCausalLM"]
128259
bfloat16
4.42.3
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,451
IlyaGusev/gemma-2-2b-it-abliterated
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.43.3
2,304
9,216
26
8
4
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
[1, 107]
null
288
2
false
null
null
false
false
false
false
false
false
false
false
true
true
1.656226
medium
long
15,452
IlyaGusev/gemma-2-9b-it-abliterated
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.42.4
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
15,453
IlyaGusev/llama_7b_ru_turbo_alpaca_lora_merged
llama
["LlamaForCausalLM"]
32000
float16
4.28.0.dev0
4,096
11,008
32
32
null
silu
null
True
null
null
0.000001
0.02
1
2
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
6.442451
medium
null
15,454
IlyaGusev/roberta-base-informal-tagger
roberta
["RobertaForTokenClassification"]
50265
float32
4.20.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,455
IlyaGusev/ruadapt_ficbook_llama2_7b
llama
["LlamaForCausalLM"]
32000
float16
4.36.2
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
15,456
IlyaGusev/saiga_aya_23_35b_sft_m1_d5
cohere
["CohereForCausalLM"]
256000
bfloat16
4.41.1
8,192
22,528
40
64
64
silu
0
True
8,192
8,000,000
null
0.02
5
255001
null
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
32.212255
large
long
15,457
IlyaGusev/saiga_aya_23_35b_sft_m1_d5_awq_4bit
cohere
["CohereForCausalLM"]
256000
float16
4.41.1
8,192
22,528
40
64
64
silu
0
True
8,192
8,000,000
null
0.02
5
255001
null
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
32.212255
large
long
15,458
IlyaGusev/saiga_gemma3_12b_sft_m2_d19_smpo_m3_d39
gemma3
["Gemma3ForConditionalGeneration"]
null
bfloat16
4.51.3
null
null
null
null
null
null
null
null
null
null
null
0.02
null
[1, 106]
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
15,459
IlyaGusev/saiga_llama3_70b_sft_m1_d5_abliterated_gptq_4bit
llama
["LlamaForCausalLM"]
128256
float16
4.41.2
8,192
28,672
80
64
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
long
15,460
IlyaGusev/saiga_llama3_8b_sft_m11_d5_abliterated_kto_m5_d3
llama
["LlamaForCausalLM"]
128256
bfloat16
4.41.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,461
IlyaGusev/saiga_llama3_8b_sft_m11_d7_abliterated_kto_m7_d7
llama
["LlamaForCausalLM"]
128256
bfloat16
4.42.0.dev0
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,462
IlyaGusev/saiga_nemo_12b_sft_m10_d16_simpo_m23_d36
mistral
["MistralForCausalLM"]
131072
bfloat16
4.46.2
5,120
14,336
40
32
8
silu
0
True
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
15,463
IlyaGusev/saiga_qwen2_7b_sft_m2_d6
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.42.0.dev0
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
15,464
IlyaGusev/saiga_qwen2_7b_sft_m2_d6_kto_m1_d5
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.42.0.dev0
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
15,465
IlyaGusev/t5-base-filler-informal
t5
["T5ForConditionalGeneration"]
32128
float32
4.20.1
null
null
null
null
null
null
null
True
null
null
null
null
2
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
15,466
IlyaGusev/vikhr_nemo_12b_gptq_8bit
mistral
["MistralForCausalLM"]
131074
float16
4.46.2
5,120
14,336
40
32
8
silu
0
True
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
15,467
IlyasMoutawwakil/sbert-all-MiniLM-L6-v2-with-pooler
bert
["BertModel"]
30522
null
4.39.3
384
1,536
6
12
null
gelu
null
True
512
null
null
0.02
null
null
null
32
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.010617
small
short
15,468
IlyasMoutawwakil/test-benchmarks
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,469
IlyasMoutawwakil/tiny-random-llama
llama
["LlamaForCausalLM"]
32000
float32
4.35.1
16
64
2
4
4
silu
null
True
2,048
10,000
0.000001
0.02
1
2
false
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
15,470
IlyasMoutawwakil/vicuna-7b-v1.5-awq-marlin
llama
["LlamaForCausalLM"]
32000
float16
4.37.1
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
15,471
ImKH/Llama-3-8B-Instruct_1562_train
llama
["LlamaForCausalLM"]
128264
float16
4.41.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,472
ImKH/Llama-3-8B-Instruct_1950_train
llama
["LlamaForCausalLM"]
128264
float16
4.41.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,473
ImKH/Llama-3-8B-Instruct_3512_train
llama
["LlamaForCausalLM"]
128264
float16
4.41.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,474
ImNobody/wav2vec2-large-xlsr-53-demo-colab
wav2vec2
["Wav2Vec2ForCTC"]
33
float32
4.28.0.dev0
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
15,475
ImSaito/bloom-chistes-spanish
bloom
["BloomForCausalLM"]
250880
float16
4.31.0
1,024
null
null
null
null
null
0
True
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,476
ImSaito/gpt2-chistes-spanish
gpt2
["GPT2LMHeadModel"]
50257
float16
4.31.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
15,477
ImSaito/llama-2-chistes-spanish
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
15,478
ImSaito/llama2es-chistes-spanish
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
2,048
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
15,479
ImSakushi/nistraal-2
mistral
["MistralForCausalLM"]
32000
float16
4.38.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
15,480
ImSota/LLama_LoRA
mistral
["MistralForCausalLM"]
32064
bfloat16
4.51.3
3,072
8,192
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
true
false
false
true
3.623879
medium
medium
15,481
ImValll/gemma-training_test_colab
gemma
["GemmaForCausalLM"]
256000
float16
4.38.0
2,048
16,384
18
8
1
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
15,482
Imad03/Prueba_Gradio
vit
["ViTForImageClassification"]
null
float32
4.22.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
15,483
ImadAziz/DialoGPT-Sheldon
gpt2
["GPT2LMHeadModel"]
50257
float32
4.22.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
15,484
Imadken/Mistral-7B-v0.1-platypus
mistral
["MistralForCausalLM"]
32000
bfloat16
4.34.0
4,096
14,336
32
32
8
silu
null
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
15,485
Imadsarvm/Sarvm-Translation
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.27.4
null
null
24
null
null
null
0
False
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
15,486
Image-editing/Emu3-Base-SFT-reasoning_super_concise-Apr17_lr1e-5-checkpoint-32000
Emu3
["Emu3ForCausalLM"]
184622
bfloat16
4.44.0
4,096
14,336
32
32
8
silu
0.1
False
4,300
1,000,000
0.00001
0.02
151849
151850
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
15,487
Image-editing/Emu3-Base-SFT-reasoning_super_concise-Apr17_rinp_lr1e-5-checkpoint-34950
Emu3
["Emu3ForCausalLM"]
184622
bfloat16
4.44.0
4,096
14,336
32
32
8
silu
0.1
False
4,300
1,000,000
0.00001
0.02
151849
151850
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
15,488
Image-editing/Emu3-Base-SFT-reasoning_verbose-May02_post_stochastic_lr1e-4-checkpoint-1600
Emu3
["Emu3ForCausalLM"]
184622
bfloat16
4.44.0
4,096
14,336
32
32
8
silu
0.1
False
4,300
1,000,000
0.00001
0.02
151849
151850
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
15,489
Image-editing/Emu3-Base-SFT-reasoning_verbose-May02_post_stochastic_lr1e-4-checkpoint-2400
Emu3
["Emu3ForCausalLM"]
184622
bfloat16
4.44.0
4,096
14,336
32
32
8
silu
0.1
False
4,300
1,000,000
0.00001
0.02
151849
151850
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
15,490
Image-editing/Emu3-Base-SFT-reasoning_verbose-May02_post_stochastic_lr1e-4-checkpoint-3200
Emu3
["Emu3ForCausalLM"]
184622
bfloat16
4.51.3
4,096
14,336
32
32
8
silu
0.1
False
4,300
1,000,000
0.00001
0.02
151849
151850
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
15,491
Image-editing/Emu3-Base-SFT-reasoning_verbose-May02_post_stochastic_lr1e-4-checkpoint-4000
Emu3
["Emu3ForCausalLM"]
184622
bfloat16
4.51.3
4,096
14,336
32
32
8
silu
0.1
False
4,300
1,000,000
0.00001
0.02
151849
151850
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
15,492
Image-editing/Emu3-Base-SFT-reasoning_verbose-May02_post_stochastic_lr1e-4-checkpoint-5600
Emu3
["Emu3ForCausalLM"]
184622
bfloat16
4.51.3
4,096
14,336
32
32
8
silu
0.1
False
4,300
1,000,000
0.00001
0.02
151849
151850
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
15,493
ImageIN/convnext-base-224_finetuned_on_ImageIn_annotations
convnext
["ConvNextForImageClassification"]
null
float32
4.22.1
null
null
null
null
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,494
ImageIN/convnext-tiny-224_finetuned
convnext
["ConvNextForImageClassification"]
null
float32
4.22.1
null
null
null
null
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,495
ImageIN/mobilevit-small_finetuned_on_unlabelled_IA_with_snorkel_labels
mobilevit
["MobileViTForImageClassification"]
null
float32
4.28.1
null
null
null
4
null
silu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,496
ImageIN/resnet-50_finetuned
resnet
["ResNetForImageClassification"]
null
float32
4.22.1
null
null
null
null
null
relu
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,497
ImagineIt/22b-Converted
mistral
["MistralForCausalLM"]
32770
bfloat16
4.41.2
6,144
16,384
56
48
8
silu
0
True
32,768
1,000,000
0.00001
0.02
32768
32769
false
128
6
false
null
null
false
false
false
false
false
false
true
false
true
true
25.367151
large
very_long
15,498
ImagineIt/Alpha-basic
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.2
4,096
14,336
32
32
8
silu
0
True
20,000
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,499
ImagineIt/alpha-r-s-v-12
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.2
4,096
14,336
32
32
8
silu
0
True
40,000
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long