index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
3,500
AndyPark/gemma3_lora_dpo_chosen
gemma3
["Gemma3ForConditionalGeneration"]
null
bfloat16
4.50.0
5,376
null
null
null
null
null
null
True
null
null
null
0.02
null
[1, 106]
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
3,501
AndyReas/GenNewsGPT
gpt2
["GPT2LMHeadModel"]
50257
float32
4.24.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,502
AndyReas/roberta-gen-news
roberta
["RobertaForMaskedLM"]
50265
float32
4.24.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,503
AndyReas/roberta-news
roberta
["RobertaForMaskedLM"]
50265
float32
4.24.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,504
AndyZijianZhang/NVILA-Lite-8B-sft-PhysInstruct-PhysPQA-hf
vila
["VILAForConditionalGeneration"]
null
bfloat16
4.52.3
3,584
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
3,505
Andyftk/a2c-PandaPickAndPlace-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,506
Andyftk/a2c-PandaReachDense-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,507
Andyhahaho/distilbert-base-uncased-finetuned-adl_hw1
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.41.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
3,508
Andyrasika/bart_tech_keywords
bart
["BartForConditionalGeneration"]
50265
float32
4.37.1
null
null
12
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
3,509
Andyrasika/dinov2-base-finetuned-oxford
dinov2
["Dinov2ForImageClassification"]
null
float32
4.38.2
768
null
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
3,510
Andyrasika/donut-base-sroie
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
4.33.0.dev0
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,511
AneeqMalik/MedScribe-Llama-2-7b-chat-finetune
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,512
Anery/bert-finetuned-ner
bert
["BertForTokenClassification"]
28996
float32
4.19.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,513
Anery/legalbert_beneficiary_single
bert
["BertForTokenClassification"]
30522
float32
4.19.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
0
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,514
Anery/legalbert_clause_combined
bert
["BertForTokenClassification"]
30522
float32
4.19.2
512
2,048
6
8
null
gelu
null
True
512
null
null
0.02
0
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.018874
small
short
3,515
AnetteHabanen/detr
detr
["DetrForObjectDetection"]
null
float32
4.40.0
null
null
6
null
null
null
0
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,516
AngLv/NoisyRewards-in-RL-RM-acc-65
qwen2
["Qwen2ForSequenceClassification"]
152064
bfloat16
4.51.0
3,584
18,944
28
28
4
silu
0
False
131,072
1,000,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
3,517
AngLv/NoisyRewards-in-RL-RM-acc-85
qwen2
["Qwen2ForSequenceClassification"]
152064
bfloat16
4.51.0
3,584
18,944
28
28
4
silu
0
False
131,072
1,000,000
0.000001
0.02
151643
151643
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
3,518
Angel-IG/distilgpt2-finetuned-mecanicos
gpt2
["GPT2LMHeadModel"]
50257
float32
4.27.4
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,519
AngelRaychev/0.5B-sos-iteration_4
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.51.2
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
3,520
AngelSlim/Deepseek_r1_distill_qwen-1_5b_fp8_static
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.51.3
1,536
8,960
28
12
2
silu
0
False
131,072
10,000
0.000001
0.02
151643
151643
false
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,521
AngelSlim/Deepseek_r1_distill_qwen-1_5b_int4_awq
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.51.3
1,536
8,960
28
12
2
silu
0
False
131,072
10,000
0.000001
0.02
151643
151643
false
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,522
AngelSlim/Deepseek_r1_distill_qwen-1_5b_int4_gptq
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.51.3
1,536
8,960
28
12
2
silu
0
False
131,072
10,000
0.000001
0.02
151643
151643
false
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,523
AngelSlim/Qwen2_5-1_5b_instruct_fp8_static
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.51.3
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,524
AngelSlim/Qwen2_5-1_5b_int4_awq
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.51.3
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,525
AngelSlim/Qwen2_5-1_5b_int4_gptq
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.51.3
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
3,526
AngelSlim/Qwen2_5-7b_instruct_fp8_static
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.52.4
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
3,527
AngelSlim/Qwen2_5-7b_int4_awq
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.51.3
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
3,528
AngelSlim/Qwen2_5-7b_int4_gptq
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.51.3
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
3,529
AngelSlim/Qwen3-0_6B_int4_awq
qwen3
["Qwen3ForCausalLM"]
151936
float16
4.51.3
1,024
3,072
28
16
8
silu
0
False
40,960
1,000,000
0.000001
0.02
151643
151645
true
64
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.352322
small
very_long
3,530
AngelSlim/Qwen3-0_6b_fp8_static
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
1,024
3,072
28
16
8
silu
0
False
40,960
1,000,000
0.000001
0.02
151643
151645
true
64
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.352322
small
very_long
3,531
AngelSlim/Qwen3-0_6b_int4_gptq
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
1,024
3,072
28
16
8
silu
0
False
40,960
1,000,000
0.000001
0.02
151643
151645
true
64
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.352322
small
very_long
3,532
AngelSlim/Qwen3-1_7B_int4_awq
qwen3
["Qwen3ForCausalLM"]
151936
float16
4.51.3
2,048
6,144
28
16
8
silu
0
False
40,960
1,000,000
0.000001
0.02
151643
151645
true
128
2
false
null
null
false
false
false
false
false
false
false
false
true
true
1.409286
medium
very_long
3,533
AngelSlim/Qwen3-1_7b_fp8_static
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
2,048
6,144
28
16
8
silu
0
False
40,960
1,000,000
0.000001
0.02
151643
151645
true
128
2
false
null
null
false
false
false
false
false
false
false
false
true
true
1.409286
medium
very_long
3,534
AngelSlim/Qwen3-1_7b_int4_gptq
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
2,048
6,144
28
16
8
silu
0
False
40,960
1,000,000
0.000001
0.02
151643
151645
true
128
2
false
null
null
false
false
false
false
false
false
false
false
true
true
1.409286
medium
very_long
3,535
AngelSlim/Qwen3-4b_fp8_static
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
2,560
9,728
36
32
8
silu
0
False
40,960
1,000,000
0.000001
0.02
151643
151645
true
80
4
false
null
null
false
false
false
false
false
false
false
false
true
true
2.831155
medium
very_long
3,536
AngelSlim/Qwen3-4b_int4_awq
qwen3
["Qwen3ForCausalLM"]
151936
float16
4.51.3
2,560
9,728
36
32
8
silu
0
False
40,960
1,000,000
0.000001
0.02
151643
151645
true
80
4
false
null
null
false
false
false
false
false
false
false
false
true
true
2.831155
medium
very_long
3,537
AngelSlim/Qwen3-4b_int4_gptq
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
2,560
9,728
36
32
8
silu
0
False
40,960
1,000,000
0.000001
0.02
151643
151645
true
80
4
false
null
null
false
false
false
false
false
false
false
false
true
true
2.831155
medium
very_long
3,538
AngelSlim/Qwen3-8b_fp8_static
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
4,096
12,288
36
32
8
silu
0
False
40,960
1,000,000
0.000001
0.02
151643
151645
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
3,539
AngelSlim/Qwen3-8b_int4_awq
qwen3
["Qwen3ForCausalLM"]
151936
float16
4.51.3
4,096
12,288
36
32
8
silu
0
False
40,960
1,000,000
0.000001
0.02
151643
151645
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
3,540
AngelSlim/Qwen3-8b_int4_gptq
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.51.3
4,096
12,288
36
32
8
silu
0
False
40,960
1,000,000
0.000001
0.02
151643
151645
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
3,541
AngelUrq/poca-SoccerTwos
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,542
AngelUrq/rl_course_vizdoom_health_gathering_supreme
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,543
Angelectronic/mBart-en-vi-test
mbart
["MBartForConditionalGeneration"]
250054
float32
4.35.2
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
3,544
AngellyCris/An-lisis_Sentimientos
roberta
["RobertaForSequenceClassification"]
30000
float32
4.51.3
768
3,072
12
12
null
gelu
null
True
130
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,545
AngellyCris/model.safetensors
roberta
["RobertaForSequenceClassification"]
30000
float32
4.51.3
768
3,072
12
12
null
gelu
null
True
130
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,546
Angelokitio/Rllm-repo
bert
["BertForSequenceClassification"]
30522
float32
4.39.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,547
Angeriod/in_car_commands_26_mdl__base_ver1
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.48.0
null
null
6
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,548
Angeriod/in_car_commands_26_mdl__base_ver2
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.48.0
null
null
6
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,549
Angeriod/in_car_commands_26_mdl__tiny_ver1
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.48.0
null
null
4
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
3,550
AngieMojica/greenguardian_model
vit
["ViTForImageClassification"]
null
float32
4.41.0
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
3,551
Anguuuuus/chordektomie-iau
wav2vec2
["Wav2Vec2ForSequenceClassification"]
32
float32
4.37.2
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
3,552
Anguuuuus/ipadsample
wav2vec2
["Wav2Vec2ForSequenceClassification"]
32
float32
4.37.2
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
3,553
Anguuuuus/mydata-iau
wav2vec2
["Wav2Vec2ForSequenceClassification"]
32
float32
4.37.2
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
3,554
Angy309/swin-tiny-patch4-window7-224-finetuned-finetuned
swin
["SwinForImageClassification"]
null
float32
4.40.2
768
null
null
null
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,555
AnhNam/nguyenkeluong_modernBert_ft3
modernbert
["ModernBertForSequenceClassification"]
50368
float32
4.48.0.dev0
768
1,152
22
12
null
null
0
null
8,192
null
null
0.02
50281
50282
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
true
0.155714
small
long
3,556
Ani5577/Llama-2-7b-ioc
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,557
Aniemore/hubert-emotion-russian-resd
hubert
["HubertForSequenceClassification"]
40
float32
4.25.1
1,024
4,096
24
16
null
gelu
0.05
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
null
3,558
Aniemore/rubert-base-emotion-russian-cedr-m7
bert
["BertForSequenceClassification"]
119547
float32
4.25.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,559
Aniemore/rubert-large-emotion-russian-cedr-m7
bert
["BertForSequenceClassification"]
120138
float32
4.25.1
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
3,560
Aniemore/rubert-tiny-emotion-russian-cedr-m7
bert
["BertForSequenceClassification"]
83828
float32
4.25.1
312
600
3
12
null
gelu
null
True
2,048
null
null
0.02
null
null
null
26
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.003504
small
medium
3,561
Aniemore/unispeech-sat-emotion-russian-resd
unispeech-sat
["UniSpeechSatForSequenceClassification"]
40
float32
4.25.1
1,024
4,096
24
16
null
gelu
0.05
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
3,562
Aniemore/wav2vec2-bert-base-s-emotion-russian-resd
null
["Wav2Vec2BertForSequenceClassification"]
null
float32
4.27.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
null
3,563
Aniemore/wav2vec2-bert-tiny2-s-emotion-russian-resd
null
["Wav2Vec2BertForSequenceClassification"]
null
float32
4.27.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
null
3,564
Aniemore/wav2vec2-emotion-russian-resd
wav2vec2
["Wav2Vec2ForSequenceClassification"]
39
float32
4.25.1
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
3,565
Aniemore/wavlm-bert-base-s-emotion-russian-resd
null
["WavLMBertForSequenceClassification"]
null
float32
4.27.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
null
3,566
Aniemore/wavlm-bert-tiny2-s-emotion-russian-resd
null
["WavLMBertForSequenceClassification"]
null
float32
4.27.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
null
3,567
Aniemore/wavlm-emotion-russian-resd
wavlm
["WavLMForSequenceClassification"]
40
float32
4.25.1
1,024
4,096
24
16
null
gelu
0.05
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
3,568
AnikaAI/dummy-model
camembert
["CamembertForMaskedLM"]
32005
float32
4.32.0
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
5
6
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,569
Aniket17/distilbert-rotten-tomatoes
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.51.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
3,570
Aniket2004/Llama_doctor_v1
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
3,571
Aniket2004/docv1
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
2,048
5,632
22
32
4
silu
null
True
2,048
10,000
0.00001
0.02
1
2
false
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
3,572
AniketArtani/ANIKET_MODEL
bart
["BartForConditionalGeneration"]
50265
float32
4.37.2
null
null
6
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
3,573
AniketArtani/server
bart
["BartForConditionalGeneration"]
50265
float32
4.39.3
null
null
6
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
3,574
AniketV/Llama-3-RSP2D
llama
["LlamaForCausalLM"]
128264
float32
4.46.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,575
AniketV/Llama-ADE
llama
["LlamaForCausalLM"]
128264
float32
4.43.4
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
3,576
Aniketsingh12/bert-finetuned-ner
bert
["BertForTokenClassification"]
28996
float32
4.41.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,577
Anil14349/gpt2_resume_modal
gpt2
["GPT2LMHeadModel"]
50257
float32
4.46.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,578
AnimateBow4809/qwen2-0.5b-math-full-train
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.42.4
896
4,864
24
14
2
silu
0
True
131,072
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
3,579
AnimateBow4809/qwen2-0.5b-math-lora
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.42.4
896
4,864
24
14
2
silu
0
True
131,072
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
3,580
AnimateBow4809/qwen2-math-0.5B-full-2
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.42.4
896
4,864
24
14
2
silu
0
True
131,072
1,000,000
0.000001
0.02
151643
151643
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
3,581
Animesh001/mistral-7b-instruct-4bit
mistral
["MistralForCausalLM"]
32000
float16
4.38.2
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
3,582
Animesh001/mistral-fine16-gguf
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
3,583
Animesh001/mistral-gg-fine04
mistral
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
true
false
false
false
null
null
null
3,584
Animesh001/un-ms-16bt-fine
mistral
["MistralForCausalLM"]
32000
float16
4.38.2
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
3,585
Aniqo/aniqo
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,586
Aniqo/zhongli
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,587
AniruddhAiyengar/Llama-3.1-8B-SOAP-notes-finetuned
llama
["LlamaForCausalLM"]
128256
float16
4.48.3
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
3,588
Aniruddha0605/qwen_sft_8k_indic
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.46.1
2,560
6,912
40
20
20
silu
0
True
32,768
5,000,000
0.000001
0.02
151643
151645
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.145728
medium
very_long
3,589
Aniruddha8699/Veterinary_ChatGPT_FineTuned
gpt2
["GPT2LMHeadModel"]
50257
float32
4.51.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,590
Anirudh25/ppo-LunarLander-v2-TEST
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,591
AnirudhLanka2002/so_ViTS_SVC_models
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,592
AnirudhVV/ACTSA-CARDIFFNLP
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.40.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,593
Anirudhmx/finetuned_gpt2
gpt2
["GPT2LMHeadModel"]
50257
float32
4.53.0
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
3,594
Anis12/swin-tiny-patch4-window7-224-finetuned-eurosat
swin
["SwinForImageClassification"]
null
float32
4.30.1
768
null
null
null
null
gelu
null
null
null
null
null
0.02
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,595
Anish13/a2c-PandaReachDense-v3
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,596
Anish13/poca-SoccerTwos
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
3,597
AnishJoshi/codellama2-finetuned-nl2bash-fin
llama
["LlamaForCausalLM"]
32016
float16
4.39.3
4,096
11,008
32
32
32
silu
0
True
16,384
1,000,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
3,598
Anishsavla2/bert-base-uncased-finetuned-wikitext2
bert
["BertLMHeadModel"]
30522
float32
4.24.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
3,599
Anishsavla2/distilgpt2-finetuned-wikitext2
gpt2
["GPT2LMHeadModel"]
50257
null
4.27.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null