index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
15,500
ImagineIt/cool-beans
llama
["LlamaForCausalLM"]
128256
bfloat16
4.40.1
4,096
14,336
32
32
8
silu
0
True
15,000
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,501
ImagineIt/merged-alpha
llama
["LlamaForCausalLM"]
128256
float16
4.40.2
4,096
14,336
32
32
8
silu
0
True
20,000
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,502
Iman-1220/Farsi-QWEN2-5-1.5B-8bit-pruned
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.43.1
1,536
8,960
28
12
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
15,503
ImanAndrea/bert-finetuned-paperQA
roberta
["RobertaForQuestionAnswering"]
50265
float32
4.45.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,504
ImanAndrea/bert2bert-en-fr-v2
encoder-decoder
["EncoderDecoderModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,505
ImanAndrea/bert2bert-en-to-fr
encoder-decoder
["EncoderDecoderModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,506
Imanbehravan/finetune_BERT
bert
["BertForQuestionAnswering"]
30522
float32
4.40.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,507
Imanbehravan/question_answering_T_5_part_v0.1
t5
["T5ForConditionalGeneration"]
32128
float32
4.46.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
15,508
Imask/Large_3000samples
bart
["BartForConditionalGeneration"]
50264
float32
4.44.2
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
15,509
Imelliot/Qwen2.5-1.5B-Open-R1-GRPO
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.49.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
15,510
Imene/rare-puppers
vit
["ViTForImageClassification"]
null
float32
4.21.1
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
15,511
ImeneT/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,512
Immanuelraja/layoutlmv3-finetuned-cord_100
layoutlmv3
["LayoutLMv3ForTokenClassification"]
50265
float32
4.22.1
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
15,513
ImperialIndians23/RobertaBaseUnprocessedAugmented
roberta
["RobertaForSequenceClassification"]
50265
float32
4.38.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,514
ImperialIndians23/RobertaBaseUnprocessedAugmentedBoth
roberta
["RobertaForSequenceClassification"]
50265
float32
4.38.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,515
ImperialIndians23/RobertaBaseUnprocessedAugmentedSynonym
roberta
["RobertaForSequenceClassification"]
50265
float32
4.38.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,516
ImperialIndians23/RobertaBaseUnprocessedTest
roberta
["RobertaForSequenceClassification"]
50265
float32
4.38.2
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,517
ImposterSyndromeIsReal/biogpt-healthcare-tuned
biogpt
["BioGptForCausalLM"]
42384
float32
4.28.1
1,024
4,096
24
16
null
gelu
null
True
1,024
null
null
0.02
0
2
null
64
null
false
null
null
false
false
true
false
false
false
false
false
false
false
0.30199
small
short
15,518
Imran1/QWEN2.5-32B-Translation
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.48.2
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
15,519
Imran1/Qwen2.5-72B-Instruct-FP8
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.45.2
8,192
29,568
80
64
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
15,520
Imran1/bloom_p560m_3
bloom
["BloomForCausalLM"]
250880
float32
4.30.0.dev0
1,024
null
null
null
null
null
0
True
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,521
Imran1/sentimen_analysis_yelp
bert
["BertForSequenceClassification"]
28996
float32
4.22.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,522
ImranzamanML/German_Semantic_V3b_Optimised
bert
["BertModel"]
31102
float32
4.48.2
1,024
4,096
24
16
null
gelu
null
True
8,192
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
long
15,523
Imxxn/AudioCourseU6-TextToSpeech
speecht5
["SpeechT5ForTextToSpeech"]
81
float32
4.32.1
768
null
null
null
null
gelu
0.1
False
null
null
null
0.02
0
2
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
15,524
Imxxn/RLCourseU5-Pyramids
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,525
Imxxn/RLCourseU5-SnowballTarget
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,526
Imxxn/RLCourseU6-AntBulleEnv-v0
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,527
Imxxn/RLCourseU6-PandaReachDense-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,528
Imxxn/model
roberta
["RobertaForSequenceClassification"]
50265
float32
4.40.1
768
3,072
3
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.021234
small
short
15,529
In2Training/FILM-7B
mistral
["MistralForCausalLM"]
32000
float32
4.37.1
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
15,530
In4Seek/Llama-3.1-8B-Instruct_MedMCQA.04.00_backup
llama
["LlamaForCausalLM"]
128256
float32
4.47.1
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
15,531
In4Seek/Llama-3.1-8B-Instruct_MedMCQA.04.01_backup
llama
["LlamaForCausalLM"]
128256
float32
4.47.1
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
15,532
In4Seek/Llama-3.1-8B-Instruct_MedMCQA.05.01_backup
llama
["LlamaForCausalLM"]
128256
float32
4.47.1
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
15,533
InHawK/jd-llama2-32bit-7b-finetune
llama
["LlamaForCausalLM"]
32000
float32
4.37.0.dev0
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
15,534
InHawK/jd-llama2-7b-finetune
llama
["LlamaForCausalLM"]
32000
float16
4.37.0.dev0
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
15,535
Inabia-AI/ark_unbranded_claims_standalone_lora_3.1_10162024
llama
["LlamaForCausalLM"]
128256
float16
4.44.2
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
15,536
Inabia-AI/continued_pretraining_model_01
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,537
InasMR/sentiment_model
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.35.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
15,538
Inasity/L3.3-San-Mai-R1-70b-4.0bpw-h8-exl2
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.3
8,192
28,672
80
64
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
15,539
Inderpreet01/Llama-3.1-8B-Instruct_rca_sft_tool_call_and_reasoning
llama
["LlamaForCausalLM"]
128256
bfloat16
4.51.3
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
15,540
Inderpreet01/Llama-3.2-8B-Instruct_rca_grpo_v2
llama
["LlamaForCausalLM"]
128256
bfloat16
4.50.1
4,096
14,336
32
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
15,541
IndexTeam/Index-1.9B-Constant-LR
index
["IndexForCausalLM"]
65029
bfloat16
4.39.2
2,048
5,888
36
16
16
silu
null
True
4,096
null
0.000001
0.01
1
2
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.811939
medium
medium
15,542
IndianChessMans/Therapy-Alpaca
gpt2
["GPT2LMHeadModel"]
50257
float32
4.41.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
15,543
IndianFood/toolcallvariant2
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.48.2
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
null
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
15,544
IndianServers/BhagavadGita
gpt2
["GPT2LMHeadModel"]
50257
float32
4.27.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
15,545
IndicAIDias/finetuned_gemma
gemma
["GemmaForCausalLM"]
256000
float16
4.38.0
2,048
16,384
18
8
1
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
15,546
Indraa99/bart_semeval_env
bart
["BartForConditionalGeneration"]
50265
float32
4.30.2
null
null
6
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
15,547
Indraa99/bert_semeval_env
bert
["BertForMaskedLM"]
30522
float32
4.30.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,548
Inesence/donut-LV-token-test
vision-encoder-decoder
["VisionEncoderDecoderModel"]
null
float32
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,549
Inespinoza/PPO-LunarLander-V0
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,550
Inespinoza/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,551
InfAI/flan-t5-text2sparql-custom-tokenizer
t5
["T5ForConditionalGeneration"]
32128
float32
4.18.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
15,552
InfZnDipl/ML_LogReg
logistic_regression
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,553
InferenceIllusionist/Excalibur-7b-DPO
mistral
["MistralForCausalLM"]
32000
float16
4.38.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
15,554
InferenceIllusionist/Magic-Dolphin-7b
mistral
["MistralForCausalLM"]
32000
float16
4.37.1
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
15,555
Inferless/deciLM-7B-GPTQ
deci_lm
["DeciLMForCausalLM"]
32000
float16
4.36.2
4,096
14,336
32
32
32
silu
null
True
8,192
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
6.442451
medium
long
15,556
Infermatic/Anubis-70B-v1-FP8-Dynamic
llama
["LlamaForCausalLM"]
128256
bfloat16
4.48.0
8,192
28,672
80
64
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128009
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
15,557
Infermatic/L3-70B-Euryale-v2.1-fp8-dynamic
llama
["LlamaForCausalLM"]
128256
bfloat16
4.43.3
8,192
28,672
80
64
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
long
15,558
Infermatic/Llama-3.1-Nemotron-70B-Instruct-HF-FP8-Dynamic
llama
["LlamaForCausalLM"]
128256
bfloat16
4.45.2
8,192
28,672
80
64
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
15,559
Infermatic/Llama-3.3-70B-Instruct-FP8-Dynamic
llama
["LlamaForCausalLM"]
128256
bfloat16
4.47.0
8,192
28,672
80
64
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
15,560
Infermatic/Miquliz-120b-v2.0-FP8-dynamic
llama
["LlamaForCausalLM"]
32000
float16
4.43.3
8,192
28,672
140
64
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
112.742892
xlarge
very_long
15,561
Infi-MM/infimm-hd
infimm-hd
["InfiMMHDModel"]
null
bfloat16
4.35.2
null
null
null
null
null
null
null
True
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
15,562
InfiniAILab/OpenR1-Qwen-1.5B-SFT-Instruct
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.49.0
1,536
8,960
28
12
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
15,563
InfiniAILab/S1-100-32B-bs8-new-tokenizer-parallel2
qwen2
["Qwen2ForCausalLM"]
151680
float32
4.46.1
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
15,564
InfiniAILab/S1-claude-1K-32B-bs16
qwen2
["Qwen2ForCausalLM"]
152064
float32
4.46.1
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
15,565
InfiniAILab/S1.1-100-32B-bs8-new-tokenizer-parallel-4.21-v0
qwen2
["Qwen2ForCausalLM"]
151680
float32
4.46.1
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
15,566
InfiniAILab/S1.1-1K-32B-bs16-new-tokenizer-mix-parallel-7.1-v3-prompted
qwen2
["Qwen2ForCausalLM"]
151680
float32
4.46.1
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
15,567
InfiniAILab/S1.1-1K-32B-bs16-new-tokenizer-parallel-4.26-v0
qwen2
["Qwen2ForCausalLM"]
151680
float32
4.46.1
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
15,568
InfiniAILab/S1.1-1K-32B-bs16-new-tokenizer-parallel-4.26-v0-2
qwen2
["Qwen2ForCausalLM"]
151680
float32
4.46.1
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
15,569
InfiniAILab/S1.1-1K-32B-bs16-new-tokenizer-parallel-4.26-v0-3
qwen2
["Qwen2ForCausalLM"]
151680
float32
4.46.1
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
15,570
InfiniAILab/S1.1-1K-32B-bs16-new-tokenizer-parallel-5.1-v0
qwen2
["Qwen2ForCausalLM"]
151680
float32
4.46.1
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
15,571
InfiniAILab/S1.1-1K-32B-bs16-new-tokenizer-sequential-4.26-v0
qwen2
["Qwen2ForCausalLM"]
151680
float32
4.46.1
5,120
27,648
64
40
8
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
5
false
null
null
false
false
false
false
false
false
false
false
true
true
20.132659
large
very_long
15,572
InfiniteEcho/poca-SoccerTwos
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,573
InfiniteEcho/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
15,574
InfinityC/test_sum_1_model
t5
["T5ForConditionalGeneration"]
32128
float32
4.39.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
15,575
InfinityC/test_sum_abs_t5_small_wasa_no_stops
t5
["T5ForConditionalGeneration"]
32128
float32
4.39.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
15,576
InfoDeskInc/InfoDeskGenericNer
xlm-roberta
["XLMRobertaForTokenClassification"]
250002
float32
4.42.4
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,577
Infomaniak-AI/smolLM2-135M-Instruct-structure-output
llama
["LlamaForCausalLM"]
49152
bfloat16
4.48.3
576
1,536
30
9
3
silu
0
True
8,192
100,000
0.00001
0.041667
1
2
true
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
15,578
Ingrid0693/mini-bravo
t5
["T5ForConditionalGeneration"]
32128
float32
4.16.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
15,579
Inhaexpress/DialoGPT-medium-harrypotter
gpt2
["GPT2LMHeadModel"]
50257
float32
4.29.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
15,580
Inishds/function_calling_qlora
phi3
["Phi3ForCausalLM"]
32032
float16
4.40.2
3,072
8,192
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
32000
false
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
medium
15,581
Injardav/detr
detr
["DetrForObjectDetection"]
null
float32
4.39.3
null
null
6
null
null
null
0
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,582
Injustice/distilbert-base-uncased-finetuned-squad
distilbert
["DistilBertForQuestionAnswering"]
30522
float32
4.30.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
15,583
InlaidOyster/RMIT_HACKATHON
bert
["BertForSequenceClassification"]
30522
float32
4.46.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,584
InnerI/A-I-0xtom-7B-slerp
llama
["LlamaForCausalLM"]
32000
bfloat16
4.35.2
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
15,585
InnerI/InnerI-bittensor-7b
llama
["LlamaForCausalLM"]
32000
bfloat16
4.40.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
15,586
InnerI/synCAI-144k-gpt2.5
gpt2
["GPT2LMHeadModel"]
50257
float32
4.40.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
15,587
Innerby/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,588
InnovationHacksAI/ofdv1
llama
["LlamaForCausalLM"]
128256
float16
4.44.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
15,589
Ino4Dev/Llama_Trained_Inolabs
llama
["LlamaForCausalLM"]
32000
float32
4.39.0
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
15,590
Inoob/Null-GPT2-Large
gpt2
["GPT2LMHeadModel"]
50257
float32
4.44.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
15,591
Inria-CEDAR/FactSpotter-DeBERTaV3-Base
deberta-v2
["DebertaV2ForSequenceClassification"]
128100
float32
4.39.1
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
15,592
Inria-CEDAR/FactSpotter-DeBERTaV3-Large
deberta-v2
["DebertaV2ForSequenceClassification"]
128100
float32
4.39.1
1,024
4,096
24
16
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
15,593
Inria-CEDAR/FactSpotter-DeBERTaV3-Small
deberta-v2
["DebertaV2ForSequenceClassification"]
128100
float32
4.39.1
768
3,072
6
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.042467
small
short
15,594
InstaDeepAI/isoformer
isoformer
["Isoformer"]
null
float32
4.29.2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
15,595
InstaDeepAI/nucleotide-transformer-v2-250m-multi-species
null
["EsmForMaskedLM", "EsmForTokenClassification", "EsmForSequenceClassification"]
4107
float32
4.32.0.dev0
768
3,072
24
16
null
null
null
False
2,050
null
null
0.02
null
null
false
48
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.169869
small
medium
15,596
InstaDeepAI/segment_nt
esm
["SegmentNT"]
4107
float32
4.33.2
1,024
4,096
29
16
null
null
null
False
2,050
null
null
0.02
null
null
false
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.364904
small
medium
15,597
InstaDeepExternalProject/llm_training_20230901_132015
llama
["LlamaForCausalLM"]
32000
float32
4.30.2
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
15,598
InstaDeepExternalProject/llm_training_20230901_132240
llama
["LlamaForCausalLM"]
32000
float32
4.30.2
5,120
13,824
40
40
40
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
15,599
InstalilyAI/3-class-deberta
deberta-v2
["DebertaV2ForSequenceClassification"]
128100
float32
4.52.4
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short