index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
6,600
Cafet/whisper-tiny-mongolian-kazah
whisper
["WhisperForConditionalGeneration"]
51865
float32
4.39.3
null
null
4
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
6,601
CaffreyR/30-rte-prune
bert
["BertForSequenceClassification"]
30522
float32
4.24.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,602
CaffreyR/sst2_lora_bert
bert
["BertForSequenceClassification"]
30522
float32
4.25.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,603
Cailucky777/text-summarization
bart
["BartForConditionalGeneration"]
50265
float32
4.27.4
null
null
6
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
6,604
Cainiao-AI/TAAS
TAAS
["TAAS"]
40000
float32
4.25.1
768
3,072
12
12
null
gelu
null
True
2,048
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
medium
6,605
CaioXapelaum/Orca-2-7b-Patent-Instruct-Llama-2
llama
["LlamaForCausalLM"]
32000
float16
4.44.1
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
6,606
Caiyun-AI/DCPythia-6.9B
dcpythia
["DCPythia"]
50257
float16
4.33.2
null
16,384
null
null
null
null
null
null
null
null
null
null
0
0
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
6,607
CakePL/calculator_model_test
encoder-decoder
["EncoderDecoderModel"]
null
float32
4.38.1
null
null
null
null
null
null
null
null
null
null
null
null
null
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,608
Cal3bd3v/llama-3-8b-Instruct-bnb-4bit-aiaustin-demo
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
6,609
CalYYY/model_test
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.51.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,610
Calam1/t5-small-finetuned-wikisql
t5
["T5ForConditionalGeneration"]
32128
float32
4.26.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,611
CalamitousFelicitousness/EVA-Qwen2.5-72B-v0.2-FP8-Dynamic
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.3
8,192
29,568
80
64
8
silu
0
False
131,072
1,000,000
0.00001
0.02
null
151643
false
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
6,612
CalamitousFelicitousness/EVA-Qwen2.5-72B-v0.2-INT8
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.3
8,192
29,568
80
64
8
silu
0
False
131,072
1,000,000
0.00001
0.02
null
151643
false
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
6,613
CalamitousFelicitousness/L3.1-70B-sunfall-v0.6.1-fp8-dynamic
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.0
8,192
28,672
80
64
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
6,614
CalamitousFelicitousness/Llama-3.05-Nemotron-Tenyxchat-Storybreaker-70B-FP8-Dynamic
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
8,192
28,672
80
64
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
6,615
CalamitousFelicitousness/Llama-3.1-70B-ArliAI-RPMax-v1.1-fp8-dynamic
llama
["LlamaForCausalLM"]
128256
bfloat16
4.44.2
8,192
28,672
80
64
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128009
false
128
8
false
null
null
true
false
false
false
false
false
false
false
true
true
64.424509
large
very_long
6,616
CalamitousFelicitousness/Yi-1.5-34B-Chat-16K-tokfix
llama
["LlamaForCausalLM"]
64000
bfloat16
4.40.0
7,168
20,480
60
56
8
silu
0
False
16,384
5,000,000
0.000001
0.02
1
2
false
128
7
false
null
null
true
false
false
false
false
false
false
false
true
true
36.993761
large
long
6,617
CalamitousFelicitousness/Yi-1.5-34B-Chat-16K-tokfix-fp8-dynamic
llama
["LlamaForCausalLM"]
64000
bfloat16
4.44.2
7,168
20,480
60
56
8
silu
0
False
16,384
5,000,000
0.000001
0.02
1
2
false
128
7
false
null
null
true
false
false
false
false
false
false
false
true
true
36.993761
large
long
6,618
CalderaAI/13B-Theseus-MK1
llama
["LlamaForCausalLM"]
32032
float32
4.31.0
5,120
13,824
40
40
40
silu
null
False
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,619
CalderaAI/30B-Lazarus-GPTQ4bit
llama
["LlamaForCausalLM"]
32000
float16
4.28.1
6,656
17,920
60
52
null
silu
null
True
2,048
null
0.000001
0.02
1
2
false
128
null
false
null
null
true
false
false
false
false
false
false
false
false
false
31.897682
large
medium
6,620
CalebE/llama3_mental_health_model_full
llama
["LlamaForCausalLM"]
128256
bfloat16
4.41.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
6,621
Calin/vit-base-patch16-224-in21k-finetuned-lora-food101
vit
["ViTForImageClassification"]
null
float32
4.32.0
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
6,622
CallComply/DeciLM-7B-Instruct-32k
deci
["DeciLMForCausalLM"]
32000
bfloat16
4.35.2
4,096
14,336
32
32
null
silu
null
True
32,768
null
0.00001
0.02
1
2
false
128
null
false
null
null
false
false
false
false
false
false
false
false
false
true
6.442451
medium
very_long
6,623
CallMeJeremy/DialoGPT-medium-THREEPIO
gpt2
["GPT2LMHeadModel"]
50257
float32
4.26.1
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
6,624
CalvinSeamons/CS301r
gpt2
["GPT2LMHeadModel"]
50257
float32
4.33.3
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
6,625
CalvinYorkCS/ppo-LunarLandar-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,626
CalvinYorkCS/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,627
Camais03/camie-tagger
pytorch
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,628
CambioMoney/mixtral-instruct-MLFlowV4
mixtral
["MixtralForCausalLM"]
32000
float16
4.36.0.dev0
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
8
2
false
false
false
false
false
false
false
true
true
true
6.442451
medium
very_long
6,629
Cambridge-KAIST/SmolLM-14m-Dolma-v0.1-Proposed-full
llama
["OutputEmbeddingSelectiveUpdate"]
50280
bfloat16
4.47.1
128
512
6
4
4
silu
0
True
2,048
100,000
0.00001
0.02
0
0
false
32
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.00118
small
medium
6,630
Cambridge-KAIST/SmolLM-14m-Dolma-v0.1-Zloss-full
llama
["LlamaForCausalLM"]
50280
bfloat16
4.46.3
128
512
6
4
4
silu
0
True
2,048
100,000
0.00001
0.02
0
0
false
32
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.00118
small
medium
6,631
Cambridge-KAIST/untied-base-checkpoint-1000
llama
["LlamaForCausalLM"]
50280
bfloat16
4.46.3
128
512
6
4
4
gelu
0
True
2,048
100,000
0.00001
0.02
0
0
false
32
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.00118
small
medium
6,632
Cambridge-KAIST2/SmolLM-14m-Dolma-v0.4-zloss
llama
["LlamaForCausalLM"]
50280
float32
4.48.1
128
512
6
4
4
silu
0
True
2,048
100,000
0.00001
0.02
0
0
false
32
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.00118
small
medium
6,633
CamelliaRan/test_medusa
null
[]
null
null
4.34.1
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,634
Camih/distilbert-base-uncased-finetuned-cola
distilbert
["DistilBertForSequenceClassification"]
30522
null
4.30.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,635
Camih/ppo-Huggy
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,636
Camilovelez1/imdbreviews_classification_distilbert_v02
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.39.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,637
CanEPecor/t5-base-finetuned-small-epochs
t5
["T5ForConditionalGeneration"]
32128
float32
4.46.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,638
CanEPecor/t5-base-finetuned_3
t5
["T5ForConditionalGeneration"]
32128
float32
4.46.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,639
CanerAI/turkishReviews-ds-mini
gpt2
["GPT2LMHeadModel"]
44208
null
4.44.2
null
null
null
null
null
null
null
True
null
null
null
0.02
0
0
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
6,640
Cansino/mi-super-modelo
bert
["BertForSequenceClassification"]
28996
float32
4.51.3
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,641
Cantaosu/my_awesome_asr_mind_model
wav2vec2
["Wav2Vec2ForCTC"]
32
float32
4.40.2
768
3,072
12
12
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
6,642
CaoHaiNam/obito-v0
roberta
["RobertaModel"]
64001
float32
4.39.1
768
3,072
12
12
null
gelu
null
True
258
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,643
CaoHaiNam/obito-v1
roberta
["RobertaModel"]
64001
float32
4.40.1
768
3,072
12
12
null
gelu
null
True
258
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,644
CaoTrungHieu/BERT_Snli
bert
["BertModel"]
30522
float32
4.34.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,645
CapEightFour/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,646
CapitainData/dyu-fr-t5-base_v3
t5
["T5ForConditionalGeneration"]
32128
null
4.38.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,647
CapitainData/dyu-fr-t5-small
t5
["T5ForConditionalGeneration"]
32128
null
4.38.2
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,648
CaptMetal/BuddAi
mistral
["MistralForCausalLM"]
32000
bfloat16
null
4,096
14,336
32
32
8
silu
0
null
32,768
1,000,000
0.00001
null
1
2
null
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
6,649
CaptainHaaz/FERmodelCNN
null
["CNN"]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,650
CaptainKoffski/guessly-yesno-classifier
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.48.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,651
CaptainLazarus/burrnesha
llama
["LlamaForCausalLM"]
32000
float16
4.34.0
3,200
8,640
26
32
32
silu
null
True
2,048
10,000
0.000001
0.02
1
2
false
100
1
false
null
null
true
false
false
false
false
false
false
false
false
true
3.19488
medium
medium
6,652
CareerNinja/T5-Base-data-v4-model-v1
t5
["T5ForConditionalGeneration"]
32128
float32
4.25.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,653
CareerNinja/T5-Small-data-v4-model-v2
t5
["T5ForConditionalGeneration"]
32128
float32
4.25.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,654
CareerNinja/t5_large_1e-4_on_V3dataset
t5
["T5ForConditionalGeneration"]
32128
float32
4.22.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,655
CarelessLee/MCQ_pooled_full_rationale_confidence_predictor
llama
["LlamaForSequenceClassification"]
128256
float32
4.43.3
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
6,656
Carick/albert-base-v2-wordnet_dataset_two-fine-tuned
albert
["AlbertForSequenceClassification"]
30000
float32
4.45.1
768
3,072
12
12
null
gelu_new
null
null
512
null
null
0.02
2
3
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,657
Carick/bart-fine-tuned-term-typing
bart
["BartForConditionalGeneration"]
50265
float32
4.45.1
null
null
6
null
null
null
0.1
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
6,658
Carick/roberta-base-wordnet_dataset_three-fine-tuned
roberta
["RobertaForSequenceClassification"]
50265
float32
4.45.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,659
Carick/roberta-base-wordnet_dataset_two-fine-tuned
roberta
["RobertaForSequenceClassification"]
50265
float32
4.45.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,660
Carick/xlm-roberta-base-wordnet_combined_one-fine-tuned
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.45.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,661
Carick/xlm-roberta-base-wordnet_dataset_three-fine-tuned
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.45.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,662
Carick/xlm-roberta-base-wordnet_dataset_two-fine-tuned
xlm-roberta
["XLMRobertaForSequenceClassification"]
250002
float32
4.45.1
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,663
Carina124/plant-vit-model-3
vit
["ViTForImageClassification"]
null
float32
4.28.0
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
6,664
CarinaCHEN/test1
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.51.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,665
CarlBrendt/layoutlmv2-base-uncased_finetuned_docvqa
layoutlmv2
["LayoutLMv2ForQuestionAnswering"]
30522
float32
4.35.0
768
3,072
12
12
null
gelu
null
null
512
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
6,666
CarlSable/distilbert-base-uncased-finetuned-squad-d5716d28
distilbert
["DistilBertForMaskedLM"]
30522
null
4.49.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,667
CarlSable/marian-finetuned-kde4-en-to-fr
marian
["MarianMTModel"]
59514
float32
4.48.3
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
6,668
CarlosApo01/Llama-2-7b-chat-finetune
llama
["LlamaForCausalLM"]
32000
bfloat16
4.35.2
4,096
11,008
32
32
32
silu
null
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
6,669
CarlosElArtista/poca-SoccerTwos
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,670
CarlosFersoft/GPBusiness0001_Q6
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
6,671
CarlosMarch/pruebaPostTagging-MarIA
roberta
["RobertaForTokenClassification"]
50262
float32
4.46.0
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
6,672
Carlosino/en2zh40
marian
["MarianMTModel"]
65001
float32
4.27.4
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
6,673
Carlosino/en2zh50
marian
["MarianMTModel"]
65001
float32
4.27.4
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
6,674
Carlosino/iwslt2017_1410
t5
["T5ForConditionalGeneration"]
32128
float32
4.27.4
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,675
Carlosino/iwslt2017_857
t5
["T5ForConditionalGeneration"]
32128
float32
4.27.4
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,676
Carlosino/zh2en40
marian
["MarianMTModel"]
65001
float32
4.27.4
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
6,677
Carlosslocar/model
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.38.0.dev0
2,048
16,384
18
8
1
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
6,678
Carlosslocar/test2
bert
["BertForSequenceClassification"]
30522
float32
4.39.3
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
6,679
Carlosslocar/test3
bert
["BertForSequenceClassification"]
30522
float32
4.39.3
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
6,680
Carlosslocar/test4
gemma
["GemmaForSequenceClassification"]
256000
float32
4.39.3
2,048
16,384
18
8
1
gelu
0
True
8,192
10,000
0.000001
0.02
2
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
6,681
CarmelaFinianos/Arabicttegypt
speecht5
["SpeechT5ForTextToSpeech"]
98
float32
4.46.2
768
null
null
null
null
gelu
0.1
False
null
null
null
0.02
0
2
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,682
Carmenvt7/roberta-base-bne-ner
roberta
["RobertaForTokenClassification"]
50262
float32
4.51.3
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,683
CarolLiu999/chinese-emotion-model
distilbert
["DistilBertForSequenceClassification"]
119547
float32
4.45.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,684
CarolLiu999/chinese-emotion-model-all
distilbert
["DistilBertForSequenceClassification"]
119547
float32
4.40.1
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,685
CarolLiu999/vivit-finetuned-6class-10epoch-5
vivit
["VivitForVideoClassification"]
null
float32
4.40.1
768
3,072
12
12
null
gelu_fast
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
6,686
CarolLiu999/vivit-finetuned-6class-3epoch-4
vivit
["VivitForVideoClassification"]
null
float32
4.39.1
768
3,072
12
12
null
gelu_fast
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
6,687
CarperAI/diff-codegen-2b-v2
codegen
["CodeGenForCausalLM"]
50295
float16
4.22.0.dev0
null
null
null
null
null
null
null
False
null
null
null
0.02
1
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,688
CarperAI/diff-codegen-350m-v2
codegen
["CodeGenForCausalLM"]
50295
float16
4.22.0.dev0
null
null
null
null
null
null
null
False
null
null
null
0.02
1
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,689
CarperAI/diff-codegen-6b-v2
codegen
["CodeGenForCausalLM"]
50295
float16
4.22.0.dev0
null
null
null
null
null
null
null
False
null
null
null
0.02
1
50256
false
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,690
CarperAI/openai_summarize_tldr_ppo
gptj
["GPTJForCausalLM"]
50400
float32
4.25.1
null
null
null
null
null
null
null
False
null
null
null
0.02
50256
50256
false
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
6,691
CarperAI/openai_summarize_tldr_sft
gptj
["GPTJForCausalLM"]
50400
float32
4.25.1
null
null
null
null
null
null
null
False
null
null
null
0.02
50256
50256
false
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
6,692
CarrotAI/Llama-3.2-Rabbit-Ko-3B-Instruct-2412
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.3
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128009
true
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
6,693
Carrymachine/model_output
bert
["BertForSequenceClassification"]
30000
float32
4.42.4
768
3,072
12
12
null
gelu
null
True
300
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,694
Cartinoe5930/CodeActAgent-Mistral-7b-v0.1_weight_diff
mistral
["MistralForCausalLM"]
32000
bfloat16
4.41.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
6,695
Cartinoe5930/Einstein-v4-7B_weight_diff
mistral
["MistralForCausalLM"]
32000
bfloat16
4.41.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
6,696
Cartinoe5930/Llama2_init_Mistral
llama
["LlamaForCausalLM"]
32000
float16
4.36.2
4,096
14,336
32
32
8
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
medium
6,697
Cartinoe5930/MAmmoTH-7B-Mistral_weight_diff
mistral
["MistralForCausalLM"]
32000
bfloat16
4.41.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
6,698
Cartinoe5930/MetaMath-Mistral-7B_weight_diff
mistral
["MistralForCausalLM"]
32000
bfloat16
4.41.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
6,699
Cartinoe5930/QAlign-1.5B-stage-2-50K
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.49.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
true
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long