index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
8
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
158 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
bool
2 classes
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
6,800
Chaanan/vicuna-7b-v1.5-W8A8-Dynamic-Per-Token
llama
["LlamaForCausalLM"]
32000
float16
4.44.2
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
6,801
Chabuno/finetuning-sentiment-model-3000-samples
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.24.0
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,802
Chaconne/llm-detect-ai-generated-text-bert
bert
["BertForSequenceClassification"]
28996
float32
4.35.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,803
Chaeeun-Kim/L_pretrain_recover_tokenizer_pooling_ep110
mt5
["MT5ForConditionalGeneration"]
50134
float32
4.43.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,804
Chaeeun-Kim/L_pretrain_recover_tokenizer_trimmed
mt5
["MT5ForConditionalGeneration"]
119196
float32
4.43.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,805
Chaeeun-Kim/L_subset_pretrain_encdec
mt5
["MT5ForConditionalGeneration"]
250113
float32
4.31.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,806
Chaeeun-Kim/L_subset_pretrain_encdec_shuffle
mt5
["MT5ForConditionalGeneration"]
250113
float32
4.31.0
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,807
Chaeeun-Kim/private_unsup_ep1
mt5
["MT5ForConditionalGeneration"]
250106
float32
4.46.3
null
null
null
null
null
null
null
True
null
null
null
null
null
1
false
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,808
Chaeeun-Kim/spider-large-pretrain-2019
dpr
["DPRContextEncoder"]
30522
float32
4.41.2
1,024
4,096
24
16
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
short
6,809
Chaeguevara/PPO-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,810
Chaeseung/log2char_Orion-14B-Base_v2
orion
["OrionForCausalLM"]
84608
float32
4.37.1
5,120
15,360
40
40
40
silu
null
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
12.582912
large
medium
6,811
Chaeseung/log2char_llama3_EN_v1
llama
["LlamaForCausalLM"]
128256
float32
4.37.1
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
6,812
Chaeyeon/first-patent-tokenizer
bert
["BertModel"]
160518
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,813
Chaeyeon/git-base-sketch
git
["GitForCausalLM"]
30522
float32
null
768
3,072
6
12
null
gelu
null
True
1,024
null
null
0.02
101
102
false
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.042467
small
short
6,814
ChafikAiEng/multilingual-e5-base-finetuned-studytours
xlm-roberta
["XLMRobertaModel"]
250002
float32
4.39.3
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,815
Chahat7874/hindi_turkish
wav2vec2
["Wav2Vec2ForCTC"]
70
float32
4.41.2
1,024
4,096
24
16
null
gelu
0
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
6,816
Chahatdatascience/config-0
wav2vec2
["Wav2Vec2ForCTC"]
106
float32
4.41.2
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
6,817
Chahatdatascience/config-1
wav2vec2
["Wav2Vec2ForCTC"]
106
float32
4.41.2
1,024
4,096
24
16
null
gelu
0.15
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
6,818
Chahatdatascience/wav2vec2-large-xls-r-300m-hindi_marathi-colab-v1-20epochs
wav2vec2
["Wav2Vec2ForCTC"]
106
float32
4.41.2
1,024
4,096
24
16
null
gelu
0.1
null
null
null
null
0.02
1
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
6,819
Chaitanya798800/videomae-large_Sports_action_recognition_5
videomae
["VideoMAEForVideoClassification"]
null
float32
4.39.3
1,024
4,096
24
16
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.30199
small
null
6,820
Chaitanya895/Sahayak_Vexa_AI_fine_tuned_model
mpnet
["MPNetForQuestionAnswering"]
30527
null
4.28.1
768
3,072
12
12
null
gelu
null
null
514
null
null
0.02
0
2
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
short
6,821
Chakshu/test-2fd961bc-aec9-4cec-8f2a-9e5ce676127a
bert
["BertForSequenceClassification"]
30522
float32
4.31.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,822
Chakshu/test-470446d9-2c78-4af9-80f1-fd17bf2c6275
bert
["BertForSequenceClassification"]
30522
float32
4.31.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,823
Chakshu/test-f3a3c123-9ecd-4ba1-8ac3-3fd8d64863dd
bert
["BertForSequenceClassification"]
30522
float32
4.31.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,824
ChakuChidiya/ppo-Lunar-Lander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,825
ChallengerSpaceShuttle/imdb-gpt-test
null
[]
50257
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,826
Chamoda/Llama-2-7b-chat-fin24
llama
["LlamaForCausalLM"]
32000
float16
4.31.0
4,096
11,008
32
32
32
silu
null
True
4,096
null
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
6,827
Chan-007/ppo-LunarLander-v2
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,828
Chan2chan1/solar_test240517_4bit
llama
["LlamaForCausalLM"]
32000
bfloat16
4.39.2
4,096
14,336
48
32
8
silu
0
False
4,096
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
9.663676
large
medium
6,829
ChanE-666/BERT-8bit
bert
["BertLMHeadModel"]
30522
float16
4.38.1
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,830
ChanE-666/phi-2-QLora
phi
["PhiForCausalLM"]
51200
float16
4.38.1
2,560
10,240
32
32
32
gelu_new
0
True
2,048
10,000
null
0.02
50256
50256
false
80
1
false
null
null
false
false
false
false
false
false
false
false
false
true
2.516582
medium
medium
6,831
ChanE-666/phi2-bnb-8bit
phi
["PhiForCausalLM"]
51200
float16
4.38.1
2,560
10,240
32
32
32
gelu_new
0
True
2,048
10,000
null
0.02
50256
50256
false
80
1
false
null
null
false
false
false
false
false
false
false
false
false
true
2.516582
medium
medium
6,832
ChanP/finetuned-th-to-en
marian
["MarianMTModel"]
62307
float32
4.15.0
null
null
6
null
null
null
0
True
512
null
null
null
0
0
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
6,833
ChandanGR/Qwen2.5-0.5B-4bit-channel
qwen2
["Qwen2ForCausalLM"]
151936
float16
4.46.3
896
4,864
24
14
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
6,834
ChandanGR/llama-3.2-1B-gptq-4bit-channel
llama
["LlamaForCausalLM"]
128256
float16
4.46.3
2,048
8,192
16
32
8
silu
0
True
131,072
500,000
0.00001
0.02
128000
[128001, 128008, 128009]
true
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
6,835
Chandana03/Llama-2-7b-chat-hf.legal-FineTune
llama
["LlamaForCausalLM"]
32000
float16
4.37.2
4,096
11,008
32
32
32
silu
0
True
4,096
10,000
0.00001
0.02
1
2
false
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
6,836
Chandans01/custom-chandan-samsum
bart
["BartForConditionalGeneration"]
50264
float32
4.44.0
null
null
12
null
null
null
0
True
1,024
null
null
null
0
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
short
6,837
ChandlerU11/t5_fine_2.0
t5
["T5ForConditionalGeneration"]
32128
float32
4.28.1
null
null
null
null
null
null
null
True
null
null
null
null
null
1
null
null
null
false
null
null
false
false
false
true
false
false
false
false
false
false
null
null
null
6,838
Chandlernnnnnn/ESM-finetuned-sequence-reconstruction
esm
["EsmForTokenClassification"]
33
float32
4.30.2
480
1,920
12
20
null
gelu
null
True
1,026
null
null
0.02
null
null
null
24
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.033178
small
short
6,839
ChandraAdhikarla/Quantised_Model
llama
["LlamaForCausalLM"]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
6,840
ChandraAdhikarla/TTSmodel
llama
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
true
false
false
false
false
false
false
false
false
false
null
null
null
6,841
Chandrababu-Namani/InLegalBERT-fine-tuned
bert
["BertForSequenceClassification"]
30522
float32
4.35.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
0
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,842
Chandrateja/lamma3-8b-oig-unsloth-merged
llama
["LlamaForCausalLM"]
128256
float16
4.38.2
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
6,843
Chang-Hoo/gemma-3-1b-cpt-dpo-0707
gemma3_text
["Gemma3ForCausalLM"]
262144
bfloat16
4.51.3
1,152
6,912
26
4
1
null
0
False
32,768
1,000,000
0.000001
0.02
2
1
null
288
4
false
null
null
false
false
false
false
false
false
false
false
true
true
0.414056
small
very_long
6,844
Chang-Hoo/gemma-3-1b-cpt-it-0707
gemma3_text
["Gemma3ForCausalLM"]
262144
bfloat16
4.51.3
1,152
6,912
26
4
1
null
0
False
32,768
1,000,000
0.000001
0.02
2
1
null
288
4
false
null
null
false
false
false
false
false
false
false
false
true
true
0.414056
small
very_long
6,845
ChangeIsKey/text-dating
roberta
["RobertaForSequenceClassification"]
50265
float32
4.36.2
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
6,846
Changgil/K2S3-27b-v0.0
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.42.3
4,608
36,864
46
32
16
gelu_pytorch_tanh
0
True
8,192
10,000
0.000001
0.02
2
1
null
144
2
false
null
null
false
false
false
false
false
false
false
false
true
true
11.720983
large
long
6,847
Changgil/K2S3-Mistral-7b-v1.2
mistral
["MistralForCausalLM"]
48000
float16
4.38.0.dev0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
6,848
Changgil/K2S3-Mistral-7b-v1.48
mistral
["MistralForCausalLM"]
48000
float16
4.38.0.dev0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
6,849
Changgil/K2S3-Mistral-7bx2-48layers_v1.2
mistral
["MistralForCausalLM"]
48000
float16
4.38.0.dev0
4,096
14,336
48
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
9.663676
large
very_long
6,850
Changgil/K2S3-v0.1
mistral
["MistralForCausalLM"]
48000
float16
4.38.0.dev0
4,096
14,336
64
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.884902
large
very_long
6,851
Changmin0816/1008model
bert
["BertForSequenceClassification"]
30000
float32
4.44.2
768
3,072
12
12
null
gelu
null
True
300
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,852
Changmin0816/ViT_beans
vit
["ViTForImageClassification"]
null
float32
4.44.2
768
3,072
12
12
null
gelu
null
null
null
null
null
0.02
null
null
null
64
null
false
null
null
false
false
false
false
false
false
false
false
false
false
0.084935
small
null
6,853
Chanjeans/tfchatbot_2
gpt2
["GPT2LMHeadModel"]
51204
float32
4.48.2
null
null
null
null
null
null
null
True
null
null
null
0.02
0
1
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
6,854
Chanjeans/tfchatbot_3
gpt2
["GPT2LMHeadModel"]
51204
float32
4.48.3
null
null
null
null
null
null
null
True
null
null
null
0.02
0
1
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
6,855
Chansik/whisper-large-v3-ko-aihub
whisper
["WhisperForConditionalGeneration"]
51866
float32
4.45.1
null
null
32
null
null
null
0
True
null
null
null
null
50257
50257
null
null
null
false
null
null
false
false
false
false
true
false
false
false
false
false
null
null
null
6,856
ChaoChao2023/RoBERTa_no_sft
roberta
["RobertaForSequenceClassification"]
50265
null
4.35.0
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
6,857
ChaoChao2023/RoBERTa_squad
roberta
["RobertaForSequenceClassification"]
50265
float32
4.38.2
1,024
4,096
24
16
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.30199
small
short
6,858
ChaoHou/ESMDance
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,859
ChaoHou/SeqDance
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,860
Chaoqi2000/BestDNA6mer
bert
["BertForMaskedLM"]
4101
float32
4.29.2
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,861
ChaoqunZhao/sd-class-butterflies-32
null
[]
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
false
null
null
null
6,862
ChaosAiVision/DeepSeek-R1-0528-Qwen3-8B-vi-sft-medical-9k
qwen3
["Qwen3ForCausalLM"]
151936
bfloat16
4.52.4
4,096
12,288
36
32
8
silu
0
True
131,072
1,000,000
0.000001
0.02
151643
151645
false
128
4
false
null
null
false
false
false
false
false
false
false
false
true
true
7.247757
large
very_long
6,863
ChaoticNeutrals/Bepis_9B
mistral
["MistralForCausalLM"]
32000
float16
4.38.1
4,096
14,336
40
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
8.053064
large
very_long
6,864
ChaoticNeutrals/BuRP_7B
mistral
["MistralForCausalLM"]
32000
bfloat16
4.38.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
6,865
ChaoticNeutrals/Captain-Eris_Violet_Toxic-Magnum-12B
mistral
["MistralForCausalLM"]
131072
bfloat16
4.48.2
5,120
14,336
40
32
8
silu
0
True
1,024,000
1,000,000
0.00001
0.02
1
2
false
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
6,866
ChaoticNeutrals/Cookie_7B
mistral
["MistralForCausalLM"]
32000
float16
4.35.2
4,096
14,336
32
32
8
silu
0
True
8,192
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
long
6,867
ChaoticNeutrals/Eris_Prime-V2-7B
mistral
["MistralForCausalLM"]
32000
bfloat16
4.38.2
4,096
14,336
32
32
8
silu
0
True
32,768
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
6,868
ChaoticNeutrals/Eris_PrimeV4-Vision-32k-7B
mistral
["MistralForCausalLM"]
32000
bfloat16
4.38.2
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
6,869
ChaoticNeutrals/Eris_PrimeV4.20-Vision-32k-7B
mistral
["MistralForCausalLM"]
32000
bfloat16
4.38.2
4,096
14,336
32
32
8
silu
0
True
32,768
1,000,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
6,870
ChaoticNeutrals/Hathor_Respawn-L3-8B-v0.8
llama
["LlamaForCausalLM"]
128256
bfloat16
4.38.1
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
6,871
ChaoticNeutrals/Pocky_9B
mistral
["MistralForCausalLM"]
32000
float16
4.37.2
4,096
14,336
40
32
8
silu
0
True
8,192
10,000
0.00001
0.02
1
2
false
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
8.053064
large
long
6,872
ChaoticNeutrals/Poppy_Porpoise-v0.6-L3-8B
llama
["LlamaForCausalLM"]
128256
bfloat16
4.39.3
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
6,873
ChaoticNeutrals/Puppy_Purpose_0.69
llama
["LlamaForCausalLM"]
128256
float16
4.39.3
4,096
14,336
32
32
8
silu
0
True
8,192
500,000
0.00001
0.02
128000
128009
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
6,874
ChaoticNeutrals/Templar_v1_8B
llama
["LlamaForCausalLM"]
128256
bfloat16
4.41.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000
128001
false
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
6,875
CharGen-Archive/CharGen-v3-beta-244-s6
mistral
["MistralForCausalLM"]
32768
float32
4.45.2
6,144
16,384
56
48
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
6
false
null
null
false
false
false
false
false
false
true
false
true
true
25.367151
large
very_long
6,876
CharGen-Archive/CharGen-v3-beta-264-s0
mistral
["MistralForCausalLM"]
32768
bfloat16
4.46.1
6,144
16,384
56
48
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1
2
false
128
6
false
null
null
false
false
false
false
false
false
true
false
true
true
25.367151
large
very_long
6,877
CharGen/CharGen-v3-mini
llama
["LlamaForCausalLM"]
128256
bfloat16
4.51.0
3,072
9,216
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000
128019
false
96
4
false
null
null
true
false
false
false
false
false
false
false
true
true
3.623879
medium
very_long
6,878
CharacterEcho/Rohit-Sharma
stablelm
["StableLmForCausalLM"]
50280
float16
4.42.3
2,560
6,912
32
32
32
silu
0
True
4,096
10,000
null
0.02
50278
50279
false
80
1
false
null
null
false
false
false
false
false
false
false
false
false
true
2.516582
medium
medium
6,879
CharanSaiVaddi/BERT-finetuned-TextSummarization
bert
["BertLMHeadModel"]
30522
float32
4.47.0
768
3,072
12
12
null
gelu
null
True
512
null
null
0.02
null
null
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,880
Charankarnati18/TASK3
roberta
["RobertaForSequenceClassification"]
50265
float32
4.48.3
768
3,072
12
12
null
gelu
null
True
514
null
null
0.02
0
2
null
64
null
false
null
null
false
true
false
false
false
false
false
false
false
false
0.084935
small
short
6,881
Charankarnati18/TASK4
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.48.3
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,882
Charles-59800/six_sentiment_distillbert_fine_tune
distilbert
["DistilBertForSequenceClassification"]
30522
float32
4.30.2
null
null
null
null
null
null
0.1
null
512
null
null
0.02
null
null
null
null
null
false
null
null
false
true
false
false
false
false
false
false
false
false
null
null
short
6,883
Charles333/gpt2_fintuned_json_data
gpt2
["GPT2LMHeadModel"]
50257
float32
4.35.2
null
null
null
null
null
null
null
True
null
null
null
0.02
50256
50256
null
null
null
false
null
null
false
false
true
false
false
false
false
false
false
false
null
null
null
6,884
CharlesLi/CoT-SFT-3B
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.51.1
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
6,885
CharlesLi/Direct-SFT-3B
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.51.1
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
6,886
CharlesLi/G1-7B
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.51.3
3,584
18,944
28
28
4
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
false
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
6,887
CharlesLi/OpenELM-1_1B-DPO-full-max-reward-least-similar
openelm
["OpenELMForCausalLM"]
32000
bfloat16
4.45.1
null
null
null
null
null
null
null
False
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
6,888
CharlesLi/OpenELM-1_1B-DPO-full-max-reward-most-similar
openelm
["OpenELMForCausalLM"]
32000
bfloat16
4.45.1
null
null
null
null
null
null
null
False
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
6,889
CharlesLi/OpenELM-1_1B-DPO-full-most-similar
openelm
["OpenELMForCausalLM"]
32000
bfloat16
4.45.1
null
null
null
null
null
null
null
False
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
6,890
CharlesLi/OpenELM-1_1B-SFT-3
openelm
["OpenELMForCausalLM"]
32000
bfloat16
4.44.2
null
null
null
null
null
null
null
False
null
null
null
0.02
1
2
null
null
null
false
null
null
false
false
false
false
false
false
false
false
false
true
null
null
null
6,891
CharlesLi/dapo_5_epoch_graph_task_3B_1150
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.51.1
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
6,892
CharlesLi/dapo_5_epoch_graph_task_3B_800
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.51.1
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
6,893
CharlesLi/graph_grpo_320
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.50.1
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
6,894
CharlesLi/grpo_5_epoch_graph_hard_base_gpu_700
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.51.1
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
6,895
CharlesLi/grpo_5_epoch_graph_hard_base_gpu_800
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.50.1
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
6,896
CharlesLi/grpo_5_epoch_graph_hard_base_start_200
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.50.1
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
6,897
CharlesLi/grpo_5_epoch_graph_hard_base_start_400
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.50.1
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
6,898
CharlesLi/grpo_5_epoch_graph_hard_ins_gpu_400
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.50.1
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long
6,899
CharlesLi/grpo_5_epoch_graph_hard_ins_gpu_700
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.51.1
2,048
11,008
36
16
2
silu
0
True
32,768
1,000,000
0.000001
0.02
151643
151645
true
128
8
false
null
null
false
false
false
false
false
false
false
false
true
true
1.811939
medium
very_long