index
int64
0
125k
modelId
stringlengths
6
115
config_model_type
stringlengths
2
46
config_architectures
stringlengths
2
91
config_vocab_size
stringlengths
1
10
config_torch_dtype
stringclasses
7 values
config_transformers_version
stringclasses
228 values
config_hidden_size
float64
0
18.4k
config_intermediate_size
float64
0
25.2M
config_num_hidden_layers
float64
-1
260
config_num_attention_heads
stringclasses
47 values
config_num_key_value_heads
float64
0
4.1k
config_hidden_act
stringclasses
19 values
config_attention_dropout
float64
0
0.5
config_use_cache
stringclasses
3 values
config_max_position_embeddings
float64
-1
10.5M
config_rope_theta
float64
256
100B
config_rms_norm_eps
float64
0
0
config_initializer_range
float64
0
2
config_bos_token_id
stringclasses
207 values
config_eos_token_id
stringclasses
339 values
config_tie_word_embeddings
stringclasses
2 values
config_head_dimension
float64
0.5
3.07k
config_gqa_ratio
float64
0.5
64
config_moe_enabled
bool
1 class
config_n_routed_experts
float64
1
384
config_num_experts_per_tok
float64
1
64
is_llama_family
bool
2 classes
is_bert_family
bool
2 classes
is_gpt_family
bool
2 classes
is_t5_family
bool
2 classes
is_whisper_family
bool
2 classes
is_deepseek_family
bool
2 classes
is_mistral_family
bool
2 classes
uses_moe
bool
2 classes
uses_gqa
bool
2 classes
uses_rope
bool
2 classes
config_approx_params_billions
float64
-0.2
606
size_category
stringclasses
4 values
context_category
stringclasses
4 values
likes
int64
0
12.5k
trendingScore
float64
0
479
downloads
int64
0
91.1M
tags
stringlengths
13
9.54k
pipeline_tag
stringclasses
51 values
library_name
stringclasses
143 values
createdAt
stringdate
2022-03-02 23:29:05
2025-07-11 13:57:16
parent_model
stringlengths
2
1.34k
finetune_parent
stringlengths
2
109
quantized_parent
stringlengths
2
95
adapter_parent
stringclasses
553 values
merge_parent
stringlengths
2
1.34k
license
stringclasses
60 values
region
stringclasses
3 values
arxiv_id
float64
0
2.51k
200
0x0xx/0x5xx
stablelm
["StableLmForCausalLM"]
100352
bfloat16
4.40.2
2,048
5,632
24
32
32
silu
0
False
4,096
10,000
null
0.02
100257.0
100278
False
64
1
false
null
null
false
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
0
0
5
['transformers', 'safetensors', 'stablelm', 'text-generation', 'conversational', 'autotrain_compatible', 'endpoints_compatible', 'region:us']
text-generation
transformers
2024-07-31T10:06:01.000Z
[]
[]
[]
[]
[]
null
us
null
201
0x1202/0066104b-f64e-4693-a892-f16d8be22d32
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000.0
128009
False
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct', 'base_model:adapter:VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct', 'license:llama3.1', 'region:us']
null
peft
2025-01-27T01:20:19.000Z
['VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct']
[]
[]
['VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct']
[]
llama3.1
us
null
202
0x1202/00b1d13e-dd1f-4249-a4d5-4c61b81a3fdb
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.46.0
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2.0
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
0
0
3
['peft', 'safetensors', 'gemma2', 'axolotl', 'generated_from_trainer', 'base_model:princeton-nlp/gemma-2-9b-it-SimPO', 'base_model:adapter:princeton-nlp/gemma-2-9b-it-SimPO', 'license:mit', 'region:us']
null
peft
2024-12-28T15:03:24.000Z
['princeton-nlp/gemma-2-9b-it-SimPO']
[]
[]
['princeton-nlp/gemma-2-9b-it-SimPO']
[]
mit
us
null
203
0x1202/058d89e6-e9a5-489c-bf57-beac76453cab
mixtral
["MixtralForCausalLM"]
32000
float32
4.46.0
1,024
3,584
2
32
8
silu
0
False
131,072
1,000,000
0.00001
0.02
1.0
2
False
32
4
false
8
2
false
false
false
false
false
false
false
true
true
true
0.025166
small
very_long
0
0
3
['peft', 'safetensors', 'mixtral', 'axolotl', 'generated_from_trainer', 'base_model:TitanML/tiny-mixtral', 'base_model:adapter:TitanML/tiny-mixtral', 'region:us']
null
peft
2025-01-16T08:58:21.000Z
['TitanML/tiny-mixtral']
[]
[]
['TitanML/tiny-mixtral']
[]
null
us
null
204
0x1202/0b63e882-2b69-418d-9a75-e697c8f85a25
llama
["LlamaForCausalLM"]
32000
float16
4.46.0
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0.0
2
False
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:fxmarty/tiny-llama-fast-tokenizer', 'base_model:adapter:fxmarty/tiny-llama-fast-tokenizer', 'region:us']
null
peft
2025-01-16T08:19:08.000Z
['fxmarty/tiny-llama-fast-tokenizer']
[]
[]
['fxmarty/tiny-llama-fast-tokenizer']
[]
null
us
null
205
0x1202/0c85f687-764e-4e9d-87b9-6368e08690f8
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
True
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
0
0
2
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:Qwen/Qwen2-0.5B-Instruct', 'base_model:adapter:Qwen/Qwen2-0.5B-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2025-01-14T12:04:03.000Z
['Qwen/Qwen2-0.5B-Instruct']
[]
[]
['Qwen/Qwen2-0.5B-Instruct']
[]
apache-2.0
us
null
206
0x1202/0ca74a48-513c-4cfe-9867-47e7787f9ddc
llama
["LlamaForCausalLM"]
32016
float16
4.46.0
4,096
11,008
32
32
32
silu
null
False
16,384
1,000,000
0.00001
0.02
1.0
2
False
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'custom_code', 'base_model:NousResearch/CodeLlama-7b-hf-flash', 'base_model:adapter:NousResearch/CodeLlama-7b-hf-flash', 'region:us']
null
peft
2024-12-28T07:48:57.000Z
['NousResearch/CodeLlama-7b-hf-flash']
[]
[]
['NousResearch/CodeLlama-7b-hf-flash']
[]
null
us
null
207
0x1202/0f98d215-a7b0-4254-a977-cacb20eab7d3
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
True
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
0
0
2
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Qwen2-0.5B-Instruct', 'base_model:adapter:unsloth/Qwen2-0.5B-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T09:25:14.000Z
['unsloth/Qwen2-0.5B-Instruct']
[]
[]
['unsloth/Qwen2-0.5B-Instruct']
[]
apache-2.0
us
null
208
0x1202/159f0ee6-96ec-4870-aa5a-a25be6bbbbe3
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
131,072
1,000,000
0.000001
0.02
null
151643
True
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
0
0
2
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:Qwen/Qwen2-0.5B', 'base_model:adapter:Qwen/Qwen2-0.5B', 'license:apache-2.0', 'region:us']
null
peft
2025-01-10T18:25:18.000Z
['Qwen/Qwen2-0.5B']
[]
[]
['Qwen/Qwen2-0.5B']
[]
apache-2.0
us
null
209
0x1202/16414734-05a3-4ad8-bf2a-71db3633669a
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
2,048
5,632
22
32
4
silu
0
False
2,048
10,000
0.00001
0.02
1.0
2
False
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:TinyLlama/TinyLlama-1.1B-Chat-v1.0', 'base_model:adapter:TinyLlama/TinyLlama-1.1B-Chat-v1.0', 'license:apache-2.0', 'region:us']
null
peft
2025-01-14T06:14:50.000Z
['TinyLlama/TinyLlama-1.1B-Chat-v1.0']
[]
[]
['TinyLlama/TinyLlama-1.1B-Chat-v1.0']
[]
apache-2.0
us
null
210
0x1202/17cf3a22-e51f-4277-896a-beb14ce43680
llama
["LlamaForCausalLM"]
32000
float16
4.46.0
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0.0
2
False
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:HuggingFaceM4/tiny-random-LlamaForCausalLM', 'base_model:adapter:HuggingFaceM4/tiny-random-LlamaForCausalLM', 'region:us']
null
peft
2025-01-19T21:27:31.000Z
['HuggingFaceM4/tiny-random-LlamaForCausalLM']
[]
[]
['HuggingFaceM4/tiny-random-LlamaForCausalLM']
[]
null
us
null
211
0x1202/1bf16ccc-d875-4de8-88f5-9a52c3846706
llama
["LlamaForCausalLM"]
46336
float16
4.46.0
4,096
11,008
32
32
32
silu
0
False
2,048
10,000
0.00001
0.02
1.0
2
False
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
0
0
1
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:oopsung/llama2-7b-koNqa-test-v1', 'base_model:adapter:oopsung/llama2-7b-koNqa-test-v1', 'region:us']
null
peft
2025-01-24T09:48:52.000Z
['oopsung/llama2-7b-koNqa-test-v1']
[]
[]
['oopsung/llama2-7b-koNqa-test-v1']
[]
null
us
null
212
0x1202/1d2564d7-388e-4114-966c-36e81231c040
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.46.0
3,072
24,576
28
16
16
gelu
0
False
8,192
10,000
0.000001
0.02
2.0
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
0
0
4
['peft', 'safetensors', 'gemma', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/gemma-7b-it', 'base_model:adapter:unsloth/gemma-7b-it', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T10:22:41.000Z
['unsloth/gemma-7b-it']
[]
[]
['unsloth/gemma-7b-it']
[]
apache-2.0
us
null
213
0x1202/1d888d48-f6d9-41d9-88cd-4db9921ba6db
qwen2
["Qwen2ForCausalLM"]
151936
float32
4.46.0
8
32
2
4
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
True
2
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.000002
small
very_long
0
0
2
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:fxmarty/tiny-dummy-qwen2', 'base_model:adapter:fxmarty/tiny-dummy-qwen2', 'license:mit', 'region:us']
null
peft
2024-12-28T08:29:26.000Z
['fxmarty/tiny-dummy-qwen2']
[]
[]
['fxmarty/tiny-dummy-qwen2']
[]
mit
us
null
214
0x1202/1f347382-481f-4f8a-a36b-e5ebaedeed9e
mistral
["MistralForCausalLM"]
131072
bfloat16
4.46.0
5,120
14,336
40
32
8
silu
0
False
1,024,000
1,000,000
0.00001
0.02
1.0
2
False
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
0
0
4
['peft', 'safetensors', 'mistral', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Mistral-Nemo-Base-2407', 'base_model:adapter:unsloth/Mistral-Nemo-Base-2407', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T03:50:58.000Z
['unsloth/Mistral-Nemo-Base-2407']
[]
[]
['unsloth/Mistral-Nemo-Base-2407']
[]
apache-2.0
us
null
215
0x1202/1fc190cf-2980-446a-bac7-773545e121aa
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000.0
128001
False
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:scb10x/llama-3-typhoon-v1.5-8b-instruct', 'base_model:adapter:scb10x/llama-3-typhoon-v1.5-8b-instruct', 'license:llama3', 'region:us']
null
peft
2025-01-14T08:58:53.000Z
['scb10x/llama-3-typhoon-v1.5-8b-instruct']
[]
[]
['scb10x/llama-3-typhoon-v1.5-8b-instruct']
[]
llama3
us
null
216
0x1202/22adb637-adc6-4d1b-8036-c577e6b0a008
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
4,096
10,000
0.000001
0.02
null
151645
True
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
medium
0
0
2
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Qwen2.5-Math-1.5B-Instruct', 'base_model:adapter:unsloth/Qwen2.5-Math-1.5B-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T06:44:02.000Z
['unsloth/Qwen2.5-Math-1.5B-Instruct']
[]
[]
['unsloth/Qwen2.5-Math-1.5B-Instruct']
[]
apache-2.0
us
null
217
0x1202/22b96257-6d8a-41ae-8f63-586c2d4b4c6c
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000.0
128009
False
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:aisingapore/Llama-SEA-LION-v2-8B-IT', 'base_model:adapter:aisingapore/Llama-SEA-LION-v2-8B-IT', 'license:llama3', 'region:us']
null
peft
2025-01-29T09:57:09.000Z
['aisingapore/Llama-SEA-LION-v2-8B-IT']
[]
[]
['aisingapore/Llama-SEA-LION-v2-8B-IT']
[]
llama3
us
null
218
0x1202/2aa3e0ca-06a8-4584-8e8a-9e33fe39fafe
mistral
["MistralForCausalLM"]
32000
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1.0
2
False
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
0
0
1
['peft', 'safetensors', 'mistral', 'axolotl', 'generated_from_trainer', 'base_model:NousResearch/Nous-Capybara-7B-V1.9', 'base_model:adapter:NousResearch/Nous-Capybara-7B-V1.9', 'license:mit', 'region:us']
null
peft
2025-01-24T17:23:24.000Z
['NousResearch/Nous-Capybara-7B-V1.9']
[]
[]
['NousResearch/Nous-Capybara-7B-V1.9']
[]
mit
us
null
219
0x1202/2bcd8539-54d7-4806-8227-c5b68e1bddb0
llama
["LlamaForCausalLM"]
32000
float16
4.46.0
4,096
11,008
32
32
32
silu
0
False
2,048
10,000
0.000001
0.02
1.0
2
False
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:lmsys/vicuna-7b-v1.3', 'base_model:adapter:lmsys/vicuna-7b-v1.3', 'region:us']
null
peft
2025-01-16T08:35:56.000Z
['lmsys/vicuna-7b-v1.3']
[]
[]
['lmsys/vicuna-7b-v1.3']
[]
null
us
null
220
0x1202/2c9bde5d-116e-4770-ac30-91b127d3833b
llama
["LlamaForCausalLM"]
46336
float16
4.46.0
4,096
11,008
32
32
32
silu
0
False
2,048
10,000
0.00001
0.02
1.0
2
False
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
0
0
5
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:oopsung/llama2-7b-n-ox-test-v1', 'base_model:adapter:oopsung/llama2-7b-n-ox-test-v1', 'region:us']
null
peft
2025-01-14T12:56:06.000Z
['oopsung/llama2-7b-n-ox-test-v1']
[]
[]
['oopsung/llama2-7b-n-ox-test-v1']
[]
null
us
null
221
0x1202/2cca3085-4f82-4284-98ac-0c45c3778369
llama
["LlamaForCausalLM"]
32000
float16
4.46.0
3,200
8,640
26
32
32
silu
0
False
2,048
10,000
0.000001
0.02
1.0
2
False
100
1
false
null
null
true
false
false
false
false
false
false
false
false
true
3.19488
medium
medium
0
0
3
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:openlm-research/open_llama_3b', 'base_model:adapter:openlm-research/open_llama_3b', 'license:apache-2.0', 'region:us']
null
peft
2024-12-28T02:07:02.000Z
['openlm-research/open_llama_3b']
[]
[]
['openlm-research/open_llama_3b']
[]
apache-2.0
us
null
222
0x1202/2cd2a320-5900-4ca8-9d58-a5c4f3b7f9f7
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
False
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
0
0
1
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Qwen2.5-Coder-7B-Instruct', 'base_model:adapter:unsloth/Qwen2.5-Coder-7B-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2025-01-26T22:50:33.000Z
['unsloth/Qwen2.5-Coder-7B-Instruct']
[]
[]
['unsloth/Qwen2.5-Coder-7B-Instruct']
[]
apache-2.0
us
null
223
0x1202/2d005e1b-a7e9-44ee-bde9-b024ad63ba89
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
131,072
1,000,000
0.000001
0.02
null
151643
True
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
0
0
2
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:Qwen/Qwen2.5-1.5B', 'base_model:adapter:Qwen/Qwen2.5-1.5B', 'license:apache-2.0', 'region:us']
null
peft
2024-12-28T17:02:52.000Z
['Qwen/Qwen2.5-1.5B']
[]
[]
['Qwen/Qwen2.5-1.5B']
[]
apache-2.0
us
null
224
0x1202/305c6cee-833e-4327-bd19-f4acad476e5e
mistral
["MistralForCausalLM"]
32002
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1.0
32000
False
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
0
0
2
['peft', 'safetensors', 'mistral', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/OpenHermes-2.5-Mistral-7B', 'base_model:adapter:unsloth/OpenHermes-2.5-Mistral-7B', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T05:15:02.000Z
['unsloth/OpenHermes-2.5-Mistral-7B']
[]
[]
['unsloth/OpenHermes-2.5-Mistral-7B']
[]
apache-2.0
us
null
225
0x1202/3099b251-c90a-4468-a064-9a6ac3ff0208
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
False
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
0
0
2
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Qwen2.5-Coder-7B', 'base_model:adapter:unsloth/Qwen2.5-Coder-7B', 'license:apache-2.0', 'region:us']
null
peft
2025-01-14T08:04:08.000Z
['unsloth/Qwen2.5-Coder-7B']
[]
[]
['unsloth/Qwen2.5-Coder-7B']
[]
apache-2.0
us
null
226
0x1202/3208cef4-bf75-4e1c-9b71-6e75d8f4e5e9
llama
["LlamaForCausalLM"]
32016
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
16,384
1,000,000
0.00001
0.02
1.0
2
False
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
0
0
4
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/codellama-7b', 'base_model:adapter:unsloth/codellama-7b', 'license:apache-2.0', 'region:us']
null
peft
2025-01-19T19:51:28.000Z
['unsloth/codellama-7b']
[]
[]
['unsloth/codellama-7b']
[]
apache-2.0
us
null
227
0x1202/3303c20d-4a21-4f5f-9a72-71a56e076b2f
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.46.0
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2.0
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
0
0
3
['peft', 'safetensors', 'gemma', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/codegemma-7b', 'base_model:adapter:unsloth/codegemma-7b', 'license:apache-2.0', 'region:us']
null
peft
2025-01-29T12:03:09.000Z
['unsloth/codegemma-7b']
[]
[]
['unsloth/codegemma-7b']
[]
apache-2.0
us
null
228
0x1202/331b9d61-8729-4aad-8e47-3a769e081873
starcoder2
["Starcoder2ForCausalLM"]
49152
null
4.46.0
3,072
12,288
30
24
2
gelu_pytorch_tanh
0.1
False
16,384
999,999.442036
null
0.018042
0.0
0
null
128
12
false
null
null
false
false
false
false
false
false
false
false
true
true
3.397386
medium
long
0
0
4
['peft', 'safetensors', 'starcoder2', 'axolotl', 'generated_from_trainer', 'base_model:bigcode/starcoder2-3b', 'base_model:adapter:bigcode/starcoder2-3b', 'license:bigcode-openrail-m', 'region:us']
null
peft
2025-01-29T16:23:05.000Z
['bigcode/starcoder2-3b']
[]
[]
['bigcode/starcoder2-3b']
[]
bigcode-openrail-m
us
null
229
0x1202/33c5edbd-7203-47c4-92c4-92cb4f0a47a5
llama
["LlamaForCausalLM"]
49153
bfloat16
4.46.0
576
1,536
30
9
3
silu
0
False
8,192
100,000
0.00001
0.041667
0.0
0
True
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/SmolLM2-135M', 'base_model:adapter:unsloth/SmolLM2-135M', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T05:48:51.000Z
['unsloth/SmolLM2-135M']
[]
[]
['unsloth/SmolLM2-135M']
[]
apache-2.0
us
null
230
0x1202/34cea818-3127-4dad-a2fa-dd1324994aee
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
False
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
0
0
2
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Qwen2.5-Coder-7B', 'base_model:adapter:unsloth/Qwen2.5-Coder-7B', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T14:06:41.000Z
['unsloth/Qwen2.5-Coder-7B']
[]
[]
['unsloth/Qwen2.5-Coder-7B']
[]
apache-2.0
us
null
231
0x1202/37e94694-3d95-483f-9307-f9c1ff1ad9f9
llama
["LlamaForCausalLM"]
49152
bfloat16
4.46.0
960
2,560
32
15
5
silu
0
False
2,048
10,000
0.00001
0.02
1.0
2
True
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.353894
small
medium
0
0
1
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/SmolLM-360M-Instruct', 'base_model:adapter:unsloth/SmolLM-360M-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2025-01-26T21:10:31.000Z
['unsloth/SmolLM-360M-Instruct']
[]
[]
['unsloth/SmolLM-360M-Instruct']
[]
apache-2.0
us
null
232
0x1202/3c9ead0e-2501-4c0e-989b-77be7f624efd
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
2,048
8,192
16
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000.0
128009
True
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Llama-3.2-1B-Instruct', 'base_model:adapter:unsloth/Llama-3.2-1B-Instruct', 'license:llama3.2', 'region:us']
null
peft
2025-01-19T15:17:21.000Z
['unsloth/Llama-3.2-1B-Instruct']
[]
[]
['unsloth/Llama-3.2-1B-Instruct']
[]
llama3.2
us
null
233
0x1202/40609a9c-1b8f-4601-834c-79ef03228960
llama
["LlamaForCausalLM"]
49152
bfloat16
4.46.0
576
1,536
30
9
3
silu
0
False
2,048
10,000
0.00001
0.02
1.0
2
True
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
medium
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/SmolLM-135M-Instruct', 'base_model:adapter:unsloth/SmolLM-135M-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T08:04:12.000Z
['unsloth/SmolLM-135M-Instruct']
[]
[]
['unsloth/SmolLM-135M-Instruct']
[]
apache-2.0
us
null
234
0x1202/44b41737-2574-4257-b517-c5119d62cbc7
llama
["LlamaForCausalLM"]
49152
bfloat16
4.46.0
2,048
8,192
24
32
32
silu
0
False
2,048
10,000
0.00001
0.02
0.0
0
True
64
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/SmolLM-1.7B', 'base_model:adapter:unsloth/SmolLM-1.7B', 'license:apache-2.0', 'region:us']
null
peft
2025-01-29T11:01:18.000Z
['unsloth/SmolLM-1.7B']
[]
[]
['unsloth/SmolLM-1.7B']
[]
apache-2.0
us
null
235
0x1202/44fb4f3c-8a2c-47ad-a8f7-84a1b51f50d7
llama
["LlamaForCausalLM"]
32000
float16
4.46.0
4,096
11,008
32
32
32
silu
0
False
2,048
10,000
0.000001
0.02
1.0
2
False
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:lmsys/vicuna-7b-v1.3', 'base_model:adapter:lmsys/vicuna-7b-v1.3', 'region:us']
null
peft
2025-01-16T09:31:52.000Z
['lmsys/vicuna-7b-v1.3']
[]
[]
['lmsys/vicuna-7b-v1.3']
[]
null
us
null
236
0x1202/4aae2d86-079c-4e47-ad41-624711748efa
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.46.0
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2.0
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
0
0
3
['peft', 'safetensors', 'gemma2', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/gemma-2-9b-it', 'base_model:adapter:unsloth/gemma-2-9b-it', 'license:gemma', 'region:us']
null
peft
2025-01-29T13:07:32.000Z
['unsloth/gemma-2-9b-it']
[]
[]
['unsloth/gemma-2-9b-it']
[]
gemma
us
null
237
0x1202/4aef5c19-2c47-4652-aaa2-9cd414c34dc3
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
True
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
0
0
2
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Qwen2-1.5B-Instruct', 'base_model:adapter:unsloth/Qwen2-1.5B-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2025-01-10T19:18:14.000Z
['unsloth/Qwen2-1.5B-Instruct']
[]
[]
['unsloth/Qwen2-1.5B-Instruct']
[]
apache-2.0
us
null
238
0x1202/4b6bee8d-2be4-4b63-b973-288ac45e304a
mixtral
["MixtralForCausalLM"]
32000
float32
4.46.0
1,024
3,584
2
32
8
silu
0
False
131,072
1,000,000
0.00001
0.02
1.0
2
False
32
4
false
8
2
false
false
false
false
false
false
false
true
true
true
0.025166
small
very_long
0
0
3
['peft', 'safetensors', 'mixtral', 'axolotl', 'generated_from_trainer', 'base_model:TitanML/tiny-mixtral', 'base_model:adapter:TitanML/tiny-mixtral', 'region:us']
null
peft
2025-01-10T16:13:10.000Z
['TitanML/tiny-mixtral']
[]
[]
['TitanML/tiny-mixtral']
[]
null
us
null
239
0x1202/4cc62169-1cb4-4a5c-ae2d-23830dca6f21
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.46.0
3,072
24,576
28
16
16
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2.0
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
0
0
1
['peft', 'safetensors', 'gemma', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/codegemma-7b', 'base_model:adapter:unsloth/codegemma-7b', 'license:apache-2.0', 'region:us']
null
peft
2025-01-24T19:42:25.000Z
['unsloth/codegemma-7b']
[]
[]
['unsloth/codegemma-7b']
[]
apache-2.0
us
null
240
0x1202/4e5cf054-cae7-4aaa-8228-7ae99474e180
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000.0
128009
False
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Llama-3.1-Storm-8B', 'base_model:adapter:unsloth/Llama-3.1-Storm-8B', 'license:llama3.1', 'region:us']
null
peft
2025-01-14T13:19:14.000Z
['unsloth/Llama-3.1-Storm-8B']
[]
[]
['unsloth/Llama-3.1-Storm-8B']
[]
llama3.1
us
null
241
0x1202/50290948-b63f-483a-b2ec-3637f259c257
mistral
["MistralForCausalLM"]
32768
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1.0
2
False
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
0
0
4
['peft', 'safetensors', 'mistral', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/mistral-7b-v0.3', 'base_model:adapter:unsloth/mistral-7b-v0.3', 'license:apache-2.0', 'region:us']
null
peft
2025-01-22T17:35:49.000Z
['unsloth/mistral-7b-v0.3']
[]
[]
['unsloth/mistral-7b-v0.3']
[]
apache-2.0
us
null
242
0x1202/535a3991-0a78-4f9f-b340-2fbb3cc351b5
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
True
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
0
0
2
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Qwen2-0.5B-Instruct', 'base_model:adapter:unsloth/Qwen2-0.5B-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T05:42:54.000Z
['unsloth/Qwen2-0.5B-Instruct']
[]
[]
['unsloth/Qwen2-0.5B-Instruct']
[]
apache-2.0
us
null
243
0x1202/572f6122-e0bd-4f05-bd76-a64dd181ba40
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.46.0
3,072
8,192
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1.0
32000
False
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
medium
0
0
2
['peft', 'safetensors', 'phi3', 'axolotl', 'generated_from_trainer', 'custom_code', 'base_model:microsoft/Phi-3-mini-4k-instruct', 'base_model:adapter:microsoft/Phi-3-mini-4k-instruct', 'license:mit', 'region:us']
null
peft
2024-12-28T13:15:52.000Z
['microsoft/Phi-3-mini-4k-instruct']
[]
[]
['microsoft/Phi-3-mini-4k-instruct']
[]
mit
us
null
244
0x1202/582f2983-5e52-4b5c-b8ad-f53ca82ebb8b
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
4,096
10,000
0.000001
0.02
null
151645
False
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
medium
0
0
1
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:Qwen/Qwen2.5-Math-7B-Instruct', 'base_model:adapter:Qwen/Qwen2.5-Math-7B-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2025-01-24T18:00:24.000Z
['Qwen/Qwen2.5-Math-7B-Instruct']
[]
[]
['Qwen/Qwen2.5-Math-7B-Instruct']
[]
apache-2.0
us
null
245
0x1202/5ac9fe7f-106b-46e7-b2f1-eb0ec0c8609a
gemma
["GemmaForCausalLM"]
256000
float32
4.46.0
32
2
1
2
1
gelu
0
False
1,024
10,000
0.000001
0.02
2.0
1
null
16
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.000012
small
short
0
0
1
['peft', 'safetensors', 'gemma', 'axolotl', 'generated_from_trainer', 'base_model:fxmarty/tiny-random-GemmaForCausalLM', 'base_model:adapter:fxmarty/tiny-random-GemmaForCausalLM', 'license:mit', 'region:us']
null
peft
2025-01-27T07:58:30.000Z
['fxmarty/tiny-random-GemmaForCausalLM']
[]
[]
['fxmarty/tiny-random-GemmaForCausalLM']
[]
mit
us
null
246
0x1202/5ade6481-31d8-4723-b9cd-2a0bf0232506
phi
["PhiForCausalLM"]
51200
float16
4.46.0
2,560
10,240
32
32
32
gelu_new
0
False
2,048
10,000
null
0.02
50256.0
50256
False
80
1
false
null
null
false
false
false
false
false
false
false
false
false
true
2.516582
medium
medium
0
0
3
['peft', 'safetensors', 'phi', 'axolotl', 'generated_from_trainer', 'base_model:microsoft/phi-2', 'base_model:adapter:microsoft/phi-2', 'license:mit', 'region:us']
null
peft
2024-12-28T01:08:07.000Z
['microsoft/phi-2']
[]
[]
['microsoft/phi-2']
[]
mit
us
null
247
0x1202/5e214cd6-a9b7-4365-b667-88e61e939ba3
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000.0
128009
False
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
0
0
3
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Llama-3.1-Storm-8B', 'base_model:adapter:unsloth/Llama-3.1-Storm-8B', 'license:llama3.1', 'region:us']
null
peft
2025-01-10T22:51:24.000Z
['unsloth/Llama-3.1-Storm-8B']
[]
[]
['unsloth/Llama-3.1-Storm-8B']
[]
llama3.1
us
null
248
0x1202/63cd5c4a-5a68-4c25-bd4d-670517664abf
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
True
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
0
0
4
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:Qwen/Qwen2.5-0.5B-Instruct', 'base_model:adapter:Qwen/Qwen2.5-0.5B-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2024-12-28T16:15:02.000Z
['Qwen/Qwen2.5-0.5B-Instruct']
[]
[]
['Qwen/Qwen2.5-0.5B-Instruct']
[]
apache-2.0
us
null
249
0x1202/65ff08e2-c60e-4c92-999a-e6285a3d2ad1
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
131,072
1,000,000
0.000001
0.02
null
151643
True
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
very_long
0
0
2
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:Qwen/Qwen2.5-1.5B', 'base_model:adapter:Qwen/Qwen2.5-1.5B', 'license:apache-2.0', 'region:us']
null
peft
2025-01-14T02:23:17.000Z
['Qwen/Qwen2.5-1.5B']
[]
[]
['Qwen/Qwen2.5-1.5B']
[]
apache-2.0
us
null
250
0x1202/67914974-084f-4c3b-9ebf-43a8089a5756
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
131,072
1,000,000
0.000001
0.02
null
151643
False
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
very_long
0
0
2
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Qwen2-7B', 'base_model:adapter:unsloth/Qwen2-7B', 'license:apache-2.0', 'region:us']
null
peft
2025-01-27T04:55:13.000Z
['unsloth/Qwen2-7B']
[]
[]
['unsloth/Qwen2-7B']
[]
apache-2.0
us
null
251
0x1202/67a992f4-04ea-4b25-84a9-4fc258302d90
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
True
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
0
0
4
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Qwen2-0.5B-Instruct', 'base_model:adapter:unsloth/Qwen2-0.5B-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T05:46:07.000Z
['unsloth/Qwen2-0.5B-Instruct']
[]
[]
['unsloth/Qwen2-0.5B-Instruct']
[]
apache-2.0
us
null
252
0x1202/6a0417a7-3e63-45bc-adbe-3ee855d97be1
starcoder2
["Starcoder2ForCausalLM"]
49152
null
4.46.0
3,072
12,288
30
24
2
gelu_pytorch_tanh
0.1
False
16,384
999,999.442036
null
0.018042
0.0
0
null
128
12
false
null
null
false
false
false
false
false
false
false
false
true
true
3.397386
medium
long
0
0
2
['peft', 'safetensors', 'starcoder2', 'axolotl', 'generated_from_trainer', 'base_model:bigcode/starcoder2-3b', 'base_model:adapter:bigcode/starcoder2-3b', 'license:bigcode-openrail-m', 'region:us']
null
peft
2025-01-19T19:22:03.000Z
['bigcode/starcoder2-3b']
[]
[]
['bigcode/starcoder2-3b']
[]
bigcode-openrail-m
us
null
253
0x1202/6fe93d12-40f8-418b-9262-220a6ab8c6e2
mistral
["MistralForCausalLM"]
32003
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
10,000
0.00001
0.02
1.0
32000
False
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
long
0
0
3
['peft', 'safetensors', 'mistral', 'axolotl', 'generated_from_trainer', 'base_model:berkeley-nest/Starling-LM-7B-alpha', 'base_model:adapter:berkeley-nest/Starling-LM-7B-alpha', 'license:apache-2.0', 'region:us']
null
peft
2025-01-14T10:46:31.000Z
['berkeley-nest/Starling-LM-7B-alpha']
[]
[]
['berkeley-nest/Starling-LM-7B-alpha']
[]
apache-2.0
us
null
254
0x1202/7aba81ee-6ef4-42c5-a18b-7b582934c391
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
2,048
5,632
22
32
4
silu
0
False
2,048
10,000
0.00001
0.02
1.0
2
False
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
0
0
1
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:TinyLlama/TinyLlama-1.1B-Chat-v0.6', 'base_model:adapter:TinyLlama/TinyLlama-1.1B-Chat-v0.6', 'license:apache-2.0', 'region:us']
null
peft
2025-01-26T19:02:17.000Z
['TinyLlama/TinyLlama-1.1B-Chat-v0.6']
[]
[]
['TinyLlama/TinyLlama-1.1B-Chat-v0.6']
[]
apache-2.0
us
null
255
0x1202/7d1d1187-2373-4514-a88c-f428c2af2fb3
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
True
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
0
0
3
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Qwen2-0.5B-Instruct', 'base_model:adapter:unsloth/Qwen2-0.5B-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2025-01-22T14:53:22.000Z
['unsloth/Qwen2-0.5B-Instruct']
[]
[]
['unsloth/Qwen2-0.5B-Instruct']
[]
apache-2.0
us
null
256
0x1202/7fff76b7-4772-4cc4-be0b-01c4b9823eee
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000.0
128001
False
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Meta-Llama-3.1-8B', 'base_model:adapter:unsloth/Meta-Llama-3.1-8B', 'license:llama3.1', 'region:us']
null
peft
2025-01-10T11:47:50.000Z
['unsloth/Meta-Llama-3.1-8B']
[]
[]
['unsloth/Meta-Llama-3.1-8B']
[]
llama3.1
us
null
257
0x1202/827cb3f7-78d9-407d-a6bb-a7034f5174e0
gpt_neox
["GPTNeoXForCausalLM"]
30080
float16
4.46.0
5,120
20,480
40
40
null
gelu
0
False
2,048
10,000
null
0.02
0.0
0
False
128
null
false
null
null
false
false
true
false
false
false
false
false
false
true
12.582912
large
medium
0
0
2
['peft', 'safetensors', 'gpt_neox', 'axolotl', 'generated_from_trainer', 'base_model:beomi/polyglot-ko-12.8b-safetensors', 'base_model:adapter:beomi/polyglot-ko-12.8b-safetensors', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T08:47:42.000Z
['beomi/polyglot-ko-12.8b-safetensors']
[]
[]
['beomi/polyglot-ko-12.8b-safetensors']
[]
apache-2.0
us
null
258
0x1202/8318414d-3946-4a84-a833-c3431b05498b
gemma
["GemmaForCausalLM"]
256000
float32
4.46.0
32
2
1
2
1
gelu
0
False
1,028
10,000
0.000001
0.02
2.0
1
null
16
2
false
null
null
false
false
false
false
false
false
false
false
true
true
0.000012
small
short
0
0
2
['peft', 'safetensors', 'gemma', 'axolotl', 'generated_from_trainer', 'base_model:fxmarty/tiny-random-GemmaForCausalLM', 'base_model:adapter:fxmarty/tiny-random-GemmaForCausalLM', 'license:mit', 'region:us']
null
peft
2024-12-28T16:19:02.000Z
['fxmarty/tiny-random-GemmaForCausalLM']
[]
[]
['fxmarty/tiny-random-GemmaForCausalLM']
[]
mit
us
null
259
0x1202/83cbf2c0-4034-4ffd-8182-6cb569c8a72e
llama
["LlamaForCausalLM"]
50964
float32
4.46.0
4,096
11,008
32
32
32
silu
0
False
4,096
10,000
0.00001
0.02
1.0
2
False
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
medium
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:Korabbit/llama-2-ko-7b', 'base_model:adapter:Korabbit/llama-2-ko-7b', 'region:us']
null
peft
2025-01-14T12:29:23.000Z
['Korabbit/llama-2-ko-7b']
[]
[]
['Korabbit/llama-2-ko-7b']
[]
null
us
null
260
0x1202/84343eff-437f-41e7-b35c-9bad1780ba57
llama
["LlamaForCausalLM"]
32000
float32
4.46.0
2,048
5,504
24
16
16
silu
0
False
4,096
10,000
0.00001
0.02
1.0
2
False
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:princeton-nlp/Sheared-LLaMA-1.3B', 'base_model:adapter:princeton-nlp/Sheared-LLaMA-1.3B', 'license:apache-2.0', 'region:us']
null
peft
2025-01-29T08:58:20.000Z
['princeton-nlp/Sheared-LLaMA-1.3B']
[]
[]
['princeton-nlp/Sheared-LLaMA-1.3B']
[]
apache-2.0
us
null
261
0x1202/84ea10e8-79fe-416d-9245-2f910f3f2c3d
llama
["LlamaForCausalLM"]
128256
float16
4.46.0
16
64
2
4
4
silu
0
False
4,096
500,000
0.00001
0.02
128000.0
128009
False
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:llamafactory/tiny-random-Llama-3', 'base_model:adapter:llamafactory/tiny-random-Llama-3', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T16:01:45.000Z
['llamafactory/tiny-random-Llama-3']
[]
[]
['llamafactory/tiny-random-Llama-3']
[]
apache-2.0
us
null
262
0x1202/88bf9116-f7df-4b49-886f-d5c730770156
gpt_neox
["GPTNeoXForCausalLM"]
30080
float16
4.46.0
5,120
20,480
40
40
null
gelu
0
False
2,048
10,000
null
0.02
0.0
0
False
128
null
false
null
null
false
false
true
false
false
false
false
false
false
true
12.582912
large
medium
0
0
2
['peft', 'safetensors', 'gpt_neox', 'axolotl', 'generated_from_trainer', 'base_model:beomi/polyglot-ko-12.8b-safetensors', 'base_model:adapter:beomi/polyglot-ko-12.8b-safetensors', 'license:apache-2.0', 'region:us']
null
peft
2025-01-29T05:59:48.000Z
['beomi/polyglot-ko-12.8b-safetensors']
[]
[]
['beomi/polyglot-ko-12.8b-safetensors']
[]
apache-2.0
us
null
263
0x1202/89bdf7de-d8a9-4d4f-b34a-43e7420ae5c9
llama
["LlamaForCausalLM"]
32001
float16
4.46.0
5,120
13,824
40
40
40
silu
0
False
2,048
10,000
0.000001
0.02
1.0
2
False
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
12.582912
large
medium
0
0
3
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:NousResearch/GPT4-x-Vicuna-13b-fp16', 'base_model:adapter:NousResearch/GPT4-x-Vicuna-13b-fp16', 'license:gpl', 'region:us']
null
peft
2025-01-14T11:03:03.000Z
['NousResearch/GPT4-x-Vicuna-13b-fp16']
[]
[]
['NousResearch/GPT4-x-Vicuna-13b-fp16']
[]
gpl
us
null
264
0x1202/8a91174b-abb5-4527-9a90-f8fb780648dc
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
4,096
10,000
0.000001
0.02
null
151645
False
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
medium
0
0
2
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:Qwen/Qwen2.5-Math-7B-Instruct', 'base_model:adapter:Qwen/Qwen2.5-Math-7B-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2024-12-28T16:45:45.000Z
['Qwen/Qwen2.5-Math-7B-Instruct']
[]
[]
['Qwen/Qwen2.5-Math-7B-Instruct']
[]
apache-2.0
us
null
265
0x1202/8b3af8ac-0219-43ad-b438-a2fd1083cbfe
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
896
4,864
24
14
2
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151645
True
64
7
false
null
null
false
false
false
false
false
false
false
false
true
true
0.231211
small
very_long
0
0
3
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Qwen2.5-0.5B-Instruct', 'base_model:adapter:unsloth/Qwen2.5-0.5B-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2025-01-19T16:36:07.000Z
['unsloth/Qwen2.5-0.5B-Instruct']
[]
[]
['unsloth/Qwen2.5-0.5B-Instruct']
[]
apache-2.0
us
null
266
0x1202/8bee46bd-8ed4-4375-b19e-b0fb21d0b02a
mistral
["MistralForCausalLM"]
32000
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1.0
2
False
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
0
0
2
['peft', 'safetensors', 'mistral', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/mistral-7b-v0.2', 'base_model:adapter:unsloth/mistral-7b-v0.2', 'license:apache-2.0', 'region:us']
null
peft
2025-01-14T11:36:47.000Z
['unsloth/mistral-7b-v0.2']
[]
[]
['unsloth/mistral-7b-v0.2']
[]
apache-2.0
us
null
267
0x1202/90b4fa7b-0f07-444d-b21b-7186cc52fb42
mistral
["MistralForCausalLM"]
131072
bfloat16
4.46.0
5,120
14,336
40
32
8
silu
0
False
131,072
1,000,000
0.00001
0.02
1.0
2
False
160
4
false
null
null
false
false
false
false
false
false
true
false
true
true
12.582912
large
very_long
0
0
2
['peft', 'safetensors', 'mistral', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Mistral-Nemo-Instruct-2407', 'base_model:adapter:unsloth/Mistral-Nemo-Instruct-2407', 'license:apache-2.0', 'region:us']
null
peft
2025-01-29T08:07:39.000Z
['unsloth/Mistral-Nemo-Instruct-2407']
[]
[]
['unsloth/Mistral-Nemo-Instruct-2407']
[]
apache-2.0
us
null
268
0x1202/95b57850-7914-4160-bf6c-84c6285d49d2
llama
["LlamaForCausalLM"]
49152
bfloat16
4.46.0
2,048
8,192
24
32
32
silu
0
False
2,048
10,000
0.00001
0.02
0.0
0
True
64
1
false
null
null
true
false
false
false
false
false
false
false
false
true
1.20796
medium
medium
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/SmolLM-1.7B', 'base_model:adapter:unsloth/SmolLM-1.7B', 'license:apache-2.0', 'region:us']
null
peft
2025-01-14T12:12:28.000Z
['unsloth/SmolLM-1.7B']
[]
[]
['unsloth/SmolLM-1.7B']
[]
apache-2.0
us
null
269
0x1202/977ff64e-732b-4dc0-a078-351b493874ab
llama
["LlamaForCausalLM"]
32256
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
16,384
100,000
0.000001
0.02
32013.0
32021
False
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:deepseek-ai/deepseek-coder-6.7b-instruct', 'base_model:adapter:deepseek-ai/deepseek-coder-6.7b-instruct', 'license:other', 'region:us']
null
peft
2025-01-22T18:09:38.000Z
['deepseek-ai/deepseek-coder-6.7b-instruct']
[]
[]
['deepseek-ai/deepseek-coder-6.7b-instruct']
[]
other
us
null
270
0x1202/97eef2dc-5580-4b0a-88af-db2892b871d3
qwen2
["Qwen2ForCausalLM"]
152064
bfloat16
4.46.0
3,584
18,944
28
28
4
silu
0
False
4,096
10,000
0.000001
0.02
null
151645
False
128
7
false
null
null
false
false
false
false
false
false
false
false
true
true
4.315939
medium
medium
0
0
2
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Qwen2.5-Math-7B-Instruct', 'base_model:adapter:unsloth/Qwen2.5-Math-7B-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2025-01-22T20:16:20.000Z
['unsloth/Qwen2.5-Math-7B-Instruct']
[]
[]
['unsloth/Qwen2.5-Math-7B-Instruct']
[]
apache-2.0
us
null
271
0x1202/9a2662ce-8c67-4f1f-b1ae-0e236740b638
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
2,048
8,192
16
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000.0
128009
True
64
4
false
null
null
true
false
false
false
false
false
false
false
true
true
0.805306
small
very_long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Llama-3.2-1B-Instruct', 'base_model:adapter:unsloth/Llama-3.2-1B-Instruct', 'license:llama3.2', 'region:us']
null
peft
2025-01-10T12:09:14.000Z
['unsloth/Llama-3.2-1B-Instruct']
[]
[]
['unsloth/Llama-3.2-1B-Instruct']
[]
llama3.2
us
null
272
0x1202/a0219e2d-3773-4daf-8a2d-cb19d4731630
phi
["PhiForCausalLM"]
1025
float32
4.46.0
32
37
2
4
4
gelu
0
False
1,024
10,000
null
0.02
0.0
0
False
8
1
false
null
null
false
false
false
false
false
false
false
false
false
true
0.000025
small
short
0
0
1
['peft', 'safetensors', 'phi', 'axolotl', 'generated_from_trainer', 'base_model:echarlaix/tiny-random-PhiForCausalLM', 'base_model:adapter:echarlaix/tiny-random-PhiForCausalLM', 'license:apache-2.0', 'region:us']
null
peft
2025-01-26T18:51:51.000Z
['echarlaix/tiny-random-PhiForCausalLM']
[]
[]
['echarlaix/tiny-random-PhiForCausalLM']
[]
apache-2.0
us
null
273
0x1202/a2a4e454-b7f3-4003-b4e6-9b81db732297
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000.0
128009
False
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:aisingapore/Llama-SEA-LION-v2-8B-IT', 'base_model:adapter:aisingapore/Llama-SEA-LION-v2-8B-IT', 'license:llama3', 'region:us']
null
peft
2025-01-29T10:31:45.000Z
['aisingapore/Llama-SEA-LION-v2-8B-IT']
[]
[]
['aisingapore/Llama-SEA-LION-v2-8B-IT']
[]
llama3
us
null
274
0x1202/a65f23a8-695e-419c-9441-05304dd5cb2b
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000.0
128001
True
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Llama-3.2-3B', 'base_model:adapter:unsloth/Llama-3.2-3B', 'license:llama3.2', 'region:us']
null
peft
2025-01-29T16:50:11.000Z
['unsloth/Llama-3.2-3B']
[]
[]
['unsloth/Llama-3.2-3B']
[]
llama3.2
us
null
275
0x1202/a671984b-9f21-4ceb-8f4f-027e57ac25f9
llama
["LlamaForCausalLM"]
49153
bfloat16
4.46.0
576
1,536
30
9
3
silu
0
False
8,192
100,000
0.00001
0.041667
0.0
0
True
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.119439
small
long
0
0
3
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/SmolLM2-135M', 'base_model:adapter:unsloth/SmolLM2-135M', 'license:apache-2.0', 'region:us']
null
peft
2025-01-22T14:11:40.000Z
['unsloth/SmolLM2-135M']
[]
[]
['unsloth/SmolLM2-135M']
[]
apache-2.0
us
null
276
0x1202/a8a15a73-b9d6-46b7-920e-39376c5a33cc
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.46.0
2,048
16,384
18
8
1
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2.0
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
0
0
2
['peft', 'safetensors', 'gemma', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/gemma-1.1-2b-it', 'base_model:adapter:unsloth/gemma-1.1-2b-it', 'license:apache-2.0', 'region:us']
null
peft
2025-01-29T12:28:05.000Z
['unsloth/gemma-1.1-2b-it']
[]
[]
['unsloth/gemma-1.1-2b-it']
[]
apache-2.0
us
null
277
0x1202/a9ccfdfc-f018-4cbc-bdd3-5edccfbfb80c
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.46.0
3,072
24,576
28
16
16
gelu
0
False
8,192
10,000
0.000001
0.02
2.0
1
null
192
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.170894
medium
long
0
0
2
['peft', 'safetensors', 'gemma', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/gemma-7b-it', 'base_model:adapter:unsloth/gemma-7b-it', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T09:11:49.000Z
['unsloth/gemma-7b-it']
[]
[]
['unsloth/gemma-7b-it']
[]
apache-2.0
us
null
278
0x1202/ad04e588-d36c-4fd5-ad55-e0ee695fd4c5
llama
["LlamaForCausalLM"]
128288
float16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000.0
128003
False
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
0
0
4
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:NousResearch/Hermes-2-Pro-Llama-3-8B', 'base_model:adapter:NousResearch/Hermes-2-Pro-Llama-3-8B', 'license:llama3', 'region:us']
null
peft
2024-12-27T23:11:21.000Z
['NousResearch/Hermes-2-Pro-Llama-3-8B']
[]
[]
['NousResearch/Hermes-2-Pro-Llama-3-8B']
[]
llama3
us
null
279
0x1202/ad6c9240-344d-44a1-8033-27496ddbe364
phi3
["Phi3ForCausalLM"]
32064
bfloat16
4.46.0
3,072
8,192
32
32
32
silu
0
False
131,072
10,000
0.00001
0.02
1.0
32000
False
96
1
false
null
null
false
false
false
false
false
false
false
false
false
true
3.623879
medium
very_long
0
0
2
['peft', 'safetensors', 'phi3', 'axolotl', 'generated_from_trainer', 'custom_code', 'base_model:microsoft/Phi-3.5-mini-instruct', 'base_model:adapter:microsoft/Phi-3.5-mini-instruct', 'license:mit', 'region:us']
null
peft
2025-01-22T17:11:12.000Z
['microsoft/Phi-3.5-mini-instruct']
[]
[]
['microsoft/Phi-3.5-mini-instruct']
[]
mit
us
null
280
0x1202/adad566c-f300-407f-a3dc-39438e8408d9
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
4,096
11,008
32
32
32
silu
0
False
32,768
1,000,000
0.000001
0.02
null
151643
False
128
1
false
null
null
false
false
false
false
false
false
false
false
false
true
6.442451
medium
very_long
0
0
1
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:Qwen/Qwen1.5-7B', 'base_model:adapter:Qwen/Qwen1.5-7B', 'license:other', 'region:us']
null
peft
2025-01-27T03:15:35.000Z
['Qwen/Qwen1.5-7B']
[]
[]
['Qwen/Qwen1.5-7B']
[]
other
us
null
281
0x1202/af2f24e3-8173-48c9-98f6-70c8457b3d90
llama
["LlamaForCausalLM"]
32016
float16
4.46.0
4,096
11,008
32
32
32
silu
null
False
16,384
1,000,000
0.00001
0.02
1.0
2
False
128
1
false
null
null
true
false
false
false
false
false
false
false
false
true
6.442451
medium
long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'custom_code', 'base_model:NousResearch/CodeLlama-7b-hf-flash', 'base_model:adapter:NousResearch/CodeLlama-7b-hf-flash', 'region:us']
null
peft
2025-01-10T21:47:24.000Z
['NousResearch/CodeLlama-7b-hf-flash']
[]
[]
['NousResearch/CodeLlama-7b-hf-flash']
[]
null
us
null
282
0x1202/b0098ce2-ce60-4093-a89e-194a6b105b08
mistral
["MistralForCausalLM"]
32768
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
1,000,000
0.00001
0.02
1.0
2
False
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
0
0
5
['peft', 'safetensors', 'mistral', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/mistral-7b-instruct-v0.3', 'base_model:adapter:unsloth/mistral-7b-instruct-v0.3', 'license:apache-2.0', 'region:us']
null
peft
2025-01-29T11:18:25.000Z
['unsloth/mistral-7b-instruct-v0.3']
[]
[]
['unsloth/mistral-7b-instruct-v0.3']
[]
apache-2.0
us
null
283
0x1202/b1a15532-d25e-44df-b90a-8083ef674b2f
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000.0
128009
False
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:Orenguteng/Llama-3-8B-Lexi-Uncensored', 'base_model:adapter:Orenguteng/Llama-3-8B-Lexi-Uncensored', 'license:llama3', 'region:us']
null
peft
2025-01-16T08:27:33.000Z
['Orenguteng/Llama-3-8B-Lexi-Uncensored']
[]
[]
['Orenguteng/Llama-3-8B-Lexi-Uncensored']
[]
llama3
us
null
284
0x1202/b2ddfd5c-19d1-41b0-aa9a-1609f614b06d
llama
["LlamaForCausalLM"]
32000
float16
4.46.0
16
64
2
4
4
silu
0
False
2,048
10,000
0.000001
0.02
0.0
2
False
4
1
false
null
null
true
false
false
false
false
false
false
false
false
true
0.000006
small
medium
0
0
4
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:HuggingFaceH4/tiny-random-LlamaForCausalLM', 'base_model:adapter:HuggingFaceH4/tiny-random-LlamaForCausalLM', 'region:us']
null
peft
2025-01-16T08:02:44.000Z
['HuggingFaceH4/tiny-random-LlamaForCausalLM']
[]
[]
['HuggingFaceH4/tiny-random-LlamaForCausalLM']
[]
null
us
null
285
0x1202/b45c1f6a-1f67-4e94-ae06-621509e932f5
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
2,048
5,632
22
32
4
silu
0
False
2,048
10,000
0.00001
0.02
1.0
2
False
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/tinyllama-chat', 'base_model:adapter:unsloth/tinyllama-chat', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T13:51:48.000Z
['unsloth/tinyllama-chat']
[]
[]
['unsloth/tinyllama-chat']
[]
apache-2.0
us
null
286
0x1202/b4e72f27-951c-4d12-8fc5-5efdf423d3de
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
131,072
500,000
0.00001
0.02
128000.0
128040
False
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
very_long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Hermes-3-Llama-3.1-8B', 'base_model:adapter:unsloth/Hermes-3-Llama-3.1-8B', 'region:us']
null
peft
2025-01-29T15:00:09.000Z
['unsloth/Hermes-3-Llama-3.1-8B']
[]
[]
['unsloth/Hermes-3-Llama-3.1-8B']
[]
null
us
null
287
0x1202/b586aee1-e5d6-4535-a91f-f0a9143e8d7c
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000.0
128001
True
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
0
0
1
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Llama-3.2-3B', 'base_model:adapter:unsloth/Llama-3.2-3B', 'license:llama3.2', 'region:us']
null
peft
2025-01-27T03:54:31.000Z
['unsloth/Llama-3.2-3B']
[]
[]
['unsloth/Llama-3.2-3B']
[]
llama3.2
us
null
288
0x1202/b6a2ad30-543d-468b-8394-d7a3c085b31e
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
3,072
8,192
28
24
8
silu
0
False
131,072
500,000
0.00001
0.02
128000.0
128001
True
128
3
false
null
null
true
false
false
false
false
false
false
false
true
true
3.170894
medium
very_long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Llama-3.2-3B', 'base_model:adapter:unsloth/Llama-3.2-3B', 'license:llama3.2', 'region:us']
null
peft
2025-01-29T07:36:31.000Z
['unsloth/Llama-3.2-3B']
[]
[]
['unsloth/Llama-3.2-3B']
[]
llama3.2
us
null
289
0x1202/b79042e9-c4aa-4947-af63-001439792f03
llama
["LlamaForCausalLM"]
128256
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
8,192
500,000
0.00001
0.02
128000.0
128009
False
128
4
false
null
null
true
false
false
false
false
false
false
false
true
true
6.442451
medium
long
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:Orenguteng/Llama-3-8B-Lexi-Uncensored', 'base_model:adapter:Orenguteng/Llama-3-8B-Lexi-Uncensored', 'license:llama3', 'region:us']
null
peft
2025-01-19T23:54:17.000Z
['Orenguteng/Llama-3-8B-Lexi-Uncensored']
[]
[]
['Orenguteng/Llama-3-8B-Lexi-Uncensored']
[]
llama3
us
null
290
0x1202/bb0a028c-6430-46cb-9cbc-505831810576
mistral
["MistralForCausalLM"]
32000
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
null
False
32,768
10,000
0.00001
0.02
1.0
2
False
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
0
0
5
['peft', 'safetensors', 'mistral', 'axolotl', 'generated_from_trainer', 'custom_code', 'base_model:NousResearch/Yarn-Mistral-7b-64k', 'base_model:adapter:NousResearch/Yarn-Mistral-7b-64k', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T06:24:42.000Z
['NousResearch/Yarn-Mistral-7b-64k']
[]
[]
['NousResearch/Yarn-Mistral-7b-64k']
[]
apache-2.0
us
null
291
0x1202/bd509448-06de-4a7b-880b-117db1374a0b
qwen2
["Qwen2ForCausalLM"]
151936
bfloat16
4.46.0
1,536
8,960
28
12
2
silu
0
False
4,096
10,000
0.000001
0.02
null
151645
True
128
6
false
null
null
false
false
false
false
false
false
false
false
true
true
0.792723
small
medium
0
0
4
['peft', 'safetensors', 'qwen2', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/Qwen2.5-Math-1.5B-Instruct', 'base_model:adapter:unsloth/Qwen2.5-Math-1.5B-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T09:00:31.000Z
['unsloth/Qwen2.5-Math-1.5B-Instruct']
[]
[]
['unsloth/Qwen2.5-Math-1.5B-Instruct']
[]
apache-2.0
us
null
292
0x1202/bd71f493-1b9e-482c-abe5-6d03c3e92109
gpt_neox
["GPTNeoXForCausalLM"]
30080
float16
4.46.0
5,120
20,480
40
40
null
gelu
0
False
2,048
10,000
null
0.02
0.0
2
False
128
null
false
null
null
false
false
true
false
false
false
false
false
false
true
12.582912
large
medium
0
0
2
['peft', 'safetensors', 'gpt_neox', 'axolotl', 'generated_from_trainer', 'base_model:tlphams/gollm-12.8b-instruct-v2.3', 'base_model:adapter:tlphams/gollm-12.8b-instruct-v2.3', 'license:cc-by-nc-4.0', 'region:us']
null
peft
2025-01-16T07:20:16.000Z
['tlphams/gollm-12.8b-instruct-v2.3']
[]
[]
['tlphams/gollm-12.8b-instruct-v2.3']
[]
cc-by-nc-4.0
us
null
293
0x1202/c21bcaac-e1aa-4bf3-ad21-61f86bde94d8
mistral
["MistralForCausalLM"]
47772
float16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1.0
2
False
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
0
0
2
['peft', 'safetensors', 'mistral', 'axolotl', 'generated_from_trainer', 'base_model:lcw99/zephykor-ko-7b-chang', 'base_model:adapter:lcw99/zephykor-ko-7b-chang', 'region:us']
null
peft
2025-01-16T16:11:55.000Z
['lcw99/zephykor-ko-7b-chang']
[]
[]
['lcw99/zephykor-ko-7b-chang']
[]
null
us
null
294
0x1202/c5979e45-c828-4b7f-b00e-8db52fcd76dc
mistral
["MistralForCausalLM"]
32002
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1.0
32000
False
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
0
0
1
['peft', 'safetensors', 'mistral', 'axolotl', 'generated_from_trainer', 'base_model:jhflow/mistral7b-lora-multi-turn-v2', 'base_model:adapter:jhflow/mistral7b-lora-multi-turn-v2', 'region:us']
null
peft
2025-01-27T04:16:19.000Z
['jhflow/mistral7b-lora-multi-turn-v2']
[]
[]
['jhflow/mistral7b-lora-multi-turn-v2']
[]
null
us
null
295
0x1202/c59ed514-7ecd-4b67-8474-30d492333633
mistral
["MistralForCausalLM"]
32000
bfloat16
4.46.0
4,096
14,336
32
32
8
silu
0
False
32,768
10,000
0.00001
0.02
1.0
2
False
128
4
false
null
null
false
false
false
false
false
false
true
false
true
true
6.442451
medium
very_long
0
0
1
['peft', 'safetensors', 'mistral', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/mistral-7b', 'base_model:adapter:unsloth/mistral-7b', 'license:apache-2.0', 'region:us']
null
peft
2025-01-24T19:03:36.000Z
['unsloth/mistral-7b']
[]
[]
['unsloth/mistral-7b']
[]
apache-2.0
us
null
296
0x1202/d1764615-54d3-4ce6-9804-2a5b3b2152b4
llama
["LlamaForCausalLM"]
32000
bfloat16
4.46.0
2,048
5,632
22
32
4
silu
0
False
2,048
10,000
0.00001
0.02
1.0
2
False
64
8
false
null
null
true
false
false
false
false
false
false
false
true
true
1.107296
medium
medium
0
0
2
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/tinyllama-chat', 'base_model:adapter:unsloth/tinyllama-chat', 'license:apache-2.0', 'region:us']
null
peft
2025-01-26T20:42:12.000Z
['unsloth/tinyllama-chat']
[]
[]
['unsloth/tinyllama-chat']
[]
apache-2.0
us
null
297
0x1202/d2ef3f1a-f989-43f3-93db-4cf9535226ef
gemma2
["Gemma2ForCausalLM"]
256000
bfloat16
4.46.0
3,584
14,336
42
16
8
gelu_pytorch_tanh
0
False
8,192
10,000
0.000001
0.02
2.0
1
null
224
2
false
null
null
false
false
false
false
false
false
false
false
true
true
6.473908
medium
long
0
0
3
['peft', 'safetensors', 'gemma2', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/gemma-2-9b-it', 'base_model:adapter:unsloth/gemma-2-9b-it', 'license:gemma', 'region:us']
null
peft
2025-01-16T10:42:13.000Z
['unsloth/gemma-2-9b-it']
[]
[]
['unsloth/gemma-2-9b-it']
[]
gemma
us
null
298
0x1202/d3fe90f7-dc3f-4beb-835d-77f437e96c14
llama
["LlamaForCausalLM"]
49153
bfloat16
4.46.0
960
2,560
32
15
5
silu
0
False
8,192
100,000
0.00001
0.02
1.0
2
True
64
3
false
null
null
true
false
false
false
false
false
false
false
true
true
0.353894
small
long
0
0
3
['peft', 'safetensors', 'llama', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/SmolLM2-360M-Instruct', 'base_model:adapter:unsloth/SmolLM2-360M-Instruct', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T08:06:32.000Z
['unsloth/SmolLM2-360M-Instruct']
[]
[]
['unsloth/SmolLM2-360M-Instruct']
[]
apache-2.0
us
null
299
0x1202/d76f3475-20a1-4bc5-8bb3-b6a89edd0437
gemma
["GemmaForCausalLM"]
256000
bfloat16
4.46.0
2,048
16,384
18
8
1
gelu
0
False
8,192
10,000
0.000001
0.02
2.0
1
null
256
8
false
null
null
false
false
false
false
false
false
false
false
true
true
0.90597
small
long
0
0
2
['peft', 'safetensors', 'gemma', 'axolotl', 'generated_from_trainer', 'base_model:unsloth/gemma-2b-it', 'base_model:adapter:unsloth/gemma-2b-it', 'license:apache-2.0', 'region:us']
null
peft
2025-01-16T06:39:28.000Z
['unsloth/gemma-2b-it']
[]
[]
['unsloth/gemma-2b-it']
[]
apache-2.0
us
null