aixk commited on
Commit
03337cd
·
verified ·
1 Parent(s): 5a866d9

Update base model with Aurora metadata and synchronized Safetensors

Browse files
aurora-volcano-f16.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18c077ad4e173218228a499e6ce69ba784f0d024766e0277aeb8dadac630b68f
3
  size 1341874880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:995b954a78590aed4fd674a5c164ee87132d66e1078c45483a1d45e6f4b32112
3
  size 1341874880
aurora-volcano-q4_k_m.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8ffa5dbf32f3c84166b394e3ab9079f404e472e38a03a41fb24bfd1b184e9a26
3
  size 411479744
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dec1e5fd5e243c8ca311f639f15d4282a9f8be7b25923a3f4437b8c41e513a44
3
  size 411479744
config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "_name_or_path": "./Aurora-0.5B-Volcano_final",
3
  "architectures": [
4
- "LlamaForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
@@ -14,7 +14,7 @@
14
  "intermediate_size": 5632,
15
  "max_position_embeddings": 2048,
16
  "mlp_bias": false,
17
- "model_type": "llama",
18
  "num_attention_heads": 16,
19
  "num_hidden_layers": 32,
20
  "num_key_value_heads": 4,
 
1
  {
2
  "_name_or_path": "./Aurora-0.5B-Volcano_final",
3
  "architectures": [
4
+ "AuroraForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
 
14
  "intermediate_size": 5632,
15
  "max_position_embeddings": 2048,
16
  "mlp_bias": false,
17
+ "model_type": "aurora",
18
  "num_attention_heads": 16,
19
  "num_hidden_layers": 32,
20
  "num_key_value_heads": 4,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:09c345b3ae835824cf78e9f79c74839becb5bce452f6d0dce41cf293c0b8169a
3
- size 2682032488
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2002ddc0550b6540de671cfa05dae4b022f5b704bd517ca28aad7d0b1c986d4b
3
+ size 1341032576