lukeleeai commited on
Commit
4029b7e
·
verified ·
1 Parent(s): 163036e

End of training

Browse files
README.md CHANGED
@@ -17,7 +17,7 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) on the openwebtext dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 1.8068
21
 
22
  ## Model description
23
 
@@ -47,7 +47,7 @@ The following hyperparameters were used during training:
47
  - total_eval_batch_size: 64
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
- - training_steps: 4
51
 
52
  ### Training results
53
 
 
17
 
18
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) on the openwebtext dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 2.7862
21
 
22
  ## Model description
23
 
 
47
  - total_eval_batch_size: 64
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
+ - training_steps: 2
51
 
52
  ### Training results
53
 
config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "_name_or_path": "mistralai/Mistral-7B-Instruct-v0.1",
3
  "architectures": [
4
- "MistralForCausalLM"
5
  ],
6
  "bos_token_id": 1,
7
  "eos_token_id": 2,
@@ -9,17 +9,21 @@
9
  "hidden_size": 4096,
10
  "initializer_range": 0.02,
11
  "intermediate_size": 14336,
12
- "max_position_embeddings": 32768,
13
  "model_type": "mistral",
14
  "num_attention_heads": 32,
15
  "num_hidden_layers": 32,
16
  "num_key_value_heads": 8,
17
- "rms_norm_eps": 1e-05,
 
 
18
  "rope_theta": 10000.0,
19
  "sliding_window": 4096,
20
  "tie_word_embeddings": false,
21
  "torch_dtype": "bfloat16",
22
  "transformers_version": "4.35.2",
23
  "use_cache": true,
 
 
24
  "vocab_size": 32000
25
  }
 
1
  {
2
  "_name_or_path": "mistralai/Mistral-7B-Instruct-v0.1",
3
  "architectures": [
4
+ "SparseMistral"
5
  ],
6
  "bos_token_id": 1,
7
  "eos_token_id": 2,
 
9
  "hidden_size": 4096,
10
  "initializer_range": 0.02,
11
  "intermediate_size": 14336,
12
+ "max_position_embeddings": 131072,
13
  "model_type": "mistral",
14
  "num_attention_heads": 32,
15
  "num_hidden_layers": 32,
16
  "num_key_value_heads": 8,
17
+ "regularization_coefficient": 10,
18
+ "regularization_type": "L1 Regularization",
19
+ "rms_norm_eps": 1e-06,
20
  "rope_theta": 10000.0,
21
  "sliding_window": 4096,
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "bfloat16",
24
  "transformers_version": "4.35.2",
25
  "use_cache": true,
26
+ "use_sparse_model": false,
27
+ "use_sparse_regularization": false,
28
  "vocab_size": 32000
29
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d09a0990c3abed9a9cddeb6d07bb7387f66f8467a5d59e043a5368fb05d53062
3
  size 4943163992
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce255e1c678bab67d40e0e7470e0bff12a4cabea3cac6dcb8f41c2adf0b57c9a
3
  size 4943163992
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5b5938632792265c598d6a194efb8494a5aca5d723162b318c71e6b2d3b56511
3
  size 4999821144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68e21ccc1b9703ae7a6ec4264f248cf3f8f9227b24723d96438e16d265d5443b
3
  size 4999821144
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b19f9ffda065181ed378b64943ea0cd914f9a62cbfe33e9551e0d2a012bcecbb
3
  size 4540517840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f99e43cac8aa500195dae0c7f6fe4fdc6c1a5353ee99b1a63ab6b5385852461
3
  size 4540517840
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:29bf7b0f483d11f7eebbccc253ff87fcc38a3c6c23e898925cb94341a5dcbc28
3
  size 6008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0edd9a9aea652105a2d8eed2ea59d50affc6d9013c74ab173064adef41bfe0fb
3
  size 6008