lukeleeai commited on
Commit
163036e
·
verified ·
1 Parent(s): a1e92d2

End of training

Browse files
README.md CHANGED
@@ -17,7 +17,7 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) on the openwebtext dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 1.9177
21
 
22
  ## Model description
23
 
@@ -47,7 +47,7 @@ The following hyperparameters were used during training:
47
  - total_eval_batch_size: 64
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
- - training_steps: 2
51
 
52
  ### Training results
53
 
 
17
 
18
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) on the openwebtext dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 1.8068
21
 
22
  ## Model description
23
 
 
47
  - total_eval_batch_size: 64
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
+ - training_steps: 4
51
 
52
  ### Training results
53
 
config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "_name_or_path": "mistralai/Mistral-7B-Instruct-v0.1",
3
  "architectures": [
4
- "SparseMistral"
5
  ],
6
  "bos_token_id": 1,
7
  "eos_token_id": 2,
@@ -9,21 +9,17 @@
9
  "hidden_size": 4096,
10
  "initializer_range": 0.02,
11
  "intermediate_size": 14336,
12
- "max_position_embeddings": 131072,
13
  "model_type": "mistral",
14
  "num_attention_heads": 32,
15
  "num_hidden_layers": 32,
16
  "num_key_value_heads": 8,
17
- "regularization_coefficient": 10,
18
- "regularization_type": "L1 Regularization",
19
- "rms_norm_eps": 1e-06,
20
  "rope_theta": 10000.0,
21
  "sliding_window": 4096,
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "bfloat16",
24
  "transformers_version": "4.35.2",
25
  "use_cache": true,
26
- "use_sparse_model": false,
27
- "use_sparse_regularization": false,
28
  "vocab_size": 32000
29
  }
 
1
  {
2
  "_name_or_path": "mistralai/Mistral-7B-Instruct-v0.1",
3
  "architectures": [
4
+ "MistralForCausalLM"
5
  ],
6
  "bos_token_id": 1,
7
  "eos_token_id": 2,
 
9
  "hidden_size": 4096,
10
  "initializer_range": 0.02,
11
  "intermediate_size": 14336,
12
+ "max_position_embeddings": 32768,
13
  "model_type": "mistral",
14
  "num_attention_heads": 32,
15
  "num_hidden_layers": 32,
16
  "num_key_value_heads": 8,
17
+ "rms_norm_eps": 1e-05,
 
 
18
  "rope_theta": 10000.0,
19
  "sliding_window": 4096,
20
  "tie_word_embeddings": false,
21
  "torch_dtype": "bfloat16",
22
  "transformers_version": "4.35.2",
23
  "use_cache": true,
 
 
24
  "vocab_size": 32000
25
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:36c02e05ed195fcba3ae7f5260df1a5f27a70f69d59fef18c8743da6f2a6873f
3
  size 4943163992
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d09a0990c3abed9a9cddeb6d07bb7387f66f8467a5d59e043a5368fb05d53062
3
  size 4943163992
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63630fde5d276c113df7f4a90006478c824032ef54d46c222e7afaf499183468
3
  size 4999821144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b5938632792265c598d6a194efb8494a5aca5d723162b318c71e6b2d3b56511
3
  size 4999821144
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e40f0e62ec2a0ffbd5979bd013bc81fdaf427c8ba293cab39d4305fc399573e2
3
  size 4540517840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b19f9ffda065181ed378b64943ea0cd914f9a62cbfe33e9551e0d2a012bcecbb
3
  size 4540517840
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0edd9a9aea652105a2d8eed2ea59d50affc6d9013c74ab173064adef41bfe0fb
3
  size 6008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29bf7b0f483d11f7eebbccc253ff87fcc38a3c6c23e898925cb94341a5dcbc28
3
  size 6008