Megnis commited on
Commit
98ed6ba
·
verified ·
1 Parent(s): 0322527

Megnis/saiga2-7b-vkr-analis

Browse files
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: NousResearch/Llama-2-7b-chat-hf
3
  tags:
4
  - generated_from_trainer
5
  model-index:
@@ -12,7 +12,7 @@ should probably proofread and complete it, then remove this comment. -->
12
 
13
  # results_modified
14
 
15
- This model is a fine-tuned version of [NousResearch/Llama-2-7b-chat-hf](https://huggingface.co/NousResearch/Llama-2-7b-chat-hf) on an unknown dataset.
16
 
17
  ## Model description
18
 
 
1
  ---
2
+ base_model: pieken/saiga2_7b_lora_merged
3
  tags:
4
  - generated_from_trainer
5
  model-index:
 
12
 
13
  # results_modified
14
 
15
+ This model is a fine-tuned version of [pieken/saiga2_7b_lora_merged](https://huggingface.co/pieken/saiga2_7b_lora_merged) on the None dataset.
16
 
17
  ## Model description
18
 
adapter_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "auto_mapping": null,
3
- "base_model_name_or_path": "NousResearch/Llama-2-7b-chat-hf",
4
  "bias": "none",
5
  "fan_in_fan_out": false,
6
  "inference_mode": true,
@@ -11,7 +11,7 @@
11
  "lora_dropout": 0.1,
12
  "modules_to_save": null,
13
  "peft_type": "LORA",
14
- "r": 32,
15
  "revision": null,
16
  "target_modules": [
17
  "q_proj",
 
1
  {
2
  "auto_mapping": null,
3
+ "base_model_name_or_path": "pieken/saiga2_7b_lora_merged",
4
  "bias": "none",
5
  "fan_in_fan_out": false,
6
  "inference_mode": true,
 
11
  "lora_dropout": 0.1,
12
  "modules_to_save": null,
13
  "peft_type": "LORA",
14
+ "r": 8,
15
  "revision": null,
16
  "target_modules": [
17
  "q_proj",
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:084bd736db7927b72ba03a1e8e931f4b551b577443db1df8210d6b1c4228998d
3
- size 67155338
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45c0aeb075099d22f5aa85dc8e9d90a2be24d0bdaf834217e0a7a33a702a45e8
3
+ size 16823434
special_tokens_map.json CHANGED
@@ -1,24 +1,7 @@
1
  {
2
- "bos_token": {
3
- "content": "<s>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "</s>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
  "pad_token": "</s>",
17
- "unk_token": {
18
- "content": "<unk>",
19
- "lstrip": false,
20
- "normalized": true,
21
- "rstrip": false,
22
- "single_word": false
23
- }
24
  }
 
1
  {
2
+ "bos_token": "<s>",
3
+ "eos_token": "</s>",
 
 
 
 
 
 
 
 
 
 
 
 
4
  "pad_token": "</s>",
5
+ "sep_token": "<s>",
6
+ "unk_token": "<unk>"
 
 
 
 
 
7
  }
tokenizer.json CHANGED
@@ -23,7 +23,7 @@
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
26
- "normalized": true,
27
  "special": true
28
  },
29
  {
@@ -32,7 +32,7 @@
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
35
- "normalized": true,
36
  "special": true
37
  },
38
  {
 
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
26
+ "normalized": false,
27
  "special": true
28
  },
29
  {
 
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
35
+ "normalized": false,
36
  "special": true
37
  },
38
  {
tokenizer_config.json CHANGED
@@ -1,4 +1,6 @@
1
  {
 
 
2
  "bos_token": {
3
  "__type": "AddedToken",
4
  "content": "<s>",
@@ -16,10 +18,12 @@
16
  "rstrip": false,
17
  "single_word": false
18
  },
19
- "legacy": false,
20
- "model_max_length": 1000000000000000019884624838656,
21
  "pad_token": null,
 
22
  "sp_model_kwargs": {},
 
23
  "tokenizer_class": "LlamaTokenizer",
24
  "unk_token": {
25
  "__type": "AddedToken",
 
1
  {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
  "bos_token": {
5
  "__type": "AddedToken",
6
  "content": "<s>",
 
18
  "rstrip": false,
19
  "single_word": false
20
  },
21
+ "legacy": true,
22
+ "model_max_length": 2048,
23
  "pad_token": null,
24
+ "padding_side": "left",
25
  "sp_model_kwargs": {},
26
+ "spaces_between_special_tokens": false,
27
  "tokenizer_class": "LlamaTokenizer",
28
  "unk_token": {
29
  "__type": "AddedToken",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e3a4ad9235e079a8f93651fb9bb0fe67d8761fb75dff3ce16db2fcec3fe4ca6c
3
  size 4472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6434e43a32d7b68b8e6b008e56d32198e9adabb62725d5b2c8ebcedf550494c5
3
  size 4472