HeroMask commited on
Commit
04ddc3d
·
verified ·
1 Parent(s): 16f75a3

Upload 108 files

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
.gitattributes CHANGED
@@ -33,3 +33,88 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ adapter_model/data/0 filter=lfs diff=lfs merge=lfs -text
37
+ adapter_model/data/1 filter=lfs diff=lfs merge=lfs -text
38
+ adapter_model/data/10 filter=lfs diff=lfs merge=lfs -text
39
+ adapter_model/data/100 filter=lfs diff=lfs merge=lfs -text
40
+ adapter_model/data/102 filter=lfs diff=lfs merge=lfs -text
41
+ adapter_model/data/104 filter=lfs diff=lfs merge=lfs -text
42
+ adapter_model/data/105 filter=lfs diff=lfs merge=lfs -text
43
+ adapter_model/data/106 filter=lfs diff=lfs merge=lfs -text
44
+ adapter_model/data/107 filter=lfs diff=lfs merge=lfs -text
45
+ adapter_model/data/108 filter=lfs diff=lfs merge=lfs -text
46
+ adapter_model/data/109 filter=lfs diff=lfs merge=lfs -text
47
+ adapter_model/data/11 filter=lfs diff=lfs merge=lfs -text
48
+ adapter_model/data/110 filter=lfs diff=lfs merge=lfs -text
49
+ adapter_model/data/111 filter=lfs diff=lfs merge=lfs -text
50
+ adapter_model/data/112 filter=lfs diff=lfs merge=lfs -text
51
+ adapter_model/data/113 filter=lfs diff=lfs merge=lfs -text
52
+ adapter_model/data/114 filter=lfs diff=lfs merge=lfs -text
53
+ adapter_model/data/116 filter=lfs diff=lfs merge=lfs -text
54
+ adapter_model/data/118 filter=lfs diff=lfs merge=lfs -text
55
+ adapter_model/data/119 filter=lfs diff=lfs merge=lfs -text
56
+ adapter_model/data/12 filter=lfs diff=lfs merge=lfs -text
57
+ adapter_model/data/120 filter=lfs diff=lfs merge=lfs -text
58
+ adapter_model/data/121 filter=lfs diff=lfs merge=lfs -text
59
+ adapter_model/data/122 filter=lfs diff=lfs merge=lfs -text
60
+ adapter_model/data/123 filter=lfs diff=lfs merge=lfs -text
61
+ adapter_model/data/124 filter=lfs diff=lfs merge=lfs -text
62
+ adapter_model/data/125 filter=lfs diff=lfs merge=lfs -text
63
+ adapter_model/data/126 filter=lfs diff=lfs merge=lfs -text
64
+ adapter_model/data/127 filter=lfs diff=lfs merge=lfs -text
65
+ adapter_model/data/128 filter=lfs diff=lfs merge=lfs -text
66
+ adapter_model/data/13 filter=lfs diff=lfs merge=lfs -text
67
+ adapter_model/data/130 filter=lfs diff=lfs merge=lfs -text
68
+ adapter_model/data/132 filter=lfs diff=lfs merge=lfs -text
69
+ adapter_model/data/133 filter=lfs diff=lfs merge=lfs -text
70
+ adapter_model/data/134 filter=lfs diff=lfs merge=lfs -text
71
+ adapter_model/data/135 filter=lfs diff=lfs merge=lfs -text
72
+ adapter_model/data/136 filter=lfs diff=lfs merge=lfs -text
73
+ adapter_model/data/137 filter=lfs diff=lfs merge=lfs -text
74
+ adapter_model/data/138 filter=lfs diff=lfs merge=lfs -text
75
+ adapter_model/data/139 filter=lfs diff=lfs merge=lfs -text
76
+ adapter_model/data/14 filter=lfs diff=lfs merge=lfs -text
77
+ adapter_model/data/140 filter=lfs diff=lfs merge=lfs -text
78
+ adapter_model/data/141 filter=lfs diff=lfs merge=lfs -text
79
+ adapter_model/data/142 filter=lfs diff=lfs merge=lfs -text
80
+ adapter_model/data/144 filter=lfs diff=lfs merge=lfs -text
81
+ adapter_model/data/146 filter=lfs diff=lfs merge=lfs -text
82
+ adapter_model/data/147 filter=lfs diff=lfs merge=lfs -text
83
+ adapter_model/data/148 filter=lfs diff=lfs merge=lfs -text
84
+ adapter_model/data/149 filter=lfs diff=lfs merge=lfs -text
85
+ adapter_model/data/15 filter=lfs diff=lfs merge=lfs -text
86
+ adapter_model/data/150 filter=lfs diff=lfs merge=lfs -text
87
+ adapter_model/data/151 filter=lfs diff=lfs merge=lfs -text
88
+ adapter_model/data/152 filter=lfs diff=lfs merge=lfs -text
89
+ adapter_model/data/153 filter=lfs diff=lfs merge=lfs -text
90
+ adapter_model/data/154 filter=lfs diff=lfs merge=lfs -text
91
+ adapter_model/data/155 filter=lfs diff=lfs merge=lfs -text
92
+ adapter_model/data/156 filter=lfs diff=lfs merge=lfs -text
93
+ adapter_model/data/158 filter=lfs diff=lfs merge=lfs -text
94
+ adapter_model/data/16 filter=lfs diff=lfs merge=lfs -text
95
+ adapter_model/data/160 filter=lfs diff=lfs merge=lfs -text
96
+ adapter_model/data/161 filter=lfs diff=lfs merge=lfs -text
97
+ adapter_model/data/162 filter=lfs diff=lfs merge=lfs -text
98
+ adapter_model/data/163 filter=lfs diff=lfs merge=lfs -text
99
+ adapter_model/data/164 filter=lfs diff=lfs merge=lfs -text
100
+ adapter_model/data/165 filter=lfs diff=lfs merge=lfs -text
101
+ adapter_model/data/166 filter=lfs diff=lfs merge=lfs -text
102
+ adapter_model/data/167 filter=lfs diff=lfs merge=lfs -text
103
+ adapter_model/data/168 filter=lfs diff=lfs merge=lfs -text
104
+ adapter_model/data/169 filter=lfs diff=lfs merge=lfs -text
105
+ adapter_model/data/170 filter=lfs diff=lfs merge=lfs -text
106
+ adapter_model/data/172 filter=lfs diff=lfs merge=lfs -text
107
+ adapter_model/data/174 filter=lfs diff=lfs merge=lfs -text
108
+ adapter_model/data/175 filter=lfs diff=lfs merge=lfs -text
109
+ adapter_model/data/176 filter=lfs diff=lfs merge=lfs -text
110
+ adapter_model/data/177 filter=lfs diff=lfs merge=lfs -text
111
+ adapter_model/data/178 filter=lfs diff=lfs merge=lfs -text
112
+ adapter_model/data/179 filter=lfs diff=lfs merge=lfs -text
113
+ adapter_model/data/18 filter=lfs diff=lfs merge=lfs -text
114
+ adapter_model/data/180 filter=lfs diff=lfs merge=lfs -text
115
+ adapter_model/data/181 filter=lfs diff=lfs merge=lfs -text
116
+ adapter_model/data/182 filter=lfs diff=lfs merge=lfs -text
117
+ adapter_model/data/183 filter=lfs diff=lfs merge=lfs -text
118
+ adapter_model/data/184 filter=lfs diff=lfs merge=lfs -text
119
+ adapter_model/data/186 filter=lfs diff=lfs merge=lfs -text
120
+ adapter_model/data/188 filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,88 +1,73 @@
1
  ---
2
  license: apache-2.0
3
- pipeline_tag: text-generation
4
  tags:
5
- - finetuned
6
- inference: true
7
- widget:
8
- - messages:
9
- - role: user
10
- content: What is your favorite condiment?
11
  ---
12
 
13
- # Model Card for Mistral-7B-Instruct-v0.2
 
14
 
15
- The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2.
 
16
 
17
- Mistral-7B-v0.2 has the following changes compared to Mistral-7B-v0.1
18
- - 32k context window (vs 8k context in v0.1)
19
- - Rope-theta = 1e6
20
- - No Sliding-Window Attention
21
 
22
- For full details of this model please read our [paper](https://arxiv.org/abs/2310.06825) and [release blog post](https://mistral.ai/news/la-plateforme/).
23
 
24
- ## Instruction format
25
 
26
- In order to leverage instruction fine-tuning, your prompt should be surrounded by `[INST]` and `[/INST]` tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.
27
 
28
- E.g.
29
- ```
30
- text = "<s>[INST] What is your favourite condiment? [/INST]"
31
- "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!</s> "
32
- "[INST] Do you have mayonnaise recipes? [/INST]"
33
- ```
34
 
35
- This format is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating) via the `apply_chat_template()` method:
36
 
37
- ```python
38
- from transformers import AutoModelForCausalLM, AutoTokenizer
39
 
40
- device = "cuda" # the device to load the model onto
41
 
42
- model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2")
43
- tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2")
44
 
45
- messages = [
46
- {"role": "user", "content": "What is your favourite condiment?"},
47
- {"role": "assistant", "content": "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!"},
48
- {"role": "user", "content": "Do you have mayonnaise recipes?"}
49
- ]
 
 
 
50
 
51
- encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")
52
 
53
- model_inputs = encodeds.to(device)
54
- model.to(device)
55
 
56
- generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
57
- decoded = tokenizer.batch_decode(generated_ids)
58
- print(decoded[0])
59
- ```
60
 
61
- ## Troubleshooting
62
- - If you see the following error:
63
- ```
64
- Traceback (most recent call last):
65
- File "", line 1, in
66
- File "/transformers/models/auto/auto_factory.py", line 482, in from_pretrained
67
- config, kwargs = AutoConfig.from_pretrained(
68
- File "/transformers/models/auto/configuration_auto.py", line 1022, in from_pretrained
69
- config_class = CONFIG_MAPPING[config_dict["model_type"]]
70
- File "/transformers/models/auto/configuration_auto.py", line 723, in getitem
71
- raise KeyError(key)
72
- KeyError: 'mistral'
73
- ```
74
 
75
- Installing transformers from source should solve the issue
76
- pip install git+https://github.com/huggingface/transformers
 
 
 
77
 
78
- This should not be required after transformers-v4.33.4.
79
 
80
- ## Limitations
 
 
 
 
 
 
 
 
 
 
81
 
82
- The Mistral 7B Instruct model is a quick demonstration that the base model can be easily fine-tuned to achieve compelling performance.
83
- It does not have any moderation mechanisms. We're looking forward to engaging with the community on ways to
84
- make the model finely respect guardrails, allowing for deployment in environments requiring moderated outputs.
85
 
86
- ## The Mistral AI Team
87
 
88
- Albert Jiang, Alexandre Sablayrolles, Arthur Mensch, Blanche Savary, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Emma Bou Hanna, Florian Bressand, Gianna Lengyel, Guillaume Bour, Guillaume Lample, Lélio Renard Lavaud, Louis Ternon, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Théophile Gervet, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed.
 
1
  ---
2
  license: apache-2.0
3
+ library_name: peft
4
  tags:
5
+ - generated_from_trainer
6
+ base_model: mistralai/Mistral-7B-Instruct-v0.1
7
+ model-index:
8
+ - name: tmp/helix/results/e9624262-34ea-4818-a31f-84692d26fc66
9
+ results: []
 
10
  ---
11
 
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
 
15
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
16
+ # tmp/helix/results/e9624262-34ea-4818-a31f-84692d26fc66
17
 
18
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) on the None dataset.
 
 
 
19
 
20
+ ## Model description
21
 
22
+ More information needed
23
 
24
+ ## Intended uses & limitations
25
 
26
+ More information needed
 
 
 
 
 
27
 
28
+ ## Training and evaluation data
29
 
30
+ More information needed
 
31
 
32
+ ## Training procedure
33
 
34
+ ### Training hyperparameters
 
35
 
36
+ The following hyperparameters were used during training:
37
+ - learning_rate: 0.0002
38
+ - train_batch_size: 6
39
+ - eval_batch_size: 1
40
+ - seed: 42
41
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
+ - lr_scheduler_type: cosine
43
+ - num_epochs: 20
44
 
45
+ ### Training results
46
 
 
 
47
 
 
 
 
 
48
 
49
+ ### Framework versions
 
 
 
 
 
 
 
 
 
 
 
 
50
 
51
+ - Transformers 4.36.0.dev0
52
+ - Pytorch 2.0.1+cu118
53
+ - Datasets 2.15.0
54
+ - Tokenizers 0.15.0
55
+ ## Training procedure
56
 
 
57
 
58
+ The following `bitsandbytes` quantization config was used during training:
59
+ - quant_method: bitsandbytes
60
+ - load_in_8bit: False
61
+ - load_in_4bit: True
62
+ - llm_int8_threshold: 6.0
63
+ - llm_int8_skip_modules: None
64
+ - llm_int8_enable_fp32_cpu_offload: False
65
+ - llm_int8_has_fp16_weight: False
66
+ - bnb_4bit_quant_type: nf4
67
+ - bnb_4bit_use_double_quant: True
68
+ - bnb_4bit_compute_dtype: bfloat16
69
 
70
+ ### Framework versions
 
 
71
 
 
72
 
73
+ - PEFT 0.6.0
adapter_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.1",
5
+ "bias": "none",
6
+ "fan_in_fan_out": null,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layers_pattern": null,
10
+ "layers_to_transform": null,
11
+ "lora_alpha": 16,
12
+ "lora_dropout": 0.1,
13
+ "modules_to_save": null,
14
+ "peft_type": "LORA",
15
+ "r": 64,
16
+ "rank_pattern": {},
17
+ "revision": null,
18
+ "target_modules": [
19
+ "v_proj",
20
+ "k_proj",
21
+ "up_proj",
22
+ "gate_proj",
23
+ "o_proj",
24
+ "q_proj",
25
+ "down_proj"
26
+ ],
27
+ "task_type": "CAUSAL_LM"
28
+ }
adapter_model/data.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:caf47cb5e4b7325d67046dac755b69313f950d1edbaea9e6db1354488b908ee5
3
+ size 73431
adapter_model/data/0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a7df746ef25c0c9c759f3a49943a66497d2a0bba1ad8eb3e297e7d3306a5f5f
3
+ size 1048576
adapter_model/data/1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c470ee5883cb1ff46d5243018a03298a274705fe3c4a1e1905de92f77707e37
3
+ size 1048576
adapter_model/data/10 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b074949f01ee2cbdf75b672a64a4d3deb719dd0a13e8df9903f71bbe448e90f
3
+ size 1048576
adapter_model/data/100 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aef739d9dfab61b8ef773921023f588570acc64b1899647cbc2b23d80bf7047c
3
+ size 1048576
adapter_model/data/101 ADDED
Binary file (262 kB). View file
 
adapter_model/data/102 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1015ce84885054db91ffe1ee8765d87b014cf238ae6b41e2e20702e0de2c93d1
3
+ size 1048576
adapter_model/data/103 ADDED
Binary file (262 kB). View file
 
adapter_model/data/104 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3cf879238c8be93f4179893e22a9f0585909885b4240fa356a728de1d4e8611
3
+ size 1048576
adapter_model/data/105 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e39233da79fc1ea4fd6ccdb5c2567b9773350ea20657a1e9f7b48c2a56f3a58
3
+ size 1048576
adapter_model/data/106 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d60d0fd8a2be97c22285b0868bcf418459432a07205697ac3237c0b1c11cacd9
3
+ size 1048576
adapter_model/data/107 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe9b629c1a36897b927c55ba61cb459988e70b8137277817ef61606d4ad8f1e2
3
+ size 3670016
adapter_model/data/108 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fde3c484b8a5f028fd0617a7cc0d5928b30a89ca27ea4159b937a1e8369e385f
3
+ size 1048576
adapter_model/data/109 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb0fc07ce979f2961318f06036b9eca74e233c3627a22ee665392a20ae92413b
3
+ size 3670016
adapter_model/data/11 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a99f185320ae78fd106c94bd66d7987b3c3e07a39cadd27e46cce65fc03177e8
3
+ size 3670016
adapter_model/data/110 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f2f097c15531a592295367efdc731dba34cdde45a6c95fef249e7af3e735474
3
+ size 3670016
adapter_model/data/111 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe88a01025078d699e34a15d2d8e4b52e99e450302d5d21267964ea4389db110
3
+ size 1048576
adapter_model/data/112 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd0687c1cf9d1d74f1bdf3218589f2b495c18d310b5ac90750ed3a28d0cd8fe0
3
+ size 1048576
adapter_model/data/113 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1530a8091101db44428f5f299ac1356b9cd58f90335f5c82b652eb9b91ba24cb
3
+ size 1048576
adapter_model/data/114 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c98d58df46a64b44c0e3de0ba2968063aa96f4986a0c4bc80ac7a611b801325c
3
+ size 1048576
adapter_model/data/115 ADDED
Binary file (262 kB). View file
 
adapter_model/data/116 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0232325393bcba6091866d054ebd207c4de3e49d4b789d909506e4d40390c85
3
+ size 1048576
adapter_model/data/117 ADDED
Binary file (262 kB). View file
 
adapter_model/data/118 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b26babfa3397adb5f6b5e9904eac8ff4fe75794bd3b5a156699885f877360b2b
3
+ size 1048576
adapter_model/data/119 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4ccd3d06fc9d2b302fea1cb66bc008445c73f292bff0666a8fc8bbae9c37aa6
3
+ size 1048576
adapter_model/data/12 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fcbe89a80a067d5d74f76528bf4facad715feee84921fb7162773081a6e7f4c4
3
+ size 3670016
adapter_model/data/120 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02167c662eee07b96ad0065e78a78dc5218a2cbcbfd1f5a3b9af4efcf3bf4b0c
3
+ size 1048576
adapter_model/data/121 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:639a96986b34544cf99ceb5f321c496e2b60993e3af8cceac53d9efe0fd84790
3
+ size 3670016
adapter_model/data/122 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa5691485b62978b81fca9c61bf396b3690d25351d6db8e5a2ba12fa3c044041
3
+ size 1048576
adapter_model/data/123 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:750221a6dcac81ba35d82f8a9b47be109ea8bd9d4cd180bca71c66500846bff4
3
+ size 3670016
adapter_model/data/124 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bd5b0d2e4b1f646a4f30824e2f261fe15fb71c3e9df75161f0c06ae1ca63938
3
+ size 3670016
adapter_model/data/125 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6174f43750eee873ee11f8fc104604aff0495de513cb048b282f0f8843abd8c
3
+ size 1048576
adapter_model/data/126 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8a1da69f2dc13802c53a834ba286658342615172e76cee2ba63ae2a764c70ba
3
+ size 1048576
adapter_model/data/127 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc950cc995fa9c4cc9990be58b050859d1f44894c0b54a03d6b2cafa9af7fb16
3
+ size 1048576
adapter_model/data/128 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f9b701881b87ef64059387805833e73b88e7c0daf75247d80de1b0bc0f6b1ac
3
+ size 1048576
adapter_model/data/129 ADDED
Binary file (262 kB). View file
 
adapter_model/data/13 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b95e69ab7613d350207320969fcb45724ac752d440923d9ac0e2d579ca8635d8
3
+ size 1048576
adapter_model/data/130 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0732ae6e14a1243decd1c1c9ef24e602620aac42a0febe9fc8729b3513cf666e
3
+ size 1048576
adapter_model/data/131 ADDED
Binary file (262 kB). View file
 
adapter_model/data/132 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cce654fbb2a5be5dae643a4545c346a5350c1ccc79ce0ea14198a1920823dfd4
3
+ size 1048576
adapter_model/data/133 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62fc616dd6509b602c58115ebf6f337f63f7c1d4ccf0984b11e202b2101af038
3
+ size 1048576
adapter_model/data/134 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ebba6a220990a02e9369ded0ed4e618f67ca5e0827a3fde37cb283ed7fe827a
3
+ size 1048576
adapter_model/data/135 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbcd0da8df1fdd77ab2296c0488f14be26693054a01b0cdc564845f1d7ed37fc
3
+ size 3670016
adapter_model/data/136 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cb4cee44aff953e18bffd1423ab64eb71abf9fc2607874288a901a55f644f07
3
+ size 1048576
adapter_model/data/137 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8bf60d7e9527c48ec14b842ae6a81e762de33059ed50d4b7d09205e530bf3f1b
3
+ size 3670016
adapter_model/data/138 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c756fe4428e006e9e2e5a9c8fc41507386e40c097ac9d25a4382947807f7636
3
+ size 3670016
adapter_model/data/139 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0c934a665ca4109351af0b3bfeacd4918b7ea87385ce84be972cc36e50f53ea
3
+ size 1048576