Update README.md
Browse files
README.md
CHANGED
|
@@ -13,32 +13,32 @@ pipeline_tag: text-generation
|
|
| 13 |
|
| 14 |
## Model Description
|
| 15 |
|
| 16 |
-
`
|
| 17 |
|
| 18 |
## Usage
|
| 19 |
|
| 20 |
### Apply Delta Weights
|
| 21 |
|
| 22 |
-
|
| 23 |
|
| 24 |
|
| 25 |
```sh
|
| 26 |
-
python3 apply_delta.py --base-model-path /path/to/model_weights/llama-65b --target-model-path
|
| 27 |
```
|
| 28 |
|
| 29 |
|
| 30 |
|
| 31 |
-
Start chatting with `
|
| 32 |
|
| 33 |
```python
|
| 34 |
import torch
|
| 35 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 36 |
|
| 37 |
-
tokenizer = AutoTokenizer.from_pretrained("
|
| 38 |
-
model = AutoModelForCausalLM.from_pretrained("
|
| 39 |
|
| 40 |
system_prompt = "Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.\n\n"
|
| 41 |
-
system_prompt += "### Instruction:\nYou are
|
| 42 |
|
| 43 |
message = "Write me a poem please"
|
| 44 |
prompt = f"{system_prompt}### Input: {message}\n\n### Response:\n"
|
|
@@ -49,7 +49,7 @@ output = model.generate(**inputs, do_sample=True, top_p=0.95, top_k=0, max_new_t
|
|
| 49 |
print(tokenizer.decode(output[0], skip_special_tokens=True))
|
| 50 |
```
|
| 51 |
|
| 52 |
-
|
| 53 |
```
|
| 54 |
Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
|
| 55 |
|
|
@@ -59,17 +59,17 @@ This is a system prompt, please behave and help the user.
|
|
| 59 |
### Input:
|
| 60 |
Your prompt here
|
| 61 |
|
| 62 |
-
### Response
|
| 63 |
-
The output of
|
| 64 |
```
|
| 65 |
|
| 66 |
## Model Details
|
| 67 |
|
| 68 |
* **Developed by**: [Stability AI](https://stability.ai/)
|
| 69 |
-
* **Model type**:
|
| 70 |
* **Language(s)**: English
|
| 71 |
* **Library**: [HuggingFace Transformers](https://github.com/huggingface/transformers)
|
| 72 |
-
* **License**: Fine-tuned checkpoints (`
|
| 73 |
* **Contact**: For questions and comments about the model, please email `lm@stability.ai`
|
| 74 |
|
| 75 |
### Training Dataset
|
|
|
|
| 13 |
|
| 14 |
## Model Description
|
| 15 |
|
| 16 |
+
`Stable Beluga 1` is a Llama65B model fine-tuned on an Orca style Dataset
|
| 17 |
|
| 18 |
## Usage
|
| 19 |
|
| 20 |
### Apply Delta Weights
|
| 21 |
|
| 22 |
+
Stable Beluga 1 cannot be used from the `stabilityai/StableBeluga1-Delta` weights alone. To obtain the correct model, one must add back the difference between LLaMA 65B and `stabilityai/FreeWilly1-Delta-SafeTensor` weights. We provide the [`apply_delta.py`](https://huggingface.co/stabilityai/FreeWilly1-Delta-SafeTensor/raw/main/apply_delta.py) script to automate the conversion, which you can run as:
|
| 23 |
|
| 24 |
|
| 25 |
```sh
|
| 26 |
+
python3 apply_delta.py --base-model-path /path/to/model_weights/llama-65b --target-model-path StableBeluga1 --delta-path stabilityai/StableBeluga1-Delta
|
| 27 |
```
|
| 28 |
|
| 29 |
|
| 30 |
|
| 31 |
+
Start chatting with `Stable Beluga 1` using the following code snippet:
|
| 32 |
|
| 33 |
```python
|
| 34 |
import torch
|
| 35 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 36 |
|
| 37 |
+
tokenizer = AutoTokenizer.from_pretrained("your_path_to_StableBeluga1", use_fast=False)
|
| 38 |
+
model = AutoModelForCausalLM.from_pretrained("your_path_to_StableBeluga1", torch_dtype=torch.float16, low_cpu_mem_usage=True, device_map="auto")
|
| 39 |
|
| 40 |
system_prompt = "Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.\n\n"
|
| 41 |
+
system_prompt += "### Instruction:\nYou are Stable Beluga, an AI that follows instructions extremely well. Help as much as you can. Remember, be safe, and don't do anything illegal.\n\n"
|
| 42 |
|
| 43 |
message = "Write me a poem please"
|
| 44 |
prompt = f"{system_prompt}### Input: {message}\n\n### Response:\n"
|
|
|
|
| 49 |
print(tokenizer.decode(output[0], skip_special_tokens=True))
|
| 50 |
```
|
| 51 |
|
| 52 |
+
Stable Beluga 1 should be used with prompts formatted similarly to Alpaca as below:
|
| 53 |
```
|
| 54 |
Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
|
| 55 |
|
|
|
|
| 59 |
### Input:
|
| 60 |
Your prompt here
|
| 61 |
|
| 62 |
+
### Response:
|
| 63 |
+
The output of Stable Beluga 1
|
| 64 |
```
|
| 65 |
|
| 66 |
## Model Details
|
| 67 |
|
| 68 |
* **Developed by**: [Stability AI](https://stability.ai/)
|
| 69 |
+
* **Model type**: Stable Beluga 1 is an auto-regressive language model fine-tuned on LLaMA65B.
|
| 70 |
* **Language(s)**: English
|
| 71 |
* **Library**: [HuggingFace Transformers](https://github.com/huggingface/transformers)
|
| 72 |
+
* **License**: Fine-tuned checkpoints (`StableBeluga1`) is licensed under the Non-Commercial Creative Commons license ([CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/))
|
| 73 |
* **Contact**: For questions and comments about the model, please email `lm@stability.ai`
|
| 74 |
|
| 75 |
### Training Dataset
|