Merge branch 'main' of https://huggingface.co/togethercomputer/RedPajama-INCITE-7B-Instruct into main
Browse files
README.md
CHANGED
|
@@ -29,15 +29,15 @@ inference:
|
|
| 29 |
max_new_tokens: 128
|
| 30 |
---
|
| 31 |
|
| 32 |
-
# RedPajama-INCITE-
|
| 33 |
|
| 34 |
-
RedPajama-INCITE-
|
| 35 |
|
| 36 |
The model was fine-tuned for few-shot applications on the data of [GPT-JT](https://huggingface.co/togethercomputer/GPT-JT-6B-v1), with exclusion of tasks that overlap with the HELM core scenarios.
|
| 37 |
|
| 38 |
-
- Base Model: [RedPajama-INCITE-
|
| 39 |
-
- Instruction-tuned Version: [RedPajama-INCITE-
|
| 40 |
-
- Chat Version: [RedPajama-INCITE-
|
| 41 |
|
| 42 |
|
| 43 |
## Model Details
|
|
@@ -66,8 +66,8 @@ MIN_TRANSFORMERS_VERSION = '4.25.1'
|
|
| 66 |
assert transformers.__version__ >= MIN_TRANSFORMERS_VERSION, f'Please upgrade transformers to version {MIN_TRANSFORMERS_VERSION} or higher.'
|
| 67 |
|
| 68 |
# init
|
| 69 |
-
tokenizer = AutoTokenizer.from_pretrained("togethercomputer/RedPajama-INCITE-
|
| 70 |
-
model = AutoModelForCausalLM.from_pretrained("togethercomputer/RedPajama-INCITE-
|
| 71 |
model = model.to('cuda:0')
|
| 72 |
# infer
|
| 73 |
prompt = "Q: The capital of France is?\nA:"
|
|
@@ -108,8 +108,8 @@ MIN_TRANSFORMERS_VERSION = '4.25.1'
|
|
| 108 |
assert transformers.__version__ >= MIN_TRANSFORMERS_VERSION, f'Please upgrade transformers to version {MIN_TRANSFORMERS_VERSION} or higher.'
|
| 109 |
|
| 110 |
# init
|
| 111 |
-
tokenizer = AutoTokenizer.from_pretrained("togethercomputer/RedPajama-INCITE-
|
| 112 |
-
model = AutoModelForCausalLM.from_pretrained("togethercomputer/RedPajama-INCITE-
|
| 113 |
|
| 114 |
# infer
|
| 115 |
prompt = "Q: The capital of France is?\nA:"
|
|
@@ -139,8 +139,8 @@ MIN_TRANSFORMERS_VERSION = '4.25.1'
|
|
| 139 |
assert transformers.__version__ >= MIN_TRANSFORMERS_VERSION, f'Please upgrade transformers to version {MIN_TRANSFORMERS_VERSION} or higher.'
|
| 140 |
|
| 141 |
# init
|
| 142 |
-
tokenizer = AutoTokenizer.from_pretrained("togethercomputer/RedPajama-INCITE-
|
| 143 |
-
model = AutoModelForCausalLM.from_pretrained("togethercomputer/RedPajama-INCITE-
|
| 144 |
# infer
|
| 145 |
prompt = "Q: The capital of France is?\nA:"
|
| 146 |
inputs = tokenizer(prompt, return_tensors='pt').to(model.device)
|
|
@@ -171,13 +171,13 @@ It is the responsibility of the end user to ensure that the model is used in a r
|
|
| 171 |
|
| 172 |
#### Out-of-Scope Use
|
| 173 |
|
| 174 |
-
RedPajama-INCITE-
|
| 175 |
For example, it may not be suitable for use in safety-critical applications or for making decisions that have a significant impact on individuals or society.
|
| 176 |
It is important to consider the limitations of the model and to only use it for its intended purpose.
|
| 177 |
|
| 178 |
#### Misuse and Malicious Use
|
| 179 |
|
| 180 |
-
RedPajama-INCITE-
|
| 181 |
Misuse of the model, such as using it to engage in illegal or unethical activities, is strictly prohibited and goes against the principles of the project.
|
| 182 |
|
| 183 |
Using the model to generate content that is cruel to individuals is a misuse of this model. This includes, but is not limited to:
|
|
@@ -194,7 +194,7 @@ Using the model to generate content that is cruel to individuals is a misuse of
|
|
| 194 |
|
| 195 |
## Limitations
|
| 196 |
|
| 197 |
-
RedPajama-INCITE-
|
| 198 |
For example, the model may not always provide accurate or relevant answers, particularly for questions that are complex, ambiguous, or outside of its training data.
|
| 199 |
We therefore welcome contributions from individuals and organizations, and encourage collaboration towards creating a more robust and inclusive chatbot.
|
| 200 |
|
|
@@ -209,7 +209,7 @@ Please refer to [togethercomputer/RedPajama-Data-1T](https://huggingface.co/data
|
|
| 209 |
- **Hardware:** 8 A100
|
| 210 |
- **Optimizer:** Adam
|
| 211 |
- **Gradient Accumulations**: 1
|
| 212 |
-
- **Num of Tokens:**
|
| 213 |
- **Learning rate:** 1e-5
|
| 214 |
|
| 215 |
## Community
|
|
|
|
| 29 |
max_new_tokens: 128
|
| 30 |
---
|
| 31 |
|
| 32 |
+
# RedPajama-INCITE-7B-Instruct
|
| 33 |
|
| 34 |
+
RedPajama-INCITE-7B-Instruct was developed by Together and leaders from the open-source AI community including Ontocord.ai, ETH DS3Lab, AAI CERC, Université de Montréal, MILA - Québec AI Institute, Stanford Center for Research on Foundation Models (CRFM), Stanford Hazy Research research group and LAION.
|
| 35 |
|
| 36 |
The model was fine-tuned for few-shot applications on the data of [GPT-JT](https://huggingface.co/togethercomputer/GPT-JT-6B-v1), with exclusion of tasks that overlap with the HELM core scenarios.
|
| 37 |
|
| 38 |
+
- Base Model: [RedPajama-INCITE-7B-Base](https://huggingface.co/togethercomputer/RedPajama-INCITE-7B-Base)
|
| 39 |
+
- Instruction-tuned Version: [RedPajama-INCITE-7B-Instruct](https://huggingface.co/togethercomputer/RedPajama-INCITE-7B-Instruct)
|
| 40 |
+
- Chat Version: [RedPajama-INCITE-7B-Chat](https://huggingface.co/togethercomputer/RedPajama-INCITE-7B-Chat)
|
| 41 |
|
| 42 |
|
| 43 |
## Model Details
|
|
|
|
| 66 |
assert transformers.__version__ >= MIN_TRANSFORMERS_VERSION, f'Please upgrade transformers to version {MIN_TRANSFORMERS_VERSION} or higher.'
|
| 67 |
|
| 68 |
# init
|
| 69 |
+
tokenizer = AutoTokenizer.from_pretrained("togethercomputer/RedPajama-INCITE-7B-Instruct")
|
| 70 |
+
model = AutoModelForCausalLM.from_pretrained("togethercomputer/RedPajama-INCITE-7B-Instruct", torch_dtype=torch.float16)
|
| 71 |
model = model.to('cuda:0')
|
| 72 |
# infer
|
| 73 |
prompt = "Q: The capital of France is?\nA:"
|
|
|
|
| 108 |
assert transformers.__version__ >= MIN_TRANSFORMERS_VERSION, f'Please upgrade transformers to version {MIN_TRANSFORMERS_VERSION} or higher.'
|
| 109 |
|
| 110 |
# init
|
| 111 |
+
tokenizer = AutoTokenizer.from_pretrained("togethercomputer/RedPajama-INCITE-7B-Instruct")
|
| 112 |
+
model = AutoModelForCausalLM.from_pretrained("togethercomputer/RedPajama-INCITE-7B-Instruct", device_map='auto', torch_dtype=torch.float16, load_in_8bit=True)
|
| 113 |
|
| 114 |
# infer
|
| 115 |
prompt = "Q: The capital of France is?\nA:"
|
|
|
|
| 139 |
assert transformers.__version__ >= MIN_TRANSFORMERS_VERSION, f'Please upgrade transformers to version {MIN_TRANSFORMERS_VERSION} or higher.'
|
| 140 |
|
| 141 |
# init
|
| 142 |
+
tokenizer = AutoTokenizer.from_pretrained("togethercomputer/RedPajama-INCITE-7B-Instruct")
|
| 143 |
+
model = AutoModelForCausalLM.from_pretrained("togethercomputer/RedPajama-INCITE-7B-Instruct", torch_dtype=torch.bfloat16)
|
| 144 |
# infer
|
| 145 |
prompt = "Q: The capital of France is?\nA:"
|
| 146 |
inputs = tokenizer(prompt, return_tensors='pt').to(model.device)
|
|
|
|
| 171 |
|
| 172 |
#### Out-of-Scope Use
|
| 173 |
|
| 174 |
+
RedPajama-INCITE-7B-Instruct is a language model and may not perform well for other use cases outside of its intended scope.
|
| 175 |
For example, it may not be suitable for use in safety-critical applications or for making decisions that have a significant impact on individuals or society.
|
| 176 |
It is important to consider the limitations of the model and to only use it for its intended purpose.
|
| 177 |
|
| 178 |
#### Misuse and Malicious Use
|
| 179 |
|
| 180 |
+
RedPajama-INCITE-7B-Instruct is designed for language modeling.
|
| 181 |
Misuse of the model, such as using it to engage in illegal or unethical activities, is strictly prohibited and goes against the principles of the project.
|
| 182 |
|
| 183 |
Using the model to generate content that is cruel to individuals is a misuse of this model. This includes, but is not limited to:
|
|
|
|
| 194 |
|
| 195 |
## Limitations
|
| 196 |
|
| 197 |
+
RedPajama-INCITE-7B-Instruct, like other language models, has limitations that should be taken into consideration.
|
| 198 |
For example, the model may not always provide accurate or relevant answers, particularly for questions that are complex, ambiguous, or outside of its training data.
|
| 199 |
We therefore welcome contributions from individuals and organizations, and encourage collaboration towards creating a more robust and inclusive chatbot.
|
| 200 |
|
|
|
|
| 209 |
- **Hardware:** 8 A100
|
| 210 |
- **Optimizer:** Adam
|
| 211 |
- **Gradient Accumulations**: 1
|
| 212 |
+
- **Num of Tokens:** 1B tokens
|
| 213 |
- **Learning rate:** 1e-5
|
| 214 |
|
| 215 |
## Community
|