Update README.md
Browse files
README.md
CHANGED
|
@@ -3,7 +3,7 @@ pipeline_tag: text-generation
|
|
| 3 |
license: other
|
| 4 |
---
|
| 5 |
|
| 6 |
-
#
|
| 7 |
|
| 8 |
LLaMA-7B is a base model for text generation. It was built and released by Meta AI alongside "[LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)".
|
| 9 |
|
|
@@ -31,7 +31,7 @@ evaluating and mitigating biases, risks, toxic and harmful content generations,
|
|
| 31 |
The primary intended users of the model are researchers in natural language processing, machine learning and artificial intelligence.
|
| 32 |
|
| 33 |
**Out-of-scope use cases**
|
| 34 |
-
LLaMA is a
|
| 35 |
|
| 36 |
## Factors
|
| 37 |
**Relevant factors**
|
|
@@ -60,12 +60,11 @@ LLaMA is a foundational model, and as such, it should not be used for downstream
|
|
| 60 |
|
| 61 |
### Setup
|
| 62 |
```python
|
| 63 |
-
# Install packages
|
| 64 |
!pip install -q -U transformers accelerate torch
|
| 65 |
```
|
| 66 |
### GPU Inference in fp16
|
| 67 |
|
| 68 |
-
This requires a GPU with at least 15GB
|
| 69 |
|
| 70 |
### First, Load the Model
|
| 71 |
|
|
@@ -103,4 +102,4 @@ _ = model.generate(
|
|
| 103 |
max_new_tokens=20,
|
| 104 |
streamer=streamer,
|
| 105 |
)
|
| 106 |
-
```
|
|
|
|
| 3 |
license: other
|
| 4 |
---
|
| 5 |
|
| 6 |
+
# 🦙 LLaMA-7B
|
| 7 |
|
| 8 |
LLaMA-7B is a base model for text generation. It was built and released by Meta AI alongside "[LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)".
|
| 9 |
|
|
|
|
| 31 |
The primary intended users of the model are researchers in natural language processing, machine learning and artificial intelligence.
|
| 32 |
|
| 33 |
**Out-of-scope use cases**
|
| 34 |
+
LLaMA is a base model, also known as a foundation model. As such, it should not be used on downstream applications without further risk evaluation, mitigation, and potential further fine-tuning (for example, on instructions and/or chats). In particular, the model has not been trained with human feedback, and can thus generate toxic or offensive content, incorrect information or generally unhelpful answers.
|
| 35 |
|
| 36 |
## Factors
|
| 37 |
**Relevant factors**
|
|
|
|
| 60 |
|
| 61 |
### Setup
|
| 62 |
```python
|
|
|
|
| 63 |
!pip install -q -U transformers accelerate torch
|
| 64 |
```
|
| 65 |
### GPU Inference in fp16
|
| 66 |
|
| 67 |
+
This requires a GPU with at least 15GB of VRAM.
|
| 68 |
|
| 69 |
### First, Load the Model
|
| 70 |
|
|
|
|
| 102 |
max_new_tokens=20,
|
| 103 |
streamer=streamer,
|
| 104 |
)
|
| 105 |
+
```
|