dfurman commited on
Commit
52f944d
·
1 Parent(s): dd6cd05

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -5
README.md CHANGED
@@ -3,7 +3,7 @@ pipeline_tag: text-generation
3
  license: other
4
  ---
5
 
6
- # 🚀 LLaMA-7B
7
 
8
  LLaMA-7B is a base model for text generation. It was built and released by Meta AI alongside "[LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)".
9
 
@@ -31,7 +31,7 @@ evaluating and mitigating biases, risks, toxic and harmful content generations,
31
  The primary intended users of the model are researchers in natural language processing, machine learning and artificial intelligence.
32
 
33
  **Out-of-scope use cases**
34
- LLaMA is a foundation model (a base model). As such, it should not be used on downstream applications without further risk evaluation and mitigation. In particular, the model has not been trained with human feedback, and can thus generate toxic or offensive content, incorrect information or generally unhelpful answers.
35
 
36
  ## Factors
37
  **Relevant factors**
@@ -60,12 +60,11 @@ LLaMA is a foundational model, and as such, it should not be used for downstream
60
 
61
  ### Setup
62
  ```python
63
- # Install packages
64
  !pip install -q -U transformers accelerate torch
65
  ```
66
  ### GPU Inference in fp16
67
 
68
- This requires a GPU with at least 15GB memory.
69
 
70
  ### First, Load the Model
71
 
@@ -103,4 +102,4 @@ _ = model.generate(
103
  max_new_tokens=20,
104
  streamer=streamer,
105
  )
106
- ```
 
3
  license: other
4
  ---
5
 
6
+ # 🦙 LLaMA-7B
7
 
8
  LLaMA-7B is a base model for text generation. It was built and released by Meta AI alongside "[LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)".
9
 
 
31
  The primary intended users of the model are researchers in natural language processing, machine learning and artificial intelligence.
32
 
33
  **Out-of-scope use cases**
34
+ LLaMA is a base model, also known as a foundation model. As such, it should not be used on downstream applications without further risk evaluation, mitigation, and potential further fine-tuning (for example, on instructions and/or chats). In particular, the model has not been trained with human feedback, and can thus generate toxic or offensive content, incorrect information or generally unhelpful answers.
35
 
36
  ## Factors
37
  **Relevant factors**
 
60
 
61
  ### Setup
62
  ```python
 
63
  !pip install -q -U transformers accelerate torch
64
  ```
65
  ### GPU Inference in fp16
66
 
67
+ This requires a GPU with at least 15GB of VRAM.
68
 
69
  ### First, Load the Model
70
 
 
102
  max_new_tokens=20,
103
  streamer=streamer,
104
  )
105
+ ```