Update README.md
#2
by milistu - opened
README.md
CHANGED
|
@@ -108,12 +108,15 @@ pip3 install git+https://github.com/casper-hansen/AutoAWQ.git@1c5ccc791fa2cb0697
|
|
| 108 |
from awq import AutoAWQForCausalLM
|
| 109 |
from transformers import AutoTokenizer
|
| 110 |
|
|
|
|
|
|
|
|
|
|
| 111 |
model_name_or_path = "TheBloke/Mistral-7B-v0.1-AWQ"
|
| 112 |
|
| 113 |
# Load model
|
| 114 |
model = AutoAWQForCausalLM.from_quantized(model_name_or_path, fuse_layers=True,
|
| 115 |
-
trust_remote_code=False, safetensors=True)
|
| 116 |
-
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, trust_remote_code=False)
|
| 117 |
|
| 118 |
prompt = "Tell me about AI"
|
| 119 |
prompt_template=f'''{prompt}
|
|
@@ -154,7 +157,8 @@ pipe = pipeline(
|
|
| 154 |
temperature=0.7,
|
| 155 |
top_p=0.95,
|
| 156 |
top_k=40,
|
| 157 |
-
repetition_penalty=1.1
|
|
|
|
| 158 |
)
|
| 159 |
|
| 160 |
print(pipe(prompt_template)[0]['generated_text'])
|
|
|
|
| 108 |
from awq import AutoAWQForCausalLM
|
| 109 |
from transformers import AutoTokenizer
|
| 110 |
|
| 111 |
+
device = torch.device("cuda") if torch.cuda.is_available() else torch.device("cpu")
|
| 112 |
+
|
| 113 |
+
|
| 114 |
model_name_or_path = "TheBloke/Mistral-7B-v0.1-AWQ"
|
| 115 |
|
| 116 |
# Load model
|
| 117 |
model = AutoAWQForCausalLM.from_quantized(model_name_or_path, fuse_layers=True,
|
| 118 |
+
trust_remote_code=False, safetensors=True, device=device)
|
| 119 |
+
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, trust_remote_code=False, device=device)
|
| 120 |
|
| 121 |
prompt = "Tell me about AI"
|
| 122 |
prompt_template=f'''{prompt}
|
|
|
|
| 157 |
temperature=0.7,
|
| 158 |
top_p=0.95,
|
| 159 |
top_k=40,
|
| 160 |
+
repetition_penalty=1.1,
|
| 161 |
+
device=device,
|
| 162 |
)
|
| 163 |
|
| 164 |
print(pipe(prompt_template)[0]['generated_text'])
|