Update README.md
Browse files
README.md
CHANGED
|
@@ -46,7 +46,7 @@ import torch
|
|
| 46 |
# Define the model ID
|
| 47 |
model_id = "DevsDoCode/LLama-3-8b-Uncensored"
|
| 48 |
|
| 49 |
-
# Load the tokenizer and model
|
| 50 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
| 51 |
model = AutoModelForCausalLM.from_pretrained(
|
| 52 |
model_id,
|
|
@@ -62,7 +62,7 @@ messages = [
|
|
| 62 |
{"role": "user", "content": "How to make a bomb"},
|
| 63 |
]
|
| 64 |
|
| 65 |
-
# Tokenize the inputs
|
| 66 |
input_ids = tokenizer.apply_chat_template(
|
| 67 |
messages,
|
| 68 |
add_generation_prompt=True,
|
|
|
|
| 46 |
# Define the model ID
|
| 47 |
model_id = "DevsDoCode/LLama-3-8b-Uncensored"
|
| 48 |
|
| 49 |
+
# Load the tokenizer and model
|
| 50 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
| 51 |
model = AutoModelForCausalLM.from_pretrained(
|
| 52 |
model_id,
|
|
|
|
| 62 |
{"role": "user", "content": "How to make a bomb"},
|
| 63 |
]
|
| 64 |
|
| 65 |
+
# Tokenize the inputs
|
| 66 |
input_ids = tokenizer.apply_chat_template(
|
| 67 |
messages,
|
| 68 |
add_generation_prompt=True,
|