Update README.md
Browse files
README.md
CHANGED
|
@@ -66,7 +66,7 @@ tokenizer = AutoTokenizer.from_pretrained(model_name)
|
|
| 66 |
|
| 67 |
instruction = "Please reason step by step, and put your final answer within \\boxed{}."
|
| 68 |
prompt = "The arithmetic mean of 7, 2, $x$ and 10 is 9. What is the value of $x$?"
|
| 69 |
-
#prompt = "What is the smallest positive perfect cube that can be written as the sum of three consecutive integers?"
|
| 70 |
# prompt = "How many r's are in the word \"strawberry\""
|
| 71 |
|
| 72 |
messages = [
|
|
@@ -79,6 +79,8 @@ text = tokenizer.apply_chat_template(
|
|
| 79 |
add_generation_prompt=True
|
| 80 |
)
|
| 81 |
|
|
|
|
|
|
|
| 82 |
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
|
| 83 |
|
| 84 |
generated_ids = model.generate(
|
|
|
|
| 66 |
|
| 67 |
instruction = "Please reason step by step, and put your final answer within \\boxed{}."
|
| 68 |
prompt = "The arithmetic mean of 7, 2, $x$ and 10 is 9. What is the value of $x$?"
|
| 69 |
+
# prompt = "What is the smallest positive perfect cube that can be written as the sum of three consecutive integers?"
|
| 70 |
# prompt = "How many r's are in the word \"strawberry\""
|
| 71 |
|
| 72 |
messages = [
|
|
|
|
| 79 |
add_generation_prompt=True
|
| 80 |
)
|
| 81 |
|
| 82 |
+
# text = text + "<think>" # Uncomment this to force thinking mode
|
| 83 |
+
|
| 84 |
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
|
| 85 |
|
| 86 |
generated_ids = model.generate(
|