Update README.md
Browse files
README.md
CHANGED
|
@@ -117,7 +117,9 @@ outputs_all = llm.generate(prompts, sampling_params)
|
|
| 117 |
evaluator_text = [output.outputs[0].text.split(tokenizer.eos_token)[0] for output in outputs_all]
|
| 118 |
|
| 119 |
print(evaluator_text[0])
|
| 120 |
-
# Explanation: Both Assistant A and Assistant B provided a numerical answer to the question, but Assistant A's answer is correct, while Assistant B's answer is incorrect. The question asks for the sum of 5 and 10, which is 15. Therefore, Assistant A's response is more accurate and reliable
|
|
|
|
|
|
|
| 121 |
```
|
| 122 |
|
| 123 |
|
|
|
|
| 117 |
evaluator_text = [output.outputs[0].text.split(tokenizer.eos_token)[0] for output in outputs_all]
|
| 118 |
|
| 119 |
print(evaluator_text[0])
|
| 120 |
+
# Explanation: Both Assistant A and Assistant B provided a numerical answer to the question, but Assistant A's answer is correct, while Assistant B's answer is incorrect. The question asks for the sum of 5 and 10, which is 15. Therefore, Assistant A's response is more accurate and reliable.
|
| 121 |
+
#
|
| 122 |
+
# Verdict: [A]
|
| 123 |
```
|
| 124 |
|
| 125 |
|