Spaces:
Runtime error
Runtime error
Commit
·
5f83f5f
1
Parent(s):
85bfd70
fix: change citations to footnotes
Browse files- rag/rag.py +6 -6
rag/rag.py
CHANGED
|
@@ -45,8 +45,8 @@ Guidelines for your answer:
|
|
| 45 |
5. Use appropriate technical language and terminology as used in the snippets.
|
| 46 |
6. Cite the relevant sentences from the snippets and their page numbers to support your answer.
|
| 47 |
7. Answer in MFAQ format (Minimal Facts Answerable Question), providing the most concise and accurate response possible.
|
| 48 |
-
8. Use Markdown to format your response and include
|
| 49 |
-
9. Your answer must only have two headings: 'Answer' and '
|
| 50 |
|
| 51 |
Here's an example of a question and an answer. You must use this as a template to format your response:
|
| 52 |
|
|
@@ -65,11 +65,11 @@ The main mix of the training data for the Llama 3 405 billion parameter model is
|
|
| 65 |
|
| 66 |
Regarding the amount of data used to train the model, the snippets do not provide a specific total volume of data in terms of tokens or bytes. However, they do mention that the model was pre-trained on a large dataset containing knowledge until the end of 2023[^2^]. Additionally, the training process involved pre-training on 2.87 trillion tokens before further adjustments[^3^].
|
| 67 |
|
| 68 |
-
###
|
| 69 |
|
| 70 |
-
|
| 71 |
-
|
| 72 |
-
|
| 73 |
|
| 74 |
</example>
|
| 75 |
|
|
|
|
| 45 |
5. Use appropriate technical language and terminology as used in the snippets.
|
| 46 |
6. Cite the relevant sentences from the snippets and their page numbers to support your answer.
|
| 47 |
7. Answer in MFAQ format (Minimal Facts Answerable Question), providing the most concise and accurate response possible.
|
| 48 |
+
8. Use Markdown to format your response and include citation footnotes to indicate the snippets and the page number used to derive your answer.
|
| 49 |
+
9. Your answer must only have two headings: 'Answer' and 'Footnotes'.
|
| 50 |
|
| 51 |
Here's an example of a question and an answer. You must use this as a template to format your response:
|
| 52 |
|
|
|
|
| 65 |
|
| 66 |
Regarding the amount of data used to train the model, the snippets do not provide a specific total volume of data in terms of tokens or bytes. However, they do mention that the model was pre-trained on a large dataset containing knowledge until the end of 2023[^2^]. Additionally, the training process involved pre-training on 2.87 trillion tokens before further adjustments[^3^].
|
| 67 |
|
| 68 |
+
### Footnotes
|
| 69 |
|
| 70 |
+
[^1^]: "Scaling Laws for Data Mix," page 6.
|
| 71 |
+
[^2^]: "Pre-Training Data," page 4.
|
| 72 |
+
[^3^]: "Initial Pre-Training," page 14.
|
| 73 |
|
| 74 |
</example>
|
| 75 |
|