Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling
Paper
•
2304.01373
•
Published
•
9
We provide a low-rank adapter for an instruction-tuned 12B-parameter GPT3-style language model.
For instructions that do not require extraneous inputs, the recommended prompt is:
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
< -- instruction goes here --- >
### Response:
For instructions that do require extraneous inputs, the recommended prompt is:
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
< -- instruction goes here -- >
### Input:
< -- extraneous input goes here -- >
### Response:
Since the model performs causal language modeling, the model's response to the prompt is the text completing the sequence beginning with the prompt.
This model was instruction-tuned on an internally-curated Alpaca-style dataset.
This model was instruction-tuned from a 12B variant from the Pythia family.
We release this adapter under the Creative Commons NonCommercial (CC BY-NC 4.0) license.