| --- |
| |
| |
| {} |
| --- |
| |
| # Model Card for Phi 2 SlimOrca |
|
|
| <!-- Provide a quick summary of what the model is/does. --> |
|
|
| Phi 2 finetuned on SlimOrca-Dedup. This model was trained with the goal of giving Phi 2 the ablity to generate the EOS token together with being capable of doing beam search. It can also follow custom system prompts as shown in the example below. |
|
|
| ## Model Details |
|
|
| ## How to Get Started with the Model |
|
|
| ```python |
| import torch |
| import transformers |
| |
| model = transformers.AutoModelForCausalLM.from_pretrained( |
| "miguelcarv/phi-2-slimorca", |
| trust_remote_code=True, |
| ).to('cuda') |
| tokenizer = transformers.AutoTokenizer.from_pretrained("microsoft/phi-2") |
| |
| |
| SYSTEM_PROMPT = "You are an AI assistant. You will be given a task. You must generate a short and concise answer." |
| input_text = f"""{SYSTEM_PROMPT} |
| |
| Instruction: Give me the first 5 prime numbers and explain what prime numbers are. |
| Output:""" |
| |
| with torch.no_grad(): |
| outputs = model.generate( |
| tokenizer(input_text, return_tensors="pt")['input_ids'].to('cuda'), |
| max_length=1024, |
| num_beams = 3, |
| eos_token_id = tokenizer.eos_token_id |
| ) |
| print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
| ``` |
|
|
| ## Training Details |
|
|
| - Trained for one epoch on SlimOrca-Dedup |
| - Learning rate: 1e-5 |
| - Cosine learning rate decay to 0 |
| - Optimizer: AdamW |
| - Batch size: 256 |
| - Trained with mixed-precision bfloat16 |
|
|