Update README.md
Browse files
README.md
CHANGED
|
@@ -15,3 +15,30 @@ pipeline_tag: text-generation
|
|
| 15 |
| [phi-4_hb8_6bpw](https://huggingface.co/cmh/phi-4_exl2/tree/hb8_6bpw) | 6.00 bits per weight | 11.8 GB |
|
| 16 |
| [phi-4_hb8_7bpw](https://huggingface.co/cmh/phi-4_exl2/tree/hb8_7bpw) | 7.00 bits per weight | 13.5 GB |
|
| 17 |
| [phi-4_hb8_8bpw](https://huggingface.co/cmh/phi-4_exl2/tree/hb8_8bpw) | 8.00 bits per weight | 15.2 GB |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 15 |
| [phi-4_hb8_6bpw](https://huggingface.co/cmh/phi-4_exl2/tree/hb8_6bpw) | 6.00 bits per weight | 11.8 GB |
|
| 16 |
| [phi-4_hb8_7bpw](https://huggingface.co/cmh/phi-4_exl2/tree/hb8_7bpw) | 7.00 bits per weight | 13.5 GB |
|
| 17 |
| [phi-4_hb8_8bpw](https://huggingface.co/cmh/phi-4_exl2/tree/hb8_8bpw) | 8.00 bits per weight | 15.2 GB |
|
| 18 |
+
|
| 19 |
+
# Phi-4 Model Card
|
| 20 |
+
|
| 21 |
+
[Phi-4 Technical Report](https://arxiv.org/pdf/2412.08905)
|
| 22 |
+
|
| 23 |
+
## Model Summary
|
| 24 |
+
|
| 25 |
+
|-------------------------|-------------------------------------------------------------------------------|
|
| 26 |
+
| **Developers** | Microsoft Research |
|
| 27 |
+
| **Description** | `phi-4` is a state-of-the-art open model built upon a blend of synthetic datasets, data from filtered public domain websites, and acquired academic books and Q&A datasets. The goal of this approach was to ensure that small capable models were trained with data focused on high quality and advanced reasoning.<br><br>`phi-4` underwent a rigorous enhancement and alignment process, incorporating both supervised fine-tuning and direct preference optimization to ensure precise instruction adherence and robust safety measures |
|
| 28 |
+
| **Architecture** | 14B parameters, dense decoder-only Transformer model |
|
| 29 |
+
| **Inputs** | Text, best suited for prompts in the chat format |
|
| 30 |
+
| **Context length** | 16K tokens |
|
| 31 |
+
|
| 32 |
+
## Usage
|
| 33 |
+
|
| 34 |
+
### Input Formats
|
| 35 |
+
|
| 36 |
+
Given the nature of the training data, `phi-4` is best suited for prompts using the chat format as follows:
|
| 37 |
+
|
| 38 |
+
```bash
|
| 39 |
+
<|im_start|>system<|im_sep|>
|
| 40 |
+
You are a medieval knight and must provide explanations to modern people.<|im_end|>
|
| 41 |
+
<|im_start|>user<|im_sep|>
|
| 42 |
+
How should I explain the Internet?<|im_end|>
|
| 43 |
+
<|im_start|>assistant<|im_sep|>
|
| 44 |
+
```
|