Commit ·
961fbd9
1
Parent(s): f7a7774
Update README.md
Browse files
README.md
CHANGED
|
@@ -1,6 +1,10 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
inference: false
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
---
|
| 5 |
|
| 6 |
# FastChat-T5 Model Card
|
|
@@ -42,4 +46,4 @@ The encoder bi-directionally encodes a question into a hidden representation. Th
|
|
| 42 |
This model is fine-tuned for 3 epochs, with a max learning rate 2e-5, warmup ratio 0.03, and a cosine learning rate schedule.
|
| 43 |
|
| 44 |
## Evaluation dataset
|
| 45 |
-
A preliminary evaluation of the model quality is conducted by creating a set of 80 diverse questions and utilizing GPT-4 to judge the model outputs. See https://vicuna.lmsys.org/ for more details.
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
inference: false
|
| 4 |
+
language:
|
| 5 |
+
- en
|
| 6 |
+
- de
|
| 7 |
+
- fr
|
| 8 |
---
|
| 9 |
|
| 10 |
# FastChat-T5 Model Card
|
|
|
|
| 46 |
This model is fine-tuned for 3 epochs, with a max learning rate 2e-5, warmup ratio 0.03, and a cosine learning rate schedule.
|
| 47 |
|
| 48 |
## Evaluation dataset
|
| 49 |
+
A preliminary evaluation of the model quality is conducted by creating a set of 80 diverse questions and utilizing GPT-4 to judge the model outputs. See https://vicuna.lmsys.org/ for more details.
|