Update README.md
Browse files
README.md
CHANGED
|
@@ -27,7 +27,7 @@ As same as 42dot-PLM, the model is built upon a Transformer decoder architecture
|
|
| 27 |
(\* unit: tokens)
|
| 28 |
### Supervised Fine-tuning
|
| 29 |
|
| 30 |
-
Fine-tuning took about
|
| 31 |
|
| 32 |
### Evaluation
|
| 33 |
Inspired by recent attempts like [Vicuna](https://lmsys.org/blog/2023-03-30-vicuna/#how-good-is-vicuna), we evaluate 42dot-PLM with other proprietary/open-sourced chatbots using GPT-4 for assessing various aspects of responses. The evaluation dataset consists of 121 prompts over 10 categories. The sample of the evaluation dataset and prompt template can be downloaded from our [GitHub repo](https://github.com/42dot/42dot_LLM).
|
|
|
|
| 27 |
(\* unit: tokens)
|
| 28 |
### Supervised Fine-tuning
|
| 29 |
|
| 30 |
+
Fine-tuning took about 112 GPU hours (in NVIDIA A100). For the training dataset, we manually constructed (question or insturuction) and response pairs, which can either be single- or multi-turn.
|
| 31 |
|
| 32 |
### Evaluation
|
| 33 |
Inspired by recent attempts like [Vicuna](https://lmsys.org/blog/2023-03-30-vicuna/#how-good-is-vicuna), we evaluate 42dot-PLM with other proprietary/open-sourced chatbots using GPT-4 for assessing various aspects of responses. The evaluation dataset consists of 121 prompts over 10 categories. The sample of the evaluation dataset and prompt template can be downloaded from our [GitHub repo](https://github.com/42dot/42dot_LLM).
|