Commit ·
84dedf9
1
Parent(s): 869d618
Update README.md
Browse files
README.md
CHANGED
|
@@ -3,5 +3,5 @@ license: apache-2.0
|
|
| 3 |
---
|
| 4 |
We try to evaluate the LLM model performance in Indonesian.
|
| 5 |
There are many ways for calculate it, for example: BLEU, Perplexity, Human Eval, and GPT4 as Judge.
|
| 6 |
-
However, we use Perplexity as it was the fastest way for big evaluation dataset.
|
| 7 |
-
If you have any faster ways for calculating BLUE or any other metrics. Feel free to contribute in this repo.
|
|
|
|
| 3 |
---
|
| 4 |
We try to evaluate the LLM model performance in Indonesian.
|
| 5 |
There are many ways for calculate it, for example: BLEU, Perplexity, Human Eval, and GPT4 as Judge.
|
| 6 |
+
However, In our opinion, we use Perplexity as it was the fastest way for big evaluation dataset.
|
| 7 |
+
If you have any faster and easier ways for calculating BLUE or any other metrics. Feel free to contribute in this repo.
|