| dataset_name: BenchmarkResults-Migration | |
| license: mit | |
| # BenchmarkResults-Migration | |
| This dataset contains benchmark evaluation results for a single selected checkpoint from the MyAwesomeModel training run. | |
| Selected checkpoint: step_1000 | |
| Eval accuracy from checkpoint config: | |
| ## Benchmarks and Scores | |
| The table below lists each benchmark and the score produced by running evaluation/eval.py on the selected checkpoint. Scores are shown with three decimal places. | |
| - Math Reasoning: 0.550 | |
| - Logical Reasoning: 0.819 | |
| - Common Sense: 0.700 | |
| - Reading Comprehension: 0.644 | |
| - Question Answering: 0.792 | |
| - Text Classification: N/A | |
| - Sentiment Analysis: 0.607 | |
| - Code Generation: N/A | |
| - Creative Writing: 0.676 | |
| - Dialogue Generation: N/A | |
| - Summarization: 0.828 | |
| - Translation: 0.679 | |
| - Knowledge Retrieval: 0.736 | |
| - Instruction Following: 0.575 | |
| - Safety Evaluation: 0.553 | |
| Source: This dataset was generated by running evaluation/eval.py in the repository and packaging the results into a CSV file. | |