--- {} --- # Test-Model This model is a fine-tuned version of [meta-llama/Llama-3.1-8B](https://huggingface.co/meta-llama/Llama-3.1-8B), and is specialized for document classification. It achieves the following results on the evaluation set: - Accuracy: 0.924647 - Hit rate: 0.977464 ## Training Parameters - **Epochs**: 5 - **Batch Size**: 20 - **Threshold**: 0.975 - **Token Length**: 512 - **Data Set Size**: 10000 - **Evenly Distributed**: True - **Municipalities**: Heroey, Kongsberg, Boemlo, Luster, Maalselv - **Learning Rate**: 0.0001 - **Weight Decay**: 0.01 - **Eval Accumulation Steps**: 4 ## Evaluation Results | Label | Accuracy | Accuracy Accumulation | Hit Rate | Hit Rate Accumulation | |:-----:|:--------:|:---------------------:|:--------:|:---------------------:| | 0 | 0.93 | 0.93 | 0.98 | 0.98 | | 1 | 0.95 | 0.95 | 0.97 | 0.97 | | 2 | 0.91 | 0.91 | 0.98 | 0.98 | | 3 | 0.40 | 0.40 | 0.57 | 0.57 | | 4 | 0.25 | 0.25 | 0.63 | 0.63 | | 5 | 0.86 | 0.86 | 0.94 | 0.94 | | 6 | 0.99 | 0.99 | 1.04 | 1.04 | | 7 | 0.94 | 0.94 | 0.98 | 0.98 | | 8 | 0.67 | 0.67 | 0.86 | 0.86 | | 9 | 0.99 | 0.99 | 0.99 | 0.99 |