|
|
--- |
|
|
license: other |
|
|
base_model: huggyllama/llama-7b |
|
|
tags: |
|
|
- alignment-handbook |
|
|
- generated_from_trainer |
|
|
model-index: |
|
|
- name: una-llama-7b |
|
|
results: [] |
|
|
--- |
|
|
|
|
|
# una-llama-7b |
|
|
**UNA: Uniform Neural Alignment** It increases 6.75% the performance of the pre-trained base LLaMA (1) 7B. |
|
|
|
|
|
This model is a fine-tuned version of [huggyllama/llama-7b](https://huggingface.co/huggyllama/llama-7b): |
|
|
- Loss: 0.5529 |
|
|
- Rewards/chosen: 0.3633 |
|
|
- Rewards/rejected: -0.1873 |
|
|
- Rewards/accuracies: 0.7230 |
|
|
- Rewards/margins: 0.5506 |
|
|
- Logps/rejected: -217.7784 |
|
|
- Logps/chosen: -235.0354 |
|
|
- Logits/rejected: -0.7752 |
|
|
- Logits/chosen: -0.5259 |
|
|
|
|
|
## Model description |
|
|
|
|
|
More information needed |
|
|
|
|
|
## Intended uses & limitations |
|
|
|
|
|
More information needed |
|
|
|
|
|
## Training and evaluation data |
|
|
|
|
|
More information needed |
|
|
|
|
|
## Training procedure |
|
|
|
|
|
### Framework versions |
|
|
|
|
|
- Transformers 4.35.0-UNA |
|
|
- Pytorch 2.1.0 |
|
|
- Datasets 2.14.6 |
|
|
- Tokenizers 0.14.1 |
|
|
|