| | --- |
| | base_model: microsoft/Phi-3-mini-4k-instruct |
| | datasets: |
| | - AlignmentLab-AI/alpaca-cot-collection |
| | language: |
| | - en |
| | library_name: peft |
| | license: apache-2.0 |
| | pipeline_tag: text-generation |
| | --- |
| | |
| | # Xenith-3B |
| | Xenith-3B is a fine-tuned language model based on the microsoft/Phi-3-mini-4k-instruct model. It has been specifically trained on the AlignmentLab-AI/alpaca-cot-collection dataset, which focuses on chain-of-thought reasoning and instruction following. |
| |
|
| | # Model Overview |
| | - Model Name: Xenith-3B |
| | - Base Model: microsoft/Phi-3-mini-4k-instruct |
| | - Fine-Tuned On: AlignmentLab-AI/alpaca-cot-collection |
| | - Model Size: 3 Billion parameters |
| | - Architecture: Transformer-based LLM |
| |
|
| | # Training Details |
| | - Objective: Fine-tune the base model to enhance its performance on tasks requiring complex reasoning and multi-step problem-solving. |
| | - Training Duration: 10 epochs |
| | - Batch Size: 8 |
| | - Learning Rate: 3e-5 |
| | - Optimizer: AdamW |
| | - Hardware Used: 2x NVIDIA L4 GPUs |
| |
|
| | # Performance |
| | Xenith-3B excels in tasks that require: |
| |
|
| | - Chain-of-thought reasoning |
| | - Instruction following |
| | - Contextual understanding |
| | - Complex problem-solving |
| | - |
| | The model has shown significant improvements in these areas compared to the base model. |