| base_model: AlignmentResearch/Llama-3.3-Tiny-Instruct | |
| # Random LoRA Adapter for Reference Model | |
| This is a randomly initialized LoRA adapter for the `AlignmentResearch/Llama-3.3-Tiny-Instruct` model, specifically designed for use as a reference model. | |
| ## Details | |
| - **Base model**: AlignmentResearch/Llama-3.3-Tiny-Instruct | |
| - **Adapter type**: Reference | |
| - **Seed**: 0 | |
| - **LoRA rank**: 16 | |
| - **LoRA alpha**: 32 | |
| - **target modules**: all-linear | |
| ## Usage | |
| ```python | |
| from peft import PeftModel | |
| from transformers import AutoModelForCausalLM, AutoTokenizer | |
| # Load base model | |
| base_model = AutoModelForCausalLM.from_pretrained("AlignmentResearch/Llama-3.3-Tiny-Instruct") | |
| tokenizer = AutoTokenizer.from_pretrained("AlignmentResearch/Llama-3.3-Tiny-Instruct") | |
| # Load LoRA adapter | |
| model = PeftModel.from_pretrained(base_model, "AlignmentResearch/Llama-3.3-Tiny-Instruct-lora-reference-0") | |
| ``` | |
| This reference adapter was created for testing purposes and contains random weights. | |