| | --- |
| | license: apache-2.0 |
| | --- |
| | # QLORA adapter trained with ORPO using Axolotl |
| |
|
| | [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl) |
| | <details><summary>See axolotl config</summary> |
| |
|
| | axolotl version: `0.4.0` |
| | ```yaml |
| | base_model: jsfs11/TemptressTensor-10.7B-v0.1a |
| | model_type: MistralForCausalLM |
| | tokenizer_type: LlamaTokenizer |
| | |
| | load_in_8bit: false |
| | load_in_4bit: true |
| | strict: false |
| | |
| | rl: orpo |
| | orpo_alpha: 0.1 |
| | remove_unused_columns: false |
| | |
| | chat_template: chatml |
| | datasets: |
| | - path: argilla/distilabel-capybara-dpo-7k-binarized |
| | type: chat_template.argilla |
| | dataset_prepared_path: last_run_prepared |
| | val_set_size: 0.1 |
| | output_dir: ./mistral-qlora-orpo-out |
| | |
| | adapter: qlora |
| | lora_model_dir: |
| | |
| | sequence_len: 4096 |
| | sample_packing: false |
| | pad_to_sequence_len: true |
| | |
| | lora_r: 32 |
| | lora_alpha: 16 |
| | lora_dropout: 0.05 |
| | lora_target_linear: true |
| | lora_fan_in_fan_out: |
| | lora_target_modules: |
| | - gate_proj |
| | - down_proj |
| | - up_proj |
| | - q_proj |
| | - v_proj |
| | - k_proj |
| | - o_proj |
| | |
| | wandb_project: tt orpo |
| | wandb_entity: |
| | wandb_watch: |
| | wandb_name: |
| | wandb_log_model: |
| | |
| | gradient_accumulation_steps: 3 |
| | micro_batch_size: 1 |
| | num_epochs: 1 |
| | optimizer: adamw_bnb_8bit |
| | lr_scheduler: cosine |
| | learning_rate: 0.0002 |
| | |
| | train_on_inputs: false |
| | group_by_length: false |
| | bf16: auto |
| | fp16: |
| | tf32: false |
| | |
| | gradient_checkpointing: false |
| | early_stopping_patience: |
| | resume_from_checkpoint: |
| | local_rank: |
| | logging_steps: 1 |
| | xformers_attention: |
| | flash_attention: true |
| | |
| | loss_watchdog_threshold: 5.0 |
| | loss_watchdog_patience: 3 |
| | |
| | warmup_steps: 10 |
| | evals_per_epoch: 4 |
| | eval_table_size: |
| | eval_max_new_tokens: 128 |
| | saves_per_epoch: 5 |
| | debug: |
| | deepspeed: |
| | weight_decay: 0.0 |
| | fsdp: |
| | fsdp_config: |
| | special_tokens: |
| | |
| | |
| | ``` |
| |
|
| | </details><br> |