| { |
| "model_name": "tat-llm-final-e4", |
| "base_model": "NousResearch/Nous-Hermes-2-Mistral-7B-DPO", |
| "tokenizer": "NousResearch/Nous-Hermes-2-Mistral-7B-DPO", |
| "adapter_type": "LoRA", |
| "adapter_config": { |
| "r": 64, |
| "alpha": 16, |
| "dropout": 0.1, |
| "bias": "none" |
| }, |
| "training": { |
| "dataset": "TAT-QA (train.json)", |
| "num_examples": 9000, |
| "num_epochs": 4, |
| "max_seq_length": 1024, |
| "batch_size_per_device": 2, |
| "learning_rate": 0.0002, |
| "lr_scheduler": "constant", |
| "fp16": false, |
| "bf16": true, |
| "optimizer": "AdamW (via Trainer)" |
| }, |
| "notes": "Instruction-tuned with simplified prompt format. No evaluation run due to memory constraints. Use .generate() for inference.", |
| "created_by": "Your Name or Team", |
| "date": "2025-07-08" |
| } |