terTATa-LLM / metadata.json
paddle12's picture
Upload 12 files
28210ef verified
raw
history blame contribute delete
784 Bytes
{
"model_name": "tat-llm-final-e4",
"base_model": "NousResearch/Nous-Hermes-2-Mistral-7B-DPO",
"tokenizer": "NousResearch/Nous-Hermes-2-Mistral-7B-DPO",
"adapter_type": "LoRA",
"adapter_config": {
"r": 64,
"alpha": 16,
"dropout": 0.1,
"bias": "none"
},
"training": {
"dataset": "TAT-QA (train.json)",
"num_examples": 9000,
"num_epochs": 4,
"max_seq_length": 1024,
"batch_size_per_device": 2,
"learning_rate": 0.0002,
"lr_scheduler": "constant",
"fp16": false,
"bf16": true,
"optimizer": "AdamW (via Trainer)"
},
"notes": "Instruction-tuned with simplified prompt format. No evaluation run due to memory constraints. Use .generate() for inference.",
"created_by": "Your Name or Team",
"date": "2025-07-08"
}