File size: 561 Bytes
c0ecc75 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 | {
"model_name": "meta-llama/Llama-3.2-1B-Instruct",
"sbert_model_name": "thenlper/gte-large",
"embedding_dim": 1024,
"llama_hidden_dim": 2048,
"prompt_length": 16,
"training_method": "dt_v7_single_theme",
"version": "2.0.0",
"labels": [
"ABSTRACT",
"OVERVIEW",
"TITLE",
"HEADLINE"
],
"best_composite_score": 0.8626318942769369,
"training_samples": 17254,
"validation_samples": 1918,
"lora_rank": 64,
"lora_alpha": 128,
"epochs": 50,
"batch_size": 8,
"gradient_accumulation_steps": 8,
"learning_rate": 5e-06
}
|