File size: 145 Bytes
6b47ad1
 
 
 
 
 
 
1
2
3
4
5
6
7
8
{
    "lora_rank": 64,
    "lora_alpha": 128,
    "lora_dropout": 0.1,
    "lora_module": ["q_proj", "v_proj", "o_proj"],
    "uselora": false
}