| { | |
| "attn_implementation": "flash_attention_2", | |
| "connection_type": "ff", | |
| "dropout": 0.1, | |
| "encoder_backbone_type": "mistral", | |
| "encoder_lora_name": "encoder_lora", | |
| "encoder_lora_target_modules": "all", | |
| "encoder_name_or_path": "mistralai/Mistral-7B-Instruct-v0.3", | |
| "is_freeze_encoder": true, | |
| "is_freeze_universal_learner": false, | |
| "loar_r": 16, | |
| "lora_alpha": 32, | |
| "num_added_tokens": 0, | |
| "pooling_method": "mean", | |
| "universal_learner_backbone_type": "xlm-r", | |
| "universal_learner_lora_name": "universal_learner_lora", | |
| "universal_learner_lora_target_modules": "all", | |
| "universal_learner_name_or_path": "FacebookAI/xlm-roberta-large" | |
| } |