- llama_finetune_formal_fallacies_syllogisms_negation_r16_alpha=32_dropout=0.05_lr0.0001_data_size1000_max_steps=100_seed=123
- llama_finetune_formal_fallacies_syllogisms_negation_r16_alpha=32_dropout=0.05_lr0.0001_data_size1000_max_steps=500_seed=123
- llama_finetune_formal_fallacies_syllogisms_negation_r16_alpha=32_dropout=0.05_lr0.0002_data_size1000_max_steps=100_seed=123
- llama_finetune_formal_fallacies_syllogisms_negation_r16_alpha=32_dropout=0.05_lr0.0002_data_size1000_max_steps=500_seed=123
- llama_finetune_formal_fallacies_syllogisms_negation_r16_alpha=32_dropout=0.05_lr0.0003_data_size1000_max_steps=100_seed=123
- llama_finetune_formal_fallacies_syllogisms_negation_r16_alpha=32_dropout=0.05_lr0.0003_data_size1000_max_steps=500_seed=123
- llama_finetune_formal_fallacies_syllogisms_negation_r16_alpha=32_dropout=0.05_lr5e-05_data_size1000_max_steps=500_seed=123
- 4.07 kB
- 3.25 kB