| datasets: | |
| - name: Scottie201/seq2seq_data | |
| split: train | |
| - name: Scottie201/seq2seq_data | |
| split: validation | |
| decision: | |
| continue_training: false | |
| reason: Model performance has not improved enough to warrant continued training. | |
| suggestions: | |
| - Consider adjusting the learning rate. | |
| - Review training data for any inconsistencies. | |
| model: | |
| architecture: seq2seq | |
| base_model: facebook/bart-large | |
| batch_size: 4 | |
| epochs: 3 | |
| learning_rate: 3e-5 | |
| output_dir: /workspaces/wakin_bake/trained_seq2seq | |
| resources: | |
| device_map: auto | |
| torch_dtype: bfloat16 | |
| use_gpu: true | |
| tokenization: | |
| max_length: 512 | |
| padding: max_length | |
| source_column: input_text | |
| target_column: target_text | |
| truncation: true | |
| training_args: | |
| eval_strategy: epoch | |
| generation_max_length: 150 | |
| generation_num_beams: 5 | |
| hub_model_id: Scottie201/seq2seq_finetuned | |
| logging_dir: ./logs | |
| logging_steps: 1 | |
| push_to_hub: true | |
| save_steps: 350 | |
| save_total_limit: 1 | |
| training_continue: false | |