| { | |
| "feature_len": 512, | |
| "num_intermediate_likelihoods": 13, | |
| "apply_delta_augmentation": false, | |
| "apply_product_augmentation": true, | |
| "conv_layer_shapes": [ | |
| [ | |
| 32, | |
| 5, | |
| 1 | |
| ], | |
| [ | |
| 64, | |
| 5, | |
| 1 | |
| ], | |
| [ | |
| 32, | |
| 3, | |
| 1 | |
| ] | |
| ], | |
| "projection_dim": 128, | |
| "lstm_hidden_dim": 256, | |
| "lstm_layers": 1, | |
| "stack_spans": 4, | |
| "hf_dataset": "TheItCrOw/MAGE-encoded-gpt2", | |
| "dataset_root_path": "/storage/projects/stoeckel/prismai/encoded/fulltext/", | |
| "models_root_path": "/storage/projects/boenisch/PrismAI/models/luminar_sequence/", | |
| "domain": null, | |
| "agent": "gpt_4o_mini_gemma2_9b", | |
| "feature_agent": "gpt2", | |
| "max_epochs": 100, | |
| "batch_size": 128, | |
| "early_stopping_patience": 8, | |
| "rescale_features": false, | |
| "kfold": 3, | |
| "learning_rate": 0.004, | |
| "seed": 42 | |
| } |