| { |
| "model_type": "ConcatModelWithRationale", |
| "hatebert_model": "GroNLP/HateBERT", |
| "rationale_model": "bert-base-uncased", |
|
|
| "training_config": { |
| "cnn_filters": 128, |
| "cnn_kernels": [2, 3, 4], |
| "cnn_dropout": 0.3, |
| "adapter_dim": 128, |
| "max_length": 128, |
| "num_labels": 2, |
|
|
| "batch_size": 8, |
| "learning_rate": 2e-5, |
| "weight_decay": 0.05, |
| "best_val_f1": 0.00 |
| }, |
|
|
| "architecture": { |
| "temporal_cnn": "TemporalCNN (max+mean pooling)", |
| "msa_cnn": "MultiScaleAttentionCNN (token attention pooling)", |
| "selector": "Gumbel-Sigmoid Token Selector", |
| "projection": "ProjectionMLP (concat CLS + CNN + rationale pooled)" |
| }, |
|
|
| "notes": { |
| "trained_with": "ConcatModelWithRationale", |
| "framework": "PyTorch + Transformers", |
| "encoder_main": "HateBERT", |
| "encoder_rationale": "BERT-base-uncased" |
| } |
| } |