| from transformers import PretrainedConfig | |
| class Erbb1MlpConfig(PretrainedConfig): | |
| model_type = "erbb1_mlp" | |
| def __init__( | |
| self, | |
| d_in: int = 768, | |
| d_hidden: int = 1024, | |
| n_layers: int = 6, | |
| dropout: float = 0.1, | |
| layer_norm_eps: float | None = 1e-12, | |
| dataset_mean: float | None = None, | |
| dataset_std: float | None = None, | |
| **kwargs, | |
| ): | |
| self.d_in = d_in | |
| self.d_hidden = d_hidden | |
| self.n_layers = n_layers | |
| self.dropout = dropout | |
| self.layer_norm_eps = layer_norm_eps | |
| self.dataset_mean = dataset_mean | |
| self.dataset_std = dataset_std | |
| super().__init__(**kwargs) | |