| from transformers import RobertaConfig | |
| class EmoAxisConfig(RobertaConfig): | |
| model_type = "emoaxis" | |
| def __init__(self, num_classes=28, freeze_upto=0, **kwargs): | |
| kwargs["max_position_embeddings"] = 514 | |
| kwargs["type_vocab_size"] = 1 | |
| super().__init__(**kwargs) | |
| self.num_classes = num_classes | |
| self.freeze_upto = freeze_upto | |