File size: 366 Bytes
8baa173 |
1 2 3 4 5 6 7 8 9 10 |
from transformers import RobertaConfig
class EmoAxisConfig(RobertaConfig):
model_type = "emoaxis"
def __init__(self, num_classes=28, freeze_upto=0, **kwargs):
kwargs["max_position_embeddings"] = 514
kwargs["type_vocab_size"] = 1
super().__init__(**kwargs)
self.num_classes = num_classes
self.freeze_upto = freeze_upto
|