File size: 487 Bytes
c6cb5bd | 1 2 3 4 5 6 7 8 9 10 11 12 | # configure.py
from transformers import PretrainedConfig
class RecombinationTransformerConfig(PretrainedConfig):
model_type = "RecombinationTransformer"
def __init__(self, embed_dim=768, num_heads=8, num_layers=4, vocab_size=50280, eos_token_id=0, **kwargs):
super().__init__(**kwargs)
self.embed_dim = embed_dim
self.num_heads = num_heads
self.num_layers = num_layers
self.vocab_size = vocab_size
self.eos_token_id = eos_token_id |