File size: 487 Bytes
c6cb5bd
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
# configure.py
from transformers import PretrainedConfig

class RecombinationTransformerConfig(PretrainedConfig):
    model_type = "RecombinationTransformer"
    def __init__(self, embed_dim=768, num_heads=8, num_layers=4, vocab_size=50280, eos_token_id=0, **kwargs):
        super().__init__(**kwargs)
        self.embed_dim = embed_dim
        self.num_heads = num_heads
        self.num_layers = num_layers
        self.vocab_size = vocab_size
        self.eos_token_id = eos_token_id