saracandu commited on
Commit
f6d552c
·
verified ·
1 Parent(s): 4eacfc4

Delete configuration.py

Browse files
Files changed (1) hide show
  1. configuration.py +0 -68
configuration.py DELETED
@@ -1,68 +0,0 @@
1
- from transformers.configuration_utils import PretrainedConfig
2
-
3
- class STLConfig(PretrainedConfig):
4
-
5
- model_type = "stldec"
6
- keys_to_ignore_at_inference = ["past_key_values"]
7
- attribute_map = {"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"}
8
-
9
- def __init__(
10
- self,
11
- vocab_size=35,
12
- decoder_vocab_size=None, # unused
13
- max_position_embeddings=1024,
14
- encoder_layers=12,
15
- encoder_ffn_dim=4096,
16
- encoder_attention_heads=16,
17
- decoder_layers=12,
18
- decoder_ffn_dim=4096,
19
- decoder_attention_heads=16,
20
- encoder_layerdrop=0.0,
21
- decoder_layerdrop=0.0,
22
- use_cache=True,
23
- is_encoder_decoder=True,
24
- activation_function="gelu",
25
- d_model=1024,
26
- dropout=0.1,
27
- attention_dropout=0.0,
28
- activation_dropout=0.0,
29
- init_std=0.02,
30
- decoder_start_token_id=3,
31
- scale_embedding=False,
32
- pad_token_id=1,
33
- eos_token_id=3,
34
- bos_token_id=2,
35
- forced_eos_token_id=3,
36
- share_encoder_decoder_embeddings=True,
37
- **kwargs,
38
- ):
39
- self.vocab_size = vocab_size
40
- self.decoder_vocab_size = decoder_vocab_size or vocab_size
41
- self.max_position_embeddings = max_position_embeddings
42
- self.d_model = d_model
43
- self.encoder_ffn_dim = encoder_ffn_dim
44
- self.encoder_layers = encoder_layers
45
- self.encoder_attention_heads = encoder_attention_heads
46
- self.decoder_ffn_dim = decoder_ffn_dim
47
- self.decoder_layers = decoder_layers
48
- self.decoder_attention_heads = decoder_attention_heads
49
- self.dropout = dropout
50
- self.attention_dropout = attention_dropout
51
- self.activation_dropout = activation_dropout
52
- self.activation_function = activation_function
53
- self.init_std = init_std
54
- self.encoder_layerdrop = encoder_layerdrop
55
- self.decoder_layerdrop = decoder_layerdrop
56
- self.use_cache = use_cache
57
- self.num_hidden_layers = encoder_layers
58
- self.scale_embedding = scale_embedding # scale factor will be sqrt(d_model) if True
59
- self.share_encoder_decoder_embeddings = share_encoder_decoder_embeddings
60
- super().__init__(
61
- bos_token_id=bos_token_id,
62
- pad_token_id=pad_token_id,
63
- eos_token_id=eos_token_id,
64
- is_encoder_decoder=is_encoder_decoder,
65
- decoder_start_token_id=decoder_start_token_id,
66
- forced_eos_token_id=forced_eos_token_id,
67
- **kwargs,
68
- )