jumelet commited on
Commit
4f9bed4
·
verified ·
1 Parent(s): 94cd4b7

Update configuration_gpt_bert.py

Browse files
Files changed (1) hide show
  1. configuration_gpt_bert.py +23 -24
configuration_gpt_bert.py CHANGED
@@ -1,30 +1,29 @@
1
-
2
  from transformers import PretrainedConfig
3
 
4
  class GPTBertConfig(PretrainedConfig):
5
  model_type = 'gpt_bert'
6
 
7
  def __init__(self, **kwargs):
8
- self.attention_probs_dropout_prob = kwargs.pop('attention_probs_dropout_prob', 0.1)
9
- self.hidden_dropout_prob = kwargs.pop('hidden_dropout_prob', 0.1)
10
- self.hidden_size = kwargs.pop('hidden_size', 768)
11
- self.intermediate_size = kwargs.pop('intermediate_size', 2560)
12
- self.max_position_embeddings = kwargs.pop('max_position_embeddings', 512)
13
- self.position_bucket_size = kwargs.pop('position_bucket_size', 32)
14
- self.num_attention_heads = kwargs.pop('num_attention_heads', 12)
15
- self.num_hidden_layers = kwargs.pop('num_hidden_layers', 12)
16
- self.vocab_size = kwargs.pop('vocab_size', 16384)
17
- self.layer_norm_eps = kwargs.pop('layer_norm_eps', 1e-5)
18
- self.force_causal_mask = kwargs.pop('force_causal_mask', True)
19
- self.classifier_dropout = kwargs.pop('classifier_dropout', 0.1)
20
- self.classifier_layer_norm_eps = kwargs.pop('classifier_layer_norm_eps', 1e-05)
21
- self.num_labels = kwargs.pop('num_labels', 2)
22
- self.problem_type = kwargs.pop('problem_type', None)
23
- self.auto_map = {
24
- 'AutoConfig': 'configuration_gpt_bert.GPTBertConfig',
25
- 'AutoModel': 'modeling_gpt_bert.GPTBertForMaskedLM',
26
- 'AutoModelForCausalLM': 'modeling_gpt_bert.GPTBertForCausalLM',
27
- 'AutoModelForMaskedLM': 'modeling_gpt_bert.GPTBertForMaskedLM',
28
- 'AutoModelForSequenceClassification': 'modeling_gpt_bert.GPTBertForSequenceClassification',
29
- }
30
- super().__init__(**kwargs)
 
 
1
  from transformers import PretrainedConfig
2
 
3
  class GPTBertConfig(PretrainedConfig):
4
  model_type = 'gpt_bert'
5
 
6
  def __init__(self, **kwargs):
7
+ kwargs.setdefault('attention_probs_dropout_prob', 0.1)
8
+ kwargs.setdefault('hidden_dropout_prob', 0.1)
9
+ kwargs.setdefault('hidden_size', 768)
10
+ kwargs.setdefault('intermediate_size', 2560)
11
+ kwargs.setdefault('max_position_embeddings', 512)
12
+ kwargs.setdefault('position_bucket_size', 32)
13
+ kwargs.setdefault('num_attention_heads', 12)
14
+ kwargs.setdefault('num_hidden_layers', 12)
15
+ kwargs.setdefault('vocab_size', 16384)
16
+ kwargs.setdefault('layer_norm_eps', 1e-5)
17
+ kwargs.setdefault('force_causal_mask', True)
18
+ kwargs.setdefault('classifier_dropout', 0.1)
19
+ kwargs.setdefault('classifier_layer_norm_eps', 1e-5)
20
+ kwargs.setdefault('num_labels', 2)
21
+ kwargs.setdefault('problem_type', None)
22
+ kwargs.setdefault('auto_map', {
23
+ 'AutoConfig': 'configuration_gpt_bert.GPTBertConfig',
24
+ 'AutoModel': 'modeling_gpt_bert.GPTBertForMaskedLM',
25
+ 'AutoModelForCausalLM': 'modeling_gpt_bert.GPTBertForCausalLM',
26
+ 'AutoModelForMaskedLM': 'modeling_gpt_bert.GPTBertForMaskedLM',
27
+ 'AutoModelForSequenceClassification': 'modeling_gpt_bert.GPTBertForSequenceClassification',
28
+ })
29
+ super().__init__(**kwargs)