jacksuuuu commited on
Commit
70a2326
·
verified ·
1 Parent(s): 992db31

Fix: Add HuggingFace standard attributes to config for generate() compatibility

Browse files
Files changed (1) hide show
  1. modeling_nanogpt.py +16 -0
modeling_nanogpt.py CHANGED
@@ -13,6 +13,15 @@ class NanoGPTConfig(PretrainedConfig):
13
  """Configuration for NanoGPT model"""
14
  model_type = "nanogpt"
15
 
 
 
 
 
 
 
 
 
 
16
  def __init__(
17
  self,
18
  vocab_size=50257,
@@ -44,6 +53,13 @@ class NanoGPTConfig(PretrainedConfig):
44
  self.attn_pdrop = attn_pdrop
45
  self.layer_norm_epsilon = layer_norm_epsilon
46
  self.initializer_range = initializer_range
 
 
 
 
 
 
 
47
 
48
 
49
  class NanoGPTAttention(nn.Module):
 
13
  """Configuration for NanoGPT model"""
14
  model_type = "nanogpt"
15
 
16
+ # Add attribute mapping for HuggingFace compatibility
17
+ attribute_map = {
18
+ "hidden_size": "n_embd",
19
+ "num_hidden_layers": "n_layer",
20
+ "num_attention_heads": "n_head",
21
+ "intermediate_size": "n_inner",
22
+ "max_position_embeddings": "n_positions",
23
+ }
24
+
25
  def __init__(
26
  self,
27
  vocab_size=50257,
 
53
  self.attn_pdrop = attn_pdrop
54
  self.layer_norm_epsilon = layer_norm_epsilon
55
  self.initializer_range = initializer_range
56
+
57
+ # Add standard HuggingFace attributes for compatibility
58
+ self.hidden_size = n_embd
59
+ self.num_hidden_layers = n_layer
60
+ self.num_attention_heads = n_head
61
+ self.intermediate_size = n_inner
62
+ self.max_position_embeddings = n_positions
63
 
64
 
65
  class NanoGPTAttention(nn.Module):