Set max_position_embeddings: 40000 for engine-builder workaround

#2
Files changed (1) hide show
  1. config.json +1 -1
config.json CHANGED
@@ -76,7 +76,7 @@
76
  "token activation 8": 8,
77
  "token activation 9": 9
78
  },
79
- "max_position_embeddings": 8192,
80
  "mlp_bias": false,
81
  "model_type": "llama",
82
  "num_attention_heads": 64,
 
76
  "token activation 8": 8,
77
  "token activation 9": 9
78
  },
79
+ "max_position_embeddings": 40000,
80
  "mlp_bias": false,
81
  "model_type": "llama",
82
  "num_attention_heads": 64,