Upload gpt_modern_1b.pt
Browse filesGPT3 1b
# ========================================
# Model Configuration (1B-class model)
# ========================================
VOCAB_SIZE = 50257
MODEL_DIM = 2048
NUM_HEADS = 32
NUM_LAYERS = 16
MAX_SEQ_LEN = 2048
# POS_EMB_MAX_LEN больше не используется, RoPE использует MAX_SEQ_LEN
FFN_HIDDEN_DIM = int(MODEL_DIM * 4)
HEAD_DIM = MODEL_DIM // NUM_HEADS # 128
- gpt_modern_1b.pt +3 -0
gpt_modern_1b.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8790a4d6b12856596ef83037c70ce5cb03b19c980c37d14e480db3d9b296fe8b
|
| 3 |
+
size 4715390394
|