duoduoyeah commited on
Commit
e6691bc
·
verified ·
1 Parent(s): 13798be

Add files using upload-large-folder tool

Browse files
pdlm_depth4_bs1_pr1_ratio40_causal_samenoisy/base_checkpoints/d4/meta_003328.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "step": 3328,
3
+ "val_bpb": 0,
4
+ "model_config": {
5
+ "sequence_len": 1024,
6
+ "pure_vocab_size": 4096,
7
+ "all_vocab_size": 4917,
8
+ "n_layer": 4,
9
+ "n_head": 2,
10
+ "n_kv_head": 2,
11
+ "n_embd": 256,
12
+ "prefix_pure_tokens": 1,
13
+ "mask_token_id": 4096
14
+ },
15
+ "user_config": {
16
+ "run": "pdlm_depth4_bs1_pr1_ratio40_causal_samenoisy",
17
+ "device_type": "",
18
+ "depth": 4,
19
+ "max_seq_len": 1024,
20
+ "block_size": 1,
21
+ "prefix_pure_tokens": 1,
22
+ "is_causal": true,
23
+ "noise_total_steps": 0,
24
+ "debug": false,
25
+ "num_iterations": -1,
26
+ "target_flops": -1.0,
27
+ "target_param_data_ratio": 40,
28
+ "device_batch_size": 64,
29
+ "total_batch_size": 65536,
30
+ "embedding_lr": 0.2,
31
+ "unembedding_lr": 0.004,
32
+ "weight_decay": 0.0,
33
+ "matrix_lr": 0.02,
34
+ "grad_clip": 1.0,
35
+ "warmup_ratio": 0.0,
36
+ "warmdown_ratio": 0.2,
37
+ "final_lr_frac": 0.0,
38
+ "resume_from_step": -1,
39
+ "eval_every": -1,
40
+ "eval_tokens": 10485760,
41
+ "core_metric_every": -1,
42
+ "core_metric_max_per_task": 500,
43
+ "sample_every": 2000,
44
+ "save_every": -1,
45
+ "model_tag": ""
46
+ },
47
+ "device_batch_size": 64,
48
+ "max_seq_len": 1024,
49
+ "dataloader_state_dict": {
50
+ "pq_idx": 3,
51
+ "rg_idx": 129
52
+ },
53
+ "loop_state": {
54
+ "min_val_bpb": Infinity,
55
+ "smooth_train_loss": 1.0520485187019877,
56
+ "total_training_time": 399.0137403011322
57
+ }
58
+ }
pdlm_depth4_bs1_pr1_ratio40_causal_samenoisy/base_checkpoints/d4/model_003328.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afa9acbfbf78bd3b4711cb80e8d2022bbc470e93ed87327a03d4ed0ce7eb2e1d
3
+ size 19305717
pdlm_depth4_bs1_pr1_ratio40_causal_samenoisy/base_checkpoints/d4/optim_003328_rank0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f601724eab5dd62ef8a582635ace16fc3c0c424e0fc81e084f67f4792638653
3
+ size 26017109
pdlm_depth4_bs1_pr1_ratio40_causal_samenoisy/report/base-model-training.md ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ## Base model training
2
+ timestamp: 2025-12-30 02:06:05
3
+
4
+ - run: pdlm_depth4_bs1_pr1_ratio40_causal_samenoisy
5
+ - device_type:
6
+ - depth: 4
7
+ - max_seq_len: 1024
8
+ - block_size: 1
9
+ - prefix_pure_tokens: 1
10
+ - is_causal: True
11
+ - noise_total_steps: 0
12
+ - debug: False
13
+ - num_iterations: -1
14
+ - target_flops: -1.0000
15
+ - target_param_data_ratio: 40
16
+ - device_batch_size: 64
17
+ - total_batch_size: 65,536
18
+ - embedding_lr: 0.2000
19
+ - unembedding_lr: 0.0040
20
+ - weight_decay: 0.0000
21
+ - matrix_lr: 0.0200
22
+ - grad_clip: 1.0000
23
+ - warmup_ratio: 0.0000
24
+ - warmdown_ratio: 0.2000
25
+ - final_lr_frac: 0.0000
26
+ - resume_from_step: -1
27
+ - eval_every: -1
28
+ - eval_tokens: 10,485,760
29
+ - core_metric_every: -1
30
+ - core_metric_max_per_task: 500
31
+ - sample_every: 2000
32
+ - save_every: -1
33
+ - model_tag:
34
+ - Number of parameters: 5,453,056
35
+ - Number of FLOPs per token: 3.774874e+07
36
+ - Calculated number of iterations: 3328
37
+ - Number of training tokens: 218,103,808
38
+ - Tokens : Params ratio: 39.9966
39
+ - DDP world size: 1
40
+ - warmup_ratio: 0.0000
41
+ - warmdown_ratio: 0.2000
42
+ - final_lr_frac: 0.0000
43
+ - Minimum validation bpb: inf
44
+ - Final validation bpb: 0
45
+ - CORE metric estimate: None
46
+ - MFU %: 2.09%
47
+ - Total training flops: 8.233143e+15
48
+ - Total training time: 6.65m
49
+ - Peak memory usage: 6892.04MiB
50
+
pdlm_depth4_bs1_pr1_ratio40_causal_samenoisy/report/header.md ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # nanochat training report
2
+
3
+ Generated: 2025-12-30 01:57:46
4
+
5
+ ## Environment
6
+
7
+ ### Git Information
8
+ - Branch: tokenizer
9
+ - Commit: 9d07067 (clean)
10
+ - Message: update cli
11
+
12
+ ### Hardware
13
+ - Platform: Linux
14
+ - CPUs: 6 cores (12 logical)
15
+ - Memory: 83.5 GB
16
+ - GPUs: 1x NVIDIA A100-SXM4-40GB
17
+ - GPU Memory: 39.6 GB total
18
+ - CUDA Version: 12.6
19
+ - Hourly Rate: $1.79/hour
20
+
21
+ ### Software
22
+ - Python: 3.12.12
23
+ - PyTorch: 2.9.0+cu126
24
+
25
+
26
+ ### Bloat
27
+ - Characters: 537,246
28
+ - Lines: 13,151
29
+ - Files: 72
30
+ - Tokens (approx): 134,311
31
+ - Dependencies (uv.lock lines): 2,218
32
+
33
+ Run started: 2025-12-30 01:57:46
34
+
35
+ ---
36
+
pdlm_depth4_bs1_pr1_ratio40_causal_samenoisy/tokenizer/token_bytes.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c8b99c0d5a1b87b87118e840f69510440302023cd514b241614fb562373d7ce
3
+ size 17961
pdlm_depth4_bs1_pr1_ratio40_causal_samenoisy/tokenizer/token_maps.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e018730860c8e77e8cbba4b57ac9c7ba6798b5926dce925743959b932a099964
3
+ size 1850237
pdlm_depth4_bs1_pr1_ratio40_causal_samenoisy/tokenizer/tokenizer.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f874c4250ec76e2e8c4f97e91c55cfdf74d9f8eedaae14cd22db36bb718ee19
3
+ size 61662