eac123 commited on
Commit
a584ba6
·
verified ·
1 Parent(s): 3b1af9a

Upload folder using huggingface_hub

Browse files
Files changed (5) hide show
  1. README.md +112 -0
  2. config.json +69 -0
  3. debug.log +7 -0
  4. generation_config.json +9 -0
  5. model.safetensors +3 -0
README.md ADDED
@@ -0,0 +1,112 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ base_model: allenai/Olmo-3-1025-7B
5
+ tags:
6
+ - generated_from_trainer
7
+ model-index:
8
+ - name: model-out
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
16
+ <details><summary>See axolotl config</summary>
17
+
18
+ axolotl version: `0.15.0`
19
+ ```yaml
20
+ # ── Continued Pretraining: 7B on 8×A40 (48GB) ──
21
+
22
+ base_model: allenai/Olmo-3-1025-7B
23
+ tokenizer_type: AutoTokenizer
24
+
25
+ # ── Data ──
26
+ datasets:
27
+ - path: data/1b/all.jsonl
28
+ type: completion
29
+ field: completion
30
+
31
+ # ── Sequence / packing ──
32
+ sequence_len: 2048
33
+ sample_packing: true
34
+ pad_to_sequence_len: true
35
+ # NOTE: do NOT enable group_by_length with sample_packing
36
+
37
+ # ── Batch sizing ──
38
+ # Per-GPU: 4 seqs × 2048 tok = 8k tokens/step/GPU
39
+ # Global: 4 × 4 accum × 8 GPUs = 128 effective seqs/step
40
+ micro_batch_size: 4
41
+ gradient_accumulation_steps: 4
42
+
43
+ # ── Training ──
44
+ train_on_inputs: true
45
+ optimizer: adamw_torch
46
+ lr_scheduler: cosine
47
+ learning_rate: 5e-5
48
+ warmup_steps: 200
49
+ max_steps: 150
50
+ weight_decay: 0.01
51
+
52
+ # ── Precision / memory ──
53
+ bf16: true
54
+ flash_attention: true
55
+ gradient_checkpointing: true
56
+
57
+ # ── DeepSpeed ZeRO Stage 2 ──
58
+ deepspeed: ds_stage2.json
59
+
60
+ # ── Logging ──
61
+ logging_steps: 10
62
+ save_strategy: steps
63
+ save_steps: 50
64
+ ```
65
+
66
+ </details><br>
67
+
68
+ # model-out
69
+
70
+ This model is a fine-tuned version of [allenai/Olmo-3-1025-7B](https://huggingface.co/allenai/Olmo-3-1025-7B) on the data/1b/all.jsonl dataset.
71
+
72
+ ## Model description
73
+
74
+ More information needed
75
+
76
+ ## Intended uses & limitations
77
+
78
+ More information needed
79
+
80
+ ## Training and evaluation data
81
+
82
+ More information needed
83
+
84
+ ## Training procedure
85
+
86
+ ### Training hyperparameters
87
+
88
+ The following hyperparameters were used during training:
89
+ - learning_rate: 5e-05
90
+ - train_batch_size: 4
91
+ - eval_batch_size: 4
92
+ - seed: 42
93
+ - distributed_type: multi-GPU
94
+ - num_devices: 8
95
+ - gradient_accumulation_steps: 4
96
+ - total_train_batch_size: 128
97
+ - total_eval_batch_size: 32
98
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
99
+ - lr_scheduler_type: cosine
100
+ - lr_scheduler_warmup_steps: 200
101
+ - training_steps: 150
102
+
103
+ ### Training results
104
+
105
+
106
+
107
+ ### Framework versions
108
+
109
+ - Transformers 5.3.0
110
+ - Pytorch 2.8.0+cu126
111
+ - Datasets 4.5.0
112
+ - Tokenizers 0.22.2
config.json ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Olmo3ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": null,
8
+ "dtype": "bfloat16",
9
+ "eos_token_id": 100257,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 4096,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 11008,
14
+ "layer_types": [
15
+ "sliding_attention",
16
+ "sliding_attention",
17
+ "sliding_attention",
18
+ "full_attention",
19
+ "sliding_attention",
20
+ "sliding_attention",
21
+ "sliding_attention",
22
+ "full_attention",
23
+ "sliding_attention",
24
+ "sliding_attention",
25
+ "sliding_attention",
26
+ "full_attention",
27
+ "sliding_attention",
28
+ "sliding_attention",
29
+ "sliding_attention",
30
+ "full_attention",
31
+ "sliding_attention",
32
+ "sliding_attention",
33
+ "sliding_attention",
34
+ "full_attention",
35
+ "sliding_attention",
36
+ "sliding_attention",
37
+ "sliding_attention",
38
+ "full_attention",
39
+ "sliding_attention",
40
+ "sliding_attention",
41
+ "sliding_attention",
42
+ "full_attention",
43
+ "sliding_attention",
44
+ "sliding_attention",
45
+ "sliding_attention",
46
+ "full_attention"
47
+ ],
48
+ "max_position_embeddings": 65536,
49
+ "model_type": "olmo3",
50
+ "num_attention_heads": 32,
51
+ "num_hidden_layers": 32,
52
+ "num_key_value_heads": 32,
53
+ "pad_token_id": 100277,
54
+ "rms_norm_eps": 1e-06,
55
+ "rope_parameters": {
56
+ "attention_factor": 1.2079441541679836,
57
+ "beta_fast": 32,
58
+ "beta_slow": 1,
59
+ "factor": 8.0,
60
+ "original_max_position_embeddings": 8192,
61
+ "rope_theta": 500000,
62
+ "rope_type": "yarn"
63
+ },
64
+ "sliding_window": 4096,
65
+ "tie_word_embeddings": false,
66
+ "transformers_version": "5.3.0",
67
+ "use_cache": false,
68
+ "vocab_size": 100278
69
+ }
debug.log ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ [2026-03-10 09:21:08,376] [INFO] [axolotl.utils.data.sft._load_raw_datasets:320] [PID:2542054] Loading raw datasets...
2
+ [2026-03-10 09:21:08,520] [INFO] [axolotl.utils.data.wrappers.get_dataset_wrapper:87] [PID:2542054] Loading dataset: data/1b/all.jsonl with base_type: completion and prompt_style: None
3
+
4
+
5
+
6
+
7
+
generation_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 100257
6
+ ],
7
+ "pad_token_id": 100277,
8
+ "transformers_version": "5.3.0"
9
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fb65ffc467e5480daf0bbac7b5647ab7fa22eb81aa16ef8b679db058682ca2d
3
+ size 14596063960