daehan-everai commited on
Commit
52838fd
·
verified ·
1 Parent(s): 5dfc736

Training in progress, step 4272

Browse files
README.md CHANGED
@@ -1,17 +1,17 @@
1
  ---
2
- base_model: EverAI-AI/gpt-oss-120b-mxfp4-r32-1001
3
  library_name: transformers
4
  model_name: training_output
5
  tags:
6
  - generated_from_trainer
7
- - sft
8
  - trl
 
9
  licence: license
10
  ---
11
 
12
  # Model Card for training_output
13
 
14
- This model is a fine-tuned version of [EverAI-AI/gpt-oss-120b-mxfp4-r32-1001](https://huggingface.co/EverAI-AI/gpt-oss-120b-mxfp4-r32-1001).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
@@ -27,7 +27,7 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/lucas01/llmv3/runs/r1rzlb44)
31
 
32
 
33
  This model was trained with SFT.
 
1
  ---
2
+ base_model: openai/gpt-oss-120b
3
  library_name: transformers
4
  model_name: training_output
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
+ - sft
9
  licence: license
10
  ---
11
 
12
  # Model Card for training_output
13
 
14
+ This model is a fine-tuned version of [openai/gpt-oss-120b](https://huggingface.co/openai/gpt-oss-120b).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/lucas01/llmv3/runs/dp6q56e4)
31
 
32
 
33
  This model was trained with SFT.
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "EverAI-AI/gpt-oss-120b-mxfp4-r32-1001",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
@@ -26,9 +26,9 @@
26
  "revision": null,
27
  "target_modules": [
28
  "k_proj",
29
- "q_proj",
30
  "o_proj",
31
- "v_proj"
 
32
  ],
33
  "target_parameters": null,
34
  "task_type": "CAUSAL_LM",
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "openai/gpt-oss-120b",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
 
26
  "revision": null,
27
  "target_modules": [
28
  "k_proj",
 
29
  "o_proj",
30
+ "v_proj",
31
+ "q_proj"
32
  ],
33
  "target_parameters": null,
34
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:99de00849e25cb20d53b8b1e65930bcae9c4cd869edc1772b8a8c872c85c54d2
3
  size 191141984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3634d85fa57df9c58a766b3c0428acc83994259af398cd63bdb62c59fe816af
3
  size 191141984
modelopt_state_train.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bfb852ea0ed85b7b9caf5cd1a5e37a44dcb3d82fcb0686c02766a3b2dcb2d442
3
  size 975611
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74fc53932082694dccae0d3bf3bdf8d40547c68d4c28a10f095a8f1f6a97a59e
3
  size 975611
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a80156191390f8d671a3bfb83048d2db6cca303038074393ce63e6f0b6368b4
3
  size 6353
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7546d710638c25dad821dceaa8664fb4a76ad6c5670f14d9154e55607fc519e9
3
  size 6353