yongtaek commited on
Commit
7dbd94f
·
verified ·
1 Parent(s): 791e230

End of training

Browse files
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  base_model: microsoft/Phi-3.5-mini-instruct
3
  library_name: transformers
4
- model_name: results
5
  tags:
6
  - generated_from_trainer
7
  - trl
@@ -9,7 +9,7 @@ tags:
9
  licence: license
10
  ---
11
 
12
- # Model Card for results
13
 
14
  This model is a fine-tuned version of [microsoft/Phi-3.5-mini-instruct](https://huggingface.co/microsoft/Phi-3.5-mini-instruct).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
@@ -20,7 +20,7 @@ It has been trained using [TRL](https://github.com/huggingface/trl).
20
  from transformers import pipeline
21
 
22
  question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
23
- generator = pipeline("text-generation", model="yongtaek/results", device="cuda")
24
  output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
25
  print(output["generated_text"])
26
  ```
@@ -35,7 +35,7 @@ This model was trained with SFT.
35
 
36
  - TRL: 0.12.0
37
  - Transformers: 4.46.2
38
- - Pytorch: 2.5.1
39
  - Datasets: 3.1.0
40
  - Tokenizers: 0.20.3
41
 
 
1
  ---
2
  base_model: microsoft/Phi-3.5-mini-instruct
3
  library_name: transformers
4
+ model_name: phi3_5_adapter_model
5
  tags:
6
  - generated_from_trainer
7
  - trl
 
9
  licence: license
10
  ---
11
 
12
+ # Model Card for phi3_5_adapter_model
13
 
14
  This model is a fine-tuned version of [microsoft/Phi-3.5-mini-instruct](https://huggingface.co/microsoft/Phi-3.5-mini-instruct).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
 
20
  from transformers import pipeline
21
 
22
  question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
23
+ generator = pipeline("text-generation", model="yongtaek/phi3_5_adapter_model", device="cuda")
24
  output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
25
  print(output["generated_text"])
26
  ```
 
35
 
36
  - TRL: 0.12.0
37
  - Transformers: 4.46.2
38
+ - Pytorch: 2.3.0
39
  - Datasets: 3.1.0
40
  - Tokenizers: 0.20.3
41
 
adapter_config.json CHANGED
@@ -20,10 +20,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "o_proj",
24
  "gate_up_proj",
25
- "down_proj",
26
- "qkv_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "qkv_proj",
24
  "gate_up_proj",
25
+ "o_proj",
26
+ "down_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:01940fef994adcd95c6c24772f60d02ca5342d2982f311dccbd56dc364928479
3
  size 402688040
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eba77701cdb28054bc433bb41335e1fab52199a74c5d7191f1fb6e8442136aa0
3
  size 402688040
runs/Dec04_04-35-02_e18105abca6d/events.out.tfevents.1733286950.e18105abca6d.4461.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8eb19555011082d000c2db969e78b7f6ec1629969b8d15a73d21984323e580f2
3
+ size 8307
runs/Dec04_04-36-45_e18105abca6d/events.out.tfevents.1733287015.e18105abca6d.4461.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e91ab1b96f13cef1173ce2b11e08aad5ab78ee325034c12e7e08361f6bb77e48
3
+ size 22567
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f4cb60d212bb9616a866a351092166fa935eee482d0789d7e47427eaba453b20
3
  size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddb13b5198f684e5efc7d98ab507d093a8c7f10719b9fe12fcf27ffd7572b842
3
  size 5560