elliotthwang commited on
Commit
94b8e06
·
verified ·
1 Parent(s): d8146ae

elliotthwang/KimLam-Ministral-4b-instruct-tw_train_outputs

Browse files
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: ministral/Ministral-4b-instruct
3
  library_name: transformers
4
  model_name: outputs
5
  tags:
@@ -11,7 +11,7 @@ licence: license
11
 
12
  # Model Card for outputs
13
 
14
- This model is a fine-tuned version of [ministral/Ministral-4b-instruct](https://huggingface.co/ministral/Ministral-4b-instruct).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
@@ -35,7 +35,7 @@ This model was trained with SFT.
35
  ### Framework versions
36
 
37
  - TRL: 0.16.1
38
- - Transformers: 4.50.3
39
  - Pytorch: 2.6.0+cu124
40
  - Datasets: 3.5.0
41
  - Tokenizers: 0.21.1
 
1
  ---
2
+ base_model: elliotthwang/Ministral-4b-instruct-tw_1k
3
  library_name: transformers
4
  model_name: outputs
5
  tags:
 
11
 
12
  # Model Card for outputs
13
 
14
+ This model is a fine-tuned version of [elliotthwang/Ministral-4b-instruct-tw_1k](https://huggingface.co/elliotthwang/Ministral-4b-instruct-tw_1k).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
35
  ### Framework versions
36
 
37
  - TRL: 0.16.1
38
+ - Transformers: 4.51.1
39
  - Pytorch: 2.6.0+cu124
40
  - Datasets: 3.5.0
41
  - Tokenizers: 0.21.1
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "ministral/Ministral-4b-instruct",
5
  "bias": "none",
6
  "eva_config": null,
7
  "exclude_modules": null,
@@ -25,9 +25,9 @@
25
  "target_modules": [
26
  "o_proj",
27
  "gate_proj",
28
- "k_proj",
29
  "q_proj",
30
- "v_proj"
 
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "elliotthwang/Ministral-4b-instruct-tw_1k",
5
  "bias": "none",
6
  "eva_config": null,
7
  "exclude_modules": null,
 
25
  "target_modules": [
26
  "o_proj",
27
  "gate_proj",
 
28
  "q_proj",
29
+ "v_proj",
30
+ "k_proj"
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:324e3d36eb24c6e28cf873842f5de93a1e86af98452e05502246d1ba4846638f
3
  size 38839760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aacb936da69c016bd8269af3f4779c532230c31c50434b0eecf02d3db87e6d9a
3
  size 38839760
runs/Apr15_02-41-38_4d29867e47ff/events.out.tfevents.1744684899.4d29867e47ff.596.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db947fa4541953706f94806e92f7543fdb9e7e445711ebcd642380386320bbdf
3
+ size 94471
tokenizer_config.json CHANGED
@@ -37,7 +37,7 @@
37
  "legacy": false,
38
  "model_max_length": 1000000000000000019884624838656,
39
  "pad_token": "</s>",
40
- "padding_side": "left",
41
  "sp_model_kwargs": {},
42
  "spaces_between_special_tokens": false,
43
  "tokenizer_class": "LlamaTokenizer",
 
37
  "legacy": false,
38
  "model_max_length": 1000000000000000019884624838656,
39
  "pad_token": "</s>",
40
+ "padding_side": "right",
41
  "sp_model_kwargs": {},
42
  "spaces_between_special_tokens": false,
43
  "tokenizer_class": "LlamaTokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7cf017de83701343051ddfd1b8bed347d97fa8b7c12ea0e9c6a0ab097a20cac3
3
- size 5624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82404c1bf537fc27d98e14da82fbebf16f5c7bb3d532cadb36de4fc1fba53217
3
+ size 5560