jda commited on
Commit
59a94ca
·
verified ·
1 Parent(s): 8ef94a8

Model save

Browse files
README.md CHANGED
@@ -1,5 +1,6 @@
1
  ---
2
  license: other
 
3
  tags:
4
  - generated_from_trainer
5
  model-index:
@@ -12,7 +13,7 @@ should probably proofread and complete it, then remove this comment. -->
12
 
13
  # opt
14
 
15
- This model is a fine-tuned version of [facebook/opt-1.3b](https://huggingface.co/facebook/opt-1.3b) on an unknown dataset.
16
 
17
  ## Model description
18
 
@@ -32,17 +33,18 @@ More information needed
32
 
33
  The following hyperparameters were used during training:
34
  - learning_rate: 0.0005
35
- - train_batch_size: 32
36
- - eval_batch_size: 64
37
  - seed: 42
38
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
39
  - lr_scheduler_type: cosine
40
  - lr_scheduler_warmup_steps: 1000
41
  - num_epochs: 1
 
42
 
43
  ### Framework versions
44
 
45
- - Transformers 4.30.2
46
- - Pytorch 1.12.1
47
- - Datasets 2.13.2
48
- - Tokenizers 0.13.3
 
1
  ---
2
  license: other
3
+ base_model: facebook/opt-350m
4
  tags:
5
  - generated_from_trainer
6
  model-index:
 
13
 
14
  # opt
15
 
16
+ This model is a fine-tuned version of [facebook/opt-350m](https://huggingface.co/facebook/opt-350m) on an unknown dataset.
17
 
18
  ## Model description
19
 
 
33
 
34
  The following hyperparameters were used during training:
35
  - learning_rate: 0.0005
36
+ - train_batch_size: 4
37
+ - eval_batch_size: 32
38
  - seed: 42
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: cosine
41
  - lr_scheduler_warmup_steps: 1000
42
  - num_epochs: 1
43
+ - mixed_precision_training: Native AMP
44
 
45
  ### Framework versions
46
 
47
+ - Transformers 4.38.2
48
+ - Pytorch 2.2.1+cu121
49
+ - Datasets 2.18.0
50
+ - Tokenizers 0.15.2
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "facebook/opt-1.3b",
3
  "_remove_final_layer_norm": false,
4
  "activation_dropout": 0.0,
5
  "activation_function": "relu",
@@ -8,24 +8,24 @@
8
  ],
9
  "attention_dropout": 0.0,
10
  "bos_token_id": 2,
11
- "do_layer_norm_before": true,
12
  "dropout": 0.1,
13
  "enable_bias": true,
14
  "eos_token_id": 2,
15
- "ffn_dim": 8192,
16
- "hidden_size": 2048,
17
  "init_std": 0.02,
18
  "layer_norm_elementwise_affine": true,
19
  "layerdrop": 0.0,
20
  "max_position_embeddings": 2048,
21
  "model_type": "opt",
22
- "num_attention_heads": 32,
23
  "num_hidden_layers": 24,
24
  "pad_token_id": 1,
25
  "prefix": "</s>",
26
  "torch_dtype": "float32",
27
- "transformers_version": "4.30.2",
28
  "use_cache": true,
29
  "vocab_size": 50265,
30
- "word_embed_proj_dim": 2048
31
  }
 
1
  {
2
+ "_name_or_path": "facebook/opt-350m",
3
  "_remove_final_layer_norm": false,
4
  "activation_dropout": 0.0,
5
  "activation_function": "relu",
 
8
  ],
9
  "attention_dropout": 0.0,
10
  "bos_token_id": 2,
11
+ "do_layer_norm_before": false,
12
  "dropout": 0.1,
13
  "enable_bias": true,
14
  "eos_token_id": 2,
15
+ "ffn_dim": 4096,
16
+ "hidden_size": 1024,
17
  "init_std": 0.02,
18
  "layer_norm_elementwise_affine": true,
19
  "layerdrop": 0.0,
20
  "max_position_embeddings": 2048,
21
  "model_type": "opt",
22
+ "num_attention_heads": 16,
23
  "num_hidden_layers": 24,
24
  "pad_token_id": 1,
25
  "prefix": "</s>",
26
  "torch_dtype": "float32",
27
+ "transformers_version": "4.38.2",
28
  "use_cache": true,
29
  "vocab_size": 50265,
30
+ "word_embed_proj_dim": 512
31
  }
generation_config.json CHANGED
@@ -3,5 +3,5 @@
3
  "bos_token_id": 2,
4
  "eos_token_id": 2,
5
  "pad_token_id": 1,
6
- "transformers_version": "4.30.2"
7
  }
 
3
  "bos_token_id": 2,
4
  "eos_token_id": 2,
5
  "pad_token_id": 1,
6
+ "transformers_version": "4.38.2"
7
  }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cb97796f804cc7bfca00f34d7c5b1b8cafd87d0d7a41601f9a7d46c6a0fd900
3
+ size 1324816544
tokenizer_config.json CHANGED
@@ -1,40 +1,30 @@
1
  {
2
  "add_bos_token": true,
3
  "add_prefix_space": false,
4
- "bos_token": {
5
- "__type": "AddedToken",
6
- "content": "</s>",
7
- "lstrip": false,
8
- "normalized": true,
9
- "rstrip": false,
10
- "single_word": false
 
 
 
 
 
 
 
 
 
 
11
  },
 
12
  "clean_up_tokenization_spaces": true,
13
- "eos_token": {
14
- "__type": "AddedToken",
15
- "content": "</s>",
16
- "lstrip": false,
17
- "normalized": true,
18
- "rstrip": false,
19
- "single_word": false
20
- },
21
  "errors": "replace",
22
  "model_max_length": 1000000000000000019884624838656,
23
- "pad_token": {
24
- "__type": "AddedToken",
25
- "content": "<pad>",
26
- "lstrip": false,
27
- "normalized": true,
28
- "rstrip": false,
29
- "single_word": false
30
- },
31
  "tokenizer_class": "GPT2Tokenizer",
32
- "unk_token": {
33
- "__type": "AddedToken",
34
- "content": "</s>",
35
- "lstrip": false,
36
- "normalized": true,
37
- "rstrip": false,
38
- "single_word": false
39
- }
40
  }
 
1
  {
2
  "add_bos_token": true,
3
  "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "1": {
6
+ "content": "<pad>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "2": {
14
+ "content": "</s>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ }
21
  },
22
+ "bos_token": "</s>",
23
  "clean_up_tokenization_spaces": true,
24
+ "eos_token": "</s>",
 
 
 
 
 
 
 
25
  "errors": "replace",
26
  "model_max_length": 1000000000000000019884624838656,
27
+ "pad_token": "</s>",
 
 
 
 
 
 
 
28
  "tokenizer_class": "GPT2Tokenizer",
29
+ "unk_token": "</s>"
 
 
 
 
 
 
 
30
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3497f6cf8f68cda25332639a2e2495ee4edce511ddd7f1394aab52b54978dd7c
3
- size 3887
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eec0bdf0c45022371a90dd8db539f779f0f5a0d6f3519058c17932f176537f84
3
+ size 4856