ssunggun2 commited on
Commit
98a8a83
·
verified ·
1 Parent(s): ce7ed2f

SKIML-ICL/Llama-3-8B-Inst_qa_squad-ans-v2_ckp-1500-3000

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -48,7 +48,7 @@ The following hyperparameters were used during training:
48
 
49
  ### Framework versions
50
 
51
- - Transformers 4.44.1
52
  - Pytorch 2.4.0
53
  - Datasets 2.19.1
54
- - Tokenizers 0.19.1
 
48
 
49
  ### Framework versions
50
 
51
+ - Transformers 4.45.1
52
  - Pytorch 2.4.0
53
  - Datasets 2.19.1
54
+ - Tokenizers 0.20.0
config.json CHANGED
@@ -24,7 +24,7 @@
24
  "rope_theta": 500000.0,
25
  "tie_word_embeddings": false,
26
  "torch_dtype": "bfloat16",
27
- "transformers_version": "4.44.1",
28
  "use_cache": false,
29
  "vocab_size": 128256
30
  }
 
24
  "rope_theta": 500000.0,
25
  "tie_word_embeddings": false,
26
  "torch_dtype": "bfloat16",
27
+ "transformers_version": "4.45.1",
28
  "use_cache": false,
29
  "vocab_size": 128256
30
  }
generation_config.json CHANGED
@@ -8,5 +8,5 @@
8
  "max_length": 4096,
9
  "temperature": 0.6,
10
  "top_p": 0.9,
11
- "transformers_version": "4.44.1"
12
  }
 
8
  "max_length": 4096,
9
  "temperature": 0.6,
10
  "top_p": 0.9,
11
+ "transformers_version": "4.45.1"
12
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bea690b172138e5cd78b7d7d0deac0850d543786b6ebb08e1c1ee051bb698129
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:912563dad4cfa4ebeb03eaacb057c0f4e0de52f942cbfd51326f1c633c89846a
3
  size 5496