nmj21c commited on
Commit
0a09ba5
Β·
verified Β·
1 Parent(s): c232d36

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -6
README.md CHANGED
@@ -6,20 +6,22 @@ language:
6
 
7
  # gemma-7b non IT 버전 μ±„νŒ… 파인 νŠœλ‹λœ 버전
8
 
9
- ## version history
10
- - 0.1 : 2024-04-05
 
 
11
 
12
  ## νŠΈλ ˆμ΄λ‹ 정보
13
- - Dataset : maywell/koVast
14
  - GPU : RTX 3090 24G x 1
15
  - optimizer : adamw_torch
16
  - lr scheduler type : cosine
17
- - trai hour : 140 hour
18
- - num of epoch : 1
19
  - train loss : 0.8991
20
  - eval loss : 0.7305
21
 
22
- ## μ‚¬μš©λ²•
23
  ```
24
  from transformers import AutoTokenizer, AutoModelForCausalLM
25
  import transformers
 
6
 
7
  # gemma-7b non IT 버전 μ±„νŒ… 파인 νŠœλ‹λœ 버전
8
 
9
+ κ°„λ‹¨ν•œ μ±„νŒ… ν˜•νƒœμ˜ λ°μ΄ν„°λ‘œ 파인 νŠœλ‹λœ λ²„μ „μž…λ‹ˆλ‹€.
10
+
11
+ ## history
12
+ - 0.1 : 2024-04-05 졜초 SFT버전 μ—…λ‘œλ“œ, DPOλŠ” κ³ λ―Ό 쀑
13
 
14
  ## νŠΈλ ˆμ΄λ‹ 정보
15
+ - μ‚¬μš©λ°μ΄ν„°μ…‹ : maywell/koVast 을 philschmid/gemma-tokenizer-chatml 에 맞게 λ³€μ‘°ν•˜μ—¬ μ‚¬μš©
16
  - GPU : RTX 3090 24G x 1
17
  - optimizer : adamw_torch
18
  - lr scheduler type : cosine
19
+ - νŠΈλ ˆμ΄λ‹ μ‹œκ°„ : 140μ‹œκ°„
20
+ - 에포크 : 1
21
  - train loss : 0.8991
22
  - eval loss : 0.7305
23
 
24
+ ## μ‚¬μš©λ²• (bfloat16, GPU λ©”λͺ¨λ¦¬ μ•½ 17κΈ°κ°€ ν•„μš”)
25
  ```
26
  from transformers import AutoTokenizer, AutoModelForCausalLM
27
  import transformers