Update readme.md

#1
by nmj21c - opened
Files changed (1) hide show
  1. README.md +54 -1
README.md CHANGED
@@ -2,4 +2,57 @@
2
  license: apache-2.0
3
  language:
4
  - ko
5
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  license: apache-2.0
3
  language:
4
  - ko
5
+ ---
6
+
7
+ # gemma-7b non IT ๋ฒ„์ „ ์ฑ„ํŒ… ํŒŒ์ธ ํŠœ๋‹๋œ ๋ฒ„์ „
8
+
9
+ ## version history
10
+ - 0.1 : 2024-04-05
11
+
12
+ ## ํŠธ๋ ˆ์ด๋‹ ์ •๋ณด
13
+ - Dataset : maywell/koVast
14
+ - GPU : RTX 3090 24G x 1
15
+ - optimizer : adamw_torch
16
+ - learning rate scheduler type : cosine
17
+ - trai hour : 140 hour
18
+ - num of epoch : 1
19
+ - train loss : 0.8991
20
+ - eval loss : 0.7305
21
+
22
+ ## ์‚ฌ์šฉ๋ฒ•
23
+ ```
24
+ from transformers import AutoTokenizer, AutoModelForCausalLM
25
+ import transformers
26
+ import torch
27
+
28
+ checkpoint = "nmj21c/gemma-7b-andj-sft"
29
+ dtype = torch.bfloat16
30
+ model = AutoModelForCausalLM.from_pretrained(checkpoint, attn_implementation="flash_attention_2", device_map={"": 0}, torch_dtype=dtype)
31
+
32
+ toknizer_checkpoint = "philschmid/gemma-tokenizer-chatml"
33
+ tokenizer = AutoTokenizer.from_pretrained(toknizer_checkpoint)
34
+
35
+ chat = [
36
+ {"role": "system", "content": ""},
37
+ {"role": "user", "content": "์„œ์šธ์˜ ๊ฐ•๋‚จ์—ญ์—์„œ ๋ง›์ง‘ ์ถ”์ฒœํ•ด์ค˜"},
38
+ ]
39
+
40
+ prompt = tokenizer.apply_chat_template(chat, tokenize=False, add_generation_prompt=True)
41
+
42
+ eos_token_str = "<|im_end|>"
43
+ eos_token = tokenizer(eos_token_str,add_special_tokens=False)["input_ids"][0]
44
+
45
+ inputs = tokenizer.encode(prompt, add_special_tokens=False, return_tensors="pt").to("cuda:0")
46
+ outputs = model.generate(
47
+ input_ids=inputs.to(model.device),
48
+ max_new_tokens=1024,
49
+ eos_token_id=eos_token,
50
+ do_sample=True,
51
+ temperature=0.7,
52
+ top_k=50,
53
+ top_p=0.95,
54
+ )
55
+
56
+ response = tokenizer.decode(outputs[0])[len(prompt):].strip().replace(eos_token_str, '')
57
+ print(response)
58
+ ```