Update README.md
#2
by
DavidLee12356
- opened
README.md
CHANGED
|
@@ -53,8 +53,8 @@ This represents approximately a 39× reduction in pretraining cost relative to `
|
|
| 53 |
|
| 54 |
```python
|
| 55 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 56 |
-
model = AutoModelForCausalLM.from_pretrained("/
|
| 57 |
-
tokenizer = AutoTokenizer.from_pretrained("/
|
| 58 |
|
| 59 |
chat = [
|
| 60 |
{"role": "tool_list", "content": ""},
|
|
|
|
| 53 |
|
| 54 |
```python
|
| 55 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 56 |
+
model = AutoModelForCausalLM.from_pretrained("naver-hyperclovax/HyperCLOVAX-SEED-Text-Instruct-0.5B")
|
| 57 |
+
tokenizer = AutoTokenizer.from_pretrained("naver-hyperclovax/HyperCLOVAX-SEED-Text-Instruct-0.5B")
|
| 58 |
|
| 59 |
chat = [
|
| 60 |
{"role": "tool_list", "content": ""},
|