Trained polyglot 5.8B with the QLORA method using the Chatbot_data_for_Korean dataset.
The hyper-parameters used for training are as follows.
- batch-size: 16
- max_steps: 3000
- Learning rate: 3e-4
- Lora r: 8
- Lora target modules: query_key_value
Prompt Template:
### ์ง๋ฌธ: {๋ฌธ์ฅ}
### ์๋ต: {๋ฌธ์ฅ}
### ์ ํ: {์ผ๋ฐ ๋๋ ์ฐ์ }
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support