| --- |
| tags: |
| - text-generation |
| license: cc-by-nc-4.0 |
| language: |
| - ko |
| base_model: beomi/OPEN-SOLAR-KO-10.7B |
| pipeline_tag: text-generation |
| --- |
| |
| # **DataVortexS-10.7B-dpo-v1.7** |
|
|
| <img src="./DataVortex.png" alt="DataVortex" style="height: 8em;"> |
|
|
| ## Our Team |
|
|
| | Research & Engineering | Product Management | |
| | :--------------------: | :----------------: | |
| | Kwangseok Yang | Seunghyun Choi | |
| | Jeongwon Choi | Hyoseok Choi | |
|
|
| ## **Model Details** |
|
|
| ### **Base Model** |
|
|
| [beomi/OPEN-SOLAR-KO-10.7B](https://huggingface.co/beomi/OPEN-SOLAR-KO-10.7B) |
|
|
| ### **Trained On** |
|
|
| - **OS**: Ubuntu 22.04 |
| - **GPU**: H100 80GB 4ea |
| - **transformers**: v4.36.2 |
|
|
| ### **Instruction format** |
|
|
| It follows **Alpaca (Chat)** format. |
|
|
| E.g. |
|
|
| ```python |
| text = """\ |
| ### System: |
| λΉμ μ μ¬λλ€μ΄ μ 보λ₯Ό μ°Ύμ μ μλλ‘ λμμ£Όλ μΈκ³΅μ§λ₯ λΉμμ
λλ€. |
| |
| ### User: |
| λνλ―Όκ΅μ μλλ μ΄λμΌ? |
| |
| ### Assistant: |
| λνλ―Όκ΅μ μλλ μμΈμ
λλ€. |
| |
| ### User: |
| μμΈ μΈκ΅¬λ μ΄ λͺ λͺ
μ΄μΌ? |
| """ |
| ``` |
|
|
| ## **Model Benchmark** |
|
|
| ### **[Ko LM Eval Harness](https://github.com/Beomi/ko-lm-evaluation-harness)** |
|
|
| | Task | 0-shot | 5-shot | 10-shot | 50-shot | |
| | :--------------- | ----------: | -----------: | ----------: | ----------: | |
| | kobest_boolq | 0.886844 | 0.917351 | 0.921621 | 0.919476 | |
| | kobest_copa | 0.805514 | 0.856782 | 0.864773 | 0.87793 | |
| | kobest_hellaswag | 0.505714 | 0.512116 | 0.50608 | 0.512187 | |
| | kobest_sentineg | 0.357169 | 0.974811 | 0.987404 | 0.979848 | |
| | **Average** | **0.63881** | **0.815265** | **0.81997** | **0.82236** | |
|
|
| ### **[Ko-LLM-Leaderboard](https://huggingface.co/spaces/upstage/open-ko-llm-leaderboard)** |
|
|
| | Average | Ko-ARC | Ko-HellaSwag | Ko-MMLU | Ko-TruthfulQA | Ko-CommonGen V2 | |
| | ------: | -----: | -----------: | ------: | ------------: | --------------: | |
| | 55.81 | 55.55 | 63.39 | 51.57 | 48.23 | 60.33 | |
|
|
| ## **Implementation Code** |
|
|
| This model contains the chat_template instruction format. |
| You can use the code below. |
| |
| ```python |
| from transformers import AutoModelForCausalLM, AutoTokenizer |
| |
| device = "cuda" # the device to load the model onto |
| |
| model = AutoModelForCausalLM.from_pretrained("Edentns/DataVortexS-10.7B-dpo-v1.7") |
| tokenizer = AutoTokenizer.from_pretrained("Edentns/DataVortexS-10.7B-dpo-v1.7") |
| |
| messages = [ |
| {"role": "system", "content": "λΉμ μ μ¬λλ€μ΄ μ 보λ₯Ό μ°Ύμ μ μλλ‘ λμμ£Όλ μΈκ³΅μ§λ₯ λΉμμ
λλ€."}, |
| {"role": "user", "content": "λνλ―Όκ΅μ μλλ μ΄λμΌ?"}, |
| {"role": "assistant", "content": "λνλ―Όκ΅μ μλλ μμΈμ
λλ€."}, |
| {"role": "user", "content": "μμΈ μΈκ΅¬λ μ΄ λͺ λͺ
μ΄μΌ?"} |
| ] |
| |
| encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt") |
|
|
| model_inputs = encodeds.to(device) |
| model.to(device) |
| |
| generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True) |
| decoded = tokenizer.batch_decode(generated_ids) |
| print(decoded[0]) |
| ``` |
| |
| ## **License** |
| |
| This model is licensed under the [cc-by-nc-4.0](https://creativecommons.org/licenses/by-nc/4.0/). which allows others to share and adapt the model for non-commercial purposes. |
| |
| <div align="center"> |
| <a href="https://edentns.com/"> |
| <img src="./Logo.png" alt="Logo" style="height: 3em;"> |
| </a> |
| </div> |
| |