Uploaded model
- Developed by: 0208suin
- License: apache-2.0
- Finetuned from model : unsloth/qwen3-14b-unsloth-bnb-4bit
This qwen3 model was trained 2x faster with Unsloth and Huggingface's TRL library.
π§ Qwen3-14B (LoRA Fine-tuned by 0208suin)
μ΄ λͺ¨λΈμ
unsloth/qwen3-14b-unsloth-bnb-4bitλ₯Ό κΈ°λ°μΌλ‘, Unsloth + TRL(Transformers Reinforcement Learning) κΈ°λ° νμΈνλμ ν΅ν΄ νκ΅μ΄ μΌκΈ°/νλ κΈ°λ‘ ν μ€νΈλ‘λΆν° ν΅μ¬ ν€μλ(5~10κ°)λ₯Ό μΆμΆν μ μλλ‘ νμ΅λ λͺ¨λΈμ λλ€.
π§© λͺ¨λΈ κ°μ
κΈ°λ° λͺ¨λΈ:
unsloth/qwen3-14b-unsloth-bnb-4bitνμΈνλ λ°©μ: PEFT + LoRA + TRL (DPO/GRPO μ€λΉ κ°λ₯)
μ§μ μΈμ΄: νκ΅μ΄ (μΆλ‘ μ€μ¬, νκ΅μ΄ μ΅μ ν ν둬ννΈ μ¬μ©)
μ μ© λΆμΌ:
- μΌκΈ° λ° μμΈμ΄ μλ λΆμ
- κ°μ± μμ½ λ° ν€μλ νκΉ
- κ΅μ‘/μλ΄/νλλΆμ λ± μΈκ° μ€μ¬ ν μ€νΈ μ΄ν΄
π¦ μ¬μ© μμ (Inference)
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
import torch
model_id = "0208suin/qwen3-14b-keyword-lora" # μ€μ λͺ¨λΈ κ²½λ‘λ‘ λ³κ²½
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", trust_remote_code=True)
messages = [
{
"role": "user",
"content": (
"λ€μ μΌκΈ°λ₯Ό μ½κ³ ν΅μ¬ ν€μλλ₯Ό 5κ° μ΄μ 10κ° μ΄νλ‘ λ
Όλ¦¬μ μΌλ‘ λμΆν΄μ€:\n\n"
"μ€λμ μ€λλ§μ νΌμ νκ°μ κ°λ€. λ²κ½μ΄ νλλ¬μ§κ² νΌμ΄ μμ΄μ μ°μ±
νλ λ΄λ΄ κΈ°λΆμ΄ μ’μλ€. "
"μ΄μ΄ν°μΌλ‘ μ’μνλ μμ
μ λ€μΌλ©΄μ κ±·λ€ λ³΄λ, μμλ μ€νΈλ μ€κ° μ‘°κΈμ ν리λ κ² κ°μλ€. "
"λ μ¨λ λ°λ»νκ³ λ°λλ μ μ ν΄μ, μ€νλ € λ§μμ΄ μ°¨λΆν΄μ‘λ€. "
"νΌμλ§μ μκ°μ΄ μ΄λ κ² μμ€νκ² λκ»΄μ§ κ±΄ μ λ§ μ€λλ§μ΄μλ€."
)
}
]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
streamer = TextStreamer(tokenizer, skip_prompt=True)
model.generate(
**tokenizer(text, return_tensors="pt").to(model.device),
max_new_tokens=256,
temperature=0.6,
top_p=0.95,
top_k=20,
streamer=streamer
)
π§ νμ΅ λ°μ΄ν° κ°μ
μμ² λ°μ΄ν°: νκ΅μ΄ μΌκΈ° λ° νλκ΄μ°° λ³΄κ³ μ ν μ€νΈ 11,000건
λΌλ²¨λ§: μ λ¬Έκ°κ° μμμ μΌλ‘ μΆμΆν ν€μλ (5~10κ°)
λ κ°μ§ λͺ¨λλ‘ νμ΅λ¨:
- β μΌλ° ν€μλ μΆμΆ (Instruction κΈ°λ°)
- β λ Όλ¦¬ κΈ°λ° ν€μλ μμ± (Reasoning λ°©μ)
π νλ ¨ λ°©μ
| νλͺ© | λ΄μ© |
|---|---|
| νλ μμν¬ | Unsloth |
| λͺ¨λΈ ν¬κΈ° | Qwen3-14B (bnb-4bit) |
| νμ΅ λ°©μ | LoRA (adapter model only) |
| 컨ν μ€νΈ κΈΈμ΄ | μ΅λ 32K tokens |
| μ¬μ© λΌμ΄λΈλ¬λ¦¬ | Hugging Face Transformers + TRL + PEFT |
π λΌμ΄μ μ€
- λ³Έ λͺ¨λΈμ Apache 2.0 λΌμ΄μ μ€λ₯Ό λ°λ¦ λλ€. μμ μ μ΄μ©, μμ λ° μ¬λ°°ν¬κ° κ°λ₯ν©λλ€.
- Qwen3 λ° Unsloth κΈ°λ° λͺ¨λΈ μμ Apache 2.0 νμ μ¬μ©λμμ΅λλ€.
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support
