Uploaded model

  • Developed by: 0208suin
  • License: apache-2.0
  • Finetuned from model : unsloth/qwen3-14b-unsloth-bnb-4bit

This qwen3 model was trained 2x faster with Unsloth and Huggingface's TRL library.


🧠 Qwen3-14B (LoRA Fine-tuned by 0208suin)

이 λͺ¨λΈμ€ unsloth/qwen3-14b-unsloth-bnb-4bitλ₯Ό 기반으둜, Unsloth + TRL(Transformers Reinforcement Learning) 기반 νŒŒμΈνŠœλ‹μ„ 톡해 ν•œκ΅­μ–΄ 일기/행동 기둝 ν…μŠ€νŠΈλ‘œλΆ€ν„° 핡심 ν‚€μ›Œλ“œ(5~10개)λ₯Ό μΆ”μΆœν•  수 μžˆλ„λ‘ ν•™μŠ΅λœ λͺ¨λΈμž…λ‹ˆλ‹€.


🧩 λͺ¨λΈ κ°œμš”

  • 기반 λͺ¨λΈ: unsloth/qwen3-14b-unsloth-bnb-4bit

  • νŒŒμΈνŠœλ‹ 방식: PEFT + LoRA + TRL (DPO/GRPO μ€€λΉ„ κ°€λŠ₯)

  • 지원 μ–Έμ–΄: ν•œκ΅­μ–΄ (μΆ”λ‘  쀑심, ν•œκ΅­μ–΄ μ΅œμ ν™” ν”„λ‘¬ν”„νŠΈ μ‚¬μš©)

  • 적용 λΆ„μ•Ό:

    • 일기 및 에세이 μžλ™ 뢄석
    • 감성 μš”μ•½ 및 ν‚€μ›Œλ“œ νƒœκΉ…
    • ꡐ윑/상담/행동뢄석 λ“± 인간 쀑심 ν…μŠ€νŠΈ 이해

πŸ“¦ μ‚¬μš© μ˜ˆμ‹œ (Inference)

from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
import torch

model_id = "0208suin/qwen3-14b-keyword-lora"  # μ‹€μ œ λͺ¨λΈ 경둜둜 λ³€κ²½
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", trust_remote_code=True)

messages = [
    {
        "role": "user",
        "content": (
            "λ‹€μŒ 일기λ₯Ό 읽고 핡심 ν‚€μ›Œλ“œλ₯Ό 5개 이상 10개 μ΄ν•˜λ‘œ λ…Όλ¦¬μ μœΌλ‘œ λ„μΆœν•΄μ€˜:\n\n"
            "μ˜€λŠ˜μ€ μ˜€λžœλ§Œμ— 혼자 ν•œκ°•μ— κ°”λ‹€. λ²šκ½ƒμ΄ νλ“œλŸ¬μ§€κ²Œ ν”Όμ–΄ μžˆμ–΄μ„œ μ‚°μ±…ν•˜λŠ” λ‚΄λ‚΄ 기뢄이 μ’‹μ•˜λ‹€. "
            "μ΄μ–΄ν°μœΌλ‘œ μ’‹μ•„ν•˜λŠ” μŒμ•…μ„ λ“€μœΌλ©΄μ„œ κ±·λ‹€ λ³΄λ‹ˆ, μŒ“μ˜€λ˜ μŠ€νŠΈλ ˆμŠ€κ°€ μ‘°κΈˆμ€ ν’€λ¦¬λŠ” 것 κ°™μ•˜λ‹€. "
            "날씨도 λ”°λœ»ν•˜κ³  λ°”λžŒλ„ μ„ μ„ ν•΄μ„œ, 였히렀 마음이 μ°¨λΆ„ν•΄μ‘Œλ‹€. "
            "혼자만의 μ‹œκ°„μ΄ μ΄λ ‡κ²Œ μ†Œμ€‘ν•˜κ²Œ λŠκ»΄μ§„ 건 정말 μ˜€λžœλ§Œμ΄μ—ˆλ‹€."
        )
    }
]

text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
streamer = TextStreamer(tokenizer, skip_prompt=True)

model.generate(
    **tokenizer(text, return_tensors="pt").to(model.device),
    max_new_tokens=256,
    temperature=0.6,
    top_p=0.95,
    top_k=20,
    streamer=streamer
)

πŸ”§ ν•™μŠ΅ 데이터 κ°œμš”

  • μ›μ²œ 데이터: ν•œκ΅­μ–΄ 일기 및 행동관찰 λ³΄κ³ μ„œ ν…μŠ€νŠΈ 11,000건

  • 라벨링: μ „λ¬Έκ°€κ°€ μˆ˜μž‘μ—…μœΌλ‘œ μΆ”μΆœν•œ ν‚€μ›Œλ“œ (5~10개)

  • 두 κ°€μ§€ λͺ¨λ“œλ‘œ ν•™μŠ΅λ¨:

    • βœ… 일반 ν‚€μ›Œλ“œ μΆ”μΆœ (Instruction 기반)
    • βœ… 논리 기반 ν‚€μ›Œλ“œ 생성 (Reasoning 방식)

πŸš€ ν›ˆλ ¨ 방식

ν•­λͺ© λ‚΄μš©
ν”„λ ˆμž„μ›Œν¬ Unsloth
λͺ¨λΈ 크기 Qwen3-14B (bnb-4bit)
ν•™μŠ΅ 방식 LoRA (adapter model only)
μ»¨ν…μŠ€νŠΈ 길이 μ΅œλŒ€ 32K tokens
μ‚¬μš© 라이브러리 Hugging Face Transformers + TRL + PEFT

πŸ“œ λΌμ΄μ„ μŠ€

  • λ³Έ λͺ¨λΈμ€ Apache 2.0 λΌμ΄μ„ μŠ€λ₯Ό λ”°λ¦…λ‹ˆλ‹€. 상업적 이용, μˆ˜μ • 및 μž¬λ°°ν¬κ°€ κ°€λŠ₯ν•©λ‹ˆλ‹€.
  • Qwen3 및 Unsloth 기반 λͺ¨λΈ μ—­μ‹œ Apache 2.0 ν•˜μ— μ‚¬μš©λ˜μ—ˆμŠ΅λ‹ˆλ‹€.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support