Model Card for Model ID

PokerBench Qwen3-8B LoRA Adapter

This repository contains a LoRA adapter for Qwen3-8B, fine-tuned on the PokerBench dataset for poker-related text generation and reasoning.

Model Details

Model Description

This is a poker-domain PEFT / LoRA adapter trained on top of Qwen3-8B using supervised fine-tuning (SFT).
It is intended to improve poker-related responses, including strategy discussion, hand analysis, poker terminology, and decision reasoning.

  • Model type: LoRA adapter for a causal language model
  • Base model: Qwen/Qwen3-8B
  • Adapter format: PEFT adapter weights only
  • Fine-tuning method: LoRA / SFT
  • Libraries: Unsloth, PEFT, TRL, Transformers

How to Get Started with the Model

This repository contains adapter weights only.
Load the base model first, then attach the adapter.

from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel
import torch

base_model_name = "Qwen/Qwen3-8B"
adapter_name = "your-username/your-adapter-repo"

tokenizer = AutoTokenizer.from_pretrained(base_model_name)

base_model = AutoModelForCausalLM.from_pretrained(
    base_model_name,
    torch_dtype=torch.float16,
    device_map="auto",
)

model = PeftModel.from_pretrained(base_model, adapter_name)

prompt = "You are on the button with AKo facing an open raise. What factors matter most?"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)

outputs = model.generate(**inputs, max_new_tokens=128)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
1
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Dominicvdb/pokerqwen

Finetuned
Qwen/Qwen3-8B
Adapter
(1226)
this model