Model Card for Model ID

PokerBench Qwen3-8B LoRA Adapter

This repository contains a LoRA adapter for Qwen3-8B, fine-tuned on the PokerBench dataset for poker-related text generation and reasoning.

Model Details

Model Description

This is a poker-domain PEFT / LoRA adapter trained on top of Qwen3-8B using supervised fine-tuning (SFT).
It is intended to improve poker-related responses, including strategy discussion, hand analysis, poker terminology, and decision reasoning.

  • Model type: LoRA adapter for a causal language model
  • Base model: Qwen/Qwen3-8B
  • Adapter format: PEFT adapter weights only
  • Fine-tuning method: LoRA / SFT
  • Libraries: Unsloth, PEFT, TRL, Transformers

Uses

Direct Use

This adapter is intended for poker-related generation tasks such as:

  • Poker strategy explanation
  • Hand and range analysis
  • Pot odds and action reasoning
  • Poker-focused assistant behavior

Downstream Use

This adapter can be used for:

  • A poker-specialized assistant
  • Domain adaptation experiments
  • Further merging or adapter research workflows

Out-of-Scope Use

This adapter is not intended for:

  • Legal, medical, or financial advice
  • High-stakes decision-making
  • Guaranteed optimal poker play
  • General-purpose QA outside the poker domain

Bias, Risks, and Limitations

Because this is a domain-specific adapter, it may:

  • Overemphasize poker interpretations
  • Produce overconfident but incorrect advice
  • Reflect limitations or biases in PokerBench
  • Underperform on non-poker tasks relative to the base model

Users should treat outputs as experimental and verify important conclusions independently.

How to Get Started with the Model

This repository contains adapter weights only.
Load the base model first, then attach the adapter.

from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel
import torch

base_model_name = "Qwen/Qwen3-8B"
adapter_name = "your-username/your-adapter-repo"

tokenizer = AutoTokenizer.from_pretrained(base_model_name)

base_model = AutoModelForCausalLM.from_pretrained(
    base_model_name,
    torch_dtype=torch.float16,
    device_map="auto",
)

model = PeftModel.from_pretrained(base_model, adapter_name)

prompt = "You are on the button with AKo facing an open raise. What factors matter most?"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)

outputs = model.generate(**inputs, max_new_tokens=128)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Dominicvdb/pokerqwen

Finetuned
Qwen/Qwen3-8B
Adapter
(898)
this model