Miner-8B

This repository hosts the Hugging Face Transformers checkpoint for MINER: Mining Intrinsic Mastery for Data-Efficient RL in Large Reasoning Models.

Model Description

Miner-8B is a reasoning model trained with MINER, a reinforcement learning method designed to improve data efficiency for large reasoning models. MINER targets the inefficiency of critic-free RL methods on positive homogeneous prompts, where all sampled rollouts are correct and standard relative-advantage training provides little or no learning signal. Instead, MINER leverages the policy’s intrinsic uncertainty as a self-supervised reward signal, without requiring auxiliary reward models or additional inference-time overhead. :contentReference[oaicite:1]{index=1}

The MINER framework introduces two central ideas:

  1. Token-level focal credit assignment, which amplifies learning on uncertain and critical tokens while suppressing overconfident ones.
  2. Adaptive advantage calibration, which integrates intrinsic and verifiable rewards in a stable way. :contentReference[oaicite:2]{index=2}

According to the paper, MINER is evaluated on six reasoning benchmarks using Qwen3-8B-Base and Qwen3-8B-Base, and reports stronger sample efficiency and accuracy than several baseline methods including GRPO variants. :contentReference[oaicite:3]{index=3}

Intended Use

This model is intended for research and experimental use in:

  • reasoning and problem solving
  • reinforcement learning for language models
  • mathematical and verifiable reasoning tasks
  • post-training and evaluation of large reasoning models

Potential use cases include:

  • academic research on RL for reasoning models
  • evaluation on reasoning benchmarks
  • ablation and reproduction studies based on the MINER framework
  • further finetuning or post-training from this checkpoint

How to Use

Transformers

from transformers import AutoTokenizer, AutoModelForCausalLM

model_name = "pixas/Miner-8B"

tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype="auto",
    device_map="auto"
)

prompt = [{"role": "user", "content": "What is 2+3?"}]
inputs = tokenizer(tokenizer.apply_chat_template(prompt, add_generation_prompt=True, tokenize=False), return_tensors='pt').to(model.device)

outputs = model.generate(
    **inputs,
    max_new_tokens=8192,
    do_sample=True
)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))

vLLM

from vllm import LLM, SamplingParams

llm = LLM(model="pixas/Miner-8B")
sampling_params = SamplingParams(
    temperature=0.6,
    max_tokens=8192
)
prompt = [{"role": "user", "content": "What is 2+3?"}]
inputs = tokenizer.apply_chat_template(prompt, add_generation_prompt=True, tokenize=False)
outputs = llm.generate(
    inputs,
    sampling_params
)

print(outputs[0].outputs[0].text)

Limitations

This model is a research checkpoint and may have several limitations:

  • It may produce incorrect, incomplete, or overconfident reasoning outputs.
  • Performance may depend heavily on prompt format and decoding setup.
  • Results reported in the paper may not transfer exactly to this released checkpoint unless the same base model, data mixture, and evaluation pipeline are used.
  • The model is not intended as a substitute for expert judgment in high-stakes domains.

Bias, Risks, and Safety

Like other large language models, this model may reflect biases present in its training data and may generate harmful, misleading, or factually incorrect outputs. Additional care is required before deployment in user-facing or safety-critical applications.

Citation

If you use this model, please cite:

@article{jiang2026miner,
  title={Miner: Mining Intrinsic Mastery for Data-Efficient RL in Large Reasoning Models},
  author={Jiang, Shuyang and Wang, Yuhao and Zhang, Ya and Wang, Yanfeng and Wang, Yu},
  journal={arXiv preprint arXiv:2601.04731},
  year={2026}
}

Acknowledgements

This model card is based on the official MINER paper and code repository:

Downloads last month
174
Safetensors
Model size
8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for pixas/Miner-8B

Finetuned
(366)
this model
Quantizations
2 models

Dataset used to train pixas/Miner-8B

Paper for pixas/Miner-8B