DualMind

Single Architecture, Dual Cognition β€” The Multi-Model Collision Array on Shared Weights

Convergent Intelligence LLC: Research Division


What This Is

DualMind is a 1.7B parameter model that implements dual-mental-modality reasoning β€” a single model with two internal voices sharing the same weights, differentiated only by role tokens:

  • <explore> β€” Unconstrained reasoning. Derivation, speculation, working through the problem freely.
  • <examine> β€” Adversarial self-response. The model reads its own explore output and critiques it. Error detection, verification, refinement.
  • <response> β€” Clean synthesis. The final answer distilled from the internal dialogue.

This is the multi-model collision array collapsed into a single architecture. The dialectical structure that produces novel insights from architectural diversity (demonstrated in our five-architecture collision experiments) is recreated through role-conditioned generation on shared weights.

Architecture

Parameter Value
Architecture Qwen3ForCausalLM
Parameters ~2.03B (1.7B effective)
Hidden Size 2048
Layers 28
Attention Heads 16 (Q) / 8 (KV) β€” GQA
Context Length 40,960 tokens
Precision BF16 (trained on H100)

Training

Base model: Disctil-Qwen3-1.7B (DISC-refined uncensored Qwen3)

Dataset: KK04/LogicInference_OA β€” Logical inference problems transformed into the DualMind cognitive loop format.

Training format: Each CoT solution is restructured into the DualMind format:

  • Derivation sentences β†’ <explore> block (reasoning phase)
  • Verification/checking sentences β†’ <examine> block (self-critique phase)
  • Final answer β†’ <response> block (synthesis)

Sentence-level splitting uses trigger detection (check, verify, however, but wait, etc.) to find the natural transition from reasoning to verification, with 70/30 positional fallback.

Hardware: Colab H100, BF16 precision. 512 steps, lr 5e-6, SFT via TRL.

Next iteration: Currently training on Crownelius/Opus-4.6-Reasoning-3300x β€” 2,160 Claude Opus 4.6 reasoning samples with pre-separated thinking/solution columns, eliminating the need for heuristic splitting.

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained(
    "reaperdoesntknow/DualMind",
    torch_dtype="auto",
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("reaperdoesntknow/DualMind")

# Start the explore block β€” the model completes the full loop
prompt = (
    "##USER:\n"
    "Prove that the sum of two even numbers is always even.\n\n"
    "<explore>\n"
)

inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
output = model.generate(
    **inputs,
    max_new_tokens=1024,
    do_sample=True,
    top_p=0.9,
    temperature=0.6,
    repetition_penalty=1.15,
)
result = tokenizer.decode(output[0], skip_special_tokens=True)
print(result)

Expected Output Structure

<explore>
[The model works through the proof freely β€” definitions, algebraic manipulation, etc.]
</explore>

<examine>
[The model critiques its own derivation β€” checks for gaps, verifies steps, catches errors]
</examine>

<response>
[Clean final answer synthesized from the internal dialogue]
</response>

Why Dual Modality

Standard CoT prompting produces a single stream of reasoning. The model has one shot to get it right. DualMind gives the model a structural mechanism for self-correction:

  1. Explore is free to make mistakes, speculate, and try approaches that might not work
  2. Examine reads the explore output adversarially β€” it's looking for errors, not confirming correctness
  3. Response has the benefit of both perspectives

This mirrors what happens in multi-model collision arrays where different architectures produce genuinely different failure modes, and the collision between them surfaces structure that neither achieves alone. DualMind recreates this dynamic within a single set of weights through role conditioning.

Distillation Chain

Qwen3-1.7B (base)
  β†’ DiStil-Qwen3-1.7B-uncensored (uncensored SFT)
    β†’ Disctil-Qwen3-1.7B (DISC refinement)
      β†’ DualMind (DualMind SFT on Opus 4.6 reasoning data) ← you are here

Related Models

Model Description Downloads
TopologicalQwen TKD + DualMind on physics CoT 622
Disctil-Qwen3-1.7B Parent model (DISC-refined) 286
Qwen3-1.7B-Thinking-Distil TKD with Thinking teacher 687

DualMind Collection β€” Dual-cognition model series

DistilQwen Collection β€” Full proof-weighted distillation series

Full methodology: Structure Over Scale (DOI: 10.57967/hf/8165)

Citation

@misc{colca2026dualmind,
  title={DualMind: Dual-Mental-Modality Reasoning via Role-Conditioned Self-Critique},
  author={Colca, Roy S.},
  year={2026},
  publisher={HuggingFace},
  url={https://huggingface.co/reaperdoesntknow/DualMind},
  note={Convergent Intelligence LLC: Research Division}
}

Convergent Intelligence LLC: Research Division "Where classical analysis fails to see, we begin."

Downloads last month
-
Safetensors
Model size
2B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Collection including reaperdoesntknow/DualMinded-Qwen3-1.7B