🧠 Super-Qwen-7B (NeuroBridge Distilled)

Super-Qwen is a series of models fine-tuned via Asymmetric Distillation using the NeuroBridge framework.

This model is designed to run on resource-constrained edge hardware (like HFT nodes and private mesh networks) while maintaining high-fidelity architectural and coding reasoning.

πŸš€ Key Features

  • Personal Distillation: Designed to be fine-tuned on your own private coding traces.
  • Workflow Capture: The NeuroBridge framework supports self-supervised learning from your professional intellectual workflow.
  • Format: GGUF (Optimized for Ollama/Llama.cpp).

πŸ“Š Evaluation Results (Base + Distillation)

Benchmark Score
Coding 88%
Safety 100%
Hallucination 50%

πŸ›  Usage with NeuroBridge

To use this model in your local NeuroBridge instance:

  1. Download the .gguf adapter.
  2. Place it in your adapters/ directory.
  3. Run:
neuro train promote my_adapter.gguf

🧠 Philosophy: Asymmetric Scaling

We believe that edge intelligence should not be a "miniature" version of the cloud, but a specialized distilled version. Super-Qwen models are built to perform specific, high-value tasks with zero network dependency.


Developed by: seyominaoto Framework: NeuroBridge

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for Seyomi/super-qwen

Base model

Qwen/Qwen2.5-7B
Finetuned
(354)
this model