Text Generation
Transformers
Safetensors
GGUF
English
qwen2
quantum-ml
hybrid-quantum-classical
quantum-kernel
research
quantum-computing
nisq
qiskit
quantum-circuits
vibe-thinker
physics-inspired-ml
quantum-enhanced
hybrid-ai
1.5b
small-model
efficient-ai
reasoning
chemistry
physics
text-generation-inference
conversational
Update README.md
Browse files
README.md
CHANGED
|
@@ -150,20 +150,11 @@ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
|
| 150 |
| gpt-oss-120B (high) | 1.1% |
|
| 151 |
| Claude 4.5 Sonnet | 1.1% |
|
| 152 |
|
| 153 |
-
### Quantum
|
| 154 |
**Sentiment Analysis Task:**
|
| 155 |
|
| 156 |
-
| Approach | Accuracy | Notes |
|
| 157 |
-
|----------|----------|-------|
|
| 158 |
-
| Classical (Linear SVM) | 100% | Traditional baseline |
|
| 159 |
-
| Chronos-1.5B (quantum kernel) | 75% | NISQ hardware noise impact |
|
| 160 |
-
|
| 161 |

|
| 162 |
|
| 163 |
-
**Why the gap?**
|
| 164 |
-
|
| 165 |
-
The 25% accuracy difference is entirely due to NISQ (Noisy Intermediate-Scale Quantum) gate errors (~1% per operation) accumulating through the quantum circuit. This is a **hardware limitation**, not an algorithmic issue.
|
| 166 |
-
|
| 167 |
**Key insight:** The quantum kernel shows learned structure (see left graph above), but current quantum hardware noise corrupts similarity computations. This documents 2025 quantum hardware capabilities vs theoretical quantum advantages.
|
| 168 |
|
| 169 |
|
|
|
|
| 150 |
| gpt-oss-120B (high) | 1.1% |
|
| 151 |
| Claude 4.5 Sonnet | 1.1% |
|
| 152 |
|
| 153 |
+
### Quantum Kernel Integration Results
|
| 154 |
**Sentiment Analysis Task:**
|
| 155 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 156 |

|
| 157 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 158 |
**Key insight:** The quantum kernel shows learned structure (see left graph above), but current quantum hardware noise corrupts similarity computations. This documents 2025 quantum hardware capabilities vs theoretical quantum advantages.
|
| 159 |
|
| 160 |
|