Text Generation
Transformers
Safetensors
GGUF
English
qwen2
quantum-ml
hybrid-quantum-classical
quantum-kernel
research
quantum-computing
nisq
qiskit
quantum-circuits
vibe-thinker
physics-inspired-ml
quantum-enhanced
hybrid-ai
1.5b
small-model
efficient-ai
reasoning
chemistry
physics
text-generation-inference
conversational
Update README.md
Browse files
README.md
CHANGED
|
@@ -31,7 +31,7 @@ base_model:
|
|
| 31 |
|
| 32 |
- **VibeThinker-1.5B** as the base transformer model for embedding extraction
|
| 33 |
- **Quantum Kernel Methods** for similarity computation
|
| 34 |
-
- **
|
| 35 |
|
| 36 |
This model demonstrates a proof-of-concept for hybrid quantum-classical machine learning applied to sentiment analysis.
|
| 37 |
|
|
@@ -44,7 +44,7 @@ Chronos 1.5B supports multiple quantum kernel execution modes:
|
|
| 44 |
| Feature | Implementation |
|
| 45 |
|------------------------------------|---------------------------------------------------------------------------------|
|
| 46 |
| Real quantum training | Quantum rotation angles were optimized on IBM **Heron r2** (`ibm_fez`) in 2025 |
|
| 47 |
-
| Saved quantum parameters | `quantum_kernel.pkl` — trained
|
| 48 |
| Quantum circuit definition | Available in `k_train_quantum.npy` / `k_test_quantum.npy` (future use) |
|
| 49 |
| Current inference | Classical simulation using the trained quantum angles (via cosine similarity) |
|
| 50 |
| True quantum execution (optional) | Possible by loading `quantum_kernel.pkl` + circuit files and running on IBM Quantum (example scripts will be added) |
|
|
@@ -61,7 +61,7 @@ Chronos 1.5B supports multiple quantum kernel execution modes:
|
|
| 61 |
- **Parameters**: ~1.5B
|
| 62 |
- **Context Length**: 131,072 tokens
|
| 63 |
- **Embedding Dimension**: 1536
|
| 64 |
-
- **Quantum Component**:
|
| 65 |
- **Training Data**: 8 sentiment examples (demonstration)
|
| 66 |
|
| 67 |
## Performance
|
|
|
|
| 31 |
|
| 32 |
- **VibeThinker-1.5B** as the base transformer model for embedding extraction
|
| 33 |
- **Quantum Kernel Methods** for similarity computation
|
| 34 |
+
- **2-qubit quantum circuits** for enhanced feature space representation
|
| 35 |
|
| 36 |
This model demonstrates a proof-of-concept for hybrid quantum-classical machine learning applied to sentiment analysis.
|
| 37 |
|
|
|
|
| 44 |
| Feature | Implementation |
|
| 45 |
|------------------------------------|---------------------------------------------------------------------------------|
|
| 46 |
| Real quantum training | Quantum rotation angles were optimized on IBM **Heron r2** (`ibm_fez`) in 2025 |
|
| 47 |
+
| Saved quantum parameters | `quantum_kernel.pkl` — trained 2-qubit gate angles (pickle) |
|
| 48 |
| Quantum circuit definition | Available in `k_train_quantum.npy` / `k_test_quantum.npy` (future use) |
|
| 49 |
| Current inference | Classical simulation using the trained quantum angles (via cosine similarity) |
|
| 50 |
| True quantum execution (optional) | Possible by loading `quantum_kernel.pkl` + circuit files and running on IBM Quantum (example scripts will be added) |
|
|
|
|
| 61 |
- **Parameters**: ~1.5B
|
| 62 |
- **Context Length**: 131,072 tokens
|
| 63 |
- **Embedding Dimension**: 1536
|
| 64 |
+
- **Quantum Component**: 2-qubit kernel
|
| 65 |
- **Training Data**: 8 sentiment examples (demonstration)
|
| 66 |
|
| 67 |
## Performance
|