QSBench commited on
Commit
82c3c62
·
verified ·
1 Parent(s): c8aba73

Update GUIDE.md

Browse files
Files changed (1) hide show
  1. GUIDE.md +96 -36
GUIDE.md CHANGED
@@ -1,62 +1,122 @@
1
  # 🌌 QSBench: Complete User Guide
2
 
3
- Welcome to **QSBench Analytics Hub**. This space is designed for exploring synthetic quantum datasets, training machine learning (ML) models, and evaluating the impact of quantum noise on expectation values.
 
4
 
5
  ---
6
 
7
- ## 📂 1. Dataset Architecture
8
 
9
- QSBench provides unified datasets for the *Quantum Machine Learning (QML)* task. In this demo, 4 core datasets are available:
10
 
11
- 1. **Core (Clean):** Base set of ideal simulations. No physical noise influence. Great starting point for testing neural network architectures.
12
- 2. **Depolarizing Noise:** Simulation of depolarizing noise (equal‑probability error on quantum gates).
13
- 3. **Amplitude Damping:** Simulation of amplitude damping (asymmetric energy loss process by qubits).
14
- 4. **Transpilation (10q):** Circuits optimised and compiled for a specific hardware topology.
15
 
16
- ### Circuit Families Covered
17
- Each set includes a balanced sample from the following circuit classes:
18
- * `QFT` Quantum Fourier Transform.
19
- * `HEA` — Hardware Efficient Ansatz (variational forms).
20
- * `RANDOM` Circuits with random gate placements.
21
- * `EFFICIENT` / `REAL_AMPLITUDES` Popular ansätze for hybrid quantum networks (VQA, QAOA).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
 
23
  ---
24
 
25
- ## 📊 2. Feature Description
26
 
27
- When you switch to the **ML Training** tab, the system automatically parses the CSV file and extracts numerical features. These metrics describe the structure and complexity of the quantum circuit.
 
28
 
29
- **Key structural metrics:**
30
- * **`n_qubits` & `depth`:** Physical size of the circuit. Depth determines the coherence time required for execution.
31
- * **`gate_entropy`:** Entropy of the gate distribution. Shows how uniformly or “chaotically” gates are distributed among qubits.
32
- * **`meyer_wallach`:** Meyer‑Wallach measure. A scalar describing the degree of global quantum entanglement in the final circuit state. A value close to 1 means maximal entanglement.
33
- * **`adjacency`:** Connectivity graph density. Shows how actively qubits interact with each other via two‑qubit gates (CX).
34
- * **Gate counters (`total_gates`, `cx_count`, `rx_count`, etc.):** Exact number of applied operations. Special attention should be paid to `cx_count`, because CNOT gates introduce the most noise on real hardware.
 
 
 
 
 
 
 
 
 
35
 
36
  ---
37
 
38
- ## 🎯 3. Target Variables
39
 
40
- In the basic experiment on the ML tab, the model is trained to predict the **Global Z‑axis Expectation Value** (`ideal_expval_Z_global`).
41
 
42
- * **What does this mean?** It is the averaged measurement outcome of the ideal quantum state (ranging from -1 to 1).
43
- * **Why are other targets hidden?** The dataset contains local expectation values (`ideal_expval_Z_q0`, `error_Z_global`, etc.). They are excluded from the list of available features to avoid data leakage when training the regressor.
44
 
45
  ---
46
 
47
- ## 🤖 4. How to Use the ML Training Module
48
 
49
- 1. **Select the dataset:** Choose one of the four packs (e.g., Amplitude Damping).
50
- 2. **Select metrics:** Pick the features on which the model will be trained. Recommended baseline set: *gate_entropy, meyer_wallach, depth, cx_count*.
51
- 3. **Training:** When you click the “Execute Baseline” button, the system splits the data (80% Train / 20% Test) and trains a **Random Forest Regressor** (100 trees, depth 10).
52
- 4. **Result analysis:**
53
- * **Parity Plot (Scatter):** Shows how accurately predictions align with the ideal diagonal line.
54
- * **Feature Importance:** Which metrics contributed most to the prediction (useful for understanding how topology affects the outcome).
55
- * **Residuals:** Distribution of prediction errors.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
56
 
57
  ---
58
 
59
  ## 🔗 5. Project Resources
60
- * 🤗 [**Hugging Face**](https://huggingface.co/QSBench)
61
- * 💻 [**GitHub**](https://github.com/QSBench)
62
- * 🌐 [**Project Website**](https://qsbench.github.io)
 
 
 
 
 
 
1
  # 🌌 QSBench: Complete User Guide
2
 
3
+ Welcome to the **QSBench Analytics Hub**.
4
+ This platform is designed to bridge the gap between quantum circuit topology and machine learning, allowing researchers to study how structural characteristics influence quantum simulation outcomes.
5
 
6
  ---
7
 
8
+ ## ⚠️ Important: Demo Dataset Notice
9
 
10
+ The datasets currently loaded in this hub are **v1.0.0-demo versions**.
11
 
12
+ - **Scale**: These are small *shards* (subsets) of the full QSBench library.
13
+ - **Accuracy**: Because the training data is limited in size, ML models trained here will show lower accuracy and higher variance compared to models trained on full-scale production datasets.
14
+ - **Purpose**: These sets are intended for **demonstration and prototyping** of analytical pipelines before moving to high-performance computing (HPC) environments.
 
15
 
16
+ ---
17
+
18
+ ## 📂 1. Dataset Architecture & Selection
19
+
20
+ QSBench provides high-fidelity simulation data for the Quantum Machine Learning (QML) community.
21
+ We provide four distinct environments to test how different noise models affect data:
22
+
23
+ ### Core (Clean)
24
+ Ideal state-vector simulations.
25
+ Used as a **"Golden Reference"** to understand the theoretical limits of a circuit's expressivity without physical interference.
26
+
27
+ ### Depolarizing Noise
28
+ Simulates the effect of qubits losing their state toward a maximally mixed state.
29
+ This is the standard **"white noise"** of quantum computing.
30
+
31
+ ### Amplitude Damping
32
+ Represents **T1 relaxation (energy loss)**.
33
+ This is an asymmetric noise model where qubits decay from ∣1⟩ to ∣0⟩, critical for studying superconducting hardware.
34
+
35
+ ### Transpilation (10q)
36
+ Circuits are mapped to a **hardware topology (heavy-hex or grid)**.
37
+ Used to study how SWAP gates and routing overhead affect final results.
38
 
39
  ---
40
 
41
+ ## 📊 2. Feature Engineering: Structural Metrics
42
 
43
+ Why do we extract these specific features?
44
+ In QML, the **structure ("shape") of a circuit directly impacts performance**.
45
 
46
+ - **gate_entropy**
47
+ Measures distribution of gates.
48
+ High entropy complex, less repetitive circuits harder for classical models to learn.
49
+
50
+ - **meyer_wallach**
51
+ Quantifies **global entanglement**.
52
+ Entanglement provides quantum advantage but increases sensitivity to noise.
53
+
54
+ - **adjacency**
55
+ Represents qubit interaction graph density.
56
+ High adjacency → faster information spread, but higher risk of cross-talk errors.
57
+
58
+ - **cx_count (Two-Qubit Gates)**
59
+ The most critical complexity metric.
60
+ On NISQ devices, CNOT gates are **10x–100x noisier** than single-qubit gates.
61
 
62
  ---
63
 
64
+ ## 🎯 3. Multi-Target Regression (The Bloch Vector)
65
 
66
+ Unlike traditional benchmarks that focus on a single observable, QSBench targets the **full global Bloch vector**:
67
 
68
+ [⟨X⟩global, ⟨Y⟩global, ⟨Z⟩global]
 
69
 
70
  ---
71
 
72
+ ### Why predict all three?
73
 
74
+ A quantum state is a point on (or inside) the **Bloch sphere**.
75
+
76
+ - Predicting only Z gives an incomplete picture
77
+ - Multi-target regression learns correlations between:
78
+ - circuit structure
79
+ - full quantum state orientation
80
+ - behavior in Hilbert space
81
+
82
+ ---
83
+
84
+ ## 🤖 4. Using the ML Analytics Module
85
+
86
+ The Hub uses a **Random Forest Regressor** to establish a baseline of predictability.
87
+
88
+ ### Workflow
89
+
90
+ 1. **Select Dataset**
91
+ Choose a noise model and observe how it affects predictability.
92
+
93
+ 2. **Select Features**
94
+ Recommended starting set:
95
+ - `gate_entropy`
96
+ - `meyer_wallach`
97
+ - `depth`
98
+ - `cx_count`
99
+
100
+ 3. **Execute Baseline**
101
+ Performs an **80/20 train-test split**.
102
+
103
+ 4. **Analyze the Triple Parity Plot**
104
+
105
+ - 🔴 **Diagonal Red Line** → perfect prediction
106
+ - 📈 **Clustering near line** → strong predictive signal
107
+ - 🔍 **Basis comparison**:
108
+ - Z often easier to predict
109
+ - X/Y depend more on circuit structure
110
+ - reveals architectural biases (HEA vs QFT, etc.)
111
 
112
  ---
113
 
114
  ## 🔗 5. Project Resources
115
+
116
+ - 🤗 Hugging Face Datasets — download dataset shards
117
+ - 💻 GitHub Repository — QSBench generator source code
118
+ - 🌐 Official Website — documentation and benchmarking leaderboards
119
+
120
+ ---
121
+
122
+ *QSBench — Synthetic Quantum Dataset Benchmarks*