DaMsTaR commited on
Commit
101d36b
Β·
verified Β·
1 Parent(s): 8f20285

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +403 -1
README.md CHANGED
@@ -1 +1,403 @@
1
- # Quantum-ML
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Quantum ML
2
+
3
+ ## Overview
4
+
5
+ This guide provides a complete, actionable roadmap to replicate the results from Kung et al. (arXiv:2503.12099), which presents a machine learning approach for automatic characterization of fluxonium superconducting qubit parameters (E_J, E_C, E_L) using a Swin Transformer V2 model trained via deep transfer learning. The paper reports ~95.6% average accuracy across all three energy parameters. The authors mention that source code will be available on GitHub after publication, but since it may not yet be released, this guide reconstructs every detail needed for a from-scratch replication.
6
+
7
+ ***
8
+
9
+ ## Phase 1: Environment Setup
10
+
11
+ ### Hardware Requirements
12
+
13
+ - A GPU with at least 8 GB VRAM (NVIDIA RTX 3060 or better recommended). Swin Transformer V2 Tiny has ~28M parameters and is relatively lightweight.[^1]
14
+ - Sufficient CPU/RAM for generating 15,000+ spectrum simulations via QuTiP.
15
+
16
+ ### Software Dependencies
17
+
18
+ Install the following Python packages:
19
+
20
+ - **PyTorch** (β‰₯1.12) with CUDA support
21
+ - **torchvision** β€” provides pre-built Swin Transformer V2 models (`swin_v2_t`, `swin_v2_b`)[^2][^1]
22
+ - **timm** (PyTorch Image Models) β€” alternative source for `swinv2_tiny_window8_256` and other variants[^3]
23
+ - **QuTiP** β€” Quantum Toolbox in Python for Hamiltonian diagonalization and spectrum computation[^4]
24
+ - **scqubits** β€” optional but helpful for fluxonium simulation and validation[^5][^6]
25
+ - **prodigyopt** β€” the Prodigy optimizer (`pip install prodigyopt`)[^7][^8]
26
+ - **scipy** β€” for `find_peaks_cwt` peak detection[^9]
27
+ - **numpy, matplotlib, PIL/Pillow**
28
+
29
+ ```
30
+ pip install torch torchvision timm qutip scqubits prodigyopt scipy numpy matplotlib pillow
31
+ ```
32
+
33
+ ***
34
+
35
+ ## Phase 2: Understanding the Fluxonium Hamiltonian
36
+
37
+ ### The Model Hamiltonian
38
+
39
+ The fluxonium qubit Hamiltonian is:
40
+
41
+ H = 4 * E_C * n^2 - E_J * cos(phi + phi_ext) + 0.5 * E_L * phi^2
42
+
43
+ where:
44
+ - **E_C** = charging energy (capacitance)
45
+ - **E_J** = Josephson energy
46
+ - **E_L** = inductive energy
47
+ - **phi** = phase operator across inductance
48
+ - **n** = displacement charge operator
49
+ - **phi_ext** = external magnetic flux (varied over one flux quantum period)
50
+
51
+ ### Parameter Ranges
52
+
53
+ The training data spans these experimentally relevant ranges:
54
+
55
+ | Parameter | Range (GHz) | Span |
56
+ |-----------|-------------|------|
57
+ | E_C | 0.5 – 3.0 | 2.5 GHz |
58
+ | E_L | 0.1 – 2.0 | 1.9 GHz |
59
+ | E_J | 2.0 – 10.0 | 8.0 GHz |
60
+
61
+ ### Transitions Considered
62
+
63
+ The energy transitions used are: **0β†’1, 0β†’2, 0β†’3, 0β†’4, 0β†’5, 1β†’2, and 1β†’3**, all within the frequency window of **4.0–8.0 GHz**.
64
+
65
+ ***
66
+
67
+ ## Phase 3: Generating Training Data
68
+
69
+ This is the most computationally intensive phase. There are two distinct datasets to generate.
70
+
71
+ ### Dataset 1: Pure Spectrum Dataset (N = 15,392)
72
+
73
+ This dataset contains only the bare transition energies (no coupling/readout effects), making it fast to compute.
74
+
75
+ **For each parameter combination (E_C, E_L, E_J):**
76
+
77
+ 1. **Sample parameters** randomly or on a grid within the ranges above. The paper uses 15,392 unique combinations.
78
+ 2. **Sweep phi_ext** with 256 points per flux period (0 to 2Ο€).
79
+ 3. **Diagonalize the Hamiltonian** at each flux point using QuTiP. Use `scqubits.Fluxonium` or build the Hamiltonian matrix directly in QuTiP with a sufficiently large cutoff (typically 110 states).[^5]
80
+ 4. **Compute transition energies** between all relevant level pairs (0-1, 0-2, ..., 1-3).
81
+ 5. **Filter transitions** to retain only those within 4.0–8.0 GHz.
82
+ 6. **Render as an image**: Plot each valid transition point as a black dot on a 2D image (x-axis = phi_ext, y-axis = frequency in GHz). The image serves as input to the Swin Transformer.
83
+
84
+ **Example code sketch for a single spectrum:**
85
+
86
+ ```python
87
+ import scqubits as scq
88
+ import numpy as np
89
+
90
+ def generate_pure_spectrum(EC, EL, EJ, n_flux=256, cutoff=110):
91
+ fluxonium = scq.Fluxonium(EJ=EJ, EC=EC, EL=EL, flux=0.0, cutoff=cutoff)
92
+ flux_vals = np.linspace(0.0, 1.0, n_flux) # in units of Phi_0
93
+
94
+ transitions = [(0,1), (0,2), (0,3), (0,4), (0,5), (1,2), (1,3)]
95
+ spectrum_points = []
96
+
97
+ for flux in flux_vals:
98
+ fluxonium.flux = flux
99
+ evals = fluxonium.eigenvals(evals_count=6)
100
+ for (i, j) in transitions:
101
+ if j < len(evals):
102
+ freq = evals[j] - evals[i]
103
+ if 4.0 <= freq <= 8.0:
104
+ spectrum_points.append((flux, freq))
105
+
106
+ return spectrum_points
107
+ ```
108
+
109
+ **Image generation**: Convert each spectrum into a fixed-resolution image (e.g., 256Γ—256 pixels). The Swin Transformer V2 Tiny expects 256Γ—256 input. Plot flux on the x-axis and frequency on the y-axis, with black dots on a white background. Save as PNG or convert directly to a tensor.[^1]
110
+
111
+ ### Dataset 2: Dispersive Readout Dataset (N = 469)
112
+
113
+ This dataset simulates a more realistic measurement scenario including dispersive readout effects:
114
+
115
+ 1. **Readout resonator** at 6.00 GHz with linewidth 7 MHz and coupling strength g = 100 MHz.
116
+ 2. **Compute the dispersive shift** for each transition using second-order perturbation theory.
117
+ 3. **Calculate voltage change** in readout response caused by dispersive shift for a saturation drive at every transition and flux value.
118
+ 4. **Threshold**: Exclude data points where readout voltage change < 10% of maximum magnitude at readout resonance.
119
+ 5. **Render as image** similarly to the pure spectrum, but now transition points carry varying intensities based on signal magnitude.
120
+
121
+ This computation is >100Γ— slower per spectrum than the pure dataset, which is why only 469 samples are used. The dispersive readout dataset is critical for the transfer learning step.
122
+
123
+ ***
124
+
125
+ ## Phase 4: Model Architecture β€” Swin Transformer V2
126
+
127
+ ### Model Selection
128
+
129
+ The paper uses **Swin Transformer V2**, chosen for its lightweight architecture compared to ResNet and DenseNet alternatives. The exact variant isn't specified, but the **Swin V2 Tiny** model is the most practical choice:[^10][^11]
130
+
131
+ | Property | Swin V2 Tiny |
132
+ |----------|-------------|
133
+ | Parameters | ~28.3M[^1] |
134
+ | Input resolution | 256 Γ— 256 |
135
+ | GFLOPs | 5.94[^1] |
136
+ | Embed dim | 96 |
137
+ | Depths | [^12][^12][^13][^12] |
138
+ | Num heads | [^14][^13][^15][^7] |
139
+ | Window size | 8 |
140
+
141
+ ### Loading the Model
142
+
143
+ ```python
144
+ import torchvision.models as models
145
+ import torch.nn as nn
146
+
147
+ # Load pretrained Swin V2 Tiny (ImageNet weights)
148
+ model = models.swin_v2_t(weights=models.Swin_V2_T_Weights.IMAGENET1K_V1)
149
+
150
+ # Modify the classification head for regression (3 outputs: EC, EL, EJ)
151
+ model.head = nn.Linear(model.head.in_features, 3)
152
+ ```
153
+
154
+ Alternatively, using `timm`:
155
+
156
+ ```python
157
+ import timm
158
+
159
+ model = timm.create_model('swinv2_tiny_window8_256', pretrained=True, num_classes=3)
160
+ ```
161
+
162
+ ### Input Preprocessing
163
+
164
+ The spectrum images should be converted to 3-channel (RGB) tensors of size 256Γ—256. Apply standard ImageNet normalization (mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]) since the model is pretrained on ImageNet.[^2][^1]
165
+
166
+ ***
167
+
168
+ ## Phase 5: Two-Stage Transfer Learning Training
169
+
170
+ This is the core methodological contribution. The training proceeds in two stages.
171
+
172
+ ### Stage 1: Pre-train on Pure Spectrum Dataset
173
+
174
+ - **Dataset**: 15,392 pure spectrum images
175
+ - **Labels**: Corresponding [E_C, E_L, E_J] vectors (continuous values)
176
+ - **Loss function**: Mean Squared Error (MSE):
177
+
178
+ Loss = (1/N) * Ξ£ (F_NN(S_E^i) - E^i)^2
179
+
180
+ - **Optimizer**: Prodigy with default lr=1.0. Prodigy is parameter-free and adaptively estimates the learning rate.[^8][^7]
181
+
182
+ ```python
183
+ from prodigyopt import Prodigy
184
+
185
+ optimizer = Prodigy(model.parameters(), lr=1.0, weight_decay=0.01)
186
+ ```
187
+
188
+ - **Training details**: Train until convergence. Use a validation split (~10-15%) from the pure dataset to monitor overfitting. The paper does not specify exact epoch counts, so train until validation loss plateaus (likely 50–200 epochs depending on batch size).
189
+ - **Batch size**: Not explicitly stated; start with 32 or 64.
190
+ - **Scheduler**: Cosine annealing is recommended with Prodigy.[^7]
191
+
192
+ ```python
193
+ scheduler = torch.optim.lr_scheduler.CosineAnnealingLR(optimizer, T_max=total_iterations)
194
+ ```
195
+
196
+ ### Stage 2: Fine-tune on Dispersive Readout Dataset
197
+
198
+ - **Dataset**: 469 dispersive readout spectrum images
199
+ - **Initialization**: Load all weights from Stage 1
200
+ - **Loss function**: Same MSE loss
201
+ - **Optimizer**: Prodigy (reinitialize for the new stage)
202
+ - **Training**: Fine-tune the entire model on the smaller, more realistic dataset. This transfer learning step is critical β€” the pure spectrum pre-training provides a strong initialization, and the dispersive dataset aligns the model with experimental conditions.
203
+ - **Caution**: With only 469 samples, overfitting is a risk. Use aggressive data augmentation (random horizontal flips, small rotations, slight noise injection) and early stopping.
204
+
205
+ ***
206
+
207
+ ## Phase 6: Evaluation and Validation
208
+
209
+ ### Test Dataset
210
+
211
+ Generate **512 test spectra** with non-repetitive parameter combinations distinct from training data, within the same parameter ranges.
212
+
213
+ ### Accuracy Metric
214
+
215
+ The paper defines accuracy per parameter as:
216
+
217
+ Acc(E_Ξ½) = (1/N_test) * Ξ£ (1 - |E_Ξ½^i - E_Ξ½^{true,i}| / R(E_Ξ½^{test}))
218
+
219
+ where R(E_Ξ½^{test}) is the training range (2.5 GHz for E_C, 1.9 GHz for E_L, 8.0 GHz for E_J). This differs from standard classification accuracy β€” it measures how close predictions are relative to the parameter range.
220
+
221
+ ### Target Accuracies
222
+
223
+ | Parameter | Target Accuracy | Implied Average Deviation |
224
+ |-----------|----------------|--------------------------|
225
+ | E_C | 94.5% | 0.125 GHz |
226
+ | E_L | 97.1% | 0.095 GHz |
227
+ | E_J | 95.3% | 0.4 GHz |
228
+ | **Overall** | **95.6%** | β€” |
229
+
230
+ These are the benchmarks from the paper.
231
+
232
+ ### Error and Cost Metrics
233
+
234
+ The combined error function:
235
+
236
+ Error = 1 - (1/3) * Ξ£ Acc(E_Ξ½) for Ξ½ = C, L, J
237
+
238
+ The cost function measures spectral fit quality:
239
+
240
+ Cost = (1/N) * Ξ£ (f(phi_i) - f_i)^2
241
+
242
+ where f(phi_i) is the transition frequency calculated from the predicted parameters.
243
+
244
+ ***
245
+
246
+ ## Phase 7: Automatic Fitting Pipeline (End-to-End)
247
+
248
+ Once the ML model is trained, the full automatic characterization pipeline works as follows:
249
+
250
+ ### Step 1: Preprocess Experimental Data
251
+
252
+ - Apply a **band-pass filter**: keep data points with signal magnitude > 2.5 standard deviations above background average and < 20% of maximum measured magnitude.
253
+ - Use **`scipy.signal.find_peaks_cwt`** to detect transition spectrum peaks at magnitude extrema.[^9]
254
+
255
+ ### Step 2: ML Initial Guess
256
+
257
+ - Feed the preprocessed spectrum image into the trained Swin Transformer V2 model.
258
+ - Obtain initial guesses: E_C^0, E_L^0, E_J^0.
259
+
260
+ ### Step 3: Transition Identification
261
+
262
+ - Simulate a spectrum using the ML-predicted parameters.
263
+ - Label each experimental data point by associating it with the nearest simulated transition, provided the nearest transition is within **0.3 GHz**.
264
+ - Exclude points that are far from any simulated transition or fall within regions where multiple transitions overlap within 0.3 GHz.
265
+
266
+ ### Step 4: Least-Squares Fitting
267
+
268
+ - Use the ML predictions as initial guesses for a least-squares fit (e.g., `scipy.optimize.least_squares` or `scipy.optimize.curve_fit`).
269
+ - Fit the labeled data points to the fluxonium Hamiltonian model.
270
+ - Constrain fitting to **5 iterations** as in the paper's benchmarks.
271
+ - Output final refined values of E_C, E_L, E_J.
272
+
273
+ ***
274
+
275
+ ## Phase 8: Reproducing Key Results
276
+
277
+ ### Result 1: Prediction Accuracy (Figure 4)
278
+
279
+ - Run inference on 512 test spectra.
280
+ - Plot predicted vs. true values for each of E_C, E_L, E_J.
281
+ - Compute average accuracy using the custom metric. Target: ~95.6% overall.
282
+
283
+ ### Result 2: Error and Cost Landscapes (Figures 5–6)
284
+
285
+ - Choose a test case, e.g., (E_C=1.28, E_J=6.50, E_L=0.70) GHz.
286
+ - Generate a 2D grid of initial parameter guesses.
287
+ - For each initial guess, run 5 fitting iterations and compute Error and Cost.
288
+ - Plot heatmaps showing that the ML prediction falls in the darkest (lowest error/cost) region.
289
+
290
+ ### Result 3: ML vs. Random Initial Guess (Table 1)
291
+
292
+ - For 60 parameter sets, compare:
293
+ - 512 random initial guesses β†’ 5 fitting iterations β†’ average Error and Cost
294
+ - ML initial guess β†’ 5 fitting iterations β†’ Error and Cost
295
+
296
+ | Method | Avg Error | Std Error | Avg Cost | Std Cost |
297
+ |--------|-----------|-----------|----------|----------|
298
+ | Random initial values | 0.218 | 0.098 | 0.146 | 0.130 |
299
+ | ML prediction | 0.037 | 0.088 | 0.024 | 0.083 |
300
+
301
+ The ML approach should yield nearly one order of magnitude improvement.
302
+
303
+ ### Result 4: Real Experimental Data (Figure 7)
304
+
305
+ - If access to real fluxonium measurement data is available, apply the full pipeline.
306
+ - The paper demonstrates successful characterization with only partial spectra (4.0–5.9 GHz instead of 4.0–8.0 GHz) and even with half-period symmetrized data.
307
+
308
+ ***
309
+
310
+ ## Phase 9: Practical Tips and Troubleshooting
311
+
312
+ ### Data Generation Optimization
313
+
314
+ - **Parallelization**: Use Python's `multiprocessing` to generate spectra in parallel. Each spectrum is independent.
315
+ - **Caching**: Save computed eigenvalues to disk (HDF5 or NumPy arrays) so you don't recompute if training is restarted.
316
+ - **scqubits cutoff**: Use cutoff=110 for the fluxonium Hilbert space. Lower cutoffs may miss higher transitions; higher cutoffs waste computation time.[^5]
317
+
318
+ ### Image Representation
319
+
320
+ - The paper plots spectra as black dots on a white background. Ensure consistent resolution (256Γ—256) and normalization.
321
+ - Consider using a fixed pixel grid: map phi_ext ∈ [0, 2Ο€] to x ∈ and frequency ∈ [4.0, 8.0] GHz to y ∈ .
322
+ - Each dot should be at least 1–2 pixels wide for visibility.
323
+
324
+ ### Training Stability
325
+
326
+ - Prodigy with lr=1.0 is recommended. If training is unstable, reduce `d_coef` to 0.5.[^7]
327
+ - For the fine-tuning stage (469 samples), consider freezing early layers of the Swin Transformer and only fine-tuning the later layers and the regression head.
328
+ - Monitor for overfitting by tracking validation loss closely in Stage 2.
329
+
330
+ ### Label Normalization
331
+
332
+ - Normalize target values to by dividing by the parameter range (e.g., E_C_normalized = (E_C - 0.5) / 2.5). This helps MSE loss treat all three parameters equally.[^16]
333
+ - At inference time, denormalize predictions back to physical units.
334
+
335
+ ***
336
+
337
+ ## Complete Replication Checklist
338
+
339
+ | Step | Task | Status |
340
+ |------|------|--------|
341
+ | 1 | Install all dependencies (PyTorch, QuTiP, scqubits, prodigyopt, timm) | ☐ |
342
+ | 2 | Implement fluxonium Hamiltonian spectrum generator | ☐ |
343
+ | 3 | Generate 15,392 pure spectrum images + labels | ☐ |
344
+ | 4 | Generate 469 dispersive readout spectrum images + labels | ☐ |
345
+ | 5 | Generate 512 test spectrum images + labels | ☐ |
346
+ | 6 | Set up Swin V2 Tiny model with 3-output regression head | ☐ |
347
+ | 7 | Stage 1: Train on pure spectrum dataset with Prodigy optimizer | ☐ |
348
+ | 8 | Stage 2: Fine-tune on dispersive readout dataset | ☐ |
349
+ | 9 | Evaluate on test set β€” target ~95.6% accuracy | ☐ |
350
+ | 10 | Implement automatic fitting pipeline (filter β†’ ML β†’ label β†’ fit) | ☐ |
351
+ | 11 | Reproduce Error/Cost comparison (Table 1) | ☐ |
352
+ | 12 | (Optional) Apply to real experimental data | ☐ |
353
+
354
+ ***
355
+
356
+ ## Key References and Resources
357
+
358
+ - **Paper**: arXiv:2503.12099 β€” Kung et al., "Automatic Characterization of Fluxonium Superconducting Qubits Parameters with Deep Transfer Learning"
359
+ - **Swin Transformer V2**: Liu et al., CVPR 2022 β€” architecture details and pretrained weights[^10]
360
+ - **Prodigy Optimizer**: Mishchenko & Defazio, arXiv:2306.06101 β€” parameter-free adaptive optimizer[^8]
361
+ - **scqubits**: Koch et al., Quantum 5, 583 (2021) β€” Python package for superconducting qubit simulation[^6]
362
+ - **QuTiP**: Quantum Toolbox in Python β€” used for Hamiltonian diagonalization[^4]
363
+ - **torchvision SwinV2**: Official PyTorch implementation with ImageNet-pretrained weights[^1]
364
+
365
+ ---
366
+
367
+ ## References
368
+
369
+ 1. [swin_v2_t β€” Torchvision main documentation](https://docs.pytorch.org/vision/main/models/generated/torchvision.models.swin_v2_t.html)
370
+
371
+ 2. [swin_v2_b β€” Torchvision main documentation](https://docs.pytorch.org/vision/main/models/generated/torchvision.models.swin_v2_b.html) - Constructs a swin_v2_base architecture from Swin Transformer V2: Scaling Up Capacity and Resolution....
372
+
373
+ 3. [Loading a pre-trained SwinV2 transformer and modifying the architecture Β· huggingface pytorch-image-models Β· Discussion #1843](https://github.com/huggingface/pytorch-image-models/discussions/1843) - I am trying to create a SwinV2 transformer model by loading pretrained weights and later modifying s...
374
+
375
+ 4. [Accelerate Qubit Research with NVIDIA cuQuantum Integrations in ...](https://developer.nvidia.com/blog/accelerate-qubit-research-with-nvidia-cuquantum-integrations-in-qutip-and-scqubits/) - The outputs of scQubits can also easily serve as inputs for analog quantum dynamics simulations usin...
376
+
377
+ 5. [Fluxonium Qubit β€” scqubits Documentation](https://scqubits.readthedocs.io/en/v2.0_a/guide/qubits/fluxonium.html) - An instance of the fluxonium qubit is created as follows: fluxonium = scqubits.Fluxonium(EJ = 8.9, E...
378
+
379
+ 6. [Scqubits: a Python package for superconducting qubits](https://arxiv.org/abs/2107.08552) - $\textbf{scqubits}$ is an open-source Python package for simulating and analyzing superconducting ci...
380
+
381
+ 7. [prodigyopt](https://pypi.org/project/prodigyopt/) - An Adam-like optimizer for neural networks with adaptive estimation of learning rate
382
+
383
+ 8. [The Prodigy optimizer and its variants for training neural ...](https://github.com/konstmish/prodigy) - The Prodigy optimizer and its variants for training neural networks. - konstmish/prodigy
384
+
385
+ 9. [find_peaks_cwt β€” SciPy v1.17.0 Manual](https://docs.scipy.org/doc/scipy/reference/generated/scipy.signal.find_peaks_cwt.html)
386
+
387
+ 10. [Swin Transformer V2: Scaling Up Capacity and Resolution](https://ieeexplore.ieee.org/document/9879380/) - We present techniques for scaling Swin Transformer [35] up to 3 billion parameters and making it cap...
388
+
389
+ 11. [Swin Transformer V2: Advancing Computer Vision with Scalable ...](https://www.raulartigues.com/en/post/swin-transformer-v2-advancing-computer-vision-with-scalable-neural-architectures) - Architecture & Functionality​​ Swin Transformer V2 retains the hierarchical structure of its predece...
390
+
391
+ 12. [SwinCNet leveraging Swin Transformer V2 and CNN for precise color correction and detail enhancement in underwater image restoration](https://www.frontiersin.org/articles/10.3389/fmars.2025.1523729/full) - Underwater image restoration confronts three major challenges: color distortion, contrast degradatio...
392
+
393
+ 13. [Retinal vessel segmentation using a swin transformer-based encoder-decoder architecture](https://link.springer.com/10.1007/s11760-025-05089-1)
394
+
395
+ 14. [DUSFormer: Dual-Swin Transformer V2 Aggregate Network for Polyp Segmentation](https://ieeexplore.ieee.org/document/10387670/) - The convolutional neural network method has certain limitations in medical image segmentation. As a ...
396
+
397
+ 15. [Leveraging Swin Transformer for Local-to-Global Weakly Supervised
398
+ Semantic Segmentation](https://arxiv.org/pdf/2401.17828.pdf) - ...a
399
+ 0.98% mAP higher localization accuracy, outperforming state-of-the-art models.
400
+ It also yields c...
401
+
402
+ 16. [An Image Denoising Method Based on Swin Transformer V2 and U-Net Architecture](https://ieeexplore.ieee.org/document/10807930/) - To address the issue of image degradation caused by noise during image acquisition and transmission,...
403
+