|
|
--- |
|
|
license: mit |
|
|
datasets: |
|
|
- ILSVRC/imagenet-1k |
|
|
- uoft-cs/cifar10 |
|
|
- uoft-cs/cifar100 |
|
|
language: |
|
|
- en |
|
|
metrics: |
|
|
- accuracy |
|
|
base_model: |
|
|
- MS-ResNet |
|
|
--- |
|
|
|
|
|
<div align="center"> |
|
|
|
|
|
<h1>I2E: Real-Time Image-to-Event Conversion for High-Performance Spiking Neural Networks</h1> |
|
|
|
|
|
[](https://arxiv.org/abs/2511.08065) |
|
|
[](https://aaai.org/) |
|
|
[](https://scholar.google.com/scholar?cluster=1814482600796011970) |
|
|
[](https://github.com/Ruichen0424/I2E) |
|
|
|
|
|
[](https://huggingface.co/papers/2511.08065) |
|
|
[](https://huggingface.co/datasets/UESTC-BICS/I2E) |
|
|
|
|
|
[](https://youtu.be/v9z0pn8kTsI?si=dkbjR6AV_RgCU3wA) |
|
|
[](https://www.bilibili.com/video/BV14XrfBjENb/) |
|
|
</div> |
|
|
|
|
|
## ๐ Introduction |
|
|
|
|
|
This repository contains the **pre-trained weights** for the paper **"I2E: Real-Time Image-to-Event Conversion for High-Performance Spiking Neural Networks"**, which has been accepted for **Oral Presentation at AAAI 2026**. |
|
|
|
|
|
**I2E** is a pioneering framework that bridges the data scarcity gap in neuromorphic computing. By simulating microsaccadic eye movements via highly parallelized convolution, I2E converts static images into high-fidelity event streams in real-time (>300x faster than prior methods). |
|
|
|
|
|
### โจ Key Highlights |
|
|
* **SOTA Performance**: Achieves **60.50%** top-1 accuracy on Event-based ImageNet. |
|
|
* **Sim-to-Real Transfer**: Pre-training on I2E data enables **92.5%** accuracy on real-world CIFAR10-DVS, setting a new benchmark. |
|
|
* **Real-Time Conversion**: Enables on-the-fly data augmentation for deep SNN training. |
|
|
|
|
|
## ๐ Model Zoo & Results |
|
|
|
|
|
We provide pre-trained models for **I2E-CIFAR** and **I2E-ImageNet**. You can download the `.pth` files directly from the [**Files and versions**](https://huggingface.co/Ruichen0424/I2E/tree/main) tab in this repository. |
|
|
|
|
|
<table border="1"> |
|
|
<tr> |
|
|
<th>Target Dataset</th> |
|
|
<th align="center">Architecture</th> |
|
|
<th align="center">Method</th> |
|
|
<th align="center">Top-1 Acc</th> |
|
|
</tr> |
|
|
<!-- CIFAR10-DVS --> |
|
|
<tr> |
|
|
<td rowspan="3" align="center" style="vertical-align: middle;"><strong>CIFAR10-DVS</strong><br>(Real)</td> |
|
|
<td align="center" style="vertical-align: middle;">MS-ResNet18</td> |
|
|
<td align="center" style="vertical-align: middle;">Baseline</td> |
|
|
<td align="center" style="vertical-align: middle;">65.6%</td> |
|
|
</tr> |
|
|
<tr> |
|
|
<td align="center" style="vertical-align: middle;">MS-ResNet18</td> |
|
|
<td align="center" style="vertical-align: middle;">Transfer-I</td> |
|
|
<td align="center" style="vertical-align: middle;">83.1%</td> |
|
|
</tr> |
|
|
<tr> |
|
|
<td align="center" style="vertical-align: middle;">MS-ResNet18</td> |
|
|
<td align="center" style="vertical-align: middle;">Transfer-II (Sim-to-Real)</td> |
|
|
<td align="center" style="vertical-align: middle;"><strong>92.5%</strong></td> |
|
|
</tr> |
|
|
<!-- I2E-CIFAR10 --> |
|
|
<tr> |
|
|
<td rowspan="3" align="center" style="vertical-align: middle;"><strong>I2E-CIFAR10</strong></td> |
|
|
<td align="center" style="vertical-align: middle;">MS-ResNet18</td> |
|
|
<td align="center" style="vertical-align: middle;">Baseline-I</td> |
|
|
<td align="center" style="vertical-align: middle;">85.07%</td> |
|
|
</tr> |
|
|
<tr> |
|
|
<td align="center" style="vertical-align: middle;">MS-ResNet18</td> |
|
|
<td align="center" style="vertical-align: middle;">Baseline-II</td> |
|
|
<td align="center" style="vertical-align: middle;">89.23%</td> |
|
|
</tr> |
|
|
<tr> |
|
|
<td align="center" style="vertical-align: middle;">MS-ResNet18</td> |
|
|
<td align="center" style="vertical-align: middle;">Transfer-I</td> |
|
|
<td align="center" style="vertical-align: middle;"><strong>90.86%</strong></td> |
|
|
</tr> |
|
|
<!-- I2E-CIFAR100 --> |
|
|
<tr> |
|
|
<td rowspan="3" align="center" style="vertical-align: middle;"><strong>I2E-CIFAR100</strong></td> |
|
|
<td align="center" style="vertical-align: middle;">MS-ResNet18</td> |
|
|
<td align="center" style="vertical-align: middle;">Baseline-I</td> |
|
|
<td align="center" style="vertical-align: middle;">51.32%</td> |
|
|
</tr> |
|
|
<tr> |
|
|
<td align="center" style="vertical-align: middle;">MS-ResNet18</td> |
|
|
<td align="center" style="vertical-align: middle;">Baseline-II</td> |
|
|
<td align="center" style="vertical-align: middle;">60.68%</td> |
|
|
</tr> |
|
|
<tr> |
|
|
<td align="center" style="vertical-align: middle;">MS-ResNet18</td> |
|
|
<td align="center" style="vertical-align: middle;">Transfer-I</td> |
|
|
<td align="center" style="vertical-align: middle;"><strong>64.53%</strong></td> |
|
|
</tr> |
|
|
<!-- I2E-ImageNet --> |
|
|
<tr> |
|
|
<td rowspan="4" align="center" style="vertical-align: middle;"><strong>I2E-ImageNet</strong></td> |
|
|
<td align="center" style="vertical-align: middle;">MS-ResNet18</td> |
|
|
<td align="center" style="vertical-align: middle;">Baseline-I</td> |
|
|
<td align="center" style="vertical-align: middle;">48.30%</td> |
|
|
</tr> |
|
|
<tr> |
|
|
<td align="center" style="vertical-align: middle;">MS-ResNet18</td> |
|
|
<td align="center" style="vertical-align: middle;">Baseline-II</td> |
|
|
<td align="center" style="vertical-align: middle;">57.97%</td> |
|
|
</tr> |
|
|
<tr> |
|
|
<td align="center" style="vertical-align: middle;">MS-ResNet18</td> |
|
|
<td align="center" style="vertical-align: middle;">Transfer-I</td> |
|
|
<td align="center" style="vertical-align: middle;">59.28%</td> |
|
|
</tr> |
|
|
<tr> |
|
|
<td align="center" style="vertical-align: middle;">MS-ResNet34</td> |
|
|
<td align="center" style="vertical-align: middle;">Baseline-II</td> |
|
|
<td align="center" style="vertical-align: middle;"><strong>60.50%</strong></td> |
|
|
</tr> |
|
|
</table> |
|
|
|
|
|
> **Method Legend:** |
|
|
> * **Baseline-I**: Training from scratch with minimal augmentation. |
|
|
> * **Baseline-II**: Training from scratch with full augmentation. |
|
|
> * **Transfer-I**: Fine-tuning from Static ImageNet (or I2E-ImageNet for CIFAR targets). |
|
|
> * **Transfer-II**: Fine-tuning from I2E-CIFAR10. |
|
|
|
|
|
## ๐๏ธ Visualization |
|
|
|
|
|
Below is the visualization of the I2E conversion process. We illustrate the high-fidelity conversion from static RGB images to dynamic event streams. |
|
|
|
|
|
More than 200 additional visualization comparisons can be found in [Visualization.md](./Visualization.md). |
|
|
|
|
|
<table border="0" style="width: 100%"> |
|
|
<tr> |
|
|
<td width="25%" align="center"><img src="./assets/original_1.jpg" alt="Original 1" style="width:100%"></td> |
|
|
<td width="25%" align="center"><img src="./assets/converted_1.gif" alt="Converted 1" style="width:100%"></td> |
|
|
<td width="25%" align="center"><img src="./assets/original_2.jpg" alt="Original 2" style="width:100%"></td> |
|
|
<td width="25%" align="center"><img src="./assets/converted_2.gif" alt="Converted 2" style="width:100%"></td> |
|
|
</tr> |
|
|
<tr> |
|
|
<td width="25%" align="center"><img src="./assets/original_3.jpg" alt="Original 3" style="width:100%"></td> |
|
|
<td width="25%" align="center"><img src="./assets/converted_3.gif" alt="Converted 3" style="width:100%"></td> |
|
|
<td width="25%" align="center"><img src="./assets/original_4.jpg" alt="Original 4" style="width:100%"></td> |
|
|
<td width="25%" align="center"><img src="./assets/converted_4.gif" alt="Converted 4" style="width:100%"></td> |
|
|
</tr> |
|
|
</table> |
|
|
|
|
|
## ๐ป Usage |
|
|
|
|
|
This repository hosts the **model weights only**. |
|
|
|
|
|
For the **I2E dataset generation code**, **training scripts**, and detailed usage instructions, please refer to our official GitHub repository. |
|
|
|
|
|
To generate the datasets (I2E-CIFAR10, I2E-CIFAR100, I2E-ImageNet) yourself using the I2E algorithm, please follow the instructions in the GitHub README. |
|
|
|
|
|
[](https://github.com/Ruichen0424/I2E) |
|
|
|
|
|
The download address for the datasets generated by the I2E algorithm is as follows. |
|
|
|
|
|
[](https://huggingface.co/datasets/UESTC-BICS/I2E) |
|
|
|
|
|
## ๐ Citation |
|
|
|
|
|
If you find this work or the models useful, please cite our AAAI 2026 paper: |
|
|
|
|
|
```bibtex |
|
|
@article{ma2025i2e, |
|
|
title={I2E: Real-Time Image-to-Event Conversion for High-Performance Spiking Neural Networks}, |
|
|
author={Ma, Ruichen and Meng, Liwei and Qiao, Guanchao and Ning, Ning and Liu, Yang and Hu, Shaogang}, |
|
|
journal={arXiv preprint arXiv:2511.08065}, |
|
|
year={2025} |
|
|
} |
|
|
``` |
|
|
|
|
|
## ๐ผ๏ธ Poster |
|
|
 |