|
|
--- |
|
|
language: |
|
|
- en |
|
|
license: mit |
|
|
tags: |
|
|
- continual-learning |
|
|
- general-continual-learning |
|
|
- online-learning |
|
|
- vision |
|
|
- image-classification |
|
|
- vit |
|
|
- prompt-tuning |
|
|
library_name: pytorch |
|
|
pipeline_tag: image-classification |
|
|
inference: false |
|
|
--- |
|
|
|
|
|
# FlyGCL Checkpoints (FlyPrompt & ViT Baselines) |
|
|
|
|
|
This repository provides **research checkpoints** for **FlyGCL**, a lightweight framework for **General Continual Learning (GCL) / online class-incremental learning** in the **Si-Blurry** setting. |
|
|
|
|
|
It is designed to be used together with the FlyGCL codebase: |
|
|
|
|
|
- Code: `https://github.com/AnAppleCore/FlyGCL` |
|
|
- Paper (arXiv): `https://www.arxiv.org/abs/2602.01976` |
|
|
- OpenReview: `https://openreview.net/forum?id=8pi1rP71qv` |
|
|
|
|
|
## What is included |
|
|
|
|
|
This model repo may contain: |
|
|
|
|
|
- **Backbone checkpoints** (ViT-B/16 variants) referenced by FlyGCL via `--backbone`. |
|
|
- **Prompt checkpoints** (optional) for DualPrompt/MISA-style prompts: |
|
|
- `g_prompt.pt` |
|
|
- `e_prompt.pt` |
|
|
|
|
|
For the exact filename mapping and where to place these files in FlyGCL, see the code repository README: |
|
|
|
|
|
- `https://github.com/AnAppleCore/FlyGCL/blob/main/README.md` |
|
|
|
|
|
## Model details |
|
|
|
|
|
- **Architecture family**: Vision Transformer (ViT-B/16) backbones + prompt-based continual learning heads. |
|
|
- **Framework**: PyTorch. |
|
|
- **Training setting**: online GCL / Si-Blurry (see paper and code for details). |
|
|
|
|
|
## Intended use |
|
|
|
|
|
These checkpoints are released for: |
|
|
|
|
|
- **Research / reproducibility** of the FlyGCL paper and baselines |
|
|
- **Benchmarking** continual learning methods in comparable settings |
|
|
|
|
|
Not intended for: |
|
|
|
|
|
- Safety-critical or medical/diagnostic use |
|
|
- Deployment without careful evaluation in your target environment |
|
|
|
|
|
## Limitations and biases |
|
|
|
|
|
- Continual learning performance depends on data ordering, hyperparameters, and backbone initialization. |
|
|
- Backbones pretrained on large-scale datasets may encode biases from their pretraining data. |
|
|
- Prompt checkpoints may not transfer to datasets/settings different from those used during training. |
|
|
|
|
|
## License |
|
|
|
|
|
- Code license: MIT (see FlyGCL `LICENSE`). |
|
|
- **Checkpoint licensing** may depend on upstream sources (e.g., DINO/iBOT/MoCo pretrained backbones). If you redistribute upstream-derived weights here, ensure the redistribution terms are compatible and include required notices. |
|
|
|
|
|
## Citation |
|
|
|
|
|
If you use FlyGCL or these checkpoints in your research, please cite: |
|
|
|
|
|
```bibtex |
|
|
@inproceedings{flyprompt2026, |
|
|
title={FlyPrompt: Brain-Inspired Random-Expanded Routing with Temporal-Ensemble Experts for General Continual Learning}, |
|
|
author={Yan, Hongwei and Sun, Guanglong and Zhou, Kanglei and Li, Qian and Wang, Liyuan and Zhong, Yi}, |
|
|
booktitle={ICLR}, |
|
|
year={2026} |
|
|
} |
|
|
``` |
|
|
|
|
|
## Contact |
|
|
|
|
|
- Maintainer: `Hongwei Yan` (`yanhw22@mails.tsinghua.edu.cn`) |
|
|
|