FlyGCL / README.md
HoraceYan's picture
Update README.md
0d554b3 verified
---
language:
- en
license: mit
tags:
- continual-learning
- general-continual-learning
- online-learning
- vision
- image-classification
- vit
- prompt-tuning
library_name: pytorch
pipeline_tag: image-classification
inference: false
---
# FlyGCL Checkpoints (FlyPrompt & ViT Baselines)
This repository provides **research checkpoints** for **FlyGCL**, a lightweight framework for **General Continual Learning (GCL) / online class-incremental learning** in the **Si-Blurry** setting.
It is designed to be used together with the FlyGCL codebase:
- Code: `https://github.com/AnAppleCore/FlyGCL`
- Paper (arXiv): `https://www.arxiv.org/abs/2602.01976`
- OpenReview: `https://openreview.net/forum?id=8pi1rP71qv`
## What is included
This model repo may contain:
- **Backbone checkpoints** (ViT-B/16 variants) referenced by FlyGCL via `--backbone`.
- **Prompt checkpoints** (optional) for DualPrompt/MISA-style prompts:
- `g_prompt.pt`
- `e_prompt.pt`
For the exact filename mapping and where to place these files in FlyGCL, see the code repository README:
- `https://github.com/AnAppleCore/FlyGCL/blob/main/README.md`
## Model details
- **Architecture family**: Vision Transformer (ViT-B/16) backbones + prompt-based continual learning heads.
- **Framework**: PyTorch.
- **Training setting**: online GCL / Si-Blurry (see paper and code for details).
## Intended use
These checkpoints are released for:
- **Research / reproducibility** of the FlyGCL paper and baselines
- **Benchmarking** continual learning methods in comparable settings
Not intended for:
- Safety-critical or medical/diagnostic use
- Deployment without careful evaluation in your target environment
## Limitations and biases
- Continual learning performance depends on data ordering, hyperparameters, and backbone initialization.
- Backbones pretrained on large-scale datasets may encode biases from their pretraining data.
- Prompt checkpoints may not transfer to datasets/settings different from those used during training.
## License
- Code license: MIT (see FlyGCL `LICENSE`).
- **Checkpoint licensing** may depend on upstream sources (e.g., DINO/iBOT/MoCo pretrained backbones). If you redistribute upstream-derived weights here, ensure the redistribution terms are compatible and include required notices.
## Citation
If you use FlyGCL or these checkpoints in your research, please cite:
```bibtex
@inproceedings{flyprompt2026,
title={FlyPrompt: Brain-Inspired Random-Expanded Routing with Temporal-Ensemble Experts for General Continual Learning},
author={Yan, Hongwei and Sun, Guanglong and Zhou, Kanglei and Li, Qian and Wang, Liyuan and Zhong, Yi},
booktitle={ICLR},
year={2026}
}
```
## Contact
- Maintainer: `Hongwei Yan` (`yanhw22@mails.tsinghua.edu.cn`)