HoraceYan commited on
Commit
0d554b3
·
verified ·
1 Parent(s): 97f2dcd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +85 -3
README.md CHANGED
@@ -1,3 +1,85 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: mit
5
+ tags:
6
+ - continual-learning
7
+ - general-continual-learning
8
+ - online-learning
9
+ - vision
10
+ - image-classification
11
+ - vit
12
+ - prompt-tuning
13
+ library_name: pytorch
14
+ pipeline_tag: image-classification
15
+ inference: false
16
+ ---
17
+
18
+ # FlyGCL Checkpoints (FlyPrompt & ViT Baselines)
19
+
20
+ This repository provides **research checkpoints** for **FlyGCL**, a lightweight framework for **General Continual Learning (GCL) / online class-incremental learning** in the **Si-Blurry** setting.
21
+
22
+ It is designed to be used together with the FlyGCL codebase:
23
+
24
+ - Code: `https://github.com/AnAppleCore/FlyGCL`
25
+ - Paper (arXiv): `https://www.arxiv.org/abs/2602.01976`
26
+ - OpenReview: `https://openreview.net/forum?id=8pi1rP71qv`
27
+
28
+ ## What is included
29
+
30
+ This model repo may contain:
31
+
32
+ - **Backbone checkpoints** (ViT-B/16 variants) referenced by FlyGCL via `--backbone`.
33
+ - **Prompt checkpoints** (optional) for DualPrompt/MISA-style prompts:
34
+ - `g_prompt.pt`
35
+ - `e_prompt.pt`
36
+
37
+ For the exact filename mapping and where to place these files in FlyGCL, see the code repository README:
38
+
39
+ - `https://github.com/AnAppleCore/FlyGCL/blob/main/README.md`
40
+
41
+ ## Model details
42
+
43
+ - **Architecture family**: Vision Transformer (ViT-B/16) backbones + prompt-based continual learning heads.
44
+ - **Framework**: PyTorch.
45
+ - **Training setting**: online GCL / Si-Blurry (see paper and code for details).
46
+
47
+ ## Intended use
48
+
49
+ These checkpoints are released for:
50
+
51
+ - **Research / reproducibility** of the FlyGCL paper and baselines
52
+ - **Benchmarking** continual learning methods in comparable settings
53
+
54
+ Not intended for:
55
+
56
+ - Safety-critical or medical/diagnostic use
57
+ - Deployment without careful evaluation in your target environment
58
+
59
+ ## Limitations and biases
60
+
61
+ - Continual learning performance depends on data ordering, hyperparameters, and backbone initialization.
62
+ - Backbones pretrained on large-scale datasets may encode biases from their pretraining data.
63
+ - Prompt checkpoints may not transfer to datasets/settings different from those used during training.
64
+
65
+ ## License
66
+
67
+ - Code license: MIT (see FlyGCL `LICENSE`).
68
+ - **Checkpoint licensing** may depend on upstream sources (e.g., DINO/iBOT/MoCo pretrained backbones). If you redistribute upstream-derived weights here, ensure the redistribution terms are compatible and include required notices.
69
+
70
+ ## Citation
71
+
72
+ If you use FlyGCL or these checkpoints in your research, please cite:
73
+
74
+ ```bibtex
75
+ @inproceedings{flyprompt2026,
76
+ title={FlyPrompt: Brain-Inspired Random-Expanded Routing with Temporal-Ensemble Experts for General Continual Learning},
77
+ author={Yan, Hongwei and Sun, Guanglong and Zhou, Kanglei and Li, Qian and Wang, Liyuan and Zhong, Yi},
78
+ booktitle={ICLR},
79
+ year={2026}
80
+ }
81
+ ```
82
+
83
+ ## Contact
84
+
85
+ - Maintainer: `Hongwei Yan` (`yanhw22@mails.tsinghua.edu.cn`)