TC-AE / README.md
nielsr's picture
nielsr HF Staff
Add metadata and improve model card
6492198 verified
|
raw
history blame
1.93 kB
---
license: mit
pipeline_tag: image-to-image
---
# TC-AE: Unlocking Token Capacity for Deep Compression Autoencoders
TC-AE is a novel Vision Transformer (ViT)-based tokenizer for deep image compression and visual generation. It addresses the challenge of latent representation collapse in high compression ratios by optimizing the token space.
<p align="center">
<a href="https://huggingface.co/papers/2604.07340"><img src="https://img.shields.io/badge/Paper-Arxiv-b31b1b.svg" alt="arXiv"></a>&nbsp;
<a href="https://github.com/inclusionAI/TC-AE"><img src="https://img.shields.io/badge/Code-GitHub-blue?logo=github" alt="GitHub"></a>
</p>
## Introduction
TC-AE achieves substantially improved reconstruction and generative performance under deep compression through two key innovations:
1. **Staged Token Compression**: Decomposes token-to-latent mapping into two stages, reducing structural information loss in the bottleneck.
2. **Semantic Enhancement**: Incorporates joint self-supervised training to produce more generative-friendly latents.
## Usage
### Environment Setup
To set up the environment for TC-AE, follow these steps:
```shell
conda create -n tcae python=3.9
conda activate tcae
pip install -r requirements.txt
```
### Image Reconstruction Demo
To use the TC-AE tokenizer for image reconstruction, you can run the following script using the pre-trained weights:
```shell
python tcae/script/demo_recon.py \
--img_folder /path/to/your/images \
--output_folder /path/to/output \
--ckpt_path results/tcae.pt \
--config configs/TC-AE-SL.yaml \
--rank 0
```
## Citation
```bibtex
@article{li2026tcae,
title={TC-AE: Unlocking Token Capacity for Deep Compression Autoencoders},
author={Li, Teng and Huang, Ziyuan and Chen, Cong and Li, Yangfu and Lyu, Yuanhuiyi and Zheng, Dandan and Shen, Chunhua and Zhang, Jun},
journal={arXiv preprint arXiv:2604.07340},
year={2026}
}
```