Sec-4B / README.md
VeryAladeen's picture
Update README.md
a12650f verified
---
license: apache-2.0
language:
- en
base_model:
- OpenIXCLab/SeC-4B
pipeline_tag: image-segmentation
library_name: transformers
---
# SeC-4B Model Files - Multiple Precision Formats
Single-file model formats for the **SeC (Segment Concept)** video object segmentation model, optimized for use with [ComfyUI SeC Nodes](https://github.com/9nate-drake/Comfyui-SecNodes).
## Model Formats
| Format | Size | Description | GPU Requirements |
|--------|------|-------------|------------------|
| **SeC-4B-fp16.safetensors** | 7.35 GB | Recommended - Best balance of quality and size | All CUDA GPUs |
| **SeC-4B-fp8.safetensors** | 3.97 GB | VRAM-constrained systems (saves 1.5-2GB VRAM) | RTX 30 series or newer |
| **SeC-4B-bf16.safetensors** | 7.35 GB | Alternative to FP16 | All CUDA GPUs |
| **SeC-4B-fp32.safetensors** | 14.14 GB | Full precision | All CUDA GPUs |
## What is SeC?
**SeC (Segment Concept)** uses Large Vision-Language Models for video object segmentation, achieving **+11.8 points** improvement over SAM 2.1 on complex semantic scenarios (SeCVOS benchmark).
Key features:
- Concept-driven tracking with semantic understanding
- Handles occlusions and appearance changes
- Bidirectional tracking support
- State-of-the-art performance on multiple benchmarks
## Usage
These models are designed for use with the [ComfyUI SeC Nodes](https://github.com/9nate-drake/Comfyui-SecNodes) custom nodes.
**Installation:**
1. Download your preferred model format
2. Place in `ComfyUI/models/sams/`
3. Install [ComfyUI SeC Nodes](https://github.com/9nate-drake/Comfyui-SecNodes)
4. The model will be automatically detected and available in the SeC Model Loader
## Original Model
These are converted single-file versions of the original model:
- **Original Repository**: [OpenIXCLab/SeC-4B](https://huggingface.co/OpenIXCLab/SeC-4B)
- **Paper**: [arXiv:2507.15852](https://arxiv.org/abs/2507.15852)
- **Official Implementation**: [github.com/OpenIXCLab/SeC](https://github.com/OpenIXCLab/SeC)
## Credits
**Original Model**: Developed by OpenIXCLab
- Model architecture and weights: Apache 2.0 License
- Paper: Zhang et al. "SeC: Advancing Complex Video Object Segmentation via Progressive Concept Construction"
**Single-File Conversions**: Created for [ComfyUI SeC Nodes](https://github.com/9nate-drake/Comfyui-SecNodes)
- Conversion script and ComfyUI integration: [9nate-drake](https://github.com/9nate-drake)
- FP8 quantization support via [torchao](https://github.com/pytorch/ao)
## License
Apache 2.0 (same as original SeC-4B model)
## Citation
If you use this model in your research, please cite the original SeC paper:
```bibtex
@article{zhang2025sec,
title = {SeC: Advancing Complex Video Object Segmentation via Progressive Concept Construction},
author = {Zhixiong Zhang and Shuangrui Ding and Xiaoyi Dong and Songxin He and
Jianfan Lin and Junsong Tang and Yuhang Zang and Yuhang Cao and
Dahua Lin and Jiaqi Wang},
journal = {arXiv preprint arXiv:2507.15852},
year = {2025}
}
```