File size: 3,048 Bytes
a9c64d4
 
 
 
 
 
 
40c894a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
---
library_name: matanyone2
tags:
- model_hub_mixin
- pytorch_model_hub_mixin
---

<div align="center">
<img src="https://github.com/pq-yang/MatAnyone2/blob/main/assets/matanyone2_logo.png?raw=true" alt="MatAnyone Logo" style="height: 52px;">
<div style="text-align: center">
    <h2>Scaling Video Matting via a Learned Quality Evaluator</h2>
</div>

<div>
    <a href='https://pq-yang.github.io/' target='_blank'>Peiqing Yang</a><sup>1</sup>&emsp;
    <a href='https://shangchenzhou.com/' target='_blank'>Shangchen Zhou</a><sup>1†</sup>&emsp;
    <a href="https://www.linkedin.com/in/kai-hao-794321382/" target='_blank'>Kai Hao</a><sup>1</sup>&emsp;
    <a href="https://scholar.google.com.sg/citations?user=fMXnSGMAAAAJ&hl=en/" target='_blank'>Qingyi Tao</a><sup>2</sup>&emsp;
</div>
<div>
    <sup>1</sup>S-Lab, Nanyang Technological University&emsp; 
    <sup>2</sup>SenseTime Research, Singapore&emsp; 
    <br>
    <sup>†</sup>Project lead
</div>


<div>
    <h4 align="center">
        <a href="https://pq-yang.github.io/projects/MatAnyone2/" target='_blank'>
        <img src="https://img.shields.io/badge/😈-Project%20Page-blue">
        </a>
        <a href="https://arxiv.org/abs/2512.11782" target='_blank'>
        <img src="https://img.shields.io/badge/arXiv-2501.14677-b31b1b.svg">
        </a>
        <a href="https://www.youtube.com/watch?v=tyi8CNyjOhc&lc=Ugw1OS7z5QbW29RZCFZ4AaABAg" target='_blank'>
        <img src="https://img.shields.io/badge/Demo%20Video-%23FF0000.svg?logo=YouTube&logoColor=white">
        </a>
        <a href="https://huggingface.co/spaces/PeiqingYang/MatAnyone" target='_blank'>
        <img src="https://img.shields.io/badge/Demo-%F0%9F%A4%97%20Hugging%20Face-blue">
        </a>
        <a href="https://colab.research.google.com/drive/1NYW_CUDf7jnzxir7tOOlY7wRRalVOifD?usp=sharing" target='_blank'>
        <img src="https://colab.research.google.com/assets/colab-badge.svg">
        </a>
    </h4>
</div>

<strong>MatAnyone 2 is a practical human video matting framework that preserves fine details by avoiding segmentation-like boundaries, while also shows enhanced robustness under challenging real-world conditions.</strong>

<div style="width: 100%; text-align: center; margin:auto;">
    <img style="width:100%" src="https://github.com/pq-yang/MatAnyone2/blob/main/assets/teaser.jpg?raw=true">
</div>

🎥 For more visual results, go checkout our <a href="https://pq-yang.github.io/projects/MatAnyone2/" target="_blank">project page</a>

</div>

---

## How to use
you can run the following commands to get started and start working with the model
```
pip install -qqU git+https://github.com/pq-yang/MatAnyone2.git
```

```python
from matanyone2 import MatAnyone2, InferenceCore
model = MatAnyone2.from_pretrained("PeiqingYang/MatAnyone2")
processor = InferenceCore(model,device="cuda:0")
# inference
processor.process_video(input_path="inputs/video/test-sample2.mp4",
                        mask_path="inputs/mask/test-sample2.png",
                        output_path="results")
```