File size: 3,310 Bytes
e5fae3c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
---
license: mit
library_name: transformers
pipeline_tag: video-classification
tags:
  - video-classification
  - zero-shot
  - vision
  - acaua
datasets:
  - kinetics-400
base_model: microsoft/xclip-base-patch32
---

# X-CLIP (base, patch 32) — acaua mirror

MIT-licensed mirror hosted under `CondadosAI/` for use with the [acaua](https://github.com/CondadosAI/acaua) computer vision library.

This is a **safetensors-only mirror** of the upstream Microsoft weights at the pinned commit shown below. The legacy `pytorch_model.bin` (pickle format) that upstream ships alongside `model.safetensors` has been deliberately removed for security hygiene — pickle loads can execute arbitrary code, and `transformers` auto-prefers safetensors when both are present, so removing it has zero functional impact on downstream users.

X-CLIP is a **zero-shot video classification** model: you provide a list of candidate text labels at inference time and the model ranks them by similarity to the video clip. It is not a closed-set softmax classifier, and it does not appear in `AutoModelForVideoClassification`.

## Provenance

| | |
|---|---|
| Upstream repo | [`microsoft/xclip-base-patch32`](https://huggingface.co/microsoft/xclip-base-patch32) |
| Upstream commit SHA | `a2e27a78a2b5d802e894b8a1ef14f3a8ce490963` |
| Upstream commit date | 2024-02-04 |
| Declared license | MIT |
| Paper | Ni et al., *"Expanding Language-Image Pretrained Models for General Video Recognition"*, ECCV 2022, arXiv:[2208.02816](https://arxiv.org/abs/2208.02816) |
| Official code | [`microsoft/VideoX`](https://github.com/microsoft/VideoX) (MIT) |
| Mirrored on | 2026-04-23 |
| Mirrored by | [CondadosAI/acaua](https://github.com/CondadosAI/acaua) |

## Usage via acaua

```python
import acaua

model = acaua.Model.from_pretrained(
    "CondadosAI/xclip_base_patch32",
    allow_non_apache=True,  # weights are MIT, not Apache-2.0
)
result = model.predict(
    "dance.mp4",
    labels=["dancing", "cooking", "running", "sleeping", "walking"],
    top_k=3,
)
for label, score in zip(result.labels, result.scores.tolist()):
    print(f"{label}: {score:.3f}")
```

## Usage via 🤗 Transformers

This mirror is drop-in compatible with the upstream repo.

```python
from transformers import XCLIPModel, XCLIPProcessor

processor = XCLIPProcessor.from_pretrained("CondadosAI/xclip_base_patch32")
model = XCLIPModel.from_pretrained("CondadosAI/xclip_base_patch32")
```

## Expected input

- **Frames:** 8 uniformly-sampled frames per clip (`vision_config.num_frames=8`).
- **Resolution:** 224 × 224 after resize + center-crop.
- **Normalization:** ImageNet mean/std (handled by `XCLIPProcessor`).
- **Text prompts:** supplied at inference time — any natural-language strings.

## License and attribution

Redistributed under MIT, consistent with the upstream declaration. See [`NOTICE`](./NOTICE) for required attribution.

## Citation

```bibtex
@inproceedings{ni2022expanding,
  title={Expanding language-image pretrained models for general video recognition},
  author={Ni, Bolin and Peng, Houwen and Chen, Minghao and Zhang, Songyang and Meng, Gaofeng and Fu, Jianlong and Xiang, Shiming and Ling, Haibin},
  booktitle={European Conference on Computer Vision (ECCV)},
  pages={1--18},
  year={2022},
  publisher={Springer}
}
```