File size: 3,660 Bytes
de80776
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
---
license: cc-by-4.0
task_categories:
- visual-question-answering
- image-text-to-text
- image-feature-extraction
language:
- en
tags:
- textvqa
- ocr
- vqa
- vision-language
- lance
- clip-embeddings
pretty_name: textvqa-lance
size_categories:
- 10K<n<100K
---
# TextVQA (Lance Format)

Lance-formatted version of [TextVQA](https://textvqa.org/) — VQA where the question requires *reading* text in the image — sourced from [`lmms-lab/textvqa`](https://huggingface.co/datasets/lmms-lab/textvqa).

Each row carries the image bytes, the question, the 10 reference answers, the OCR tokens detected by the dataset's pre-processing, and CLIP image + question embeddings.

## Splits

| Split | Rows |
|-------|------|
| `validation.lance` | 5,000 |
| `train.lance`      | 34,602 |

## Schema

| Column | Type | Notes |
|---|---|---|
| `id` | `int64` | Row index within split |
| `image` | `large_binary` | Inline JPEG bytes |
| `image_id` | `string?` | TextVQA image id |
| `question_id` | `string?` | TextVQA question id |
| `question` | `string` | The question text |
| `answers` | `list<string>` | 10 annotator answers |
| `answer` | `string` | First answer — used as canonical / FTS target |
| `ocr_tokens` | `list<string>` | OCR tokens detected on the image |
| `image_classes` | `list<string>` | OpenImages-style scene tags from the source |
| `set_name` | `string?` | Source partition (`train`, `val`) |
| `image_emb` | `fixed_size_list<float32, 512>` | OpenCLIP image embedding (cosine-normalized) |
| `question_emb` | `fixed_size_list<float32, 512>` | OpenCLIP text embedding of the question |

## Pre-built indices

- `IVF_PQ` on `image_emb` and `question_emb``metric=cosine`
- `INVERTED` (FTS) on `question` and `answer`
- `BTREE` on `image_id`, `question_id`, `set_name`

## Quick start

```python
import lance
ds = lance.dataset("hf://datasets/lance-format/textvqa-lance/data/validation.lance")
print(ds.count_rows(), ds.schema.names, ds.list_indices())
```

## Cross-modal text→image search

```python
import lance, pyarrow as pa, open_clip, torch

model, _, _ = open_clip.create_model_and_transforms("ViT-B-32", pretrained="laion2b_s34b_b79k")
tokenizer = open_clip.get_tokenizer("ViT-B-32")
model = model.eval().cuda().half()
with torch.no_grad():
    q = model.encode_text(tokenizer(["what brand is on this billboard?"]).cuda())
    q = (q / q.norm(dim=-1, keepdim=True)).float().cpu().numpy()[0]

ds = lance.dataset("hf://datasets/lance-format/textvqa-lance/data/validation.lance")
emb_field = ds.schema.field("image_emb")
hits = ds.scanner(
    nearest={"column": "image_emb", "q": pa.array([q.tolist()], type=emb_field.type)[0], "k": 10},
    columns=["question", "answer", "ocr_tokens"],
).to_table().to_pylist()
```

## Why Lance?

- One dataset for images + questions + answers + OCR + dual embeddings + indices — no JSON/feature folders.
- Cross-modal search and OCR-text filtering work on the same dataset on the Hub.
- Schema evolution: add columns (alternate OCR systems, model predictions) without rewriting the data.

## Source & license

Converted from [`lmms-lab/textvqa`](https://huggingface.co/datasets/lmms-lab/textvqa). TextVQA is released under [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) by Singh et al. (Facebook AI Research).

## Citation

```
@inproceedings{singh2019towards,
  title={Towards VQA models that can read},
  author={Singh, Amanpreet and Natarajan, Vivek and Shah, Meet and Jiang, Yu and Chen, Xinlei and Batra, Dhruv and Parikh, Devi and Rohrbach, Marcus},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
  year={2019}
}
```