| --- |
| license: apache-2.0 |
| pretty_name: USearchWiki |
| language: |
| - multilingual |
| size_categories: |
| - 10M<n<100M |
| task_categories: |
| - feature-extraction |
| - sentence-similarity |
| - text-retrieval |
| source_datasets: |
| - HuggingFaceFW/finewiki |
| tags: |
| - embeddings |
| - wikipedia |
| - ann-benchmarks |
| - vector-search |
| - colbert |
| --- |
| |
| # USearchWiki |
|
|
| Multi-model embedding dataset built on [HuggingFace FineWiki](https://huggingface.co/datasets/HuggingFaceFW/finewiki), designed for approximate nearest neighbor (ANN) search benchmarking with [USearch](https://github.com/unum-cloud/usearch) and other vector search engines. |
|
|
| The same Wikipedia corpus — chunked, cleaned, and enriched with graph metadata — is embedded by multiple models spanning dense BERT-like encoders, GPT-style decoder-based LLMs, and late-interaction ColBERT-style architectures. |
| Each model's embeddings ship with precomputed ground-truth k-nearest neighbors, enabling reproducible recall and throughput benchmarks without re-running expensive exact search. |
|
|
| ## Why USearchWiki? |
|
|
| Existing ANN benchmarks suffer from three gaps: |
|
|
| 1. __Stale descriptors.__ |
| The most popular benchmarks (SIFT-1B, Deep-1B, GloVe) use features from 2014-2021 — image descriptors and word vectors, not modern text embeddings. |
| 2. __Single-model datasets.__ |
| Each benchmark is produced by one model. |
| You cannot compare how the _same_ retrieval engine handles different vector distributions without re-embedding. |
| 3. __No decoder embeddings.__ |
| State-of-the-art embedding models (GTE-Qwen, Llama-Embed-Nemotron, Qwen3-Embedding) are decoder-based LLMs, yet no ANN benchmark uses their outputs. |
|
|
| USearchWiki fixes all three: one corpus, multiple models, modern architectures, with graph-structured metadata for filtered search. |
|
|
| ## Source Corpus |
|
|
| [HuggingFaceFW/finewiki](https://huggingface.co/datasets/HuggingFaceFW/finewiki) — August 2025 snapshot, 325 languages, 61.5M articles. |
|
|
| FineWiki is extracted from Wikimedia's __Enterprise HTML dumps__ (not raw wikitext), so templates are fully rendered by MediaWiki's own engine. |
| This avoids the well-known content loss that plagues wikitext-based parsers, as `mwparserfromhell` cannot expand templates. |
| Section headings, tables, math, and lists are preserved as Markdown. |
| Bot-generated stubs, disambiguation pages, and cross-language leakage are filtered out. |
|
|
| ### Text processing |
|
|
| No chunking is used. |
| Short-content models only process the abstract. |
| Long-context models are prioritized and receive the whole document in the original form. |
|
|
| ### Scale |
|
|
| | Scope | Articles | Parquet, GB | Avg Bytes/Article | |
| | :------------------------ | -------: | ----------: | ----------------: | |
| | English | 6.6M | 38 | 5,700 | |
| | Top 5: EN, DE, FR, ES, RU | 15.7M | 86 | 5,460 | |
| | Top 10 by text volume¹ | 22.9M | 120 | 5,250 | |
| | Top 20 | 41.6M | 149 | 3,580 | |
| | All 325 languages | 61.6M | ~170 | 2,740 | |
|
|
| ¹ EN, DE, FR, ES, RU, IT, JA, ZH, PL, UK — excluding bot-generated wikis (Cebuano, Swedish, Waray, Egyptian Arabic) which inflate article counts with minimal text. |
|
|
| Parquet weight includes both `text` and `wikitext` columns; pure text is roughly half. |
| Average bytes/article drops at wider scope because smaller wikis are dominated by stubs. |
|
|
| ### Corpus Structure |
|
|
| Measured by scanning every parquet shard, including rendered Markdown `text` and raw `wikitext` columns: |
|
|
| | Quantity | Total | Per article | |
| | :----------------------------------------------- | -------: | ----------: | |
| | Articles | 61.55M | — | |
| | Rendered text bytes: `text` column | 195.2 GB | 3.2 KB | |
| | Wikitext bytes: `wikitext` column | 337.1 GB | 5.5 KB | |
| | Markdown paragraphs: blank-line-separated blocks | 254.2M | 4.13 | |
| | Section headings: `#`, `##`, `###` | 206.3M | 3.35 | |
|
|
| 35% of articles have a single paragraph (stubs), 65% have ≤ 3, only 10% have 8+. |
| Paragraph length distribution: |
|
|
| - 12%: under 50 bytes, mostly headings or one-liners, |
| - 38%: in 200–800 bytes, the prose sweet spot, |
| - 4%: over 3.2 KB, long lists or tables rendered as one block. |
|
|
| Annotation density extracted from raw `wikitext`: |
|
|
| | Annotation kind | Total | Articles touched | |
| | :---------------------------------------- | --------: | ---------------: | |
| | Plain `[[wikilinks]]` | 1.42 B | 99.4% | |
| | Templates `{{...}}` | 0.998 B | 98.6% | |
| | Piped links `[[T\|d]]` | 0.648 B | 89.3% | |
| | Citations `<ref>...` | 0.400 B | 71.0% | |
| | External URLs `[https://...]` | 84M | 41.8% | |
| | Categories `[[Category:...]]` | 55M | 16.3% | |
| | Tables `{...}` | 19M | 14.6% | |
| | Files / images `[[File:...]]` | 14M | 7.2% | |
| | __Section anchors `[[Article#Section]]`__ | __11.5M__ | __6.4%__ | |
| | Math `<math>...` | 6.4M | 0.5% | |
| | Self anchors `[[#Section]]` | 2.6M | 0.7% | |
| | Galleries `<gallery>` | 2.5M | 3.2% | |
| | Inline interwiki `[[lang:...]]` | 0.83M | 1.1% | |
|
|
| Section anchors deserve attention: they form a 11.5M-edge __paragraph-level link graph__ already curated by editors — a built-in supervision signal for sub-article retrieval evaluation. |
|
|
| ## Embedding Models |
|
|
| Each model embeds the same article corpus independently. |
| No chunking is applied — short-context models see truncated articles, long-context models see the full text. |
| Dense models produce one vector per article. |
| ColBERT models produce one vector per token (~2,000 vectors per average article). |
|
|
| | Model | Year | Type | Dims | Context | Params | License | Base / Fine-tuned by | Perf | |
| | :------------------------------------------------------------------------------------ | ---: | :---------------- | ---: | ------: | -----: | :--------- | :--------------------------------------- | :------------- | |
| | [Qwen3-Embedding-0.6B](https://huggingface.co/Qwen/Qwen3-Embedding-0.6B) | 2025 | Dense (decoder) | 1024 | 32 K | 600 M | Apache 2.0 | Qwen3 (Alibaba) | 70.7 MTEB v2 | |
| | [GTE-ModernColBERT-v1](https://huggingface.co/lightonai/GTE-ModernColBERT-v1) | 2025 | ColBERT (encoder) | 128 | 8-32 K | 139 M | Apache 2.0 | ModernBERT (Answer.AI) / LightOn | 88.4 LongEmbed | |
| | [arctic-embed-l-v2.0](https://huggingface.co/Snowflake/snowflake-arctic-embed-l-v2.0) | 2024 | Dense (encoder) | 1024 | 8 K | 568 M | Apache 2.0 | XLM-R (Meta) → BGE-M3 (BAAI) / Snowflake | 55.6 BEIR | |
| | [nomic-embed-text-v1.5](https://huggingface.co/nomic-ai/nomic-embed-text-v1.5) | 2024 | Dense (encoder) | 768 | 8 K | 137 M | Apache 2.0 | NomicBERT (Nomic) | 62.3 MTEB v1 | |
| | [e5-mistral-7b-instruct](https://huggingface.co/intfloat/e5-mistral-7b-instruct) | 2023 | Dense (decoder) | 4096 | 4 K | 7.1 B | MIT | Mistral-7B (Mistral AI) / Microsoft | 66.6 MTEB v1 | |
|
|
| ### Compute Estimates |
|
|
| All embeddings are computed and stored in half-precision - Float16, to maximize space efficiency and compatibility with off-the-shelf tools like NumPy, which have `np.float16`, but not the brein-float variant. |
| FP8 quantization can improve throughput ~1.5× with negligible quality loss. |
|
|
| Encoder models use [TEI](https://github.com/huggingface/text-embeddings-inference) with the Hopper Docker image. |
| Decoder models use [vLLM](https://github.com/vllm-project/vllm) with `--task embed`. |
|
|
| Token counts vary by tokenizer — CJK text produces ~1 token per 2-3 bytes, Latin/Cyrillic ~1 per 4-5 bytes. |
| Average article length across all languages is ~400 tokens, but this is dragged down by millions of stubs in smaller wikis; English articles average ~2,700 tokens. |
|
|
| | Model | Throughput | Total tokens | Time | Vectors | Storage | Notes | |
| | :----------------------------------- | ---------: | -----------: | -----: | ------: | ------: | :------------------------------------ | |
| | Qwen3-Embedding-0.6B | 500 doc/s | 24 B | 1.4 d | 61.6 M | 126 GB | Full articles, one vector per article | |
| | GTE-ModernColBERT-v1, section-pooled | 800 doc/s | 24 B | 0.9 d | 206.3 M | 53 GB | Mean-pool tokens within each section | |
| | arctic-embed-l-v2.0 | 800 doc/s | 28 B | 0.9 d | 61.6 M | 126 GB | Truncated at 8K tokens | |
| | nomic-embed-text-v1.5 | 1200 doc/s | 21 B | 0.6 d | 61.6 M | 95 GB | Truncated at 8K tokens | |
| | e5-mistral-7b-instruct | 50 doc/s | 21 B | 14.3 d | 61.6 M | 505 GB | Truncated at 4K tokens | |
|
|
| ## Dataset Layout |
|
|
| Layout mirrors [FineWiki's](https://huggingface.co/datasets/HuggingFaceFW/finewiki) `data/<wiki>/<group>_<shard>.parquet` structure: one directory per Wikipedia language, with shard filenames preserved 1:1. |
| Each `.f16bin` is row-aligned with its source parquet — `.f16bin` row N is the embedding of parquet row N, in native order. |
| If the source text was empty or null the row is a zero vector (`norm == 0`); the parquet's `id` column provides the doc identifier, so no separate ids file is needed. |
|
|
| Binary format: `u32` rows count, `u32` columns count, then `rows × cols` little-endian `f16` values — directly compatible with [USearch](https://github.com/unum-cloud/USearch)'s and the Big-ANN benchmark ecosystem. |
|
|
| `.body.f16bin` is the article-body embedding; `.title.f16bin` is the title-only embedding (short-context, useful for title-vs-body retrieval studies). |
|
|
| ``` |
| unum-cloud/USearchWiki/ |
| ├── README.md |
| ├── LICENSE |
| ├── .gitattributes |
| ├── usearchwiki.py # consumer module: load_lang, read_bin, discover_collection, ... |
| ├── embed_articles.py # one dense vector per article, via TEI |
| ├── embed_sections.py # late-chunking ColBERT: one vector per section |
| ├── late_chunking.py # section-aware windowing primitives |
| ├── ground_truth.py # exact global k-NN via tiled CuPy GEMMs |
| ├── build_index.py # build a USearch HNSW index from per-shard f16bin |
| ├── eval_recall.py # measure recall@k of an index against the ground truth |
| │ |
| ├── qwen3-embedding-0.6b/ # 1024-dim, decoder, float16 |
| │ ├── enwiki/ |
| │ │ ├── 000_00000.body.f16bin # mirrors enwiki/000_00000.parquet |
| │ │ ├── 000_00000.title.f16bin |
| │ │ ├── 000_00001.body.f16bin |
| │ │ ├── 000_00001.title.f16bin |
| │ │ └── ... |
| │ ├── dewiki/ |
| │ │ └── ... |
| │ └── ... # one dir per Wikipedia language |
| │ |
| ├── snowflake-arctic-embed-l-v2.0/ # 1024-dim, encoder, float16 |
| │ └── <wiki>/<group>_<shard>.{body,title}.f16bin |
| │ |
| ├── nomic-embed-text-v1.5/ # 768-dim, encoder, float16 |
| │ └── <wiki>/<group>_<shard>.{body,title}.f16bin |
| │ |
| ├── e5-mistral-7b-instruct/ # 4096-dim, decoder, float16 (planned) |
| │ └── <wiki>/<group>_<shard>.{body,title}.f16bin |
| │ |
| └── gte-moderncolbert-v1/ # 128-dim per token, ColBERT (planned) |
| └── <wiki>/<group>_<shard>.{body,title}.f16bin |
| ``` |
|
|
| ## Downloading |
|
|
| USearchWiki uses an unusual distribution policy. |
| Single repository, no separation of code and data. |
| USearchWiki lives on three coordinated mirrors, all sharing the same single-branch Git history: |
|
|
| | Mirror | Holds | Best for | |
| | :-------------- | :----------------------------- | :------------------------------- | |
| | HuggingFace Hub | code + LFS bytes (canonical) | `git clone`, `hf` CLI, streaming | |
| | GitHub | code + LFS pointers (no bytes) | reading the code, contributing | |
| | Nebius S3 | flat byte mirror of LFS blobs | bulk downloads, batch jobs | |
|
|
| `.f16bin` files are tracked via [Git LFS](https://git-lfs.com); on GitHub, the LFS server is rerouted to HuggingFace, so GitHub clones receive only ~200-byte pointer files. |
|
|
| ### From HuggingFace |
|
|
| The default and the simplest path — full code, full data, single command: |
|
|
| ```sh |
| git clone https://huggingface.co/datasets/unum-cloud/USearchWiki |
| ``` |
|
|
| To skip the ~600 GB of binaries and get only code + pointers: |
|
|
| ```sh |
| GIT_LFS_SKIP_SMUDGE=1 git clone https://huggingface.co/datasets/unum-cloud/USearchWiki |
| ``` |
|
|
| ### From GitHub |
|
|
| The GitHub repo holds only code and LFS pointers; the actual binaries live on HuggingFace. After cloning, point Git LFS at HuggingFace and pull: |
|
|
| ```sh |
| git clone https://github.com/unum-cloud/USearchWiki |
| cd USearchWiki |
| git config lfs.url https://huggingface.co/datasets/unum-cloud/USearchWiki.git/info/lfs |
| git lfs pull |
| ``` |
|
|
| ### From Nebius S3 |
|
|
| The fastest path for bulk downloads — pulls byte-identical LFS objects directly from object storage, then materializes the `.f16bin` files into the working tree. |
| Via [`s5cmd`](https://github.com/peak/s5cmd) - a parallel, single Go binary, often ~5–10× faster than `aws s3 sync` for many-files workloads: |
|
|
| ```sh |
| # One-time install |
| curl -sL https://github.com/peak/s5cmd/releases/download/v2.3.0/s5cmd_2.3.0_linux_amd64.deb -o /tmp/s5cmd.deb |
| sudo dpkg -i /tmp/s5cmd.deb |
| |
| # Sync the byte mirror, then materialize the working tree |
| s5cmd --endpoint-url https://storage.us-central1.nebius.cloud --no-sign-request \ |
| sync 's3://usearch-wiki/lfs/*' ./.git/lfs/objects/ |
| git lfs checkout |
| ``` |
|
|
| The bucket is configured for anonymous read access, so no Nebius account or credentials are needed — `--no-sign-request` tells `s5cmd` to skip request signing. |
| `aws s3 sync --no-sign-request` works equivalently with the same endpoint. |
|
|
| ### Loading embeddings in Python |
|
|
| ```python |
| from usearchwiki import read_bin |
| matrix = read_bin("qwen3-embedding-0.6b/enwiki/000_00000.body.f16bin", dtype="f16") |
| # matrix.shape == (rows_in_shard, 1024) |
| ``` |
|
|
| Or pull just one model's embeddings for a single language: |
|
|
| ```sh |
| hf download unum-cloud/USearchWiki \ |
| --repo-type dataset \ |
| --include "qwen3-embedding-0.6b/enwiki/*" |
| ``` |
|
|
| ### Workflow |
|
|
| The embedding pipeline is designed for multi-day runs on GPU servers with checkpoint/resume: |
|
|
| ```sh |
| # 1. Download FineWiki articles |
| python corpus.py --lang en --output corpus/ |
| |
| # 2. Embed with each model (resume-safe — rerun after interruptions) |
| python embed_articles.py --model qwen3-0.6b --input corpus/ --output embeddings/ --resume |
| python embed_articles.py --model e5-mistral-7b --input corpus/ --output embeddings/ --resume |
| python embed_articles.py --model arctic-embed-l-v2 --input corpus/ --output embeddings/ --resume |
| python embed_articles.py --model nomic-v1.5 --input corpus/ --output embeddings/ --resume |
| # Section-pooled ColBERT uses a different pipeline (late chunking) |
| python embed_sections.py --model gte-moderncolbert --input corpus/ --output embeddings/ --resume |
| |
| # 3. Extract graph metadata |
| python graph.py --lang en --output graph/ |
| |
| # 4. Compute ground truth for each model |
| python ground_truth.py --embeddings embeddings/qwen3-0.6b/ --k 100 --queries 10000 |
| python ground_truth.py --embeddings embeddings/e5-mistral-7b/ --k 100 --queries 10000 |
| |
| # 5. Upload to HuggingFace |
| python upload.py --repo unum-cloud/USearchWiki |
| ``` |
|
|
| Each step is idempotent. |
| Progress is tracked in `state/*.json` files — if a job dies (OOM, SSH drop, GPU error), rerunning the same command picks up from the last checkpoint. |
| Adding a new embedding model requires only step 2 + step 4 — the corpus and graph are shared. |
|
|
| ## Hosting |
|
|
| | Location | Storage/mo (1 TB) | Egress/GB | Notes | |
| | ---------------------------------------------------------------------------------- | ----------------- | --------- | ---------------------------------------------------------------- | |
| | [HuggingFace Hub](https://huggingface.co/unum-cloud/USearchWiki) | Free | Free | Primary. Xet storage, unlimited public downloads | |
| | [AWS S3](https://aws.amazon.com/s3/pricing/) Standard | $23.00 | $0.09 | S3-compatible mirror. Egress adds up fast for popular datasets | |
| | [Nebius Object Storage](https://docs.nebius.com/object-storage/resources/pricing/) | $15.05 | $0.015 | S3-compatible. ~35% cheaper storage, ~6× cheaper egress than AWS | |
|
|
| ## License |
|
|
| The embedding pipeline code in this repository is licensed under [Apache 2.0](LICENSE). |
|
|
| Dataset licensing depends on the components: |
|
|
| - __Wikipedia text__: [CC BY-SA 4.0](https://creativecommons.org/licenses/by-sa/4.0/) |
| - __FineWiki extraction__: [Apache 2.0](https://huggingface.co/datasets/HuggingFaceFW/finewiki) |
| - __Embeddings__: Governed by each model's license (see table above — all selected models use Apache 2.0 or MIT) |
| - __Graph metadata__: Derived from Wikimedia/Wikidata dumps ([CC0](https://creativecommons.org/publicdomain/zero/1.0/) for Wikidata, [CC BY-SA 4.0](https://creativecommons.org/licenses/by-sa/4.0/) for Wikipedia) |
|
|