Dataset Viewer
Auto-converted to Parquet Duplicate
query-id
string
corpus-id
string
score
int64
query_0
Nelia Lezak
1
query_0
Mansfield Dekle
1
query_1
Jerimy Kopstein
1
query_1
Harlan Sommerer
1
query_2
Deane Bouy
1
query_2
Tal Mejiahernandez
1
query_3
Lynnette Trundy
1
query_3
Tracey Copeman
1
query_4
Tollie Vanwoert
1
query_4
Asha Spohr
1
query_5
Isai Robel
1
query_5
Shanta Lester
1
query_6
Shaquille Birkby
1
query_6
Katelynn Scovel
1
query_7
Reyna Kurti
1
query_7
Maribeth Gentzler
1
query_8
Sie Desilus
1
query_8
Kesha Carrizal
1
query_9
Wm Keester
1
query_9
Alla Conlon
1
query_10
Laureen Presha
1
query_10
Mattie Poret
1
query_11
Ishmael Hinmon
1
query_11
Del Vaculin
1
query_12
Victorine Monnahan
1
query_12
Shanta Lester
1
query_13
Melodie Eades
1
query_13
Ike Kovalenko
1
query_14
Almyra Kureshi
1
query_14
Rico Klock
1
query_15
Julie Rosensteel
1
query_15
Machelle Audi
1
query_16
Jaylah Favour
1
query_16
Nyla Pratts
1
query_17
Janel Rowin
1
query_17
Les Dinn
1
query_18
Dionicio Folkerth
1
query_18
Eduardo Dunkley
1
query_19
Norine Dazzo
1
query_19
Che Roskilly
1
query_20
Billy Strenger
1
query_20
Granville Gaerte
1
query_21
Thomas Delatore
1
query_21
Nonie Venugopal
1
query_22
Angelina Maksim
1
query_22
Charline Vinogradov
1

LIMIT-small-random

LIMIT-small-random is a self-produced dataset created for the reproduction of the paper On the Theoretical Limitations of Embedding-Based Retrieval.
The reproduction codebase is available at https://github.com/gabor-hosu/embedding_dimension_limit.

The dataset is derived from the name and attribute distribution of the original LIMIT dataset, while following the same underlying construction principles described in the paper.

Due to hardware constraints, the dataset size was scaled down compared to the original experimental setup. The released version contains 2,000 documents and 23 queries.

Dataset Generation

The dataset was generated using the following procedure.

First, a random binary relevance matrix is constructed, defining which documents are relevant to which queries:

import numpy as np
from itertools import combinations
import random

def random_matrix(num_of_queries: int, num_of_docs: int, k: int = 2, seed: int = 42) -> np.array:
    random.seed(seed)
    all_indexes = np.arange(num_of_docs)
    combos_sample = random.sample(list(combinations(all_indexes, k)), num_of_queries)

    A = np.zeros((num_of_queries, num_of_docs), dtype=bool)
    for row, combo in zip(range(num_of_queries), combos_sample):
        A[row, combo] = True

    return A

Next, the binary relevance structure is converted into a natural-language corpus, queries, and relevance judgments following the MTEB format:

import pandas as pd
import random

def generate_dataset(
    liked_items: list[str],
    names: list[str],
    qrel_matrix: np.ndarray,
    items_per_person: int = 20,
    total_num_of_docs: int = 2000,
    seed: int = 42,
):
    num_of_queries, num_of_docs = qrel_matrix.shape
    random.seed(seed)

    query_items = random.sample(liked_items, num_of_queries)
    remaining_items = list(set(liked_items) - set(query_items))
    
    doc_ids = np.array(random.sample(names, num_of_docs))
    remaining_doc_ids = list(set(names) - set(doc_ids))
    
    docs = {}
    qrels_data = []
    
    # fill up the binary qrel structure with natural language
    for query_idx, (mask, item) in enumerate(zip(qrel_matrix, query_items)):
        selected_doc_ids = doc_ids[mask]
        for doc_id in selected_doc_ids:
            doc = docs.get(doc_id)
            if doc is None:
                docs[doc_id] = []
                doc = docs[doc_id]

            doc.append(item)
            qrels_data.append({
                "query-id": f"query_{query_idx}",
                "corpus-id": doc_id,
                "score": 1
            })
    
    # add remaining items to the docs
    for doc_id in docs:
        num_new_items_per_docs = items_per_person - len(docs[doc_id])
        new_items = random.sample(remaining_items, num_new_items_per_docs)
        docs[doc_id].extend(new_items)
        
    num_new_docs = total_num_of_docs - len(docs)
    if num_new_docs > 0:
        new_doc_ids = random.sample(remaining_doc_ids, num_new_docs)
        docs |= {
            doc_id: random.sample(remaining_items, items_per_person)
            for doc_id in new_doc_ids
        }
    
    # build and return the proper mteb format
    corpus = pd.DataFrame(
        [
            {
                "_id": doc_id,
                "title": "",
                "text": f"{doc_id} likes {', '.join(random.sample(docs[doc_id], len(docs[doc_id])))}."
            }
            for doc_id in docs
        ]
    )

    queries = pd.DataFrame(
        [
            {
                "_id": f"query_{query_idx}",
                "text": f"Who likes {item}?"
            }
            for query_idx, item in enumerate(query_items)
        ]
    )
    
    qrels = pd.DataFrame(qrels_data)
    
    return corpus, queries, qrels
A = random_matrix(num_of_queries=23, num_of_docs=2000)

corpus, queries, qrels = generate_dataset(
    liked_items=liked_items,
    names=names,
    qrel_matrix=A,
)

corpus.to_json("corpus.jsonl", orient="records", lines=True)
queries.to_json("queries.jsonl", orient="records", lines=True)
qrels.to_json("qrels.jsonl", orient="records", lines=True)

The resulting dataset consists of a synthetic natural-language corpus, corresponding queries, and dense relevance judgments designed to stress-test embedding-based retrieval under constrained dimensionality.

Downloads last month
24

Collection including gabor-hosu/LIMIT-small-random

Paper for gabor-hosu/LIMIT-small-random