hc99's picture
Add files using upload-large-folder tool
011bd7a verified
from __future__ import annotations
import os
from mteb import MTEB
HEADER = "| Name | Hub URL | Description | Type | Category | #Languages | Train #Samples | Dev #Samples | Test #Samples | Avg. chars / train | Avg. chars / dev | Avg. chars / test"
SEP = "|:-----|:-----|:-----|:-----|:-----|-----:|-----:|-----:|-----:|-----:|-----:|-----:|"
ONE_LINE = "| {} | {} | {} | {} | {} | {} | {} | {} | {} | {} | {} | {} |"
TABLE_STRING = "\n".join([HEADER, SEP])
LEN_KEYS = {
"text",
"sentences",
"sentence1",
"sentence2",
"sent1",
"sent2" "query",
"positive",
"negative" "queries",
"corpus",
"machine_summaries",
"human_summaries",
}
DATAPATH = "/gpfsscratch/rech/six/commun/commun/experiments/muennighoff/mteb"
def load_data(hf_hub_name, subset=None):
"""Load dataset from Hub via cloning for easy offline usage with HF_DATASETS_OFFLINE=1
Can be replaced with just `load_dataset(hf_hub_name, subset)` if preferred
"""
from datasets import load_dataset
path = os.path.join(DATAPATH, hf_hub_name)
if os.path.exists(path):
dataset = load_dataset(path, subset)
else:
from git import Repo
Repo.clone_from("https://huggingface.co/datasets/" + hf_hub_name, path)
dataset = load_dataset(path, subset)
return dataset
def get_ds_stats_beir_hub(hf_hub_name):
"""Not used as some BEIR datasets are still missing on the Hub"""
lens = {}
for subset in ["corpus", "queries"]:
ds = load_data("mteb/hfbeir" + hf_hub_name.replace("BeIR", ""), subset)
splits = list(ds.keys())
len_keys = set(ds[splits[-1]].features.keys()) & LEN_KEYS
for split in splits:
if split not in ds:
continue
lens.setdefault(split, [])
for k in len_keys:
if isinstance(ds[split][k][0], str):
lens[split] += [len(x) for x in ds[split][k]]
elif isinstance(ds[split][k][0], list):
assert isinstance(ds[split][k][0][0], str), f"Too nested: {k}"
lens[split] += [len(y) for x in ds[split][k] for y in x]
else:
raise ValueError(f"Unknown type {type(ds[split][k])}")
all_lens = [x for y in lens.values() for x in y]
avg_len = sum(all_lens) / len(all_lens)
return ["TODO"] * 3 + [round(avg_len, 1)] * 3
def get_ds_stats_beir(hf_hub_name):
from beir.datasets.data_loader import GenericDataLoader as BeirDataLoader
path = os.path.join(DATAPATH, hf_hub_name)
if not os.path.exists(path):
from beir import util
url = f"https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/{hf_hub_name}.zip"
util.download_and_unzip(url, DATAPATH)
lens = {"train": [], "dev": [], "test": []}
for split in lens.keys():
try:
corpus, queries, relevant_docs = BeirDataLoader(path).load(split=split)
except: # split does not exist # noqa: E722
continue
# + 1 for space added between Title & Text by default in BEIR
avg_lens_c = [len(v["text"]) + len(v["title"]) + 1 for v in corpus.values()]
avg_lens_q = [len(v) for v in queries.values()]
lens[split].extend(avg_lens_c)
lens[split].extend(avg_lens_q)
avg_lens = {
k: round(sum(lens[k]) / len(lens[k]), 1) if lens[k] else 0 for k in lens
}
return (
len(lens["train"]),
len(lens["dev"]),
len(lens["test"]),
avg_lens["train"],
avg_lens["dev"],
avg_lens["test"],
)
def get_ds_stats(hf_hub_name):
ds = load_data(hf_hub_name)
assert "test" in ds, f"No test set for {hf_hub_name}"
len_keys = set(ds["test"].features.keys()) & LEN_KEYS
dev_key = "dev" if "dev" in ds else "validation"
lens = {"train": [], dev_key: [], "test": []}
for split in lens.keys():
if split not in ds:
continue
for k in len_keys:
if isinstance(ds[split][k][0], str):
lens[split] += [len(x) for x in ds[split][k]]
elif isinstance(ds[split][k][0], list):
assert isinstance(ds[split][k][0][0], str), f"Too nested: {k}"
lens[split] += [len(y) for x in ds[split][k] for y in x]
else:
raise ValueError(f"Unknown type {type(ds[split][k])}")
avg_lens = {
k: round(sum(lens[k]) / len(lens[k]), 1) if lens[k] else 0 for k in lens
}
return (
len(lens["train"]),
len(lens[dev_key]),
len(lens["test"]),
avg_lens["train"],
avg_lens[dev_key],
avg_lens["test"],
)
# Select all tasks
for task in MTEB().tasks:
print("Task: ", task)
if "dataset" in task.metadata_dict:
hub_name = hub_url = task.metadata_dict["dataset"]["path"]
ds_stats = get_ds_stats(hub_name.split("/")[-1])
elif "beir_name" in task.metadata_dict:
hub_name = hub_url = "BeIR/" + task.metadata_dict.get("beir_name")
ds_stats = get_ds_stats_beir("/".join(hub_name.split("/")[1:]))
if "cqadupstack" in hub_name:
hub_url = "BeIR/cqadupstack-qrels"
TABLE_STRING += "\n" + ONE_LINE.format(
f"[{task.metadata_dict['name']}]({task.metadata_dict['reference']})",
f"[{hub_name}](https://huggingface.co/datasets/{hub_url})",
task.metadata_dict["description"],
task.metadata_dict["type"],
task.metadata_dict["category"],
len(task.metadata_dict["eval_langs"]),
*ds_stats,
)
with open("./mdtable.md", "w") as f:
f.write(TABLE_STRING)
# Convert to latex
for line in TABLE_STRING.split("\n")[2:]:
if line:
cols = line.split(" | ")
idx = cols[0].index("]")
cols[0] = cols[0][3:idx]
cols[-1] = cols[-1][:-1]
out = " & ".join(cols[:1] + cols[3:]) + " \\\\"
print(out)