Datasets:

Modalities:
Tabular
Text
Formats:
parquet
Libraries:
Datasets
Dask
License:
lotte-streams-for-murr / sampling.py
eugene-yang's picture
update readme
27e0d3b
import json
from pathlib import Path
import numpy as np
import pandas as pd
import scipy.stats as sts
import ir_datasets as irds
import ir_measures as irms
from tqdm import tqdm
domains = ['science', 'recreation', 'technology', 'lifestyle', 'writing']
# Define the distribution
domain_distributions = pd.concat({
"D1": pd.DataFrame({
domains[i]: [sts.distributions.betabinom(a=(i+1), b=(5-i), n=4).pmf(j) for j in range(5) ]
for i in range(5)
}),
"D2": pd.DataFrame({
domains[i]: ([sts.distributions.betabinom(a=1, b=10, n=5).pmf(j) for j in range(5) ]*2)[5-i:10-i]
for i in range(5)
}),
"D3": pd.DataFrame({
domains[i]: ([0]*i + [sts.distributions.betabinom(a=1, b=(5-i)*2, n=4-i).pmf(j) for j in range(5) ])[:5]
for i in range(5)
})
}, names=['dist_type', 'session'], axis=0).rename_axis('domains', axis=1)
sampling_cdf = domain_distributions.unstack('dist_type').cumsum(axis=0).pipe(lambda x: x/x.loc[4])
# Sampling for queries
np.random.seed(123)
fps = {
dt: { i: open(f'./stream_distribution_dss/queries/test_{dt}_{i}.jsonl', 'w') for i in sampling_cdf.index }
for dt in sampling_cdf.columns.get_level_values('dist_type').unique()
}
for dom in tqdm(domains, desc='queries'):
cdfs = sampling_cdf[dom]
for query in tqdm(irds.load(f'lotte/{dom}/test/forum').queries, desc=dom):
query = { 'query_id': dom+query.query_id, 'text': query.text, 'randval': np.random.random() }
belong = (query['randval'] < cdfs).agg(lambda x: x.tolist().index(True)).to_dict()
for dt, session in belong.items():
fps[dt][session].write(json.dumps(query) + '\n')
for l in fps.values():
for fp in l.values():
fp.close()
# Assigning matching qrels for the sampled queries
qrels_pool = [ q._replace(query_id=d+q.query_id, doc_id=d+q.doc_id) for d in domains for q in irds.load(f'lotte/{d}/test/forum').qrels ]
qrels_pool_grouped = {}
for q in qrels_pool:
if q.query_id not in qrels_pool_grouped:
qrels_pool_grouped[q.query_id] = []
qrels_pool_grouped[q.query_id].append(q)
for fn in tqdm(Path("./queries/").glob("*.jsonl"), desc='qrels'):
_, dt, i = fn.stem.split("_")
contains_query_ids = [ json.loads(q)['query_id'] for q in fn.open() ]
with open(f'./qrels/{fn.stem}.qrels', 'w') as fw:
for query_id in contains_query_ids:
for qr in qrels_pool_grouped[query_id]:
fw.write(f"{query_id} {qr.iteration} {qr.doc_id} {qr.relevance}\n")
doc_latest_appear_session = {
dt: {
d.doc_id: i
for i in range(5)
for d in irms.read_trec_qrels(f'./qrels/test_{dt}_{i}.qrels')
}
for dt in sampling_cdf.columns.get_level_values('dist_type').unique()
}
def _marginalize_cdf(cdf):
return cdf/cdf.iloc[-1]
# Sampling and assigning documents to sessions
np.random.seed(123)
doc_lists = {
dt: {i: [] for i in sampling_cdf.index}
for dt in sampling_cdf.columns.get_level_values('dist_type').unique()
}
for dom in tqdm(domains, desc='docs'):
cdfs = sampling_cdf[dom]
for doc in tqdm(irds.load(f'lotte/{dom}/test/forum').docs, desc=dom):
doc = {
'docid': dom+doc.doc_id, # for tevatron
'text': doc.text,
'randval': np.random.random(),
}
for dt in cdfs.columns:
appear_before = doc_latest_appear_session[dt].get(doc['docid'], 99)
session = (doc['randval'] < cdfs[dt].loc[:appear_before+1].pipe(_marginalize_cdf)).tolist().index(True)
doc_lists[dt][session].append({**doc, 'before': appear_before})
for dt, ll in doc_lists.items():
for i, l in ll.items():
pd.DataFrame(l).to_parquet(f'./docs/test_{dt}_{i}.parquet')