from __future__ import annotations import os import pandas as pd import requests from huggingface_hub import create_repo, upload_file MAX_OUTPUT = 10000 NB_RESULTS = 100000 ORGANIZATION = "lyon-nlp/" REPO_NAME = "clustering-hal-s2s" SAVE_PATH = "test.jsonl" df_papers = pd.DataFrame(columns=["hal_id", "title", "domain"]) start_index = 0 while start_index < NB_RESULTS: response = requests.request( "GET", f"https://api.archives-ouvertes.fr/search/?q=*:*&wt=json&fl=halId_s,title_s,level0_domain_s&fq=language_s:fr&fq=submittedDateY_i:[2000%20TO%20*]&rows={MAX_OUTPUT}&start={start_index}", ) if "response" in response.json(): papers = response.json()["response"]["docs"] for paper in papers: if ("title_s" in paper) and ("level0_domain_s" in paper): paper_info = { "hal_id": paper["halId_s"], "title": paper["title_s"][0], "domain": paper["level0_domain_s"][0], } df_papers = pd.concat( [df_papers, pd.DataFrame([paper_info])], ignore_index=True ) start_index += MAX_OUTPUT df_papers = df_papers.drop_duplicates() df_papers.to_json(SAVE_PATH, orient="records", lines=True) create_repo(ORGANIZATION + REPO_NAME, repo_type="dataset") upload_file( path_or_fileobj=SAVE_PATH, path_in_repo=SAVE_PATH, repo_id=ORGANIZATION + REPO_NAME, repo_type="dataset", ) os.system(f"rm {SAVE_PATH}")