File size: 1,512 Bytes
011bd7a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
from __future__ import annotations

import os

import pandas as pd
import requests
from huggingface_hub import create_repo, upload_file

MAX_OUTPUT = 10000
NB_RESULTS = 100000
ORGANIZATION = "lyon-nlp/"
REPO_NAME = "clustering-hal-s2s"
SAVE_PATH = "test.jsonl"

df_papers = pd.DataFrame(columns=["hal_id", "title", "domain"])


start_index = 0
while start_index < NB_RESULTS:
    response = requests.request(
        "GET",
        f"https://api.archives-ouvertes.fr/search/?q=*:*&wt=json&fl=halId_s,title_s,level0_domain_s&fq=language_s:fr&fq=submittedDateY_i:[2000%20TO%20*]&rows={MAX_OUTPUT}&start={start_index}",
    )
    if "response" in response.json():
        papers = response.json()["response"]["docs"]
        for paper in papers:
            if ("title_s" in paper) and ("level0_domain_s" in paper):
                paper_info = {
                    "hal_id": paper["halId_s"],
                    "title": paper["title_s"][0],
                    "domain": paper["level0_domain_s"][0],
                }
                df_papers = pd.concat(
                    [df_papers, pd.DataFrame([paper_info])], ignore_index=True
                )
    start_index += MAX_OUTPUT

df_papers = df_papers.drop_duplicates()
df_papers.to_json(SAVE_PATH, orient="records", lines=True)

create_repo(ORGANIZATION + REPO_NAME, repo_type="dataset")

upload_file(
    path_or_fileobj=SAVE_PATH,
    path_in_repo=SAVE_PATH,
    repo_id=ORGANIZATION + REPO_NAME,
    repo_type="dataset",
)
os.system(f"rm {SAVE_PATH}")