| | import json, re |
| | import requests |
| | from urlextract import URLExtract |
| | import sys, gzip |
| |
|
| |
|
| | utid = 'snidiff1' |
| | base= { 'model':'https://huggingface.co/', 'data': 'https://huggingface.co/datasets/', 'source': 'https://' } |
| | post = '/raw/main/README.md' |
| | postGH = '/blob/master/README.md' |
| | postGHalt = '/blob/main/README.md' |
| |
|
| | extU = URLExtract() |
| | DOIpattern = r'\b(10\.\d{4,9}\/[-._;()/:A-Z0-9]+)\b/i' |
| | |
| |
|
| | def extractURLs (c): |
| | res = extU.find_urls (c) |
| | return res |
| |
|
| | def extractDOIs (c): |
| | res = re.findall (DOIpattern, c) |
| | return res |
| |
|
| | fo = gzip.open(f"output/{utid}.json.gz", 'w') |
| |
|
| | def run (tp): |
| | post0 = post |
| | |
| | with open(f"input/{utid}_{tp}", 'r', errors='ignore') as f: |
| | for line in f: |
| | line = line.strip () |
| | if tp == 'source': |
| | (npapers,line) = line.split(';') |
| | post0 = postGH |
| | print(line) |
| |
|
| | url = base[tp] + f"{line}{post0}" |
| | print(url) |
| | r = requests.get (url) |
| | |
| | |
| | if r.status_code < 200 or r.status_code > 299: |
| | print("error, trying main") |
| | |
| | |
| | url = base[tp] + f"{line}{postGHalt}" |
| | print(url) |
| | r = requests.get (url) |
| | if r.status_code < 200 or r.status_code > 299: |
| | print("error code returned") |
| | continue |
| | |
| | content = r.text |
| | urls = extractURLs(content) |
| | dois = extractDOIs(content) |
| | res = { 'ID': line, 'type': tp, 'url': url, 'content': content, 'links': urls, 'dois': dois } |
| | out = json.dumps(res, ensure_ascii=False) |
| | fo.write((out+"\n").encode()) |
| |
|
| | run('model') |
| | run('data') |
| | run('source') |
| |
|
| | fo.close() |