Datasets:
License:
| import multiprocessing | |
| import pathlib | |
| import traceback | |
| from bs4 import BeautifulSoup | |
| import orjson | |
| import unidecode | |
| from RojaStringRemixer import MultilangWikipediaProcessor | |
| mp = MultilangWikipediaProcessor() | |
| class WikitonaryProcessor(MultilangWikipediaProcessor): | |
| def soup_section(self, input_soup: BeautifulSoup, title:str): | |
| # Wiktionary is typically single layered | |
| languages = [] | |
| lang_buffer = [] | |
| sub_cat_buffer = [] | |
| # print(input_soup) | |
| # for html_element in | |
| for sect in input_soup.select(".mw-body-content section"): | |
| sect.unwrap() | |
| [meta.decompose() for meta in input_soup.select(".mw-body-content meta")] | |
| [meta.decompose() for meta in input_soup.select(".mw-body-content link")] | |
| [meta.decompose() for meta in input_soup.select(".disambig-see-also")] | |
| # print(input_soup.selec("section")) | |
| # print([i.name for i in list(input_soup.select_one(".mw-body-content").children)]) | |
| for html_element in input_soup.select_one(".mw-body-content").children: | |
| if html_element.decomposed or html_element is None: | |
| continue | |
| if html_element.name == "h1": | |
| print("[?] Expecting lowest header element to be h1") | |
| continue | |
| if html_element.name == "h2": | |
| # lang_buffer.append(sub_cat_buffer) | |
| if sub_cat_buffer: | |
| lang_buffer.append(sub_cat_buffer) | |
| sub_cat_buffer = [self.md.convert_soup(html_element).strip()] | |
| else: | |
| sub_cat_buffer = [self.md.convert_soup(html_element).strip()] | |
| if lang_buffer: | |
| languages.append(lang_buffer) | |
| lang_buffer = [] | |
| elif html_element.name == "h3": | |
| lang_buffer.append(sub_cat_buffer) | |
| sub_cat_buffer = [self.md.convert_soup(html_element).strip()] | |
| else: | |
| if html_element.get_text().strip(): | |
| sub_cat_buffer.append(self.md.convert_soup(html_element).strip()) | |
| if sub_cat_buffer: | |
| lang_buffer.append(sub_cat_buffer) | |
| if lang_buffer: | |
| languages.append(lang_buffer) | |
| # print(languages) | |
| final_txt = {} | |
| for root_lang in languages: | |
| lang_data = [] | |
| for sub_pair in root_lang[1:]: | |
| concat = " \n".join(sub_pair[1:]) | |
| lang_data.append(f"## {sub_pair[0]}\n\n{concat}") | |
| lang_final = f"# {title.title()}: {root_lang[0][0]}\n\n" + "\n\n".join(lang_data) | |
| txt = unidecode.unidecode(lang_final.replace("“","\"").replace("”","\"").replace("\t"," ")).encode('utf-8', errors='replace').decode('unicode_escape', errors='ignore') | |
| final_txt[root_lang[0][0].lower()] = txt | |
| # print(final_txt) | |
| return final_txt | |
| all_selectors = [ | |
| "style", # Remove styling | |
| "sup.reference", # Seems to still exist across | |
| "table.nomobile", # Seems to still exist across | |
| "div.sister-wikipedia", | |
| ".floatright", | |
| ] | |
| def convert_soup(self, input_soup: BeautifulSoup): | |
| # Remove Generic notices | |
| [i.unwrap() for i in input_soup.select('[data-mw^="interface"]')] | |
| [i.decompose() for i in input_soup.select(", ".join(self.all_selectors))] | |
| [i.decompose() for i in input_soup.select("[class*=\"NavHead\"]")] | |
| title = input_soup.select_one("title").extract() | |
| for i in input_soup.select(".mw-collapsible"): | |
| hidden = i.select_one("div.hidden-content") | |
| if hidden: | |
| # Expose collapsed content | |
| hidden["class"].remove("hidden-content") | |
| # Cleanup meta | |
| for i in input_soup.select("[data-mw]"): | |
| del i["data-mw"] | |
| for i in input_soup.select("[rel]"): | |
| del i["rel"] | |
| for i in input_soup.select("link[href]"): | |
| del i["href"] | |
| for i in input_soup.select("link[typeof]"): | |
| del i["typeof"] | |
| for i in input_soup.select("[id]"): | |
| del i["id"] | |
| for i in input_soup.select("[about]"): | |
| del i["about"] | |
| for i in input_soup.select("[lang]"): | |
| del i["lang"] | |
| ifbs = [i.extract() for i in input_soup.select("table.infobox")] | |
| ifbs += [i.extract() for i in input_soup.select("table.sidebar.vcard.hlist")] | |
| ifbs += [i.extract() for i in input_soup.select("table.infobox.vcard")] | |
| ifbs = [self.process_infobox(ifb) for ifb in ifbs] | |
| [fig.decompose ()for fig in input_soup.select('figure[typeof^="mw:File/Thumb"]')] | |
| return input_soup, ifbs, None, title | |
| def convert(self, wiki_data: bytes): | |
| data = orjson.loads(wiki_data.rstrip(b"\n")) | |
| try: | |
| templates = [ | |
| ":".join(template["name"].split(":")[1:]) | |
| for template in data.get("templates", []) | |
| ] | |
| categories = [ | |
| ":".join(category["name"].split(":")[1:]) | |
| for category in data.get("categories", []) | |
| ] | |
| if not data["article_body"].get("wikitext"): | |
| return None | |
| soup = BeautifulSoup(data["article_body"]["html"], "lxml") | |
| is_stub = self.is_stub(soup) | |
| soup, infobox, figures, title = self.convert_soup(soup) | |
| sections = self.soup_section(soup, data["name"]) | |
| text = list(sections.values())[0][2:] | |
| return orjson.dumps( | |
| { | |
| "id": data["identifier"], | |
| "title": data["name"], | |
| "url": data["url"], | |
| "text": text, | |
| "stub": is_stub, | |
| "template": templates, | |
| "category": categories, | |
| "license": [lic["name"] for lic in data["license"]], | |
| "wikitext": data["article_body"].get("wikitext"), | |
| "lang": data["in_language"]["identifier"], | |
| "abstract": data.get("abstract", ""), | |
| "sections": sections, | |
| "infobox_html": infobox, | |
| "figures_dict": figures, | |
| } | |
| ) | |
| except Exception as e: | |
| print(f"Exception at Soup Conversion: {e} [{data['name']}]\n{traceback.format_exception(e)[-1]}") | |
| return None | |
| mp = WikitonaryProcessor() | |
| fs = [ | |
| pathlib.Path(f"wiktionary/enwiktionary_namespace_0_{i}.ndjson") for i in range(16) | |
| ] | |
| # for line in pathlib.Path("") | |
| def main(): | |
| with open("en-wiktionary.jsonl","wb") as fout: | |
| with multiprocessing.Pool(32) as pool: | |
| for file in fs: | |
| tasks = [] | |
| with open(file,"rb") as f: | |
| for line in f: | |
| tasks.append(pool.apply_async(mp.convert,(line,))) | |
| if len(tasks) % 5000 == 0: | |
| for task in tasks: | |
| r = task.get() | |
| if r is None: | |
| continue | |
| fout.write(r + b"\n") | |
| tasks = [] | |
| for task in tasks: | |
| r = task.get() | |
| if r is None: | |
| continue | |
| fout.write(r + b"\n") | |
| # mp.convert(pathlib.Path("misbriefs.json").read_bytes()) | |
| main() |