SuperWiki-1.5 / Scripts /RojaStringRemixer.py
KaraKaraWitch's picture
Inital Commit
365cd4e
import multiprocessing
import pathlib
import re
import time
import traceback
import orjson
import tqdm
import typer
from bs4 import BeautifulSoup, Tag
from markdownify import MarkdownConverter, chomp
CONCURRENT = 64
class WikiConverter(MarkdownConverter):
def convert_a(self, el, text, convert_as_inline):
prefix, suffix, text = chomp(text)
if not text:
return ""
return "%s%s%s" % (prefix, text, suffix)
integer_rgx = re.compile("^[0-9]*$")
@staticmethod
def is_intable(string: str):
if not string or not string.isdigit():
return False
if WikiConverter.integer_rgx.match(string):
return True
def convert_img(self, el, text, convert_as_inline):
convert_as_inline = True
alt = el.attrs.get("alt", None) or ""
if (
convert_as_inline
and el.parent.name not in self.options["keep_inline_images_in"]
):
return alt
return alt
def convert_li(self, el, text, convert_as_inline):
parent = el.parent
if parent is not None and parent.name == "ol":
start = parent.get("start")
if start and WikiConverter.is_intable(start.strip()):
start = int(start.strip())
else:
start = 1
bullet = "%s." % (start + parent.index(el))
else:
depth = -1
while el:
if el.name == "ul":
depth += 1
el = el.parent
bullets = self.options["bullets"]
bullet = bullets[depth % len(bullets)]
return "%s %s\n" % (bullet, (text or "").strip())
class MultilangWikipediaProcessor:
def __init__(self) -> None:
self.md = WikiConverter()
def is_stub(self, soup: BeautifulSoup):
for plainlinks in soup.select(".metadata.plainlinks"):
if "stub" in plainlinks.get("id", "") or "stub" in plainlinks.get(
"class", []
):
return True
return False
def rital_ambox(self, input_soup: BeautifulSoup):
ambox_classes = []
for ambox in selects:
if ambox is not None:
ambox_classes.append(ambox.get("class"))
for ambox in selects:
ambox.decompose()
return input_soup, ambox_classes
def rital_ombox(self, input_soup: BeautifulSoup):
ombox_classes = []
selects = input_soup.select('table[class~="ombox"]')
for ombox in selects:
if ombox is not None:
ombox_classes.append(ombox["class"])
for ombox in selects:
ombox.decompose()
return input_soup, ombox_classes
def table_filtration(self, input_soup: BeautifulSoup, title):
for table in input_soup.select("table"):
tds = len(table.find_all("td"))
texsize = len(table.get_text().replace(" ", ""))
if tds >= texsize and texsize < 50:
print(table.get_text().replace(" ", ""))
print("Removing table from", title, ". TD exceeds Content")
return input_soup
all_selectors = [
"style", # Remove styling
"sup.reference", # Seems to still exist across
"table.nomobile", # Seems to still exist across
]
def process_infobox(self, infobox: Tag):
return str(infobox)
def process_figures(self, figure: Tag):
figure_data = {}
fig_a = figure.find("a")
fig_cap = figure.find("figcaption")
if fig_a:
figure_data["file_url"] = fig_a.get("href", None)
else:
figure_data["file_url"] = None
if fig_cap:
figure_data["caption"] = fig_cap.get_text()
else:
figure_data["caption"] = None
if figure_data["caption"] == figure_data["file_url"] is None:
return None
return figure_data
def convert_soup(self, input_soup: BeautifulSoup):
[i.unwrap() for i in input_soup.select('[data-mw^="interface"]')]
[i.decompose() for i in input_soup.select(", ".join(self.all_selectors))]
title = input_soup.select_one("title").extract()
# titletext = title.get_text()
for i in input_soup.select(".mw-collapsible"):
hidden = i.select_one("div.hidden-content")
if hidden:
# Expose collapsed content
hidden["class"].remove("hidden-content")
for i in input_soup.select("[data-mw]"):
i["data-mw"] = ""
ifbs = [i.extract() for i in input_soup.select("table.infobox")]
ifbs += [i.extract() for i in input_soup.select("table.sidebar.vcard.hlist")]
ifbs += [i.extract() for i in input_soup.select("table.infobox.vcard")]
ifbs = [self.process_infobox(ifb) for ifb in ifbs]
figures = [
self.process_figures(fig.extract())
for fig in input_soup.select('figure[typeof^="mw:File/Thumb"]')
]
return input_soup, ifbs, figures, title
def convert(self, wiki_data: bytes):
data = orjson.loads(wiki_data.rstrip(b"\n"))
try:
templates = [
":".join(template["name"].split(":")[1:])
for template in data.get("templates", [])
]
categories = [
":".join(category["name"].split(":")[1:])
for category in data.get("categories", [])
]
if not data["article_body"].get("wikitext"):
return None
soup = BeautifulSoup(data["article_body"]["html"], "lxml")
is_stub = self.is_stub(soup)
soup, infobox, figures, title = self.convert_soup(soup)
# soup, issues, issue_selectors = self.remove_templates(soup, all_templates)
soup, amboxes = self.rital_ambox(soup)
soup, omboxes = self.rital_ombox(soup)
soup = self.table_filtration(soup, title)
text = (
self.md.convert_soup(soup)
.strip()
.replace("\n\n", "\n")
.replace("\n\n\n", "\n\n")
)
return orjson.dumps(
{
"id": data["identifier"],
"title": data["name"],
"url": data["url"],
"stub": is_stub,
"template": templates,
"category": categories,
"license": [lic["name"] for lic in data["license"]],
"text": text,
"wikitext": data["article_body"].get("wikitext"),
"lang": data["in_language"]["identifier"],
"abstract": data.get("abstract", ""),
"boxes_filters": amboxes + omboxes,
"infobox_html": infobox,
"figures_dict": figures,
}
)
except Exception as e:
print(f"Exception at Soup Conversion: {e} [{data['name']}]")
return None
queue = multiprocessing.Queue(maxsize=1024 * 1024)
def worker(worker_idx: int, prefix_path: pathlib.Path):
if prefix_path.parent:
prefix_path.parent.mkdir(exist_ok=True, parents=True)
processor = MultilangWikipediaProcessor()
prefix_path = prefix_path.with_name(
f"{prefix_path.name}-{str(worker_idx).zfill(2)}.jsonl"
)
with open(prefix_path, "wb") as f:
while True:
data = queue.get()
if data is None:
break
# print(data[:16])
parsed = processor.convert(data)
if parsed:
f.write(parsed)
f.write(b"\n")
app = typer.Typer()
def err_cb(err: Exception):
print(traceback.format_exception(err)[0])
@app.command()
def main(folder: pathlib.Path, prefix_path: pathlib.Path, processes: int = 64):
# orjson.loads()
with multiprocessing.Pool(processes=processes) as pool:
workers = [
pool.apply_async(
worker, args=(worker_idx, prefix_path), error_callback=err_cb
)
for worker_idx in range(processes)
]
for file in pathlib.Path(folder).glob("*.ndjson"):
with open(file, "rb") as fp:
pbar = tqdm.tqdm(desc=f"{file.name}")
for line in fp:
# print(line[:16])
queue.put(line)
pbar.update(1)
print("processed", file.name)
pbar.close()
for _ in range(processes):
queue.put(None)
while not queue.empty():
print("Waiting for empty queue")
time.sleep(5)
print("Queue is empty. Waiting for threads to finish.")
patience = 120
# processor_thr = 0.05
while True:
workers_done = 0
for running_worker in workers:
if running_worker.ready():
workers_done += 1
if workers_done == processes:
break
if patience <= 0:
# Screw it.
pool.terminate()
break
else:
if workers_done > processes // 1.5:
print("Waiting for dangling processes to complete.")
time.sleep(10)
patience -= 1
# pool.join()
if __name__ == "__main__":
app()