nixos-wiki-infra/checks/linkcheck/main.py
2024-09-17 10:57:42 +02:00

139 lines
4.7 KiB
Python

import json
import csv
import re
import sys
import argparse
import xml.etree.ElementTree as ET
def get_revision_timestamp(revision: ET.Element, ns: dict[str, str]) -> str:
timestamp = revision.find("mw:timestamp", ns)
if timestamp is None:
print(
f"Timestamp tag not found in revision: {ET.tostring(revision)}",
file=sys.stderr,
)
return ""
if timestamp.text is None:
print(
f"Timestamp text doesn't exist in revision: {ET.tostring(revision)}",
file=sys.stderr,
)
return ""
return timestamp.text
# filter out unimportant pages like Talk:, User:, and old revisions of posts
def process_dump( args: argparse.Namespace) -> None:
tree = ET.parse(args.dump_file)
root = tree.getroot()
ns = {"mw": "http://www.mediawiki.org/xml/export-0.11/"}
ET.register_namespace("", ns["mw"])
for page in root.findall("mw:page", ns):
title_tag = page.find("mw:title", ns)
if title_tag is None:
print(f"Title tag not found in page: {ET.tostring(page)}", file=sys.stderr)
continue
title = title_tag.text
if title is None:
print(
f"Title text doesn't exist in page: {ET.tostring(page)}",
file=sys.stderr,
)
continue
if title.startswith("User:") or title.startswith("Talk:"):
root.remove(page)
continue
revisions = page.findall("mw:revision", ns)
if len(revisions) > 1:
latest_revision = max(
revisions, key=lambda revison: get_revision_timestamp(revison, ns)
)
# Remove all revisions except the latest one
for revision in revisions:
if revision != latest_revision:
page.remove(revision)
tree.write(args.out_file, encoding="utf-8", xml_declaration=False)
def badlinks_print(args: argparse.Namespace) -> None:
# known_file: str, outfile: str) -> None:
with open(args.known_file, "r") as infile, open(args.outfile, "w") as of:
for line in infile:
stripped_line = line.strip()
if stripped_line and not stripped_line.startswith("#"):
of.write(f"--exclude {stripped_line} ")
def dump_link_map(args: argparse.Namespace) -> None:
with open(args.jsonfile, "r") as json_file:
fail_data = json.load(json_file)
with open(args.dumpfile, mode="w", newline="", encoding="utf-8") as csv_file:
csv_writer = csv.writer(csv_file, delimiter="\t", quotechar='"')
csv_writer.writerow(["STATUS", "URL", "WIKIURL"])
for xml_file, failed_url_entries in fail_data["fail_map"].items():
with open(xml_file, "r", encoding="utf-8") as xmlf:
root = ET.fromstring(f"<root>{xmlf.read()}</root>")
for doc in root.findall("doc"):
title = doc.attrib.get("title")
if title is None:
print(
f"Title not found in doc: {ET.tostring(doc)}", file=sys.stderr
)
continue
title = re.sub(r"\s+", "_", title)
content = doc.text
for entry in failed_url_entries:
url = entry["url"]
status = entry.get("status", {}).get("code", 403)
if url in content:
csv_writer.writerow(
[
status,
url,
f"https://wiki.nixos.org/wiki/{title}",
]
)
def main() -> None:
parser = argparse.ArgumentParser(description="Process wiki dump files")
subparsers = parser.add_subparsers()
parser_filter = subparsers.add_parser("filter", help="Filter out unimportant pages")
parser_filter.add_argument("dump_file", type=str)
parser_filter.add_argument("out_file", type=str)
parser_filter.set_defaults(func=process_dump)
parser_badlinks = subparsers.add_parser(
"badlinks", help="Parse and print known allowed links"
)
parser_badlinks.add_argument("known_file", type=str)
parser_badlinks.add_argument("out_file", type=str)
parser_badlinks.set_defaults(func=badlinks_print)
parser_dumplinkmap = subparsers.add_parser(
"dumplinkmap", help="Dump a map of url and nixos article where it is present"
)
parser_dumplinkmap.add_argument("jsonfile", type=str)
parser_dumplinkmap.add_argument("dumpfile", type=str)
parser_dumplinkmap.set_defaults(func=dump_link_map)
args = parser.parse_args()
args.func(args)
if __name__ == "__main__":
main()