#!/usr/bin/python3 import sys, re import bz2, html, urllib.parse import sqlite3 usageInfo = f""" Usage: {sys.argv[0]} For some set of page IDs, looks up their content in the wiki dump, and tries to parse infobox image names, storing them into a database. The program can be re-run with an updated set of page IDs, and will skip already-processed page IDs. """ if len(sys.argv) > 1: print(usageInfo, file=sys.stderr) sys.exit(1) def getInputPageIds(): pageIds = set() dbCon = sqlite3.connect("../data.db") dbCur = dbCon.cursor() for (pageId,) in dbCur.execute("SELECT id from wiki_ids"): pageIds.add(pageId) dbCon.close() return pageIds dumpFile = "enwiki-20220501-pages-articles-multistream.xml.bz2" indexDb = "dumpIndex.db" imgDb = "imgData.db" # The database to create idLineRegex = re.compile(r"(.*)") imageLineRegex = re.compile(r".*\| *image *= *([^|]*)") bracketImageRegex = re.compile(r"\[\[(File:[^|]*).*]]") imageNameRegex = re.compile(r".*\.(jpg|jpeg|png|gif|tiff|tif)", flags=re.IGNORECASE) cssImgCropRegex = re.compile(r"{{css image crop\|image *= *(.*)", flags=re.IGNORECASE) # In testing, got about 360k image names print("Getting input page-ids") pageIds = getInputPageIds() print(f"Found {len(pageIds)}") print("Opening databases") indexDbCon = sqlite3.connect(indexDb) indexDbCur = indexDbCon.cursor() imgDbCon = sqlite3.connect(imgDb) imgDbCur = imgDbCon.cursor() print("Checking tables") if imgDbCur.execute("SELECT name FROM sqlite_master WHERE type='table' AND name='page_imgs'").fetchone() == None: # Create tables if not present imgDbCur.execute("CREATE TABLE page_imgs (page_id INT PRIMARY KEY, img_name TEXT)") # img_name may be NULL imgDbCur.execute("CREATE INDEX page_imgs_idx ON page_imgs(img_name)") else: # Check for already-processed page IDs numSkipped = 0 for (pid,) in imgDbCur.execute("SELECT page_id FROM page_imgs"): if pid in pageIds: pageIds.remove(pid) numSkipped += 1 else: print(f"WARNING: Found already-processed page ID {pid} which was not in input set") print(f"Will skip {numSkipped} already-processed page IDs") print("Getting dump-file offsets") offsetToPageids = {} offsetToEnd = {} # Maps chunk-start offsets to their chunk-end offsets iterNum = 0 for pageId in pageIds: iterNum += 1 if iterNum % 1e4 == 0: print(f"At iteration {iterNum}") # query = "SELECT offset, next_offset FROM offsets WHERE id = ?" row = indexDbCur.execute(query, (pageId,)).fetchone() if row == None: print(f"WARNING: Page ID {pageId} not found") continue (chunkOffset, endOffset) = row offsetToEnd[chunkOffset] = endOffset if chunkOffset not in offsetToPageids: offsetToPageids[chunkOffset] = [] offsetToPageids[chunkOffset].append(pageId) print(f"Found {len(offsetToEnd)} chunks to check") print("Iterating through chunks in dump file") def getImageName(content): " Given an array of text-content lines, tries to return an infoxbox image name, or None " # Doesn't try and find images in outside-infobox [[File:...]] and sections for line in content: match = imageLineRegex.match(line) if match != None: imageName = match.group(1).strip() if imageName == "": return None imageName = html.unescape(imageName) # Account for {{... if imageName.startswith("{"): match = cssImgCropRegex.match(imageName) if match == None: return None imageName = match.group(1) # Account for [[File:...|...]] if imageName.startswith("["): match = bracketImageRegex.match(imageName) if match == None: return None imageName = match.group(1) # Account for