diff options
| author | Terry Truong <terry06890@gmail.com> | 2022-05-04 01:17:06 +1000 |
|---|---|---|
| committer | Terry Truong <terry06890@gmail.com> | 2022-05-04 01:17:06 +1000 |
| commit | 90a5e15bb824b84e5bb60337d6a57a1394090dc6 (patch) | |
| tree | 661ea356c8d83b74d16f19d3555b0a1d3eb6eb56 /backend/data/genEnwikiData.py | |
| parent | ec29e5731136c74a1991e2f93b5e233747f2a230 (diff) | |
Add scripts for obtaining/sending/displaying wikipedia descriptions
Add backend/data/enwiki/ directory containing scripts and instructive
READMEs. Adjust some other scripts to generate 'eol_ids' sqlite table
separate from 'names'. Make server respond to /data/desc requests,
and have client TileInfo component display response data.
Also adjust .gitignore entries to be root-relative.
Diffstat (limited to 'backend/data/genEnwikiData.py')
| -rwxr-xr-x | backend/data/genEnwikiData.py | 64 |
1 files changed, 64 insertions, 0 deletions
diff --git a/backend/data/genEnwikiData.py b/backend/data/genEnwikiData.py new file mode 100755 index 0000000..f1490b6 --- /dev/null +++ b/backend/data/genEnwikiData.py @@ -0,0 +1,64 @@ +#!/usr/bin/python3 + +import sys, re +import sqlite3 + +usageInfo = f"usage: {sys.argv[0]}\n" +usageInfo += "Reads Wikimedia enwiki data from enwiki/, along with node and name data\n" +usageInfo += "from a sqlite database, associates nodes with enwiki pages, and adds\n" +usageInfo += "alt-name and description information for those nodes.\n" +if len(sys.argv) > 1: + print(usageInfo, file=sys.stderr) + sys.exit(1) + +enwikiDb = "enwiki/enwikiData.db" +dbFile = "data.db" + +# Open dbs +enwikiCon = sqlite3.connect(enwikiDb) +enwikiCur = enwikiCon.cursor() +dbCon = sqlite3.connect(dbFile) +dbCur = dbCon.cursor() +# Find page id for each node name +nodeToPageId = {} +print("Getting node page-ids") +iterationNum = 0 +for row in dbCur.execute("SELECT name from nodes"): + iterationNum += 1 + if iterationNum % 1e4 == 0: + print("At iteration {}".format(iterationNum)) + # + name = row[0] + row = enwikiCur.execute("SELECT id FROM pages where pages.title = ? COLLATE nocase", (name,)).fetchone() + if row != None: + nodeToPageId[name] = row[0] +# Resolve redirects +print("Resolving redirects") +redirectingNames = set() +iterationNum = 0 +for (name, pageId) in nodeToPageId.items(): + iterationNum += 1 + if iterationNum % 1e4 == 0: + print("At iteration {}".format(iterationNum)) + # + row = enwikiCur.execute("SELECT target_id FROM redirects where redirects.id = ?", (pageId,)).fetchone() + if row != None: + nodeToPageId[name] = row[0] + redirectingNames.add(name) +# Add descriptions for each node +print("Adding node description data") +dbCur.execute("CREATE TABLE descs (name TEXT PRIMARY KEY, desc TEXT, redirected INT)") +iterationNum = 0 +for (name, pageId) in nodeToPageId.items(): + iterationNum += 1 + if iterationNum % 1e4 == 0: + print("At iteration {}".format(iterationNum)) + # + row = enwikiCur.execute("SELECT desc FROM descs where descs.id = ?", (pageId,)).fetchone() + if row != None: + dbCur.execute("INSERT INTO descs VALUES (?, ?, ?)", (name, row[0], 1 if name in redirectingNames else 0)) +# Close dbs +dbCon.commit() +dbCon.close() +enwikiCon.commit() +enwikiCon.close() |
