diff options
Diffstat (limited to 'backend/hist_data/enwiki/gen_img_data.py')
| -rwxr-xr-x | backend/hist_data/enwiki/gen_img_data.py | 23 |
1 files changed, 13 insertions, 10 deletions
diff --git a/backend/hist_data/enwiki/gen_img_data.py b/backend/hist_data/enwiki/gen_img_data.py index 762952e..b4ade9f 100755 --- a/backend/hist_data/enwiki/gen_img_data.py +++ b/backend/hist_data/enwiki/gen_img_data.py @@ -34,7 +34,8 @@ def genData(pageIds: set[int], dumpFile: str, indexDb: str, imgDb: str) -> None: print('Checking tables') if imgDbCur.execute('SELECT name FROM sqlite_master WHERE type="table" AND name="page_imgs"').fetchone() is None: # Create tables if not present - imgDbCur.execute('CREATE TABLE page_imgs (page_id INT PRIMARY KEY, img_name TEXT)') # img_name may be NULL + imgDbCur.execute('CREATE TABLE page_imgs (page_id INT PRIMARY KEY, title TEXT UNIQUE, img_name TEXT)') + # 'img_name' may be NULL imgDbCur.execute('CREATE INDEX page_imgs_idx ON page_imgs(img_name)') else: # Check for already-processed page IDs @@ -48,24 +49,26 @@ def genData(pageIds: set[int], dumpFile: str, indexDb: str, imgDb: str) -> None: print(f'Will skip {numSkipped} already-processed page IDs') # print('Getting dump-file offsets') - offsetToPageids: dict[int, list[int]] = {} + offsetToPageId: dict[int, list[int]] = {} offsetToEnd: dict[int, int] = {} # Maps chunk-start offsets to their chunk-end offsets + pageIdToTitle: dict[int, str] = {} iterNum = 0 for pageId in pageIds: iterNum += 1 if iterNum % 1e4 == 0: print(f'At iteration {iterNum}') # - query = 'SELECT offset, next_offset FROM offsets WHERE id = ?' - row: tuple[int, int] | None = indexDbCur.execute(query, (pageId,)).fetchone() + query = 'SELECT offset, next_offset, title FROM offsets WHERE id = ?' + row = indexDbCur.execute(query, (pageId,)).fetchone() if row is None: print(f'WARNING: Page ID {pageId} not found') continue - chunkOffset, endOffset = row + chunkOffset, endOffset, title = row offsetToEnd[chunkOffset] = endOffset - if chunkOffset not in offsetToPageids: - offsetToPageids[chunkOffset] = [] - offsetToPageids[chunkOffset].append(pageId) + if chunkOffset not in offsetToPageId: + offsetToPageId[chunkOffset] = [] + offsetToPageId[chunkOffset].append(pageId) + pageIdToTitle[pageId] = title print(f'Found {len(offsetToEnd)} chunks to check') # print('Iterating through chunks in dump file') @@ -76,7 +79,7 @@ def genData(pageIds: set[int], dumpFile: str, indexDb: str, imgDb: str) -> None: if iterNum % 100 == 0: print(f'At iteration {iterNum}') # - chunkPageIds = offsetToPageids[pageOffset] + chunkPageIds = offsetToPageId[pageOffset] # Jump to chunk file.seek(pageOffset) compressedData = file.read(None if endOffset == -1 else endOffset - pageOffset) @@ -121,7 +124,7 @@ def genData(pageIds: set[int], dumpFile: str, indexDb: str, imgDb: str) -> None: content.append(line[:line.rfind('</text>')]) # Look for image-filename imageName = getImageName(content) - imgDbCur.execute('INSERT into page_imgs VALUES (?, ?)', (pageId, imageName)) + imgDbCur.execute('INSERT into page_imgs VALUES (?, ?, ?)', (pageId, None if imageName is None else pageIdToTitle[pageId], imageName)) break if not foundTextEnd: print(f'WARNING: Did not find </text> for page id {pageId}') |
