diff --git a/.gitignore b/.gitignore
index f377651..2bcc7a3 100644
--- a/.gitignore
+++ b/.gitignore
@@ -3,4 +3,3 @@ __pycache__
commons_contributions/thumbnail_cache.json
commons_contributions/sent_mail_index.json
flickr_mail.db
-download_sent_mail.local.json
diff --git a/AGENTS.md b/AGENTS.md
index ef5ddb7..60e1614 100644
--- a/AGENTS.md
+++ b/AGENTS.md
@@ -85,20 +85,16 @@ for the Flickr mail URL. Scrapes the user's profile page for embedded params.
Shows recent Wikimedia Commons uploads on the home page, filtered to only
those obtained via Flickr mail requests.
-**Database tables used by the app**:
-- `sent_messages`: downloaded from Flickr sent mail, includes extracted Flickr
- URL and Wikipedia URL from message body
-- `contributions`: downloaded from Commons `usercontribs`
-- `flickr_uploads`: derived table built by `update_flickr_uploads.py` by
- matching Commons uploads to Flickr URLs
-- `thumbnail_cache`: cached Commons API thumbnail URLs (7-day TTL)
-- `interaction_log`: written by the web app to record searches and message
- generation events (see below)
+**Data files** (in `commons_contributions/`):
+- `flickr_uploads.json`: List of Commons uploads from Flickr with metadata
+- `thumbnail_cache.json`: Cached Commons API thumbnail URLs (7-day TTL)
+- `sent_mail_index.json`: Index of sent mail messages (flickr_url → wikipedia_url)
**Key functions**:
+- `build_sent_mail_index()`: Parses sent mail JSON files, extracts Flickr and
+ Wikipedia URLs from message bodies, caches the index
- `get_recent_commons_uploads()`: Loads uploads, filters by sent mail match,
- joins `flickr_uploads` with `sent_messages`, and fetches thumbnails from
- Commons API
+ fetches thumbnails from Commons API
- `normalize_flickr_url()`: Normalizes URLs for matching (removes protocol, www, trailing slash)
**CommonsUpload dataclass**:
@@ -108,32 +104,9 @@ those obtained via Flickr mail requests.
- `wiki_link_url`, `wiki_link_label`: Handles Wikidata vs Wikipedia links
**Maintenance script** (`update_flickr_uploads.py`):
-Builds/updates `flickr_uploads` from `contributions` and links to
-`sent_messages`.
-- Scans file contributions containing `UploadWizard` in the comment
-- Supports both comment styles:
- - `User created page with UploadWizard` (older)
- - `Uploaded a work by ... with UploadWizard` (newer; often includes URL)
-- Extracts Flickr URL from contribution comment when present
-- Falls back to Commons `extmetadata.Credit` lookup when comment has no URL
-
-### Interaction Logging (`log_interaction`)
-
-The `log_interaction()` helper writes a row to `interaction_log` on each
-meaningful user action:
-
-- `"search_article"` – user submits a Wikipedia article search (page 1 only,
- to avoid logging every pagination hit)
-- `"search_category"` – user submits a Wikipedia category search
-- `"generate_message"` – a non-free CC message is generated after clicking a photo
-
-Each row captures: Unix `timestamp`, `interaction_type`, `ip_address`
-(prefers `X-Forwarded-For` for proxy setups), `user_agent`, `query` (article
-title or category name), and optionally `flickr_url` / `wikipedia_url`.
-
-The table is created by `init_db()` (called via `python3 -c "from
-flickr_mail.database import init_db; init_db()"` or any of the maintenance
-scripts). The web app never calls `init_db()` itself.
+Run to find Flickr uploads from UploadWizard contributions that don't have
+the Flickr URL in the edit comment. Queries Commons API for image metadata
+and checks the Credit field for Flickr URLs.
### Category Search (`/category` route)
@@ -152,7 +125,7 @@ to allow back-navigation to the category.
### Previous Message Detection (`get_previous_messages`)
-Checks the `sent_messages` database table for previous messages to a Flickr user.
+Checks `sent_mail/messages_index.json` for previous messages to a Flickr user.
Matches by both display name and username (case-insensitive). Results shown as
an info alert on the message page.
@@ -186,24 +159,6 @@ print(f"{len(result.photos)} photos, {result.total_pages} pages")
print(result.photos[0].title, result.photos[0].license_name)
```
-## Data Sync Workflow
-
-To refresh "recent Commons uploads obtained via Flickr mail", run scripts in
-this order:
-
-1. `./download_sent_mail.py`
-2. `./download_commons_contributions.py`
-3. `./update_flickr_uploads.py`
-
-Notes:
-- `download_sent_mail.py` reads Flickr auth cookies from
- `download_sent_mail.local.json` (`cookies_str` key). Copy
- `download_sent_mail.example.json` to create local config.
-- `main.py` does not populate `flickr_uploads`; it only reads from it.
-- `download_commons_contributions.py` intentionally stops after several
- consecutive fully-known API batches (overlap window) to avoid full-history
- scans while still catching shallow gaps.
-
## Potential Improvements
- Cache search results to reduce Flickr requests
diff --git a/README.md b/README.md
index 3848a3e..848e89d 100644
--- a/README.md
+++ b/README.md
@@ -1,100 +1,89 @@
-# Flickr Mail
+# Flickr Photo Finder for Wikipedia Articles
Tool lives here:
-Flickr Mail is a Flask app that helps find Flickr photos for Wikipedia articles
-and contact photographers to request Wikipedia-compatible licensing.
+This tool is designed to help you find photos on Flickr for Wikipedia articles
+and contact the photographer. It's a Python application that leverages the Flask
+framework for web development.
-## What It Does
+## Table of Contents
+- [Introduction](#introduction)
+- [Usage](#usage)
+- [Error Handling](#error-handling)
+- [Running the Application](#running-the-application)
-- Searches Flickr from a Wikipedia article title/URL
-- Shows license status for each result (free vs non-free CC variants)
-- Builds a ready-to-send Flickr message for non-free licenses
-- Finds image-less articles in a Wikipedia category
-- Shows recent Commons uploads that came from Flickr mail outreach
+## Introduction
-## Project Layout
+This tool is developed and maintained by Edward Betts (edward@4angle.com). Its
+primary purpose is to simplify the process of discovering and contacting
+photographers on Flickr whose photos can be used to enhance Wikipedia articles.
-- `main.py`: Flask app routes and core logic
-- `templates/`: UI templates
-- `download_sent_mail.py`: sync Flickr sent messages into DB
-- `download_commons_contributions.py`: sync Commons contributions into DB
-- `update_flickr_uploads.py`: derive `flickr_uploads` from contributions/sent mail
-- `flickr_mail.db`: SQLite database
+### Key Features
+- **Integrated Flickr search**: Enter a Wikipedia article title and see Flickr
+ photos directly in the interface - no need to visit Flickr's search page.
+- **Photo grid with metadata**: Search results display as a grid of thumbnails
+ showing the user's name and license for each photo.
+- **License handling**: Photos with Wikipedia-compatible licenses (CC BY,
+ CC BY-SA, CC0, Public Domain) are highlighted with a green badge and link
+ directly to the Commons UploadWizard. Non-free CC licenses (NC/ND) show a
+ tailored message explaining Wikipedia's requirements. Supports both CC 2.0
+ and CC 4.0 license codes.
+- **One-click message composition**: Click any photo to compose a permission
+ request message with the photo displayed alongside, showing the user's Flickr
+ profile and current license.
+- **Previous message detection**: The message page checks sent mail history and
+ warns if you have previously contacted the user.
+- **Category search**: Find Wikipedia articles without images in a given
+ category, with links to search Flickr for each article.
+- **Pagination**: Browse through thousands of search results with page navigation.
+- **Recent uploads showcase**: The home page displays recent Wikimedia Commons
+ uploads that were obtained via Flickr mail requests, with links to the
+ Wikipedia article and user's Flickr profile.
+- Handle exceptions gracefully and provide detailed error information.
-## Database Pipeline
+## Usage
-The recent uploads section depends on a 3-step pipeline:
+To use the tool, follow these steps:
-1. `./download_sent_mail.py` updates `sent_messages`
-2. `./download_commons_contributions.py` updates `contributions`
-3. `./update_flickr_uploads.py` builds/updates `flickr_uploads`
+1. Start the tool by running the script.
+2. Access the tool through a web browser.
+3. Enter a Wikipedia article title or URL, or use "Find articles by category"
+ to discover articles that need images.
+4. Browse the Flickr search results displayed in the interface.
+5. Click on a photo to select it. If the license is Wikipedia-compatible, you'll
+ be linked to the Commons UploadWizard. Otherwise, a message is composed to
+ request a license change.
+6. Copy the subject and message, then click "Send message on Flickr" to contact
+ the user.
-`main.py` only reads `flickr_uploads`; it does not populate it.
+## Error Handling
-## UploadWizard Detection
+The application includes error handling to ensure a smooth user experience. If
+an error occurs, it will display a detailed error message with traceback
+information. The error handling is designed to provide valuable insights into
+any issues that may arise during use.
-`update_flickr_uploads.py` supports both Commons UploadWizard comment styles:
+## Running the Application
-- `User created page with UploadWizard` (older)
-- `Uploaded a work by ... with UploadWizard` (newer)
+To run the application, ensure you have Python 3 installed on your system. You
+will also need to install the required Python modules mentioned in the script,
+including Flask, requests, and others.
-It first tries to extract a Flickr URL directly from the contribution comment.
-If absent, it falls back to Commons `extmetadata.Credit`.
-
-## Local Run
-
-Install dependencies (example):
-
-```bash
-pip install flask requests beautifulsoup4 sqlalchemy
-```
-
-Start the app:
+1. Clone this repository to your local machine.
+2. Navigate to the project directory.
+3. Run the following command to start the application:
```bash
python3 main.py
```
-Then open:
+4. Access the application by opening a web browser and visiting the provided URL
+ (usually `http://localhost:5000/`).
-- `http://localhost:5000/`
+That's it! You can now use the Flickr Photo Finder tool to streamline the
+process of finding and contacting photographers for Wikipedia articles.
-## Refresh Data
+If you encounter any issues or have questions, feel free to contact Edward Betts
+(edward@4angle.com).
-Run in this order:
-
-```bash
-./download_sent_mail.py
-./download_commons_contributions.py
-./update_flickr_uploads.py
-```
-
-Before running `./download_sent_mail.py`, create local auth config:
-
-```bash
-cp download_sent_mail.example.json download_sent_mail.local.json
-```
-
-Then edit `download_sent_mail.local.json` and set `cookies_str` to your full
-Flickr `Cookie` header value.
-
-## Interaction Logging
-
-The app logs searches and message generation to the `interaction_log` table:
-
-- `search_article`: when a user searches for a Wikipedia article title (page 1 only)
-- `search_category`: when a user searches a Wikipedia category
-- `generate_message`: when a non-free CC message is generated for a photo
-
-Each row records the timestamp, interaction type, client IP (from
-`X-Forwarded-For` if present), User-Agent, query, and (for message events)
-the Flickr and Wikipedia URLs.
-
-## Notes
-
-- `download_commons_contributions.py` uses an overlap window of known-only
- batches before stopping to avoid full-history scans while still catching
- shallow gaps.
-- If a known Commons upload is missing from `flickr_uploads`, re-run the full
- 3-step pipeline above.
+Happy photo hunting!
diff --git a/download_commons_contributions.py b/download_commons_contributions.py
index a2dea6e..1f8f508 100755
--- a/download_commons_contributions.py
+++ b/download_commons_contributions.py
@@ -12,7 +12,6 @@ from flickr_mail.models import Contribution
API_URL = "https://commons.wikimedia.org/w/api.php"
USERNAME = "Edward"
-CONSECUTIVE_KNOWN_BATCHES_TO_STOP = 3
# Identify ourselves properly to Wikimedia
USER_AGENT = "CommonsContributionsDownloader/0.1 (edward@4angle.com)"
@@ -49,8 +48,12 @@ def fetch_contributions(
return contributions, new_continue
-def insert_contribution(session, c: dict) -> None:
- """Insert a contribution row (caller must ensure revid is new)."""
+def upsert_contribution(session, c: dict) -> None:
+ """Insert or update a contribution by revid."""
+ existing = session.query(Contribution).filter_by(revid=c["revid"]).first()
+ if existing:
+ return # Already have this revision
+
session.add(Contribution(
userid=c.get("userid"),
user=c.get("user"),
@@ -94,7 +97,6 @@ def main() -> None:
batch_num = 0
new_count = 0
continue_token = None
- consecutive_known_batches = 0
while True:
batch_num += 1
@@ -106,24 +108,13 @@ def main() -> None:
print("no results")
break
- # One DB query per batch to identify already-known revisions.
- revids = [c["revid"] for c in contributions if "revid" in c]
- existing_revids = {
- row[0]
- for row in (
- session.query(Contribution.revid)
- .filter(Contribution.revid.in_(revids))
- .all()
- )
- }
-
batch_new = 0
for c in contributions:
- revid = c.get("revid")
- if revid in existing_revids:
+ # Stop if we've reached contributions we already have
+ existing = session.query(Contribution).filter_by(revid=c["revid"]).first()
+ if existing:
continue
-
- insert_contribution(session, c)
+ upsert_contribution(session, c)
batch_new += 1
new_count += batch_new
@@ -132,18 +123,7 @@ def main() -> None:
session.commit()
if batch_new == 0:
- consecutive_known_batches += 1
- print(
- " Batch fully known "
- f"({consecutive_known_batches}/"
- f"{CONSECUTIVE_KNOWN_BATCHES_TO_STOP})"
- )
- else:
- consecutive_known_batches = 0
-
- if consecutive_known_batches >= CONSECUTIVE_KNOWN_BATCHES_TO_STOP:
- # Stop after a small overlap window of known-only batches.
- # This catches recent historical gaps without full-history scans.
+ # All contributions in this batch already exist, we're caught up
print(" Caught up with existing data")
break
diff --git a/download_sent_mail.example.json b/download_sent_mail.example.json
deleted file mode 100644
index 4b305c6..0000000
--- a/download_sent_mail.example.json
+++ /dev/null
@@ -1,3 +0,0 @@
-{
- "cookies_str": "paste your full Flickr Cookie header value here"
-}
diff --git a/download_sent_mail.py b/download_sent_mail.py
index 1951882..c3ac224 100755
--- a/download_sent_mail.py
+++ b/download_sent_mail.py
@@ -1,9 +1,7 @@
#!/usr/bin/env python3
"""Download sent FlickrMail messages for backup."""
-import json
import time
-from pathlib import Path
import requests
from bs4 import BeautifulSoup
@@ -19,9 +17,6 @@ from flickr_mail.url_utils import (
BASE_URL = "https://www.flickr.com"
SENT_MAIL_URL = f"{BASE_URL}/mail/sent/page{{page}}"
MESSAGE_URL = f"{BASE_URL}/mail/sent/{{message_id}}"
-MAX_SENT_MAIL_PAGES = 29 # Fallback upper bound if we need to backfill everything
-CONFIG_FILE = Path(__file__).with_name("download_sent_mail.local.json")
-EXAMPLE_CONFIG_FILE = Path(__file__).with_name("download_sent_mail.example.json")
HEADERS = {
"User-Agent": "Mozilla/5.0 (X11; Linux x86_64; rv:147.0) Gecko/20100101 Firefox/147.0",
@@ -38,23 +33,7 @@ HEADERS = {
"Priority": "u=0, i",
}
-def load_cookie_string() -> str:
- """Load Flickr cookies string from local JSON config."""
- if not CONFIG_FILE.exists():
- raise RuntimeError(
- f"Missing config file: {CONFIG_FILE}. "
- f"Copy {EXAMPLE_CONFIG_FILE.name} to {CONFIG_FILE.name} and set cookies_str."
- )
-
- try:
- data = json.loads(CONFIG_FILE.read_text())
- except json.JSONDecodeError as exc:
- raise RuntimeError(f"Invalid JSON in {CONFIG_FILE}: {exc}") from exc
-
- cookie_str = data.get("cookies_str", "").strip()
- if not cookie_str:
- raise RuntimeError(f"{CONFIG_FILE} must contain a non-empty 'cookies_str' value")
- return cookie_str
+COOKIES_STR = """ccc=%7B%22needsConsent%22%3Atrue%2C%22managed%22%3A0%2C%22changed%22%3A0%2C%22info%22%3A%7B%22cookieBlock%22%3A%7B%22level%22%3A2%2C%22blockRan%22%3A1%7D%7D%7D; _sp_ses.df80=*; _sp_id.df80=968931de-089d-4576-b729-6662c2c13a65.1770187027.1.1770187129..adf2374b-b85c-4899-afb7-63c2203d0c44..9422de57-9cdf-49c9-ac54-183eaa1ec457.1770187027101.24; TAsessionID=7f373c97-e9f8-46cb-bc1a-cb4f164ce46b|NEW; notice_behavior=expressed,eu; usprivacy=1---; acstring=3~550.1942.3126.3005.3077.1329.196.1725.1092; euconsent-v2=CQfGXgAQfGXgAAvACDENCQFsAP_gAEPgAAAALktB9G5cSSFBYCJVYbtEYAQDwFhg4oAhAgABEwAATBoAoIwGBGAoIAiAICACAAAAIARAIAEECAAAQAAAIIABAAAMAEAAIAACIAAACAABAgAACEAIAAggWAAAAEBEAFQAgAAAQBIACFAAAgABAUABAAAAAACAAQAAACAgQAAAAAAAAAAAkAhAAAAAAAAAABAMAAABIAAAAAAAAAAAAAAAAAAABAAAAICBAAAAQAAAAAAAAAAAAAAAAAAAAgqY0H0blxJIUFgIFVhu0QgBBPAWADigCEAAAEDAABMGgCgjAIUYCAgSIAgIAAAAAAgBEAgAQAIAABAAAAAgAEAAAwAQAAgAAAAAAAAAAECAAAAQAgACCBYAAAAQEQAVACBAABAEgAIUAAAAAEBQAEAAAAAAIABAAAAICBAAAAAAAAAAACQCEAAAAAAAAAAEAwBAAEgAAAAAAAAAAAAAAAAAAAEABAAgIEAAABAA.YAAAAAAAAAAA.ILktB9G5cSSFBYCJVYbtEYAQTwFhg4oAhAgABEwAATBoAoIwGFGAoIEiAICACAAAAIARAIAEECAAAQAAAIIABAAAMAEAAIAACIAAACAABAgAACEAIAAggWAAAAEBEAFQAgQAAQBIACFAAAgABAUABAAAAAACAAQAAACAgQAAAAAAAAAAAkAhAAAAAAAAAABAMAQABIAAAAAAAAAAAAAAAAAAABAAQAICBAAAAQAAAAAAAAAAAAAAAAAAAAgA; notice_preferences=2:; notice_gdpr_prefs=0,1,2:; cmapi_gtm_bl=; cmapi_cookie_privacy=permit 1,2,3; AMCV_48E815355BFE96970A495CD0%40AdobeOrg=281789898%7CMCMID%7C44859851125632937290373504988866174366%7CMCOPTOUT-1770194232s%7CNONE%7CvVersion%7C4.1.0; AMCVS_48E815355BFE96970A495CD0%40AdobeOrg=1; xb=646693; localization=en-us%3Buk%3Bgb; flrbp=1770187037-cfbf3914859af9ef68992c8389162e65e81c86c4; flrbgrp=1770187037-8e700fa7d73b4f2d43550f40513e7c6f507fd20f; flrbgdrp=1770187037-9af21cc74000b5f3f0943243608b4284d5f60ffd; flrbgmrp=1770187037-53f7bfff110731954be6bdfb2f587d59a8305670; flrbrst=1770187037-440e42fcee9b4e8e81ba8bc3eb3d0fc8b62e7083; flrtags=1770187037-7b50035cb956b9216a2f3372f498f7008d8e26a8; flrbrp=1770187037-c0195dc99caa020d4e32b39556131add862f26a0; flrb=34; session_id=2693fb01-87a0-42b1-a426-74642807b534; cookie_session=834645%3A29f2a9722d8bac88553ea1baf7ea11b4; cookie_accid=834645; cookie_epass=29f2a9722d8bac88553ea1baf7ea11b4; sa=1775371036%3A79962317%40N00%3A8fb60f4760b4840f37af3ebc90a8cb57; vp=2075%2C1177%2C1%2C0; flrbfd=1770187037-88a4e436729c9c5551794483fbd9c80e9dac2354; flrbpap=1770187037-18adaacf3a389df4a7bdc05cd471e492c54ef841; liqpw=2075; liqph=672"""
def parse_cookies(cookie_str: str) -> dict[str, str]:
@@ -71,7 +50,7 @@ def create_session() -> requests.Session:
"""Create a requests session with authentication."""
session = requests.Session()
session.headers.update(HEADERS)
- session.cookies.update(parse_cookies(load_cookie_string()))
+ session.cookies.update(parse_cookies(COOKIES_STR))
return session
@@ -187,41 +166,22 @@ def main() -> None:
http_session = create_session()
+ # Scrape all pages to find new messages
+ total_pages = 29
new_messages: list[dict] = []
- stop_fetching = False
- print("Fetching message list until we reach existing messages...")
- for page in range(1, MAX_SENT_MAIL_PAGES + 1):
+ print("Fetching message list from all pages...")
+ for page in range(1, total_pages + 1):
url = SENT_MAIL_URL.format(page=page)
- print(f" Fetching page {page}...")
+ print(f" Fetching page {page}/{total_pages}...")
try:
soup = fetch_page(http_session, url)
page_messages = extract_messages_from_list_page(soup)
- if not page_messages:
- print(" No messages found on this page, stopping")
- break
-
- page_new_messages = 0
for msg in page_messages:
- msg_id = msg.get("message_id")
- if not msg_id:
- continue
- if msg_id in existing_ids:
- stop_fetching = True
- break
-
- new_messages.append(msg)
- page_new_messages += 1
-
- if stop_fetching:
- print(" Reached messages already in the database, stopping pagination")
- break
-
- if page_new_messages == 0:
- print(" No new messages on this page, stopping pagination")
- break
+ if msg["message_id"] not in existing_ids:
+ new_messages.append(msg)
time.sleep(1) # Be polite to the server
diff --git a/flickr_mail/models.py b/flickr_mail/models.py
index 0b2bedc..090b37a 100644
--- a/flickr_mail/models.py
+++ b/flickr_mail/models.py
@@ -12,20 +12,20 @@ class Contribution(Base):
__tablename__ = "contributions"
id: Mapped[int] = mapped_column(primary_key=True)
- userid: Mapped[int]
- user: Mapped[str]
- pageid: Mapped[int]
- revid: Mapped[int] = mapped_column(unique=True)
- parentid: Mapped[int]
- ns: Mapped[int]
- title: Mapped[str]
- timestamp: Mapped[str]
+ userid: Mapped[int | None]
+ user: Mapped[str | None]
+ pageid: Mapped[int | None]
+ revid: Mapped[int | None] = mapped_column(unique=True)
+ parentid: Mapped[int | None]
+ ns: Mapped[int | None]
+ title: Mapped[str | None]
+ timestamp: Mapped[str | None]
minor: Mapped[str | None]
top: Mapped[str | None]
- comment: Mapped[str] = mapped_column(Text)
- size: Mapped[int]
- sizediff: Mapped[int]
- tags: Mapped[str] = mapped_column(Text) # JSON array stored as text
+ comment: Mapped[str | None] = mapped_column(Text)
+ size: Mapped[int | None]
+ sizediff: Mapped[int | None]
+ tags: Mapped[str | None] = mapped_column(Text) # JSON array stored as text
__table_args__ = (
Index("ix_contributions_timestamp", "timestamp"),
@@ -37,16 +37,16 @@ class SentMessage(Base):
__tablename__ = "sent_messages"
message_id: Mapped[str] = mapped_column(primary_key=True)
- subject: Mapped[str]
- url: Mapped[str]
- recipient: Mapped[str]
- date: Mapped[str]
- body: Mapped[str] = mapped_column(Text)
- body_html: Mapped[str] = mapped_column(Text)
- flickr_url: Mapped[str]
- normalized_flickr_url: Mapped[str]
- wikipedia_url: Mapped[str]
- creator_profile_url: Mapped[str]
+ subject: Mapped[str | None]
+ url: Mapped[str | None]
+ recipient: Mapped[str | None]
+ date: Mapped[str | None]
+ body: Mapped[str | None] = mapped_column(Text)
+ body_html: Mapped[str | None] = mapped_column(Text)
+ flickr_url: Mapped[str | None]
+ normalized_flickr_url: Mapped[str | None]
+ wikipedia_url: Mapped[str | None]
+ creator_profile_url: Mapped[str | None]
flickr_uploads: Mapped[list["FlickrUpload"]] = relationship(
back_populates="sent_message"
@@ -62,15 +62,15 @@ class FlickrUpload(Base):
__tablename__ = "flickr_uploads"
id: Mapped[int] = mapped_column(primary_key=True)
- pageid: Mapped[int]
- revid: Mapped[int]
- title: Mapped[str]
- timestamp: Mapped[str]
- flickr_url: Mapped[str]
- normalized_flickr_url: Mapped[str]
+ pageid: Mapped[int | None]
+ revid: Mapped[int | None]
+ title: Mapped[str | None]
+ timestamp: Mapped[str | None]
+ flickr_url: Mapped[str | None]
+ normalized_flickr_url: Mapped[str | None]
creator: Mapped[str | None]
- wikipedia_url: Mapped[str]
- creator_profile_url: Mapped[str]
+ wikipedia_url: Mapped[str | None]
+ creator_profile_url: Mapped[str | None]
sent_message_id: Mapped[str | None] = mapped_column(
ForeignKey("sent_messages.message_id")
)
@@ -89,23 +89,5 @@ class ThumbnailCache(Base):
__tablename__ = "thumbnail_cache"
title: Mapped[str] = mapped_column(primary_key=True)
- thumb_url: Mapped[str]
- fetched_at: Mapped[int] # Unix timestamp
-
-
-class InteractionLog(Base):
- __tablename__ = "interaction_log"
-
- id: Mapped[int] = mapped_column(primary_key=True)
- timestamp: Mapped[int] # Unix timestamp
- interaction_type: Mapped[str] # "search_article", "search_category", "generate_message"
- ip_address: Mapped[str | None]
- user_agent: Mapped[str | None] = mapped_column(Text)
- query: Mapped[str | None] # search term or category name
- flickr_url: Mapped[str | None]
- wikipedia_url: Mapped[str | None]
-
- __table_args__ = (
- Index("ix_interaction_log_timestamp", "timestamp"),
- Index("ix_interaction_log_type", "interaction_type"),
- )
+ thumb_url: Mapped[str | None]
+ fetched_at: Mapped[int | None] # Unix timestamp
diff --git a/main.py b/main.py
index 2a6d238..25279ff 100755
--- a/main.py
+++ b/main.py
@@ -18,7 +18,7 @@ from sqlalchemy import func
from werkzeug.debug.tbtools import DebugTraceback
from flickr_mail.database import get_session
-from flickr_mail.models import FlickrUpload, InteractionLog, SentMessage, ThumbnailCache
+from flickr_mail.models import FlickrUpload, SentMessage, ThumbnailCache
from flickr_mail.url_utils import extract_urls_from_message, normalize_flickr_url
import re
@@ -348,14 +348,6 @@ class ArticleWithoutImage:
return f"/?enwp={quote(self.title)}"
-@dataclasses.dataclass
-class CategoryResult:
- """Result of a paginated category search."""
-
- articles: list[ArticleWithoutImage]
- gcmcontinue: str | None
-
-
# Common non-content images to ignore when checking if an article has images
NON_CONTENT_IMAGE_PATTERNS = [
"OOjs UI icon",
@@ -387,15 +379,14 @@ def has_content_image(images: list[dict]) -> bool:
def get_articles_without_images(
- category: str,
- limit: int = 200,
- gcmcontinue: str | None = None,
-) -> CategoryResult:
+ category: str, limit: int = 100
+) -> tuple[list[ArticleWithoutImage], str | None]:
"""Get articles in a category that don't have images.
Uses generator=categorymembers with prop=images to efficiently check
- multiple articles in a single API request, following continuation until
- the limit is reached or all category members have been processed.
+ multiple articles in a single API request.
+
+ Returns a tuple of (articles_list, continue_token).
"""
params = {
"action": "query",
@@ -403,73 +394,49 @@ def get_articles_without_images(
"gcmtitle": category,
"gcmtype": "page", # Only articles, not subcategories or files
"gcmnamespace": "0", # Main namespace only
- "gcmlimit": "50", # Small batches so images fit in one response
+ "gcmlimit": str(limit),
"prop": "images",
- "imlimit": "max",
+ "imlimit": "max", # Need enough to check all pages in batch
"format": "json",
}
headers = {"User-Agent": WIKIMEDIA_USER_AGENT}
+
+ try:
+ response = requests.get(
+ WIKIPEDIA_API, params=params, headers=headers, timeout=30
+ )
+ response.raise_for_status()
+ data = response.json()
+ except (requests.RequestException, json.JSONDecodeError) as e:
+ print(f"Wikipedia API error: {e}")
+ return [], None
+
articles_without_images: list[ArticleWithoutImage] = []
- seen_pageids: set[int] = set()
- next_gcmcontinue: str | None = None
- # Build initial continue params from the external pagination token
- continue_params: dict[str, str] = {}
- if gcmcontinue:
- continue_params = {"gcmcontinue": gcmcontinue, "continue": "gcmcontinue||"}
+ pages = data.get("query", {}).get("pages", {})
+ for page in pages.values():
+ images = page.get("images", [])
- while True:
- request_params = params.copy()
- request_params.update(continue_params)
+ # Skip if page has content images (not just UI icons)
+ if has_content_image(images):
+ continue
- try:
- response = requests.get(
- WIKIPEDIA_API, params=request_params, headers=headers, timeout=30
+ title = page.get("title", "")
+ pageid = page.get("pageid", 0)
+
+ if title and pageid:
+ articles_without_images.append(
+ ArticleWithoutImage(title=title, pageid=pageid)
)
- response.raise_for_status()
- data = response.json()
- except (requests.RequestException, json.JSONDecodeError) as e:
- print(f"Wikipedia API error: {e}")
- break
-
- pages = data.get("query", {}).get("pages", {})
- for page in pages.values():
- pageid = page.get("pageid", 0)
- if not pageid or pageid in seen_pageids:
- continue
- seen_pageids.add(pageid)
-
- images = page.get("images", [])
-
- # Skip if page has content images (not just UI icons)
- if has_content_image(images):
- continue
-
- title = page.get("title", "")
- if title:
- articles_without_images.append(
- ArticleWithoutImage(title=title, pageid=pageid)
- )
-
- api_continue = data.get("continue")
- if not api_continue:
- break
-
- # Only stop at generator boundaries where we have a resumable token
- gcmc = api_continue.get("gcmcontinue")
- if gcmc and len(articles_without_images) >= limit:
- next_gcmcontinue = gcmc
- break
-
- continue_params = api_continue
# Sort by title for consistent display
articles_without_images.sort(key=lambda a: a.title)
- return CategoryResult(
- articles=articles_without_images,
- gcmcontinue=next_gcmcontinue,
- )
+
+ # Get continue token if there are more results
+ continue_token = data.get("continue", {}).get("gcmcontinue")
+
+ return articles_without_images, continue_token
def is_valid_flickr_image_url(url: str) -> bool:
@@ -491,7 +458,7 @@ def is_valid_flickr_image_url(url: str) -> bool:
def search_flickr(search_term: str, page: int = 1) -> SearchResult:
"""Search Flickr for photos matching the search term."""
- encoded_term = quote(search_term)
+ encoded_term = quote(f'"{search_term}"')
url = f"https://flickr.com/search/?view_all=1&text={encoded_term}&page={page}"
response = requests.get(url, headers=BROWSER_HEADERS)
@@ -616,33 +583,6 @@ def parse_flickr_search_results(html: str, page: int = 1) -> SearchResult:
)
-def log_interaction(
- interaction_type: str,
- query: str | None = None,
- flickr_url: str | None = None,
- wikipedia_url: str | None = None,
-) -> None:
- """Log a user interaction to the database."""
- forwarded_for = flask.request.headers.get("X-Forwarded-For")
- ip_address = forwarded_for.split(",")[0].strip() if forwarded_for else flask.request.remote_addr
- user_agent = flask.request.headers.get("User-Agent")
- session = get_session()
- try:
- entry = InteractionLog(
- timestamp=int(time.time()),
- interaction_type=interaction_type,
- ip_address=ip_address,
- user_agent=user_agent,
- query=query,
- flickr_url=flickr_url,
- wikipedia_url=wikipedia_url,
- )
- session.add(entry)
- session.commit()
- finally:
- session.close()
-
-
@app.errorhandler(werkzeug.exceptions.InternalServerError)
def exception_handler(e: werkzeug.exceptions.InternalServerError) -> tuple[str, int]:
"""Handle exception."""
@@ -711,24 +651,18 @@ def start() -> str:
# Get category param if coming from category search
cat = flask.request.args.get("cat")
- # Allow overriding the Flickr search term (default includes quotes for phrase search)
- flickr_search = flask.request.args.get("flickr_search") or f'"{name}"'
-
flickr_url = flask.request.args.get("flickr")
if not flickr_url:
# Search Flickr for photos
page = flask.request.args.get("page", 1, type=int)
page = max(1, page) # Ensure page is at least 1
- if page == 1:
- log_interaction("search_article", query=flickr_search, wikipedia_url=wikipedia_url)
- search_result = search_flickr(flickr_search, page)
+ search_result = search_flickr(name, page)
return flask.render_template(
"combined.html",
name=name,
enwp=enwp,
search_result=search_result,
cat=cat,
- flickr_search=flickr_search,
)
if "/in/" in flickr_url:
@@ -782,16 +716,8 @@ def start() -> str:
flickr_user_url=flickr_user_url,
cat=cat,
previous_messages=previous_messages,
- flickr_search=flickr_search,
)
- log_interaction(
- "generate_message",
- query=name,
- flickr_url=flickr_url,
- wikipedia_url=wikipedia_url,
- )
-
msg = flask.render_template(
"message.jinja",
flickr_url=flickr_url,
@@ -823,7 +749,6 @@ def start() -> str:
flickr_user_url=flickr_user_url,
cat=cat,
previous_messages=previous_messages,
- flickr_search=flickr_search,
)
@@ -843,9 +768,7 @@ def category_search() -> str:
cat=cat,
)
- log_interaction("search_category", query=category)
- gcmcontinue = flask.request.args.get("gcmcontinue") or None
- result = get_articles_without_images(category, gcmcontinue=gcmcontinue)
+ articles, continue_token = get_articles_without_images(category)
# Get the display name (without Category: prefix)
category_name = category.replace("Category:", "")
@@ -855,8 +778,8 @@ def category_search() -> str:
cat=cat,
category=category,
category_name=category_name,
- articles=result.articles,
- gcmcontinue=result.gcmcontinue,
+ articles=articles,
+ continue_token=continue_token,
)
diff --git a/templates/category.html b/templates/category.html
index 62fcb1a..da52980 100644
--- a/templates/category.html
+++ b/templates/category.html
@@ -33,7 +33,7 @@
diff --git a/update_flickr_uploads.py b/update_flickr_uploads.py
index 72c70e8..06b85e0 100644
--- a/update_flickr_uploads.py
+++ b/update_flickr_uploads.py
@@ -2,12 +2,8 @@
"""
Find UploadWizard contributions that are from Flickr and add them to the database.
-Supports both UploadWizard comment styles:
-- "User created page with UploadWizard" (older)
-- "Uploaded a work by ... with UploadWizard" (newer, often includes Flickr URL)
-
-If a Flickr URL is not present in the contribution comment, queries Commons API
-to check if the image source is Flickr (by checking the Credit field).
+For contributions with comment 'User created page with UploadWizard', queries the
+Commons API to check if the image source is Flickr (by checking the Credit field).
"""
import json
@@ -31,13 +27,6 @@ def extract_flickr_url_from_credit(credit: str) -> str | None:
return match.group(0) if match else None
-def extract_flickr_url_from_comment(comment: str) -> str | None:
- """Extract Flickr URL directly from a contribution comment."""
- pattern = r'https?://(?:www\.)?flickr\.com/photos/[^/\s]+/\d+'
- match = re.search(pattern, comment or "")
- return match.group(0) if match else None
-
-
def get_image_metadata(titles: list[str]) -> dict[str, dict]:
"""Fetch image metadata from Commons API for multiple titles."""
if not titles:
@@ -108,12 +97,10 @@ def main():
)
url_to_message = {msg.normalized_flickr_url: msg for msg in sent_messages}
- # Find UploadWizard file uploads.
- # Old format: "User created page with UploadWizard"
- # New format: "Uploaded a work by ... with UploadWizard"
+ # Find UploadWizard contributions (page creations only)
upload_wizard = (
session.query(Contribution)
- .filter(Contribution.comment.contains("UploadWizard"))
+ .filter(Contribution.comment == "User created page with UploadWizard")
.filter(Contribution.title.startswith("File:"))
.all()
)
@@ -140,10 +127,7 @@ def main():
credit = meta.get("credit", "")
artist = meta.get("artist", "")
- # Prefer URL directly in comment; fall back to extmetadata Credit.
- flickr_url = extract_flickr_url_from_comment(c.comment or "")
- if not flickr_url:
- flickr_url = extract_flickr_url_from_credit(credit)
+ flickr_url = extract_flickr_url_from_credit(credit)
if not flickr_url:
continue