2019-08-18 15:56:53 +01:00
|
|
|
#!/usr/bin/python3
|
2023-11-17 15:23:03 +00:00
|
|
|
"""Reverse geocode: convert lat/lon to Wikidata item & Wikimedia Commons category."""
|
2019-08-18 15:56:53 +01:00
|
|
|
|
|
|
|
import random
|
2023-11-24 15:35:54 +00:00
|
|
|
import socket
|
2023-10-10 10:11:23 +01:00
|
|
|
import typing
|
|
|
|
|
2023-11-17 15:21:07 +00:00
|
|
|
import sqlalchemy.exc
|
2023-10-10 10:11:23 +01:00
|
|
|
from flask import Flask, jsonify, redirect, render_template, request, url_for
|
2023-11-04 21:35:33 +00:00
|
|
|
from sqlalchemy.orm.query import Query
|
2023-10-10 10:11:23 +01:00
|
|
|
from werkzeug.wrappers import Response
|
|
|
|
|
|
|
|
import geocode
|
|
|
|
from geocode import database, model, scotland, wikidata
|
2021-04-15 12:28:02 +01:00
|
|
|
|
2021-04-16 16:21:38 +01:00
|
|
|
city_of_london_qid = "Q23311"
|
2019-08-18 15:56:53 +01:00
|
|
|
app = Flask(__name__)
|
2021-04-16 21:24:59 +01:00
|
|
|
app.config.from_object("config.default")
|
2021-04-17 17:02:53 +01:00
|
|
|
database.init_app(app)
|
2019-08-18 15:56:53 +01:00
|
|
|
|
2023-11-04 20:49:36 +00:00
|
|
|
Tags = typing.Mapping[str, str]
|
2023-11-24 15:35:54 +00:00
|
|
|
logging_enabled = True
|
2023-11-04 20:49:36 +00:00
|
|
|
|
2019-08-18 15:56:53 +01:00
|
|
|
|
2023-10-10 10:11:23 +01:00
|
|
|
def get_random_lat_lon() -> tuple[float, float]:
|
|
|
|
"""Select random lat/lon within the UK."""
|
2021-04-16 16:14:12 +01:00
|
|
|
south, east = 50.8520, 0.3536
|
|
|
|
north, west = 53.7984, -2.7296
|
|
|
|
|
|
|
|
mul = 10000
|
2019-08-18 15:56:53 +01:00
|
|
|
lat = random.randrange(int(south * mul), int(north * mul)) / mul
|
|
|
|
lon = random.randrange(int(west * mul), int(east * mul)) / mul
|
|
|
|
|
2021-04-16 16:14:12 +01:00
|
|
|
return lat, lon
|
|
|
|
|
|
|
|
|
2023-10-10 10:11:23 +01:00
|
|
|
def do_lookup(
|
2023-11-04 21:35:33 +00:00
|
|
|
elements: "Query[model.Polygon]", lat: str | float, lon: str | float
|
2023-10-10 10:11:23 +01:00
|
|
|
) -> wikidata.WikidataDict:
|
2023-11-04 20:49:36 +00:00
|
|
|
"""Do lookup."""
|
2021-04-15 12:28:02 +01:00
|
|
|
try:
|
|
|
|
hit = osm_lookup(elements, lat, lon)
|
2021-04-17 17:02:53 +01:00
|
|
|
except wikidata.QueryError as e:
|
2021-04-15 12:28:02 +01:00
|
|
|
return {
|
2021-04-16 16:21:38 +01:00
|
|
|
"query": e.query,
|
|
|
|
"error": e.r.text,
|
|
|
|
"query_url": "https://query.wikidata.org/#" + e.query,
|
2021-04-15 12:28:02 +01:00
|
|
|
}
|
|
|
|
|
2021-04-17 18:29:09 +01:00
|
|
|
return wikidata.build_dict(hit, lat, lon)
|
2021-04-15 12:28:02 +01:00
|
|
|
|
2021-04-16 16:14:32 +01:00
|
|
|
|
2023-10-10 10:11:23 +01:00
|
|
|
def lat_lon_to_wikidata(lat: str | float, lon: str | float) -> dict[str, typing.Any]:
|
2023-11-04 20:49:36 +00:00
|
|
|
"""Lookup lat/lon and find most appropriate Wikidata item."""
|
2021-04-17 17:02:53 +01:00
|
|
|
scotland_code = scotland.get_scotland_code(lat, lon)
|
2021-04-15 12:28:02 +01:00
|
|
|
|
2023-10-10 10:11:23 +01:00
|
|
|
elements: typing.Any
|
2021-04-15 12:28:02 +01:00
|
|
|
if scotland_code:
|
2021-04-17 18:29:09 +01:00
|
|
|
rows = wikidata.lookup_scottish_parish_in_wikidata(scotland_code)
|
|
|
|
hit = wikidata.commons_from_rows(rows)
|
2021-04-15 12:28:02 +01:00
|
|
|
elements = []
|
2021-04-17 18:29:09 +01:00
|
|
|
result = wikidata.build_dict(hit, lat, lon)
|
2021-04-15 12:28:02 +01:00
|
|
|
|
2021-04-16 16:22:39 +01:00
|
|
|
return {"elements": elements, "result": result}
|
2021-04-15 12:28:02 +01:00
|
|
|
|
2021-04-17 17:31:58 +01:00
|
|
|
elements = model.Polygon.coords_within(lat, lon)
|
2021-04-15 12:28:02 +01:00
|
|
|
result = do_lookup(elements, lat, lon)
|
|
|
|
|
|
|
|
# special case because the City of London is admin_level=6 in OSM
|
2022-05-18 14:12:34 +01:00
|
|
|
if result.get("wikidata") == city_of_london_qid:
|
2021-04-16 16:22:39 +01:00
|
|
|
return {"elements": elements, "result": result}
|
2021-04-15 12:28:02 +01:00
|
|
|
|
2022-05-18 14:12:34 +01:00
|
|
|
admin_level = result.get("admin_level")
|
2023-11-04 20:49:36 +00:00
|
|
|
if not admin_level:
|
|
|
|
return {"elements": elements, "result": result}
|
2021-04-15 12:28:02 +01:00
|
|
|
|
2023-11-04 20:49:36 +00:00
|
|
|
assert isinstance(admin_level, int)
|
|
|
|
if admin_level >= 7:
|
2021-04-16 16:22:39 +01:00
|
|
|
return {"elements": elements, "result": result}
|
2021-04-15 12:28:02 +01:00
|
|
|
|
2021-04-17 18:29:09 +01:00
|
|
|
row = wikidata.geosearch(lat, lon)
|
2021-04-15 12:28:02 +01:00
|
|
|
if row:
|
2021-04-17 18:29:09 +01:00
|
|
|
hit = wikidata.commons_from_rows([row])
|
2021-04-15 12:28:02 +01:00
|
|
|
elements = []
|
2021-04-17 18:29:09 +01:00
|
|
|
result = wikidata.build_dict(hit, lat, lon)
|
2021-04-15 12:28:02 +01:00
|
|
|
|
2021-04-16 16:22:39 +01:00
|
|
|
return {"elements": elements, "result": result}
|
2019-08-18 15:56:53 +01:00
|
|
|
|
2021-04-16 16:14:32 +01:00
|
|
|
|
2023-11-04 20:49:36 +00:00
|
|
|
def get_admin_level(tags: Tags) -> int | None:
|
|
|
|
"""Read admin_level from tags."""
|
|
|
|
admin_level_tag = tags.get("admin_level")
|
|
|
|
return (
|
|
|
|
int(admin_level_tag) if admin_level_tag and admin_level_tag.isdigit() else None
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
def hit_from_wikidata_tag(tags: Tags) -> wikidata.Hit | None:
|
|
|
|
"""Check element for a wikidata tag."""
|
|
|
|
return (
|
|
|
|
{
|
|
|
|
"wikidata": qid,
|
|
|
|
"commons_cat": commons,
|
|
|
|
}
|
|
|
|
if "wikidata" in tags
|
|
|
|
and (commons := wikidata.qid_to_commons_category(qid := tags["wikidata"]))
|
|
|
|
else None
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
def hit_from_ref_gss_tag(tags: Tags) -> wikidata.Hit | None:
|
|
|
|
"""Check element for rss:gss tag."""
|
|
|
|
gss = tags.get("ref:gss")
|
|
|
|
return wikidata.get_commons_cat_from_gss(gss) if gss else None
|
|
|
|
|
|
|
|
|
|
|
|
def hit_from_name(
|
|
|
|
tags: Tags, lat: str | float, lon: str | float
|
|
|
|
) -> wikidata.Hit | None:
|
|
|
|
"""Use name to look for hit."""
|
|
|
|
if not (name := tags.get("name")):
|
|
|
|
return None
|
|
|
|
if name.endswith(" CP"): # civil parish
|
|
|
|
name = name[:-3]
|
|
|
|
|
|
|
|
rows = wikidata.lookup_wikidata_by_name(name, lat, lon)
|
|
|
|
return wikidata.commons_from_rows(rows) if len(rows) == 1 else None
|
|
|
|
|
|
|
|
|
2023-10-10 10:11:23 +01:00
|
|
|
def osm_lookup(
|
2023-11-04 21:35:33 +00:00
|
|
|
elements: "Query[model.Polygon]", lat: str | float, lon: str | float
|
2023-10-10 10:11:23 +01:00
|
|
|
) -> wikidata.Hit | None:
|
|
|
|
"""OSM lookup."""
|
|
|
|
ret: wikidata.Hit | None
|
2021-04-15 12:28:02 +01:00
|
|
|
for e in elements:
|
2023-10-10 10:11:23 +01:00
|
|
|
assert isinstance(e, model.Polygon)
|
|
|
|
assert e.tags
|
2023-11-04 20:49:36 +00:00
|
|
|
tags: typing.Mapping[str, str] = e.tags
|
|
|
|
admin_level: int | None = get_admin_level(tags)
|
2023-11-04 21:22:09 +00:00
|
|
|
if not admin_level and tags.get("boundary") not in ("political", "place"):
|
2021-04-15 12:28:02 +01:00
|
|
|
continue
|
2023-11-04 20:49:36 +00:00
|
|
|
if not (
|
|
|
|
(hit := hit_from_wikidata_tag(tags))
|
|
|
|
or (hit := hit_from_ref_gss_tag(tags))
|
|
|
|
or (hit := hit_from_name(tags, lat, lon))
|
|
|
|
):
|
2019-08-18 15:56:53 +01:00
|
|
|
continue
|
2023-11-04 20:49:36 +00:00
|
|
|
hit["admin_level"] = admin_level
|
|
|
|
hit["element"] = e.osm_id
|
|
|
|
hit["geojson"] = typing.cast(str, e.geojson_str)
|
|
|
|
return hit
|
|
|
|
|
|
|
|
has_wikidata_tag = [e for e in elements if e.tags.get("wikidata")]
|
2021-04-15 12:28:02 +01:00
|
|
|
if len(has_wikidata_tag) != 1:
|
2023-10-10 10:11:23 +01:00
|
|
|
return None
|
2021-04-15 12:28:02 +01:00
|
|
|
|
2023-11-04 20:49:36 +00:00
|
|
|
e = has_wikidata_tag[0]
|
|
|
|
assert e.tags
|
|
|
|
qid = e.tags["wikidata"]
|
|
|
|
return {
|
|
|
|
"wikidata": qid,
|
|
|
|
"element": e.osm_id,
|
|
|
|
"geojson": typing.cast(str, e.geojson_str),
|
|
|
|
"commons_cat": wikidata.qid_to_commons_category(qid),
|
|
|
|
"admin_level": admin_level,
|
|
|
|
}
|
2023-10-10 10:11:23 +01:00
|
|
|
|
|
|
|
|
|
|
|
def redirect_to_detail(q: str) -> Response:
|
|
|
|
"""Redirect to detail page."""
|
|
|
|
lat, lon = [v.strip() for v in q.split(",", 1)]
|
|
|
|
return redirect(url_for("detail_page", lat=lat, lon=lon))
|
2019-08-18 15:56:53 +01:00
|
|
|
|
|
|
|
|
2023-11-17 15:21:07 +00:00
|
|
|
@app.errorhandler(sqlalchemy.exc.OperationalError)
|
|
|
|
def handle_database_error(error: Exception) -> tuple[str, int]:
|
|
|
|
"""Show error screen on database error."""
|
|
|
|
return render_template("database_error.html"), 500
|
|
|
|
|
|
|
|
|
2021-04-16 21:30:25 +01:00
|
|
|
@app.route("/")
|
2023-10-10 10:11:23 +01:00
|
|
|
def index() -> str | Response:
|
|
|
|
"""Index page."""
|
2023-11-17 15:21:07 +00:00
|
|
|
database.session.execute("SELECT 1")
|
2021-04-16 21:30:25 +01:00
|
|
|
q = request.args.get("q")
|
|
|
|
if q and q.strip():
|
2023-10-10 10:11:23 +01:00
|
|
|
return redirect_to_detail(q)
|
2021-04-16 21:30:25 +01:00
|
|
|
|
|
|
|
lat, lon = request.args.get("lat"), request.args.get("lon")
|
|
|
|
|
2023-11-24 15:13:00 +00:00
|
|
|
if lat is None or lon is None:
|
|
|
|
samples = sorted(geocode.samples, key=lambda row: row[2])
|
|
|
|
return render_template("index.html", samples=samples)
|
|
|
|
|
|
|
|
result = lat_lon_to_wikidata(lat, lon)["result"]
|
|
|
|
result.pop("element", None)
|
|
|
|
result.pop("geojson", None)
|
2023-11-24 15:35:54 +00:00
|
|
|
if logging_enabled:
|
|
|
|
remote_addr = request.headers.get("X-Forwarded-For", request.remote_addr)
|
|
|
|
log = model.LookupLog(
|
|
|
|
lat=lat,
|
|
|
|
lon=lon,
|
|
|
|
remote_addr=remote_addr,
|
|
|
|
fqdn=socket.getfqdn(remote_addr),
|
|
|
|
result=result,
|
|
|
|
)
|
|
|
|
database.session.add(log)
|
|
|
|
database.session.commit()
|
2023-11-24 15:13:00 +00:00
|
|
|
return jsonify(result)
|
2021-04-16 21:30:25 +01:00
|
|
|
|
|
|
|
|
|
|
|
@app.route("/random")
|
2023-11-04 21:36:05 +00:00
|
|
|
def random_location() -> str | Response:
|
2023-10-10 10:11:23 +01:00
|
|
|
"""Return detail page for random lat/lon."""
|
2021-04-16 21:30:25 +01:00
|
|
|
lat, lon = get_random_lat_lon()
|
2023-11-04 21:36:05 +00:00
|
|
|
return build_detail_page(lat, lon)
|
2021-04-16 21:30:25 +01:00
|
|
|
|
|
|
|
|
|
|
|
@app.route("/wikidata_tag")
|
2023-10-10 10:11:23 +01:00
|
|
|
def wikidata_tag() -> str:
|
|
|
|
"""Lookup Wikidata tag for lat/lon."""
|
|
|
|
lat_str, lon_str = request.args["lat"], request.args["lon"]
|
|
|
|
lat, lon = float(lat_str), float(lon_str)
|
2021-04-16 21:30:25 +01:00
|
|
|
|
2021-04-17 17:02:53 +01:00
|
|
|
scotland_code = scotland.get_scotland_code(lat, lon)
|
2021-04-16 21:30:25 +01:00
|
|
|
|
2023-10-10 10:11:23 +01:00
|
|
|
elements: typing.Any
|
2021-04-16 21:30:25 +01:00
|
|
|
if scotland_code:
|
2021-04-17 18:29:09 +01:00
|
|
|
rows = wikidata.lookup_scottish_parish_in_wikidata(scotland_code)
|
|
|
|
hit = wikidata.commons_from_rows(rows)
|
2021-04-16 21:30:25 +01:00
|
|
|
elements = []
|
2021-04-17 18:29:09 +01:00
|
|
|
result = wikidata.build_dict(hit, lat, lon)
|
2021-04-16 21:30:25 +01:00
|
|
|
else:
|
2021-04-17 17:31:58 +01:00
|
|
|
elements = model.Polygon.coords_within(lat, lon)
|
2021-04-16 21:30:25 +01:00
|
|
|
result = do_lookup(elements, lat, lon)
|
|
|
|
|
|
|
|
return render_template(
|
|
|
|
"wikidata_tag.html", lat=lat, lon=lon, result=result, elements=elements
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2023-11-04 21:36:05 +00:00
|
|
|
def build_detail_page(lat: float, lon: float) -> str:
|
|
|
|
"""Run lookup and build detail page."""
|
2022-05-18 14:12:34 +01:00
|
|
|
try:
|
|
|
|
reply = lat_lon_to_wikidata(lat, lon)
|
|
|
|
except wikidata.QueryError as e:
|
|
|
|
query, r = e.args
|
2023-10-10 10:08:59 +01:00
|
|
|
return render_template("query_error.html", lat=lat, lon=lon, query=query, r=r)
|
2022-05-18 14:12:34 +01:00
|
|
|
|
2023-10-13 16:50:27 +01:00
|
|
|
element = reply["result"].pop("element", None)
|
2023-11-03 10:32:18 +00:00
|
|
|
geojson = reply["result"].pop("geojson", None)
|
2023-10-13 16:50:27 +01:00
|
|
|
|
|
|
|
return render_template(
|
2023-11-03 10:32:18 +00:00
|
|
|
"detail.html",
|
|
|
|
lat=lat,
|
|
|
|
lon=lon,
|
|
|
|
str=str,
|
|
|
|
element_id=element,
|
|
|
|
geojson=geojson,
|
2023-11-04 20:49:36 +00:00
|
|
|
**reply,
|
2023-10-13 16:50:27 +01:00
|
|
|
)
|
2021-04-16 21:30:25 +01:00
|
|
|
|
|
|
|
|
2023-11-04 21:36:05 +00:00
|
|
|
@app.route("/detail")
|
|
|
|
def detail_page() -> Response | str:
|
|
|
|
"""Detail page."""
|
2023-11-17 15:21:07 +00:00
|
|
|
database.session.execute("SELECT 1")
|
2023-11-04 21:36:05 +00:00
|
|
|
try:
|
|
|
|
lat_str, lon_str = request.args["lat"], request.args["lon"]
|
|
|
|
lat, lon = float(lat_str), float(lon_str)
|
|
|
|
except TypeError:
|
|
|
|
return redirect(url_for("index"))
|
|
|
|
|
|
|
|
return build_detail_page(lat, lon)
|
|
|
|
|
|
|
|
|
2021-04-16 16:21:38 +01:00
|
|
|
if __name__ == "__main__":
|
|
|
|
app.run(host="0.0.0.0")
|