fix missing alpr_cache.py

This commit is contained in:
Will Freeman
2025-12-05 11:38:46 -07:00
parent 8700286aef
commit 940e385dbe
2 changed files with 145 additions and 1 deletions

2
.gitignore vendored
View File

@@ -44,7 +44,7 @@ terraform.tfstate*
# Lambda Python Stuff
serverless/*/lambda.zip
serverless/*/src/*
!serverless/*/src/alpr_counts.py
!serverless/*/src/alpr_cache.py
!serverless/*/src/alpr_clusters.py
!serverless/*/src/requirements.txt
!serverless/*/src/Dockerfile

View File

@@ -0,0 +1,144 @@
import json
from collections import defaultdict
from typing import Any
import boto3
import os
import time
import math
import requests
import re
from concurrent.futures import ThreadPoolExecutor
def terraform_rate_expression_to_minutes(rate_expression: str) -> int:
match = re.match(r"rate\((\d+)\s*(day|hour|minute)s?\)", rate_expression)
if not match:
raise ValueError(f"Invalid rate expression: {rate_expression}")
value, unit = int(match.group(1)), match.group(2)
if unit == "day":
return value * 24 * 60
elif unit == "hour":
return value * 60
elif unit == "minute":
return value
else:
raise ValueError(f"Unsupported time unit: {unit}")
UPDATE_RATE_MINS = terraform_rate_expression_to_minutes(os.getenv("UPDATE_RATE_MINS", "rate(60 minutes)"))
GRACE_PERIOD_MINS = int(2) # XXX: set expiration a few minutes after in case put object takes a while
TILE_SIZE_DEGREES = int(20)
WHITELISTED_TAGS = [
"operator",
"manufacturer",
"direction",
"brand",
"camera:direction",
"surveillance:brand",
"surveillance:operator",
"surveillance:manufacturer"
]
def get_all_nodes():
# Set up the Overpass API query
query = """
[out:json];
node["man_made"="surveillance"]["surveillance:type"="ALPR"];
out body;
"""
url = "http://overpass-api.de/api/interpreter"
response = requests.get(
url, params={"data": query}, headers={"User-Agent": "DeFlock/1.0"}
)
response.raise_for_status()
return response.json()["elements"]
def segment_regions(nodes: Any, tile_size_degrees: int) -> dict[str, list[Any]]:
print("Segmenting regions...")
tile_dict = defaultdict(list)
for node in nodes:
lat, lon = node["lat"], node["lon"]
tile_lat = math.floor(lat / tile_size_degrees) * tile_size_degrees
tile_lon = math.floor(lon / tile_size_degrees) * tile_size_degrees
bare_node = {
"id": node["id"],
"lat": lat,
"lon": lon,
"tags": {k: v for k, v in node["tags"].items() if k in WHITELISTED_TAGS},
}
tile_dict[f"{tile_lat}/{tile_lon}"].append(bare_node)
print("Region segmentation complete.")
return tile_dict
def lambda_handler(event, context):
nodes = get_all_nodes()
regions_dict = segment_regions(nodes=nodes, tile_size_degrees=TILE_SIZE_DEGREES)
print("Uploading data to S3...")
s3 = boto3.client("s3")
bucket_new = os.getenv("OUTPUT_BUCKET", "cdn.deflock.me")
# TODO: handle outdated index files when their referenced files are deleted
epoch = int(time.time())
tile_index = {
"expiration_utc": epoch + (UPDATE_RATE_MINS + GRACE_PERIOD_MINS) * 60,
"regions": list(regions_dict.keys()),
"tile_url": "https://cdn.deflock.me/regions/{lat}/{lon}.json?v=" + str(epoch),
"tile_size_degrees": TILE_SIZE_DEGREES,
}
print("Uploading regions to S3...")
def upload_json_to_s3(bucket, key, body):
s3.put_object(
Bucket=bucket,
Key=key,
Body=body,
ContentType="application/json",
)
# Use ThreadPoolExecutor for concurrent uploads
with ThreadPoolExecutor(max_workers=10) as executor:
futures = []
for latlng_string, elements in regions_dict.items():
lat, lon = latlng_string.split("/")
key = f"regions/{lat}/{lon}.json"
body = json.dumps(elements)
futures.append(executor.submit(upload_json_to_s3, bucket_new, key, body))
# add index file
futures.append(executor.submit(upload_json_to_s3, bucket_new, "regions/index.json", json.dumps(tile_index)))
# Wait for all futures to complete
for future in futures:
future.result()
print("Regions uploaded to S3. Done!")
return {
"statusCode": 200,
"body": "Successfully cached OSM nodes",
}
if __name__ == "__main__":
from pathlib import Path
nodes_file_path = Path("nodes.json")
if nodes_file_path.exists():
nodes = json.load(nodes_file_path.open())
else:
nodes = get_all_nodes()
with nodes_file_path.open("w") as f:
json.dump(nodes, f)
regions_dict = segment_regions(nodes=nodes, tile_size_degrees=5)
with open("regions_dict.json", "w") as f:
json.dump(regions_dict, f)