mirror of
https://github.com/BigBodyCobain/Shadowbroker.git
synced 2026-04-24 03:26:06 +02:00
668ce16dc7
Gate messages now propagate via the Infonet hashchain as encrypted blobs — every node syncs them through normal chain sync while only Gate members with MLS keys can decrypt. Added mesh reputation system, peer push workers, voluntary Wormhole opt-in for node participation, fork recovery, killwormhole scripts, obfuscated terminology, and hardened the self-updater to protect encryption keys and chain state during updates. New features: Shodan search, train tracking, Sentinel Hub imagery, 8 new intelligence layers, CCTV expansion to 11,000+ cameras across 6 countries, Mesh Terminal CLI, prediction markets, desktop-shell scaffold, and comprehensive mesh test suite (215 frontend + backend tests passing). Community contributors: @wa1id, @AlborzNazari, @adust09, @Xpirix, @imqdcr, @csysp, @suranyami, @chr0n1x, @johan-martensson, @singularfailure, @smithbh, @OrfeoTerkuci, @deuza, @tm-const, @Elhard1, @ttulttul
256 lines
7.8 KiB
Python
256 lines
7.8 KiB
Python
"""Geocoding proxy for Nominatim with caching and proper headers."""
|
|
|
|
from __future__ import annotations
|
|
|
|
import json
|
|
import os
|
|
import time
|
|
import threading
|
|
from typing import Any, Dict, List
|
|
from pathlib import Path
|
|
from urllib.parse import urlencode
|
|
|
|
from services.network_utils import fetch_with_curl
|
|
from services.fetchers.geo import cached_airports
|
|
|
|
_CACHE_TTL_S = 900
|
|
_CACHE_MAX = 1000
|
|
_cache: Dict[str, Dict[str, Any]] = {}
|
|
_cache_lock = threading.Lock()
|
|
_local_search_cache: List[Dict[str, Any]] | None = None
|
|
_local_search_lock = threading.Lock()
|
|
|
|
_USER_AGENT = os.environ.get(
|
|
"NOMINATIM_USER_AGENT", "ShadowBroker/1.0 (https://github.com/BigBodyCobain/Shadowbroker)"
|
|
)
|
|
|
|
|
|
def _get_cache(key: str):
|
|
now = time.time()
|
|
with _cache_lock:
|
|
entry = _cache.get(key)
|
|
if not entry:
|
|
return None
|
|
if now - entry["ts"] > _CACHE_TTL_S:
|
|
_cache.pop(key, None)
|
|
return None
|
|
return entry["value"]
|
|
|
|
|
|
def _set_cache(key: str, value):
|
|
with _cache_lock:
|
|
if len(_cache) >= _CACHE_MAX:
|
|
# Simple eviction: drop ~10% oldest keys
|
|
keys = list(_cache.keys())[: max(1, _CACHE_MAX // 10)]
|
|
for k in keys:
|
|
_cache.pop(k, None)
|
|
_cache[key] = {"ts": time.time(), "value": value}
|
|
|
|
|
|
def _load_local_search_cache() -> List[Dict[str, Any]]:
|
|
global _local_search_cache
|
|
with _local_search_lock:
|
|
if _local_search_cache is not None:
|
|
return _local_search_cache
|
|
|
|
results: List[Dict[str, Any]] = []
|
|
cache_path = Path(__file__).resolve().parents[1] / "data" / "geocode_cache.json"
|
|
try:
|
|
if cache_path.exists():
|
|
raw = json.loads(cache_path.read_text(encoding="utf-8"))
|
|
if isinstance(raw, dict):
|
|
for label, coords in raw.items():
|
|
if (
|
|
isinstance(label, str)
|
|
and isinstance(coords, list)
|
|
and len(coords) == 2
|
|
and all(isinstance(v, (int, float)) for v in coords)
|
|
):
|
|
results.append(
|
|
{
|
|
"label": label,
|
|
"lat": float(coords[0]),
|
|
"lng": float(coords[1]),
|
|
}
|
|
)
|
|
except Exception:
|
|
results = []
|
|
|
|
_local_search_cache = results
|
|
return _local_search_cache
|
|
|
|
|
|
def _search_local_fallback(query: str, limit: int) -> List[Dict[str, Any]]:
|
|
q = query.strip().lower()
|
|
if not q:
|
|
return []
|
|
|
|
matches: List[Dict[str, Any]] = []
|
|
seen: set[tuple[float, float, str]] = set()
|
|
|
|
for item in cached_airports:
|
|
haystacks = [
|
|
str(item.get("name", "")).lower(),
|
|
str(item.get("iata", "")).lower(),
|
|
str(item.get("id", "")).lower(),
|
|
]
|
|
if any(q in h for h in haystacks):
|
|
label = f'{item.get("name", "Airport")} ({item.get("iata", "")})'
|
|
key = (float(item["lat"]), float(item["lng"]), label)
|
|
if key not in seen:
|
|
seen.add(key)
|
|
matches.append(
|
|
{
|
|
"label": label,
|
|
"lat": float(item["lat"]),
|
|
"lng": float(item["lng"]),
|
|
}
|
|
)
|
|
if len(matches) >= limit:
|
|
return matches
|
|
|
|
for item in _load_local_search_cache():
|
|
label = str(item.get("label", ""))
|
|
if q in label.lower():
|
|
key = (float(item["lat"]), float(item["lng"]), label)
|
|
if key not in seen:
|
|
seen.add(key)
|
|
matches.append(item)
|
|
if len(matches) >= limit:
|
|
break
|
|
|
|
return matches
|
|
|
|
|
|
def _reverse_geocode_offline(lat: float, lng: float) -> Dict[str, Any]:
|
|
try:
|
|
import reverse_geocoder as rg
|
|
|
|
hit = rg.search((lat, lng), mode=1)[0]
|
|
city = hit.get("name") or ""
|
|
state = hit.get("admin1") or ""
|
|
country = hit.get("cc") or ""
|
|
parts = [city, state, country]
|
|
label = ", ".join([p for p in parts if p]) or "Unknown"
|
|
return {"label": label}
|
|
except Exception:
|
|
return {"label": "Unknown"}
|
|
|
|
|
|
def search_geocode(query: str, limit: int = 5, local_only: bool = False) -> List[Dict[str, Any]]:
|
|
q = query.strip()
|
|
if not q:
|
|
return []
|
|
limit = max(1, min(int(limit or 5), 10))
|
|
key = f"search:{q.lower()}:{limit}:{int(local_only)}"
|
|
cached = _get_cache(key)
|
|
if cached is not None:
|
|
return cached
|
|
if local_only:
|
|
results = _search_local_fallback(q, limit)
|
|
_set_cache(key, results)
|
|
return results
|
|
|
|
params = urlencode({"q": q, "format": "json", "limit": str(limit)})
|
|
url = f"https://nominatim.openstreetmap.org/search?{params}"
|
|
try:
|
|
res = fetch_with_curl(
|
|
url,
|
|
headers={
|
|
"User-Agent": _USER_AGENT,
|
|
"Accept-Language": "en",
|
|
},
|
|
timeout=6,
|
|
)
|
|
except Exception:
|
|
results = _search_local_fallback(q, limit)
|
|
_set_cache(key, results)
|
|
return results
|
|
|
|
results: List[Dict[str, Any]] = []
|
|
if res and res.status_code == 200:
|
|
try:
|
|
data = res.json() or []
|
|
for item in data:
|
|
try:
|
|
results.append(
|
|
{
|
|
"label": item.get("display_name"),
|
|
"lat": float(item.get("lat")),
|
|
"lng": float(item.get("lon")),
|
|
}
|
|
)
|
|
except (TypeError, ValueError):
|
|
continue
|
|
except Exception:
|
|
results = []
|
|
if not results:
|
|
results = _search_local_fallback(q, limit)
|
|
|
|
_set_cache(key, results)
|
|
return results
|
|
|
|
|
|
def reverse_geocode(lat: float, lng: float, local_only: bool = False) -> Dict[str, Any]:
|
|
key = f"reverse:{lat:.4f},{lng:.4f}:{int(local_only)}"
|
|
cached = _get_cache(key)
|
|
if cached is not None:
|
|
return cached
|
|
if local_only:
|
|
payload = _reverse_geocode_offline(lat, lng)
|
|
_set_cache(key, payload)
|
|
return payload
|
|
|
|
params = urlencode(
|
|
{
|
|
"lat": f"{lat}",
|
|
"lon": f"{lng}",
|
|
"format": "json",
|
|
"zoom": "10",
|
|
"addressdetails": "1",
|
|
}
|
|
)
|
|
url = f"https://nominatim.openstreetmap.org/reverse?{params}"
|
|
try:
|
|
res = fetch_with_curl(
|
|
url,
|
|
headers={
|
|
"User-Agent": _USER_AGENT,
|
|
"Accept-Language": "en",
|
|
},
|
|
timeout=6,
|
|
)
|
|
except Exception:
|
|
payload = _reverse_geocode_offline(lat, lng)
|
|
_set_cache(key, payload)
|
|
return payload
|
|
|
|
label = "Unknown"
|
|
if res and res.status_code == 200:
|
|
try:
|
|
data = res.json() or {}
|
|
addr = data.get("address") or {}
|
|
city = (
|
|
addr.get("city")
|
|
or addr.get("town")
|
|
or addr.get("village")
|
|
or addr.get("county")
|
|
or ""
|
|
)
|
|
state = addr.get("state") or addr.get("region") or ""
|
|
country = addr.get("country") or ""
|
|
parts = [city, state, country]
|
|
label = ", ".join([p for p in parts if p]) or (
|
|
data.get("display_name", "") or "Unknown"
|
|
)
|
|
except Exception:
|
|
label = "Unknown"
|
|
if label == "Unknown":
|
|
payload = _reverse_geocode_offline(lat, lng)
|
|
_set_cache(key, payload)
|
|
return payload
|
|
|
|
payload = {"label": label}
|
|
_set_cache(key, payload)
|
|
return payload
|