I was reading the Proximity Service chapter in System Design Volume 2 by Alex Xu, where one of the approaches for locating businesses near a specific location involved using "GeoHashes".
However, when using this approach to find nearby businesses, a LIKE
query is necessary, such as
WHERE geohash LIKE '9qbwac%'
Despite having an index on the geohash column in Postgres, this query will trigger a full table scan. As the number of businesses grows, a full table scan could significantly slow down the process. Hence, is this approach truly better?
I tried building a dummy table in Postgres with index on the geoHash column and inserted 1 million random entries. Running a WHERE
clause with LIKE
operator does full table scan.
My guess is you didn't set up the right index that would be able to support queries based on
like
or^@
, as already suggested by @jjanes. Here's an example demo at db<>fiddle, comparing operations based on geohash vs regular PostGIS geometries. For prefix checks,text_pattern_ops
will be enough, but I went withpg_trgm
to also compare<->
distance/similarity measurements (ignoring the fact that's not really the way to compute distance based on a geohash).PostGIS:
Geohash: