11-30, 10:40–11:00 (Asia/Seoul), Workshop Room
The Uber H3 library is a powerful geospatial indexing system that offers a versatile and efficient way to index and query geospatial data. It provides static indexing scheme that allows for fast and accurate calculations of geospatial distances, as well as easy partitioning of data into regions. In this proposal, we suggest using the Uber H3 indexing library in Postgres for geospatial data analytics.
Postgres is an open-source relational database management system that provides robust support for geospatial data processing through the PostGIS extension. PostGIS enables the storage, indexing, and querying of geospatial data in Postgres, and it offers a range of geospatial functions to manipulate and analyze geospatial data.
However, the performance of PostGIS can be limited in some cases when dealing with large datasets or complex queries. This is where the Uber H3 library can be of great use. By integrating Uber H3 indexing with Postgres, we can improve the performance of PostGIS, especially for operations that involve partitioning of data, distance calculations and zonal statistics
This talk will demonstrate the use of Uber H3 indexing library in Postgres for geospatial data processing through a series of examples and benchmarks. The proposed presentation will showcase the benefits of using Uber H3 indexing for geospatial data processing in Postgres, such as improved query performance and better partitioning of data.
The proposed presentation will be of interest to developers, data scientists, and geospatial analysts who work with geospatial data in Postgres. It will provide a practical guide to integrating Uber H3 indexing with Postgres, and offer insights into the performance gains and applications of this integration.
I have been honing my skills in the geospatial domain, gaining diverse experience in Climate tech startups, Agritech solutions, and more.
My experience extends to working with satellite data (including Sentinel and Landsat), geospatial data modeling, and handling large datasets at scale in the cloud using Docker, Python, S3, etc
I am most interested in building a generic spatial-temporal database that can handle a wide variety of data and use cases.