Show HN: Pontoon – Open-source customer data syncs

Hacker News (score: 11)
Found: August 01, 2025
ID: 626

Description

Other
Show HN: Pontoon – Open-source customer data syncs Hi HN,

We’re Alex and Kalan, the creators of Pontoon (https://github.com/pontoon-data/Pontoon). Pontoon is an open-source data export platform that makes it really easy to create data syncs and send data to your enterprise customers. Check out our demo here: https://app.storylane.io/share/onova7c23ai6 or try it out with docker: https://pontoon-data.github.io/Pontoon/getting-started/quick...

While at our prior roles as data engineers, we’ve both felt the pain of data APIs. We either had to spend weeks building out data pipelines in house or spend a lot on ETL tools like Fivetran (https://www.fivetran.com/). However, there were a few companies that offered data syncs that would sync directly to our data warehouse (eg. Redshift, Snowflake, etc.), and when that was an option, we always chose it. This led us to wonder “Why don’t more companies offer data syncs?”. It turns out, building reliable cross-cloud data syncs is difficult. That’s why we built Pontoon.

We designed Pontoon to be:

- Easily deployed: we provide a single, self-contained Docker image for easy deployment and Docker Compose for larger workloads (https://pontoon-data.github.io/Pontoon/getting-started/quick...)

- Support modern data warehouses: we support syncing to/from Snowflake, BigQuery, Redshift, and Postgres.

- Sync cross cloud: sync from BigQuery to Redshift, Snowflake to BigQuery, Postgres to Redshift, etc.

- Developer friendly: data syncs can also be built via the API

- Open source: Pontoon is free to use by anyone

Under the hood, we use Apache Arrow (https://arrow.apache.org/) to move data between sources and destinations. Arrow is very performant - we wanted to use a library that could handle the scale of moving millions of records per minute.

In the shorter-term, there are several improvements we want to make, like:

- Adding support for DBT models to make adding data models easier

- UX improvements like better error messaging and monitoring of data syncs

- More sources and destinations (S3, GCS, Databricks, etc.)

- Improve the API for a more developer friendly experience (it’s currently tied pretty closely to the front end)

In the longer-term, we want to make data sharing as easy as possible. As data engineers, we sometimes felt like second class citizens with how we were told to get the data we needed - “just loop through this api 1000 times”, “you probably won’t get rate limited” (we did), “we can schedule an email to send you a csv every day”. We want to change how modern data sharing is done and make it simple for everyone.

Give it a try https://github.com/pontoon-data/Pontoon. Cheers!

More from Hacker

Poking holes into bytecode with peephole optimisations

Poking holes into bytecode with peephole optimisations

Show HN: Run LLMs in Docker for any language without prebuilding containers

Show HN: Run LLMs in Docker for any language without prebuilding containers I&#x27;ve been looking for a way to run LLMs safely without needing to approve every command. There are plenty of projects out there that run the agent in docker, but they don&#x27;t always contain the dependencies that I need.<p>Then it struck me. I already define project dependencies with mise. What if we could build a container on the fly for any project by reading the mise config?<p>I&#x27;ve been using agent-en-place for a couple of weeks now, and it&#x27;s working great! I&#x27;d love to hear what y&#x27;all think

Bare metal programming with RISC-V guide (2023)

Bare metal programming with RISC-V guide (2023)

Fly's Sprites.dev addresses dev environment sandboxes and API sandboxes together

Fly's Sprites.dev addresses dev environment sandboxes and API sandboxes together

No other tools from this source yet.