MAY 13, 2026

4 MIN READ

YANNICK UTARD

Introducing Altertable Workers

Introducing Altertable Workers

Most data platforms ask you to send the data to them. Altertable Workers flip that model; the runtime moves to where the data already lives.

Blog

Today, we are introducing Altertable Workers in early access: a way to run the Altertable data plane inside your own cloud, while keeping altertable.ai as the control plane.

The old bargain

For the last decade, cloud data products have offered the same deal: send us the data, we run the compute, you use the UI. That model made adoption easy and worked well enough when the primary job was scheduled ingestion, dashboards, and human-driven analysis.

It breaks when data becomes an active input into products, operations, and agents. The problem is no longer storage or visualization; it is topology. The data is in one place, the runtime is somewhere else, and every new use case adds another network path, another permission boundary, another source of latency, and another bill that grows with usage.

These are symptoms of one architectural assumption: the provider owns the data plane. In the AI era, that default is too rigid.

BYOC is hard to ship

The control-plane/data-plane split is not a new idea. The hard part is shipping it without creating a second product. Most platforms' runtimes are too large, too coupled to their own cloud, or too dependent on assumptions that disappear inside a customer cluster. They end up with half-products: different code paths, delayed features, heavier operations.

Altertable is different because the unit of execution is already a worker; small enough to run close to your data, powerful enough to execute real analytical workloads, and integrated enough to remain part of the same product.

How Altertable Workers work

Altertable Workers

altertable.ai is the control plane: catalog, authentication, connection configuration, UI, billing, audit, metadata, and orchestration. Altertable Workers are the data plane: source connectivity, query execution, Parquet reads and writes, object storage access, local cache, and result streaming.

A self-hosted worker is installed with Helm or as a Docker container. During installation it receives a one-time enrollment token, phones home to altertable.ai, and appears online in the UI. The worker fetches its configuration, connects to sources over your private network, and executes work inside your environment.

The important part is what does not have to happen. Your database does not need a public IP. You do not need a bastion or a vendor-facing endpoint. Your object storage credentials do not leave your infrastructure. Raw data does not round-trip through the control plane.

Under the hood, the worker is a Rust binary built around DuckDB. It speaks to the control plane, executes analytical work locally, and operates on data in object storage. The runtime is intentionally small; a BYOC data plane only works if the thing you ship is something customers can actually run: understandable, observable, upgradeable, and boring enough to fit existing infrastructure practices.

Currently, communication between clients and altertable.ai, and between altertable.ai and workers, runs over Arrow Flight SQL. Yesterday, the DuckDB team announced Quack, a native client-server protocol for DuckDB. We are very excited to adopt it; Quack will make this layer simpler and more performant by letting DuckDB instances talk to each other directly.

What this changes

Security asks for less. In the traditional model, the provider asks you to open network paths, allowlist IPs, export data, trust an external runtime. With a self-hosted worker, the architecture inverts: credentials stay local, object storage stays customer-controlled, raw data never enters Altertable's data path.

Compute becomes yours. Scaling is an infrastructure choice, not a vendor conversation. You choose the machines, memory, and isolation model. This matters especially for AI workloads; agents are exploratory by nature, and if every step is metered through a vendor-owned data plane, curiosity becomes a budget item.

One runtime, many placements. A hosted worker is a worker we provision. A self-hosted worker is a worker you run. Same control plane, same UI, same runtime. Hybrid setups are natural: managed workers for shared exploration, a self-hosted worker inside a production VPC for data that cannot leave your environment.

Early access

Altertable Workers are available in early access.

If your production data sits behind private networks, if raw data cannot leave your infrastructure, or if your AI workflows need customer-controlled compute, we should talk.

Book a demo and we will show you how to bring the Altertable data plane into your own cloud.

Share

Yannick Utard, Co-Founder at Altertable

Yannick Utard

Co-Founder

Experienced software & site reliability engineer with a passion for building scalable and efficient systems. Previously at Sorare & Front.

Stay Updated

Get the latest insights on data, AI, and modern infrastructure delivered to your inbox

For more information, please consult our Privacy Policy

Related Articles

Continue exploring topics related to this article

Lakehouse table formats in 2026
APRIL 14TH, 2026
Sylvain Utard

Lakehouse table formats in 2026

Product, Engineering, Data Stack

There is no single “winning” lakehouse table format in 2026. What has emerged instead is a more interesting split.

READ ARTICLE
The Data Stack Is Broken
MAY 20TH, 2025
Sylvain Utard

The Data Stack Is Broken

Data Stack, Architecture, Product

Most data sits idle—trapped behind complexity, bloated budgets, and brittle tooling. The modern data stack promised agility but delivered a slow, siloed maze.

READ ARTICLE
Rethinking the Lakehouse
JULY 30TH, 2025
Yannick Utard

Rethinking the Lakehouse

Architecture, Performance, Data Stack

Breaking down our storage and query architecture: why we're leaning into Apache Iceberg and why DuckDB is emerging as our real-time query engine of choice.

READ ARTICLE
Grep your lakehouse
MARCH 27TH, 2026
Sylvain Utard

Grep your lakehouse

Product, Performance, Engineering

Agents do not fail because they lack SQL generation. They fail because they lack a native way to retrieve the right slice of data before writing precise queries.

READ ARTICLE
AI's Event Backbone
MARCH 10TH, 2026
Sylvain Utard

AI's Event Backbone

Product, Performance, Engineering

AI-native products generate a new kind of infrastructure problem. Here's how to build the event backbone for your AI system.

READ ARTICLE
Altertable Logo

A lakehouse your apps, BI, and agents share

DuckDB workers on open formats, federated SQL across your existing systems,
and an MCP server for agents — at flat monthly pricing.