AI-native products generate a new kind of infrastructure problem.
Every prompt, response, tool call, retry, and fallback produces data. Each model invocation carries metadata about latency, cost, tokens, and execution paths. When agents orchestrate multiple steps, they produce traces: trees of decisions, tool invocations, and intermediate outputs.
What used to be a product emitting a manageable stream of clicks and backend events is now a system continuously producing high-volume structured event data.
And once these systems reach production, a simple question quickly becomes difficult:
Where does all this event data go?
Not in theory. In practice. Where do you push millions of traces, tool invocations, and behavioral events so they land quickly, stay affordable to query, and remain easy to store?
This question matters more than it seems.
The new class of product telemetry
Traditional SaaS applications produced a predictable stream of events:
- page views
- clicks
- backend actions
- job logs
- billing usage
AI-native products produce all of those, and a second layer on top:
- prompts and responses
- model calls and token counts
- tool selections and tool outputs
- retries and fallbacks
- step-by-step workflow traces
- latency and cost measurements per invocation
This data is not fundamentally exotic. Most of it is still timestamped, structured, and append-only, just like event logs.
What has changed is the volume, the burstiness, and the importance of the data.
AI systems are probabilistic. Improving them requires observing how they behave across thousands of real interactions. Every conversation becomes a potential signal. Event data stops being passive telemetry and becomes the primary way to understand whether the product actually works. Not in the infrastructure sense (is the service up?), but in the business sense: which workflows deliver value, which models cost too much, and where the real opportunities for improvement lie.
Written like logs, queried like analytics
The challenge is not simply collecting these events. Human time should not be spent designing ingestion pipelines, operating storage infrastructure, or scaling systems to handle growing data volumes. It should be spent extracting business value from that data.
From the ingestion side, the workload looks like logging. Events should be easy to emit, cheap to store, and resilient to spikes in volume. Developers should not have to design complex ingestion pipelines just to record system behavior.
From the consumption side, the requirements look much closer to analytics. Teams want to explore this data with SQL, correlate it with product metrics, and analyze it over long time horizons.
They want to answer questions like:
- Which agent workflows lead to successful outcomes?
- Which tool calls introduce the most latency?
- Which model configuration delivers the best outcome per dollar?
- Where do retries or fallbacks occur most frequently?
- Which users or accounts generate the most expensive agent workloads?
These are analytical questions. They require joins, aggregations, filtering, segmentation, and exploration. They also require documentation: trusted, shared definitions of what each event means, what its fields represent, and how it was recorded; so that the answers different teams reach from the same data can actually be compared.
In practice, engineers want the ease of pushing data into a logging system with the analytical flexibility of a database.
Why existing stacks feel broken
There are already ways to build infrastructure for high-volume events. Most teams end up combining streaming systems, analytical databases, and data warehouses.
In practice, it tends to look like this: events go into a message broker (Kafka, RabbitMQ, SQS), a consumer sinks them to ClickHouse, another pipeline syncs a subset to the data warehouse for business reporting, and then someone builds a separate service that aggregates metrics for the product dashboard. Four systems, three pipelines, and a cron job holding it together.
This works until it doesn't. Consumer lag silently grows. Schema changes break the sink. The warehouse query that used to take two seconds now takes twenty. The dashboard team and the engineering team are looking at different numbers.
Message brokers and streaming platforms are not the problem. They do what they were designed to do: decouple services, distribute events across consumers, and enable real-time processing. If you need pub/sub, event sourcing, or inter-service messaging, that infrastructure earns its complexity.
But many teams reach for a broker not because they need distributed messaging, but because they have no other obvious way to handle high-volume event ingestion. Their actual goal is simpler: get events out of the application and into a system where they can be queried with SQL. For that, operating a full streaming tier (managing consumers, partitions, backpressure, and schema evolution) is overhead that does not serve the end goal.
Cloud data warehouses like BigQuery or Snowflake excel at large-scale analytical queries and centralized reporting. Both offer streaming ingest — Snowflake through Snowpipe, BigQuery through Datastream — but continuous small writes remain a cost and operational concern rather than a first-class experience. Per-query pricing creates anxiety about exploration. Teams end up building caching layers, aggregation pipelines, and summary tables just to make interactive queries affordable, which reintroduces the batch architecture they were trying to avoid.
The friction appears when a single stream of product and AI events must simultaneously support high-frequency ingestion, interactive analytics, and real-time product visibility. None of the existing tools were designed to do all three.
Product analytics was the first signal
The industry has already seen this pattern once before.
When product analytics became a serious discipline, teams discovered that behavioral events (page views, clicks, funnels) were genuinely difficult to store and query with existing infrastructure. Traditional databases were not built for append-heavy write patterns at consumer scale. Warehouses were too slow and expensive for interactive exploration. BI tools expected pre-aggregated data.
That's exactly why Amplitude, Mixpanel, and Segment emerged: not because the data was technically exotic, but because no existing system made it easy to ingest raw events continuously and explore them interactively without building a custom pipeline.
AI workloads are now amplifying the same tension. Agent traces, tool calls, and workflow telemetry generate far more event data than product analytics ever did, and make real-time understanding even more critical.
What product analytics revealed first, AI systems are now making unavoidable.
The need for an event backbone
What modern AI products require is not simply a logging pipeline or an analytics warehouse. They require a system that can act as an event backbone: a single layer capable of absorbing large volumes of append-only events, making them queryable immediately, and keeping them accessible alongside the rest of the company's data.
Most importantly, it must allow event data to connect with business context. Agent traces only become meaningful when they can be joined with users, accounts, product features, and cost data.
That query (which model configuration costs the most, broken down by plan and workflow) is impossible to answer if your traces live in one system and your user data lives in another.
When those connections exist, event streams transform from raw telemetry into actionable intelligence.
Why this matters for AI products
AI systems introduce a new operational loop. Instead of shipping deterministic code and observing whether it runs, teams deploy systems whose behavior evolves through prompts, models, and tool integrations.
Improving these systems requires analyzing how they behave across thousands or millions of interactions. Developers need to understand where workflows succeed, where they fail, and which parts of the system create unnecessary cost or latency.
The only reliable way to answer those questions is through event data.
Agent traces become the record of how the system actually behaves. By analyzing them alongside product and business data, teams can see which workflows deliver value and which ones require refinement.
That feedback loop does not work if the data is fragmented, slow to query, or expensive to explore.
Where Altertable fits
At Altertable, we have been building toward this architecture from the beginning.
Altertable is an operational lakehouse: an analytical database with built-in event ingestion, product analytics, and AI agents. Event data can be pushed at scale, queried interactively, and combined with the rest of the company's analytical context.
For teams that already run a message broker or streaming platform, Altertable serves as the analytical destination, replacing the ClickHouse or warehouse sink at the end of the pipeline. For teams that don't need that infrastructure, Altertable's ingestion API eliminates the streaming tier entirely. Events pushed through the API become queryable in seconds, with no consumers, no sinks, and no pipelines to maintain.
Either way, the same system that stores product events can serve interactive analytics, support SQL exploration, and power AI agents analyzing data in the background. And because streaming is a native state of the system, agents can analyze and act on data the moment it arrives, regardless of the ingestion path.
The event data generated by modern applications should not be fragmented across multiple tools and pipelines. It should form a coherent, queryable foundation that both humans and machines can explore.
AI-native products are making that need clearer every day.
As agents generate larger volumes of traces and telemetry, the systems that capture and analyze those events will become as fundamental as the databases that store application state.
AI systems need intelligence to function, but they also need a backbone for the events that intelligence produces.





