Click Below to Get the Code

Browse, clone, and build from real-world templates powered by Harper.
Blog
GitHub Logo

Pub/Sub for AI: The New Requirements for Real-Time Data

Harper’s unified pub/sub architecture delivers real-time data, low-latency replication, and multi-protocol streaming for AI and edge applications. Learn how database-native MQTT, WebSockets, and SSE replace legacy brokers and pipelines, enabling millisecond decisions, resilient edge deployments, and globally consistent state for next-generation intelligent systems.
A.I.
Blog
A.I.

Pub/Sub for AI: The New Requirements for Real-Time Data

By
Ivan R. Judson, Ph.D.
November 18, 2025
By
Ivan R. Judson, Ph.D.
November 18, 2025
By
Ivan R. Judson, Ph.D.
November 18, 2025
November 18, 2025
Harper’s unified pub/sub architecture delivers real-time data, low-latency replication, and multi-protocol streaming for AI and edge applications. Learn how database-native MQTT, WebSockets, and SSE replace legacy brokers and pipelines, enabling millisecond decisions, resilient edge deployments, and globally consistent state for next-generation intelligent systems.
Ivan R. Judson, Ph.D.
Distinguished Solution Architect

AI systems increasingly depend on real-time pub/sub, MQTT event streams, and immediate access to fresh data—especially at the edge. From detecting fraud in financial transactions to spotting anomalies in industrial equipment to powering dynamic personalization, modern models must react to events in milliseconds, not seconds.

But today’s architectures force teams to glue together separate systems: an MQTT broker, a message queue, a feature store, a database, an API layer, and a replication engine. Every boundary introduces latency, inconsistency, and operational overhead. These problems become existential when deploying AI at the edge or across unstable networks.

Harper takes a different approach: it unifies pub/sub, MQTT, WebSockets, SSE, storage, replication, and vector search directly inside the database, creating a real-time platform purpose-built for AI.

This article explores why edge AI requires a new data architecture, why legacy pub/sub systems fall short, and how Harper transforms tables into real-time event streams that replicate globally in 50–100 ms.

AI Needs Real-Time Infrastructure, and Today’s Stack Isn’t Built for It

AI is moving from dashboards to decisions. Models are no longer grading what happened yesterday; they’re deciding what to do now.

How that plays out:

Predictive maintenance

Industrial assets stream high-frequency vibration and temperature data. Edge AI systems are expected to catch anomalies and trigger interventions in milliseconds, not minutes, as factories adopt predictive maintenance to avoid costly downtime.

Fraud detection

Card networks already run fraud scoring in about 50 ms per transaction, and research shows most real-time fraud systems operate at millisecond latency. Every extra round-trip risks either blocking a good customer or letting fraud through.

Personalization

Modern sites stream clickstream data into real-time pipelines so recommendations can update in seconds as users browse. If features lag behind behavior, relevance — and revenue — drop fast.

The common requirement

Across use cases, the bar looks similar:

  • Low latency for decisions on fresh events
  • Durable state feeding models
  • Reliable replication across sites
  • Operation during network interruptions
  • Multiple protocols feeding the same system
  • Edge-local inference that runs without cloud in the hot path

The problem: today, this is assembled from MQTT brokers, Kafka clusters, databases, sync layers, edge gateways, and more. Guides for building predictive maintenance or fraud platforms all assume a patchwork of services. It works, but only because teams pour time into stitching it together.

We need an architecture where real-time isn’t an add-on, it’s the default.

Cloud Deprecations Exposed the Cracks

Cloud-managed IoT and event services have given teams a harsh lesson in platform risk:

  • Google’s Cloud IoT Core was retired on August 16, 2023, with device connections to its HTTP and MQTT bridges shut down after that date.
  • AWS IoT Analytics ends support on December 15, 2025; after that, its console and resources will no longer be accessible.
    AWS IoT Events ends support on May 20, 2026, with access to its console and resources disabled after the deadline.

These services weren’t fringe tools, they sat in the middle of industrial IoT, retail telemetry, connected vehicles, and automation. As they were retired, customers were told to rebuild architectures and migrate pipelines on tight timelines, often at their own expense.

That exposes a fundamental truth: critical real-time and edge systems cannot depend on cloud services that can disappear with a product notice and a migration guide.

There’s another crack: hyperscale architectures still assume stable connectivity, an assumption that simply doesn’t hold in the real world. Factories, retail stores, vehicle fleets, and remote installations operate in conditions where networks are intermittent by default. As the Pratexo white paper notes, “intermittent connectivity is a hallmark challenge in edge environments,” especially those operating in remote or mobile contexts where links are “sporadic or unstable.” Systems in these settings must keep operating offline and then re-sync cleanly once connectivity returns—a capability widely cited as a core benefit of edge computing.

Harper is built around one principle: your real-time architecture should run anywhere—cloud, on-prem, or edge—with identical behavior, regardless of which managed services come and go.

The Architectural Reset: Database-Native Pub/Sub

Real-time pipelines traditionally separate the message broker from the database:

  • Events go through MQTT, Kafka, or Kinesis
  • State lives in a database
  • Sync logic glues them together

This creates the notorious “broker/database gap”—a latency and consistency window between the time an event arrives and the time it is durable.

Harper eliminates this gap.

Tables are topics. The database is the pub/sub system.

A record behaves like a retained message:

  • Publishing with retain atomically updates the record and notifies subscribers
  • Publishing without retain broadcasts ephemeral events
  • New subscribers automatically receive the current state from the table

No double-writes. No cross-system sync. No two sources of truth.

State and stream become the same thing.

Replication Designed for the Real World

Replication is the backbone of real-time systems, especially for AI at the edge. Traditional replication is often bolted on, brittle, or outsourced to a broker.

Harper’s replication engine, Plexus, merges the transaction log and replication pipeline into a single fast, efficient system.

Plexus enables:

  • Reliable, ordered replication
  • Automatic catch-up after disconnects
  • Peer-to-peer, self-healing mesh
  • Efficient batching + compression
  • Selective table or database replication
  • Bidirectional or upstream-only sync for tiered edge topologies

Replication typically completes in 50–100 ms, depending on network distance. Same-region deployments often perform even faster.

For AI pipelines where milliseconds matter, this reliability is essential.

Unified Protocol Support: MQTT, WebSockets, SSE

Real-time systems involve diverse clients:

  • IoT devices → MQTT
  • Browser apps → WebSockets
  • Monitoring dashboards → SSE

In legacy systems, each requires a separate backend service.

Harper flips the model:

All protocols subscribe to the same underlying tables.

Integrated support includes:

  • MQTT v3.1/v5 with wildcards, QoS 0–1, mTLS, retained messages
  • WebSockets with JSON, MessagePack, or CBOR
  • SSE for lightweight push streams

Every protocol sees the same state and publishes into the same system.
No routing rules, no protocol-specific brokers, no duplication.

Deploy Anywhere, Run Everywhere

AI doesn’t live in one environment, and real-time systems can’t either. Some workloads run in hyperscale clouds, others in industrial facilities, retail stores, vehicle fleets, or remote sites. The hard part isn’t computing; it’s the inconsistency between environments, where each location demands a different stack, sync model, and operational footprint.

Harper removes that divide.

Whether running on AWS, Azure, GCP, bare-metal servers, or embedded edge devices, Harper behaves the same. The APIs are the same. The pub/sub model is the same. The replication engine is the same.

Your architecture stays portable instead of being tied to a specific cloud vendor or the lifespan of a particular service.

This continuity isn’t just convenient, it’s strategic. When cloud providers retire key services or edge deployments need long-term stability, Harper offers a platform designed to last.

Operational Simplicity Is a Superpower

The modern cloud offers incredible tools, but stitching them together is often where reliability breaks down. Real-time architectures frequently rely on an MQTT broker, a message queue, an event bus, a streaming pipeline, a database, a feature store, and custom glue in between.

Harper replaces this patchwork with a single system.

When pub/sub is native to your database, you don’t synchronize events with state. When replication is built in, you don’t need a separate system to push deltas around the world. When your edge and cloud environments share one data model, you don’t need divergent code paths or fragile sync layers.

Simplicity becomes reliability.
Reliability becomes speed.
Speed becomes a competitive advantage.

And your engineering teams get to focus on what creates value, not on integrating half a dozen services.

A Unified Real-Time Data Layer for the AI Era

As AI moves to the edge and decisions shift from batch windows to milliseconds, organizations need an architecture that treats real-time as the default. They need a system where:

  • tables behave like topics
  • events and state are one operation
  • replication is automatic and resilient
  • protocols converge instead of fragment
  • deployments span cloud, on-prem, and edge without friction

That’s the world Harper built: applications that react instantly, operate anywhere, and scale globally without the complexity tax of conventional pub/sub architectures.

Harper Pub/Sub isn’t just a feature; it’s the foundation for the next generation of intelligent, distributed systems.

Ready to Build the Future of Real-Time AI?

If you’re exploring real-time pipelines, deploying AI at the edge, modernizing legacy pub/sub systems, or evaluating how to unify your data and streaming stack, now is the time to try Harper.

Spin up a free distributed cluster (no credit card required), subscribe to a table, publish an event, and watch your system respond instantly, anywhere in the world.

AI systems increasingly depend on real-time pub/sub, MQTT event streams, and immediate access to fresh data—especially at the edge. From detecting fraud in financial transactions to spotting anomalies in industrial equipment to powering dynamic personalization, modern models must react to events in milliseconds, not seconds.

But today’s architectures force teams to glue together separate systems: an MQTT broker, a message queue, a feature store, a database, an API layer, and a replication engine. Every boundary introduces latency, inconsistency, and operational overhead. These problems become existential when deploying AI at the edge or across unstable networks.

Harper takes a different approach: it unifies pub/sub, MQTT, WebSockets, SSE, storage, replication, and vector search directly inside the database, creating a real-time platform purpose-built for AI.

This article explores why edge AI requires a new data architecture, why legacy pub/sub systems fall short, and how Harper transforms tables into real-time event streams that replicate globally in 50–100 ms.

AI Needs Real-Time Infrastructure, and Today’s Stack Isn’t Built for It

AI is moving from dashboards to decisions. Models are no longer grading what happened yesterday; they’re deciding what to do now.

How that plays out:

Predictive maintenance

Industrial assets stream high-frequency vibration and temperature data. Edge AI systems are expected to catch anomalies and trigger interventions in milliseconds, not minutes, as factories adopt predictive maintenance to avoid costly downtime.

Fraud detection

Card networks already run fraud scoring in about 50 ms per transaction, and research shows most real-time fraud systems operate at millisecond latency. Every extra round-trip risks either blocking a good customer or letting fraud through.

Personalization

Modern sites stream clickstream data into real-time pipelines so recommendations can update in seconds as users browse. If features lag behind behavior, relevance — and revenue — drop fast.

The common requirement

Across use cases, the bar looks similar:

  • Low latency for decisions on fresh events
  • Durable state feeding models
  • Reliable replication across sites
  • Operation during network interruptions
  • Multiple protocols feeding the same system
  • Edge-local inference that runs without cloud in the hot path

The problem: today, this is assembled from MQTT brokers, Kafka clusters, databases, sync layers, edge gateways, and more. Guides for building predictive maintenance or fraud platforms all assume a patchwork of services. It works, but only because teams pour time into stitching it together.

We need an architecture where real-time isn’t an add-on, it’s the default.

Cloud Deprecations Exposed the Cracks

Cloud-managed IoT and event services have given teams a harsh lesson in platform risk:

  • Google’s Cloud IoT Core was retired on August 16, 2023, with device connections to its HTTP and MQTT bridges shut down after that date.
  • AWS IoT Analytics ends support on December 15, 2025; after that, its console and resources will no longer be accessible.
    AWS IoT Events ends support on May 20, 2026, with access to its console and resources disabled after the deadline.

These services weren’t fringe tools, they sat in the middle of industrial IoT, retail telemetry, connected vehicles, and automation. As they were retired, customers were told to rebuild architectures and migrate pipelines on tight timelines, often at their own expense.

That exposes a fundamental truth: critical real-time and edge systems cannot depend on cloud services that can disappear with a product notice and a migration guide.

There’s another crack: hyperscale architectures still assume stable connectivity, an assumption that simply doesn’t hold in the real world. Factories, retail stores, vehicle fleets, and remote installations operate in conditions where networks are intermittent by default. As the Pratexo white paper notes, “intermittent connectivity is a hallmark challenge in edge environments,” especially those operating in remote or mobile contexts where links are “sporadic or unstable.” Systems in these settings must keep operating offline and then re-sync cleanly once connectivity returns—a capability widely cited as a core benefit of edge computing.

Harper is built around one principle: your real-time architecture should run anywhere—cloud, on-prem, or edge—with identical behavior, regardless of which managed services come and go.

The Architectural Reset: Database-Native Pub/Sub

Real-time pipelines traditionally separate the message broker from the database:

  • Events go through MQTT, Kafka, or Kinesis
  • State lives in a database
  • Sync logic glues them together

This creates the notorious “broker/database gap”—a latency and consistency window between the time an event arrives and the time it is durable.

Harper eliminates this gap.

Tables are topics. The database is the pub/sub system.

A record behaves like a retained message:

  • Publishing with retain atomically updates the record and notifies subscribers
  • Publishing without retain broadcasts ephemeral events
  • New subscribers automatically receive the current state from the table

No double-writes. No cross-system sync. No two sources of truth.

State and stream become the same thing.

Replication Designed for the Real World

Replication is the backbone of real-time systems, especially for AI at the edge. Traditional replication is often bolted on, brittle, or outsourced to a broker.

Harper’s replication engine, Plexus, merges the transaction log and replication pipeline into a single fast, efficient system.

Plexus enables:

  • Reliable, ordered replication
  • Automatic catch-up after disconnects
  • Peer-to-peer, self-healing mesh
  • Efficient batching + compression
  • Selective table or database replication
  • Bidirectional or upstream-only sync for tiered edge topologies

Replication typically completes in 50–100 ms, depending on network distance. Same-region deployments often perform even faster.

For AI pipelines where milliseconds matter, this reliability is essential.

Unified Protocol Support: MQTT, WebSockets, SSE

Real-time systems involve diverse clients:

  • IoT devices → MQTT
  • Browser apps → WebSockets
  • Monitoring dashboards → SSE

In legacy systems, each requires a separate backend service.

Harper flips the model:

All protocols subscribe to the same underlying tables.

Integrated support includes:

  • MQTT v3.1/v5 with wildcards, QoS 0–1, mTLS, retained messages
  • WebSockets with JSON, MessagePack, or CBOR
  • SSE for lightweight push streams

Every protocol sees the same state and publishes into the same system.
No routing rules, no protocol-specific brokers, no duplication.

Deploy Anywhere, Run Everywhere

AI doesn’t live in one environment, and real-time systems can’t either. Some workloads run in hyperscale clouds, others in industrial facilities, retail stores, vehicle fleets, or remote sites. The hard part isn’t computing; it’s the inconsistency between environments, where each location demands a different stack, sync model, and operational footprint.

Harper removes that divide.

Whether running on AWS, Azure, GCP, bare-metal servers, or embedded edge devices, Harper behaves the same. The APIs are the same. The pub/sub model is the same. The replication engine is the same.

Your architecture stays portable instead of being tied to a specific cloud vendor or the lifespan of a particular service.

This continuity isn’t just convenient, it’s strategic. When cloud providers retire key services or edge deployments need long-term stability, Harper offers a platform designed to last.

Operational Simplicity Is a Superpower

The modern cloud offers incredible tools, but stitching them together is often where reliability breaks down. Real-time architectures frequently rely on an MQTT broker, a message queue, an event bus, a streaming pipeline, a database, a feature store, and custom glue in between.

Harper replaces this patchwork with a single system.

When pub/sub is native to your database, you don’t synchronize events with state. When replication is built in, you don’t need a separate system to push deltas around the world. When your edge and cloud environments share one data model, you don’t need divergent code paths or fragile sync layers.

Simplicity becomes reliability.
Reliability becomes speed.
Speed becomes a competitive advantage.

And your engineering teams get to focus on what creates value, not on integrating half a dozen services.

A Unified Real-Time Data Layer for the AI Era

As AI moves to the edge and decisions shift from batch windows to milliseconds, organizations need an architecture that treats real-time as the default. They need a system where:

  • tables behave like topics
  • events and state are one operation
  • replication is automatic and resilient
  • protocols converge instead of fragment
  • deployments span cloud, on-prem, and edge without friction

That’s the world Harper built: applications that react instantly, operate anywhere, and scale globally without the complexity tax of conventional pub/sub architectures.

Harper Pub/Sub isn’t just a feature; it’s the foundation for the next generation of intelligent, distributed systems.

Ready to Build the Future of Real-Time AI?

If you’re exploring real-time pipelines, deploying AI at the edge, modernizing legacy pub/sub systems, or evaluating how to unify your data and streaming stack, now is the time to try Harper.

Spin up a free distributed cluster (no credit card required), subscribe to a table, publish an event, and watch your system respond instantly, anywhere in the world.

Harper’s unified pub/sub architecture delivers real-time data, low-latency replication, and multi-protocol streaming for AI and edge applications. Learn how database-native MQTT, WebSockets, and SSE replace legacy brokers and pipelines, enabling millisecond decisions, resilient edge deployments, and globally consistent state for next-generation intelligent systems.

Download

White arrow pointing right
Harper’s unified pub/sub architecture delivers real-time data, low-latency replication, and multi-protocol streaming for AI and edge applications. Learn how database-native MQTT, WebSockets, and SSE replace legacy brokers and pipelines, enabling millisecond decisions, resilient edge deployments, and globally consistent state for next-generation intelligent systems.

Download

White arrow pointing right
Harper’s unified pub/sub architecture delivers real-time data, low-latency replication, and multi-protocol streaming for AI and edge applications. Learn how database-native MQTT, WebSockets, and SSE replace legacy brokers and pipelines, enabling millisecond decisions, resilient edge deployments, and globally consistent state for next-generation intelligent systems.

Download

White arrow pointing right

Explore Recent Resources

Blog
GitHub Logo

Pub/Sub for AI: The New Requirements for Real-Time Data

Harper’s unified pub/sub architecture delivers real-time data, low-latency replication, and multi-protocol streaming for AI and edge applications. Learn how database-native MQTT, WebSockets, and SSE replace legacy brokers and pipelines, enabling millisecond decisions, resilient edge deployments, and globally consistent state for next-generation intelligent systems.
A.I.
Blog
Harper’s unified pub/sub architecture delivers real-time data, low-latency replication, and multi-protocol streaming for AI and edge applications. Learn how database-native MQTT, WebSockets, and SSE replace legacy brokers and pipelines, enabling millisecond decisions, resilient edge deployments, and globally consistent state for next-generation intelligent systems.
A man with short dark hair, glasses, and a goatee smiles slightly, wearing a black shirt in front of a nature background.
Ivan R. Judson, Ph.D.
Distinguished Solution Architect
Blog

Pub/Sub for AI: The New Requirements for Real-Time Data

Harper’s unified pub/sub architecture delivers real-time data, low-latency replication, and multi-protocol streaming for AI and edge applications. Learn how database-native MQTT, WebSockets, and SSE replace legacy brokers and pipelines, enabling millisecond decisions, resilient edge deployments, and globally consistent state for next-generation intelligent systems.
Ivan R. Judson, Ph.D.
Nov 2025
Blog

Pub/Sub for AI: The New Requirements for Real-Time Data

Harper’s unified pub/sub architecture delivers real-time data, low-latency replication, and multi-protocol streaming for AI and edge applications. Learn how database-native MQTT, WebSockets, and SSE replace legacy brokers and pipelines, enabling millisecond decisions, resilient edge deployments, and globally consistent state for next-generation intelligent systems.
Ivan R. Judson, Ph.D.
Blog

Pub/Sub for AI: The New Requirements for Real-Time Data

Harper’s unified pub/sub architecture delivers real-time data, low-latency replication, and multi-protocol streaming for AI and edge applications. Learn how database-native MQTT, WebSockets, and SSE replace legacy brokers and pipelines, enabling millisecond decisions, resilient edge deployments, and globally consistent state for next-generation intelligent systems.
Ivan R. Judson, Ph.D.
Blog
GitHub Logo

Deliver Performance and Simplicity with Distributed Microliths

Distributed microliths unify data, logic, and execution into one high-performance runtime, eliminating microservice latency and complexity. By replicating a single coherent process across regions, they deliver sub-millisecond responses, active-active resilience, and edge-level speed. Platforms like Harper prove this model reduces infrastructure, simplifies operations, and scales globally with ease.
System Design
Blog
Distributed microliths unify data, logic, and execution into one high-performance runtime, eliminating microservice latency and complexity. By replicating a single coherent process across regions, they deliver sub-millisecond responses, active-active resilience, and edge-level speed. Platforms like Harper prove this model reduces infrastructure, simplifies operations, and scales globally with ease.
A man with short dark hair, glasses, and a goatee smiles slightly, wearing a black shirt in front of a nature background.
Ivan R. Judson, Ph.D.
Distinguished Solution Architect
Blog

Deliver Performance and Simplicity with Distributed Microliths

Distributed microliths unify data, logic, and execution into one high-performance runtime, eliminating microservice latency and complexity. By replicating a single coherent process across regions, they deliver sub-millisecond responses, active-active resilience, and edge-level speed. Platforms like Harper prove this model reduces infrastructure, simplifies operations, and scales globally with ease.
Ivan R. Judson, Ph.D.
Nov 2025
Blog

Deliver Performance and Simplicity with Distributed Microliths

Distributed microliths unify data, logic, and execution into one high-performance runtime, eliminating microservice latency and complexity. By replicating a single coherent process across regions, they deliver sub-millisecond responses, active-active resilience, and edge-level speed. Platforms like Harper prove this model reduces infrastructure, simplifies operations, and scales globally with ease.
Ivan R. Judson, Ph.D.
Blog

Deliver Performance and Simplicity with Distributed Microliths

Distributed microliths unify data, logic, and execution into one high-performance runtime, eliminating microservice latency and complexity. By replicating a single coherent process across regions, they deliver sub-millisecond responses, active-active resilience, and edge-level speed. Platforms like Harper prove this model reduces infrastructure, simplifies operations, and scales globally with ease.
Ivan R. Judson, Ph.D.