Click Below to Get the Code

Browse, clone, and build from real-world templates powered by Harper.
Blog
GitHub Logo

How One Retailer Served 2 Million Product Pages During a Major Outage

When a major U.S. retailer's origin system went down, Harper seamlessly stepped in as a resilience layer, serving 2 million product page requests from a pre-warmed cache with ultra-low latency (P95 ~2ms), ensuring most customers never noticed the outage. Designed for high availability, Harper’s architecture—featuring pre-rendered full-page HTML, a fused stack for low latency, and geographically distributed nodes—allowed the retailer to preserve 80% of traffic and revenue during the incident. This event highlights the critical role of robust infrastructure in maintaining user experience and business continuity during backend failures.
Blog

How One Retailer Served 2 Million Product Pages During a Major Outage

By
Aleks Haugom
April 29, 2025
By
Aleks Haugom
April 29, 2025
By
Aleks Haugom
April 29, 2025
April 29, 2025
When a major U.S. retailer's origin system went down, Harper seamlessly stepped in as a resilience layer, serving 2 million product page requests from a pre-warmed cache with ultra-low latency (P95 ~2ms), ensuring most customers never noticed the outage. Designed for high availability, Harper’s architecture—featuring pre-rendered full-page HTML, a fused stack for low latency, and geographically distributed nodes—allowed the retailer to preserve 80% of traffic and revenue during the incident. This event highlights the critical role of robust infrastructure in maintaining user experience and business continuity during backend failures.
Aleks Haugom
Senior Manager of GTM & Marketing

When a top U.S. retailer’s origin system went down, 2 million shoppers never noticed.

Their product pages kept loading. Revenue kept flowing. And the engineering team could breathe a little easier.

Why? Harper was quietly running in the background, acting as a resilience layer built to handle exactly this kind of chaos.

In the video below, I sit down with Daniel Abbott, Technical Account Manager at Harper, to break down exactly how the system responded in real-time, how it was architected to handle failure gracefully, and what lessons your team can take away to prepare for the next outage you can’t afford to have.

What Happened: The Quick Version

  • A major U.S. retailer experienced a critical outage at their origin layer.
  • Harper instantly took over product page delivery with 40 million pre-rendered pages in cache.
  • Over the span of one hour, Harper served 2 million requests at a P95 latency of approximately 2 milliseconds.
  • The result: 80% of traffic was preserved during the downtime, and most customers never knew there was an issue.

Why Harper Was Built for This

Full-Page HTML, Pre-Warmed and Ready: The retailer preloads critical product pages into Harper via periodic cache warm-ups. Each page is stored as fully rendered HTML, so when the origin is unavailable, Harper serves the exact same experience—no degraded templates, no missing content.

Low Latency by Design: Harper’s fused stack approach means there are no extra network hops between compute, data, and cache. The result is consistently fast page delivery, even under heavy load. During the outage, Harper’s processing latency remained at or below 2 milliseconds for 95% of site visitors.

Distributed, Redundant, and Built for Failover: For this deployment, Harper was running across six geographically distributed locations (2 nodes at each location). Each node holds its own copy of the dataset, enabling local reads and eliminating the need to route across regions. Smart load balancing ensures only the healthiest, fastest nodes handle traffic.

Real Business Impact: This wasn’t just an engineering win. By keeping product pages live, the retailer preserved revenue, avoided customer frustration, and gained time to resolve the underlying issue. That’s the true value of resilience at the infrastructure layer.

Lessons for Engineering Teams

This incident is a clear reminder that the cloud doesn't guarantee resilience. Without proper architecture, origin failures still result in downtime, lost sales, and reputational damage.

Harper can mitigate this risk by eliminating the origin dependency at the moment it matters most. Its fully distributed design, combined with tight data coupling and smart page caching, allows teams to deliver consistent user experiences even when their backend systems are under stress.

Looking to Build Something Similar?

If your team supports a high-traffic retail website—or any application where uptime equals revenue—it's worth asking: how would your system handle a full origin failure?

If the answer isn’t crystal clear, Harper can help.

Book a demo—we’ll walk you through exactly how to implement this level of protection in your environment.

When a top U.S. retailer’s origin system went down, 2 million shoppers never noticed.

Their product pages kept loading. Revenue kept flowing. And the engineering team could breathe a little easier.

Why? Harper was quietly running in the background, acting as a resilience layer built to handle exactly this kind of chaos.

In the video below, I sit down with Daniel Abbott, Technical Account Manager at Harper, to break down exactly how the system responded in real-time, how it was architected to handle failure gracefully, and what lessons your team can take away to prepare for the next outage you can’t afford to have.

What Happened: The Quick Version

  • A major U.S. retailer experienced a critical outage at their origin layer.
  • Harper instantly took over product page delivery with 40 million pre-rendered pages in cache.
  • Over the span of one hour, Harper served 2 million requests at a P95 latency of approximately 2 milliseconds.
  • The result: 80% of traffic was preserved during the downtime, and most customers never knew there was an issue.

Why Harper Was Built for This

Full-Page HTML, Pre-Warmed and Ready: The retailer preloads critical product pages into Harper via periodic cache warm-ups. Each page is stored as fully rendered HTML, so when the origin is unavailable, Harper serves the exact same experience—no degraded templates, no missing content.

Low Latency by Design: Harper’s fused stack approach means there are no extra network hops between compute, data, and cache. The result is consistently fast page delivery, even under heavy load. During the outage, Harper’s processing latency remained at or below 2 milliseconds for 95% of site visitors.

Distributed, Redundant, and Built for Failover: For this deployment, Harper was running across six geographically distributed locations (2 nodes at each location). Each node holds its own copy of the dataset, enabling local reads and eliminating the need to route across regions. Smart load balancing ensures only the healthiest, fastest nodes handle traffic.

Real Business Impact: This wasn’t just an engineering win. By keeping product pages live, the retailer preserved revenue, avoided customer frustration, and gained time to resolve the underlying issue. That’s the true value of resilience at the infrastructure layer.

Lessons for Engineering Teams

This incident is a clear reminder that the cloud doesn't guarantee resilience. Without proper architecture, origin failures still result in downtime, lost sales, and reputational damage.

Harper can mitigate this risk by eliminating the origin dependency at the moment it matters most. Its fully distributed design, combined with tight data coupling and smart page caching, allows teams to deliver consistent user experiences even when their backend systems are under stress.

Looking to Build Something Similar?

If your team supports a high-traffic retail website—or any application where uptime equals revenue—it's worth asking: how would your system handle a full origin failure?

If the answer isn’t crystal clear, Harper can help.

Book a demo—we’ll walk you through exactly how to implement this level of protection in your environment.

When a major U.S. retailer's origin system went down, Harper seamlessly stepped in as a resilience layer, serving 2 million product page requests from a pre-warmed cache with ultra-low latency (P95 ~2ms), ensuring most customers never noticed the outage. Designed for high availability, Harper’s architecture—featuring pre-rendered full-page HTML, a fused stack for low latency, and geographically distributed nodes—allowed the retailer to preserve 80% of traffic and revenue during the incident. This event highlights the critical role of robust infrastructure in maintaining user experience and business continuity during backend failures.

Download

White arrow pointing right
When a major U.S. retailer's origin system went down, Harper seamlessly stepped in as a resilience layer, serving 2 million product page requests from a pre-warmed cache with ultra-low latency (P95 ~2ms), ensuring most customers never noticed the outage. Designed for high availability, Harper’s architecture—featuring pre-rendered full-page HTML, a fused stack for low latency, and geographically distributed nodes—allowed the retailer to preserve 80% of traffic and revenue during the incident. This event highlights the critical role of robust infrastructure in maintaining user experience and business continuity during backend failures.

Download

White arrow pointing right
When a major U.S. retailer's origin system went down, Harper seamlessly stepped in as a resilience layer, serving 2 million product page requests from a pre-warmed cache with ultra-low latency (P95 ~2ms), ensuring most customers never noticed the outage. Designed for high availability, Harper’s architecture—featuring pre-rendered full-page HTML, a fused stack for low latency, and geographically distributed nodes—allowed the retailer to preserve 80% of traffic and revenue during the incident. This event highlights the critical role of robust infrastructure in maintaining user experience and business continuity during backend failures.

Download

White arrow pointing right

Explore Recent Resources

News
GitHub Logo

Harper 5.0 Is Here: Open Source, RocksDB, and a Runtime Built for the Agentic Era

Harper 5.0 launches with a fully open-source core under Apache 2.0, RocksDB as a native storage engine alongside LMDB, and source-available Harper Pro. This release delivers a unified runtime purpose-built for agentic engineering, from prototype to production.
Product Update
News
Harper 5.0 launches with a fully open-source core under Apache 2.0, RocksDB as a native storage engine alongside LMDB, and source-available Harper Pro. This release delivers a unified runtime purpose-built for agentic engineering, from prototype to production.
Person with short dark hair and moustache, wearing a colorful plaid shirt, smiling outdoors in a forested mountain landscape.
Aleks Haugom
Senior Manager of GTM & Marketing
News

Harper 5.0 Is Here: Open Source, RocksDB, and a Runtime Built for the Agentic Era

Harper 5.0 launches with a fully open-source core under Apache 2.0, RocksDB as a native storage engine alongside LMDB, and source-available Harper Pro. This release delivers a unified runtime purpose-built for agentic engineering, from prototype to production.
Aleks Haugom
Apr 2026
News

Harper 5.0 Is Here: Open Source, RocksDB, and a Runtime Built for the Agentic Era

Harper 5.0 launches with a fully open-source core under Apache 2.0, RocksDB as a native storage engine alongside LMDB, and source-available Harper Pro. This release delivers a unified runtime purpose-built for agentic engineering, from prototype to production.
Aleks Haugom
News

Harper 5.0 Is Here: Open Source, RocksDB, and a Runtime Built for the Agentic Era

Harper 5.0 launches with a fully open-source core under Apache 2.0, RocksDB as a native storage engine alongside LMDB, and source-available Harper Pro. This release delivers a unified runtime purpose-built for agentic engineering, from prototype to production.
Aleks Haugom
Podcast
GitHub Logo

Maintaining Momentum: Versioning, Stability & the Road to Nuxt 5 with Daniel Roe

In this podcast episode, Daniel Roe, lead of the Nuxt framework, shares insights on Nuxt 3, 4, and the upcoming Nuxt 5 release. We discuss open-source development, upgrading Nuxt apps, Vue-powered full-stack web apps, version maintenance, and the future of modern web development.
Select*
Podcast
In this podcast episode, Daniel Roe, lead of the Nuxt framework, shares insights on Nuxt 3, 4, and the upcoming Nuxt 5 release. We discuss open-source development, upgrading Nuxt apps, Vue-powered full-stack web apps, version maintenance, and the future of modern web development.
Person with short hair wearing a light blue patterned shirt, smiling widely outdoors with blurred greenery and trees in the background.
Austin Akers
Head of Developer Relations
Podcast

Maintaining Momentum: Versioning, Stability & the Road to Nuxt 5 with Daniel Roe

In this podcast episode, Daniel Roe, lead of the Nuxt framework, shares insights on Nuxt 3, 4, and the upcoming Nuxt 5 release. We discuss open-source development, upgrading Nuxt apps, Vue-powered full-stack web apps, version maintenance, and the future of modern web development.
Austin Akers
Apr 2026
Podcast

Maintaining Momentum: Versioning, Stability & the Road to Nuxt 5 with Daniel Roe

In this podcast episode, Daniel Roe, lead of the Nuxt framework, shares insights on Nuxt 3, 4, and the upcoming Nuxt 5 release. We discuss open-source development, upgrading Nuxt apps, Vue-powered full-stack web apps, version maintenance, and the future of modern web development.
Austin Akers
Podcast

Maintaining Momentum: Versioning, Stability & the Road to Nuxt 5 with Daniel Roe

In this podcast episode, Daniel Roe, lead of the Nuxt framework, shares insights on Nuxt 3, 4, and the upcoming Nuxt 5 release. We discuss open-source development, upgrading Nuxt apps, Vue-powered full-stack web apps, version maintenance, and the future of modern web development.
Austin Akers
Blog
GitHub Logo

Most LLM Calls Are Waste. Here's the Math.

Semantic caching for LLMs can reduce API costs by 20–70% by reusing similar responses. Combined with deterministic routing and improved retrieval, enterprises can significantly lower LLM usage, though effectiveness varies by workload and improves over time.
Blog
Semantic caching for LLMs can reduce API costs by 20–70% by reusing similar responses. Combined with deterministic routing and improved retrieval, enterprises can significantly lower LLM usage, though effectiveness varies by workload and improves over time.
Person with short dark hair and moustache, wearing a colorful plaid shirt, smiling outdoors in a forested mountain landscape.
Aleks Haugom
Senior Manager of GTM & Marketing
Blog

Most LLM Calls Are Waste. Here's the Math.

Semantic caching for LLMs can reduce API costs by 20–70% by reusing similar responses. Combined with deterministic routing and improved retrieval, enterprises can significantly lower LLM usage, though effectiveness varies by workload and improves over time.
Aleks Haugom
Apr 2026
Blog

Most LLM Calls Are Waste. Here's the Math.

Semantic caching for LLMs can reduce API costs by 20–70% by reusing similar responses. Combined with deterministic routing and improved retrieval, enterprises can significantly lower LLM usage, though effectiveness varies by workload and improves over time.
Aleks Haugom
Blog

Most LLM Calls Are Waste. Here's the Math.

Semantic caching for LLMs can reduce API costs by 20–70% by reusing similar responses. Combined with deterministic routing and improved retrieval, enterprises can significantly lower LLM usage, though effectiveness varies by workload and improves over time.
Aleks Haugom
Blog
GitHub Logo

Build a Conversational AI Agent on Harper in 5 Minutes

Build a conversational AI agent in minutes using Harper’s unified platform. This guide shows how to create, deploy, and scale real-time AI agents with built-in database, vector search, and APIs—eliminating infrastructure complexity for faster development.
Blog
Build a conversational AI agent in minutes using Harper’s unified platform. This guide shows how to create, deploy, and scale real-time AI agents with built-in database, vector search, and APIs—eliminating infrastructure complexity for faster development.
A smiling man with a beard and salt-and-pepper hair stands outdoors with arms crossed, wearing a white button-down shirt.
Stephen Goldberg
CEO & Co-Founder
Blog

Build a Conversational AI Agent on Harper in 5 Minutes

Build a conversational AI agent in minutes using Harper’s unified platform. This guide shows how to create, deploy, and scale real-time AI agents with built-in database, vector search, and APIs—eliminating infrastructure complexity for faster development.
Stephen Goldberg
Apr 2026
Blog

Build a Conversational AI Agent on Harper in 5 Minutes

Build a conversational AI agent in minutes using Harper’s unified platform. This guide shows how to create, deploy, and scale real-time AI agents with built-in database, vector search, and APIs—eliminating infrastructure complexity for faster development.
Stephen Goldberg
Blog

Build a Conversational AI Agent on Harper in 5 Minutes

Build a conversational AI agent in minutes using Harper’s unified platform. This guide shows how to create, deploy, and scale real-time AI agents with built-in database, vector search, and APIs—eliminating infrastructure complexity for faster development.
Stephen Goldberg
Podcast
GitHub Logo

Inside PixiJS, AT Protocol, and Modern Game Development with Trezy Who

Trezy shares his journey from professional drummer and filmmaker to software engineer and open source maintainer. Learn about PixieJS, game development, AT Proto, BlueSky, data sovereignty, and how developers can confidently contribute to open source projects.
Select*
Podcast
Trezy shares his journey from professional drummer and filmmaker to software engineer and open source maintainer. Learn about PixieJS, game development, AT Proto, BlueSky, data sovereignty, and how developers can confidently contribute to open source projects.
Person with short hair wearing a light blue patterned shirt, smiling widely outdoors with blurred greenery and trees in the background.
Austin Akers
Head of Developer Relations
Podcast

Inside PixiJS, AT Protocol, and Modern Game Development with Trezy Who

Trezy shares his journey from professional drummer and filmmaker to software engineer and open source maintainer. Learn about PixieJS, game development, AT Proto, BlueSky, data sovereignty, and how developers can confidently contribute to open source projects.
Austin Akers
Mar 2026
Podcast

Inside PixiJS, AT Protocol, and Modern Game Development with Trezy Who

Trezy shares his journey from professional drummer and filmmaker to software engineer and open source maintainer. Learn about PixieJS, game development, AT Proto, BlueSky, data sovereignty, and how developers can confidently contribute to open source projects.
Austin Akers
Podcast

Inside PixiJS, AT Protocol, and Modern Game Development with Trezy Who

Trezy shares his journey from professional drummer and filmmaker to software engineer and open source maintainer. Learn about PixieJS, game development, AT Proto, BlueSky, data sovereignty, and how developers can confidently contribute to open source projects.
Austin Akers