In decentralized networks, nodes come and go.

This isn’t an edge case—it’s the default.

Most storage systems treat churn as a rare event. When too many nodes leave or join, they either:

  • Pause writes


  • Slow down reads


  • Or trigger massive data reshuffling across the network


All three hurt reliability.

Walrus was designed with a different assumption: churn is constant.

The challenge isn’t just replacing nodes. It’s doing so without breaking availability while large volumes of data are already stored.

Walrus solves this with a multi-stage epoch change protocol.

Instead of switching reads and writes at a single fragile moment, Walrus splits responsibility:

  • New writes are directed to the incoming committee


  • Reads continue from the current committee


  • Each blob carries metadata indicating which epoch it belongs to


This simple separation prevents a resource race where outgoing nodes are forced to both serve users and migrate data at the same time.

Meanwhile, Red Stuff’s self-healing recovery ensures new nodes don’t need full blobs to catch up. They recover only the slivers they’re responsible for, using bandwidth proportional to the missing data—not the total stored data.

The result:

  • No downtime during reconfiguration


  • No rewrite of existing blobs


  • No network-wide bandwidth spikes

Walrus doesn’t freeze when the network changes.

It adapts while staying live.

This is what makes Walrus viable as long-term infrastructure—not just a research idea.

Next: why Walrus only needs ~4.5× replication and what that means for cost and scalability.

@Walrus 🦭/acc

$WAL

#Walrus