Part of the original design goal of’s Triple Redundant Architecture was to ensure scalability in times of load spikes outside of the bounds of the original traffic specs. Because the cluster is configured as an N+1 architecture, we can respond to legitimate traffic events by removing a node from the cluster, upsizing it, returning it into rotation, and then repeating the process on the next node in turn.

Scaling Process and Procedure 

The scaling process isn’t automatic and requires manual effort. It may be initiated in two ways.

  • On customer request via a ticket. We strongly recommend notifying us ahead of time if you know a large traffic event is coming (a major product launch, Black Friday, etc.) We can’t guarantee a turnaround time on a resizing unless given prior notice.
  • High load incidents detected by our monitoring system.

If the load is diagnosed to be due to a bot or crawler that we can block, we attempt to block it. This prevents unnecessary scaling, which prevents unnecessary costs to you. If it isn’t a bot or can’t be blocked, then the upscaling process detailed above is started. Be advised that this process may take up to 60-90 minutes depending on the diagnostic steps needed.

We open a support ticket to notify you of such changes, but we don’t wait for your response before upscaling your cluster. The uptime of your application is our top priority and reactive scaling events are part of how we ensure that we meet the obligations of our Service Level Agreement.