The runtime environment's compute resources mainly include Worker node resources, Proxy node resources, and Controller node resources. When your cluster resource utilization is high or the load is heavy, you can scale the cluster based on the following dimensions.
Cache instances are deployed on worker nodes. When worker node load is high or memory usage is high, you can increase the number of worker node instances.
Montplex Control components are deployed on the Controller node group. When the Controller node's load is high or memory usage is high, you can expand the number of instances in the Controller node group.
The Proxy component is deployed in the Proxy node group. The Proxy node group uses mixed instance deployment (On-Demand and Spot instances), with a default configuration of running up to 50% Spot instances.
The Proxy component uses HPA (Horizontal Pod Autoscaler) to monitor CPU resource utilization. It automatically scales elastically based on resource usage, adjusting the number of Proxy component replicas.
As the number of Proxy component replicas increases or decreases, the resource utilization of the Proxy node group will also increase or decrease. When the Proxy node group's resources are nearly exhausted, the cluster will activate the automatic scaling service for Proxy nodes, increasing the number of nodes in the Proxy group. When the Proxy node group's resources are excessive, the cluster will activate the automatic scaling-in service for Proxy nodes, reducing the number of nodes in the Proxy group.
When there is a business need to deploy in a new, previously unused availability zone, you can add a Worker node group for the new availability zone.
This process will launch the specified number of EC2 instances in the new availability zone and add them to the cluster. If the VPC doesn't have enough available subnets, the creation will fail. This process is expected to take about 3-5 minutes.