Creating Clusters

The following steps describe the process of creating new clusters as well as performing Day-2 management operations such as scaling, resizing, reconfiguring etc. on various clouds.


The following is the deployment architecture for an AWS cluster.

The Kubernetes nodes are distributed across multiple AZs to achieve high availability. For each of the AZ that you choose, a public subnet and a private subnet is created.

All the control plane nodes and worker nodes are created within the private subnets so there is no direct public access available.

A NAT gateway is created in the public subnet of each AZ, to allow nodes in the private subnet to be able to go out to the internet or call other AWS services.

An Internet gateway is created for each VPC, to allow SSH access to the bastion node for debugging purposes. SSH into Kubernetes nodes is only available through the Bastion node. A bastion node helps to provide access to the ec2 instances. This is because the ec2 instances are created in a private subnet and the bastion node operates as a secure, single point of entry into the infrastructure. The bastion node can be accessed via SSH or RDP.

The APIServer endpoint is accessible through an ELB, which load balancing across all the control plane nodes.



Spectro Cloud creates compute, network, and storage resources on AWS during the provisioning of Kubernetes clusters. The following pre-requisites should be met for the successful creation of clusters.

Resource Capacity

Sufficient capacity in the desired AWS region should exist for the creation of the following resources:

  • vCpu
  • VPC
  • Elastic IP
  • Internet Gateway
  • Elastic Load Balancers
  • NAT Gateway

AWS Cloud Account Permissions

To create an AWS cloud account, an access key as well as a secret access key will be needed.

Ensure that the IAM user or the ROOT user has the following minimum permissions:

"Version": "2012-10-17",
"Statement": [
"Effect": "Allow",
"Action": [
"Resource": [
"Effect": "Allow",
"Action": [
"Resource": [
"Condition": {
"StringLike": {
"iam:AWSServiceName": ""
"Effect": "Allow",
"Action": [
"Resource": [
"Effect": "Allow",
"Action": [
"Resource": [
The policy below cannot be used as an inline policy, as it exceeds the 2048 non-whitespace character limit by AWS.
The following warning on this policy is expected:

This policy defines some actions, resources, or conditions that do not provide permissions. To grant access, policies must have an action that has an applicable resource or condition.

Create an AWS Cluster

The following steps need to be performed to provision a new AWS cluster:

  • Provide basic cluster information like name, description, and tags. Tags on a cluster are propagated to the VMs deployed on the cloud/data center environments.

  • Select a cluster profile created for AWS cloud. The profile definition will be used as the cluster construction template.

  • Review and override pack parameters as desired. By default, parameters for all packs are set with values defined in the cluster profile.

  • Provide the AWS Cloud account and placement information.

    • Cloud Account - Select the desired cloud account. AWS cloud accounts with AWS credentials need to be pre-configured in project settings.
    • Region - Choose the desired AWS region where you would like the clusters to be provisioned.
    • SSH Key Pair Name - Choose the desired SSH Key pair. SSH key pairs need to be pre-configured on AWS for the desired regions. The selected key is inserted into the VMs provisioned.
    • Static Placement - By default, Spectro Cloud uses dynamic placement wherein a new VPC with a public and private subnet is created to place cluster resources for every cluster. These resources are fully managed by Spectro Cloud and deleted when the corresponding cluster is deleted. Turn on the Static Placement option if its desired to place resources into preexisting VPCs and subnets.
  • Configure the master and worker node pools. A master and a worker node pool are configured by default.

    • Name - a descriptive name for the node pool.

    • Size - Number of VMs to be provisioned for the node pool. For the master pool, this number can be 1, 3, or 5.

    • Allow worker capability (master pool) - Select this option for allowing workloads to be provisioned on master nodes.

    • Instance type - Select the AWS instance type to be used for all nodes in the node pool.

    • Availability Zones - Choose one or more availability zones. Spectro Cloud provides fault tolerance to guard against failures like hardware failures, network failures, etc. by provisioning nodes across availability

      zones if multiple zones are selected.

    • By default, worker pools are configured to use On-Demand instances. Optionally, to take advantage of discounted spot instance pricing, the ‘On-Spot’ option can be selected. This option allows you to specify a maximum bid price for the nodes as a percentage of the on-demand price. Spectro Cloud tracks the current price for spot instances and launches nodes when the spot pricefalls in the specified range.

  • Review settings and deploy the cluster. Provisioning status with details of ongoing provisioning tasks is available to track progress.

New worker pools may be added if its desired to customize certain worker nodes to run specialized workloads. As an example, the default worker pool may be configured with the ‘m3.large’ instance types for general-purpose workloads, and another worker pool with instance type ‘g2.2xlarge’ can be configured to run GPU workloads.

Scaling an AWS Cluster

Scaling a cluster up or down involves changing the size of node pools. The following steps need to be performed to scale up/down an AWS cluster.

  • Access the ‘Nodes’ view of the cluster.
  • For the desired node pool change the size directly from the nodes panel or by editing node pool settings.
  • After the node pool configuration is updated, the scale-up/down operation is initiated in a few minutes.
  • Provisioning status is updated with the ongoing progress of the scale operation.
Master node pool may be scaled from 1 to 3 or 3 to 5 nodes. Scale down operation is not supported for master nodes.

Add an AWS worker pool

The following steps need to be performed to add a new worker node pool to a cluster:-

  • Invoke the option to ‘Add Node Pool’ from the cluster’s node information page.

  • Provide node pool settings as follows:

    • A descriptive name for the node pool.
    • The number of nodes in the node pool.
    • One or more availability zones. Nodes are distributed across availability zones when multiple zones are selected.
    • The instance type to be used for all the nodes launched in the node pool.
    • Save the node pool settings. New worker pool settings are updated and cluster updates begin within a few minutes. The provisioning status is updated with the ongoing progress of tasks related to the addition of new nodes.

Remove an AWS worker pool

The following steps need to be performed to remove a worker pool from the cluster:-

  • Access the 'Nodes' view of the cluster.
  • Delete the desired worker pool and confirm the deletion.
  • Upon confirmation, the worker node deletion begins in a few minutes.

Reconfigure AWS nodes

The following steps need to be performed to reconfigure worker pool nodes:-

  • Access the 'Nodes' view of the cluster.
  • Edit the settings of the desired node pool.
  • Change the instance type to the desired instance type.
  • Save the node pool settings. After the node pool settings are updated, the node pool reconfiguration begins within a few minutes. The older nodes in the node pool are deleted one by one and replaced by new nodes launched with the new instance type configured.
  • The provisioning status is updated with the ongoing progress of nodes being deleted and added.