Resource Quota

Palette allocates quota to users using the Palette cluster group(s) to try out our free tier offering. Users may launch (vanilla) Palette Virtual Clusters or deploy apps to new Palette Virtual Clusters or existing ones. Uses will use the quota defined at the system level while exploring the capabilities offered by Palette (SaaS). s At the same time, our Enterprise users may offer the capabilities mentioned earlier to their internal developers, i.e., enterprise developers. These enterprise users (in theory) could use the system-level quota offered to every other developer. Still, in all likelihood, the tenant admin will set up different cluster groups(s) and quotas for enterprise developers. You may view this as the enterprise developer's quota offered at the tenant level. Enterprise developers may not exceed their system-level quota while launching Palette Virtual Clusters in system-level cluster group(s). In addition, they may not exceed their tenant-level quota while launching Palette Virtual Clusters in their tenant-level cluster groups.

Manage the developer Quota

The developer quota can be managed through the Palette Platform as below:

  1. Login to the Palette console as Tenant Admin, go to Tenant Settings and select `Developer Settings.
  1. In the wizard to Manage Developer Settings, there is an option to Hide system-level cluster groups from tenant users using the toggle button. You will have to create your cluster groups by hiding the system-level cluster groups.
  1. Set the desired User Quotas values. The user quotas have the following default values:
Resource Requirement for two Palette Virtual ClustersDefault QuotaMinimum Value
CPU124
Memory16 GiB4 GiB
Storage20 GiB2 GiB

Note:

Palette users can track the status of the resource quota from the Left Main Menu of Palette Dev Engine console.

The Quota allocation can be interpreted as each user in this tenant may create up to two Palette Virtual Clusters with a cumulative CPU/memory/storage quota across all their Palette Virtual Clusters. Each virtual cluster may consume up to 2.5 vCPU and 2.5 GiB memory from the allocated quota.