Autopilot is a new mode of operation in Google Kubernetes Engine (GKE) that is designed to reduce the operational cost of managing clusters, optimize your clusters for production, and yield higher workload availability. The mode of operation refers to the level of flexibility, responsibility, and control that you have over your cluster. In addition to the benefits of a fully managed control plane and node automations, GKE offers two modes of operation:

  • Autopilot: GKE provisions and manages the cluster’s underlying infrastructure, including nodes and node pools, giving you an optimized cluster with a hands-off experience.
  • Standard: You manage the cluster’s underlying infrastructure, giving you node configuration flexibility.

With Autopilot, you no longer have to monitor the health of your nodes or calculate the amount of compute capacity that your workloads require. Autopilot supports most Kubernetes APIs, tools, and its rich ecosystem. You stay within GKE without having to interact with the Compute Engine APIs, CLIs, or UI, as the nodes are not accessible through Compute Engine, like they are in Standard mode. You pay only for the CPU, memory, and storage that your Pods request while they are running.

Autopilot clusters are pre-configured with an optimized cluster configuration that is ready for production workloads. This streamlined configuration follows GKE best practices and recommendations for cluster and workload setup and security. Some of these built-in settings (detailed in the table below) are immutable and other optional settings can be turned on or off.

Autopilot comes with a SLA that covers both the control plane and your Pods. With Autopilot, as the underlying infrastructure is abstracted away, you can focus on the Kubernetes API and your deployments. Autopilot uses the resource requirements that you define in your PodSpec and provisions the resources for the deployment such as CPU, memory, and persistent disks.

There are two main reasons why you might want to use the Standard mode of operation instead of Autopilot:

  • You require a higher level of control over your cluster configuration.
  • Your clusters must run workloads that do not meet Autopilot constraints.

Comparing Autopilot and Standard modes

With Autopilot, GKE manages many complexities of the lifecycle of your cluster. The following table shows options that are available depending on the mode of operation for the cluster:

  • Pre-configured: This setting is built-in and you cannot change it.
  • Default: This setting is turned on but you can override it.
  • Optional: This setting is turned off but you can turn it on.
OptionsAutopilot modeStandard mode
Basic cluster typeAvailability and version:

Pre-configuredRegional
DefaultRegular release channel
Availability and version:

Optional:
Regional or zonalRelease channel or static version
Nodes and node poolsManaged by GKE.Managed, configured, and specified by you.
Provisioning resourcesGKE dynamically provisions resources based on your Pod specification.You manually provision additional resources and set overall cluster size. Configure cluster autoscaling and node auto-provisioning to help automate the process.
Image typePre-configuredContainer-Optimized OS with containerdChoose one of the following:
Container-Optimized OS with containerdContainer-Optimized OS with DockerUbuntu with containerdUbuntu with DockerWindows Server LTSCWindows Server SAC
BillingPay per Pod resource requests (CPU, memory, and ephemeral storage)Pay per node (CPU, memory, boot disk)
SecurityPre-configured:
Workload IdentityShielded nodesSecure bootOptional:
Customer-managed encryption keys (CMEK)Application-layer secrets encryptionGoogle Groups for RBAC
Optional:
Workload IdentityShielded nodesSecure bootApplication-layer secrets encryptionBinary authorizationCustomer-managed encryption keys (CMEK)Google Groups for RBAC
NetworkingPre-configured:
VPC-native (alias IP)Maximum 32 Pods per nodeIntranode visibilityDefault:
Public clusterDefault CIDR ranges
Note: Ensure that you review your CIDR ranges to factor in expected cluster growth.Network name/subnetOptional:
Private clusterCloud NAT1 (private clusters only)Authorized networks
Optional:
VPC-native (alias IP)Maximum 110 Pods per nodeIntranode visibilityCIDR ranges and max cluster sizeNetwork name/subnetPrivate clusterCloud NAT1Network policyAuthorized networks
Upgrades, repair, and maintenancePre-configured:
Node auto-repairNode auto-upgradeMaintenance windowsSurge upgrades
Optional:
Node auto-repairNode auto-upgradeMaintenance windowsSurge upgrades
Authentication credentialsPre-configuredWorkload IdentityOptional:
Compute Engine service accountWorkload Identity
ScalingPre-configured: Autopilot handles all the scaling and configuring of your nodes.

Default:You configure Horizontal pod autoscaling (HPA)You configure Vertical Pod autoscaling (VPA)
Optional:
Node auto-provisioningYou configure cluster autoscaling.HPAVPA
LoggingPre-configuredSystem and workload loggingDefaultSystem and workload logging

Optional: System-only logging
MonitoringPre-configuredSystem monitoring

Optional: System and workload monitoring
DefaultSystem monitoring

Optional: System and workload monitoring
RoutingPre-configured: Pod-based routing. Network endpoint groups (NEGs) enabled.Choose node-based packet routing (default), or Pod-based routing.
Cluster add-onsPre-configured:
HTTP load balancingDefault:
Compute Engine persistent disk CSI DriverCompute Engine Filestore CSI DriverNodeLocal DNSCacheOptional:
Managed Anthos Service Mesh (Preview)Istio (Use Managed Anthos Service Mesh (Preview))
Optional:
Compute Engine persistent disk CSI DriverCompute Engine Filestore CSI DriverHTTP load balancingNodeLocal DNSCacheCloud BuildCloud RunCloud TPUConfig ConnectorManaged Anthos Service MeshKalmUsage metering

1Further configuration required to enable Cloud NAT on a cluster.

Unsupported cluster features

The following GKE cluster features are not supported for Autopilot clusters:

Compute Engine instances

For GKE versions prior to 1.21.4, the following instance types are not supported in Autopilot clusters:

Security

Storage

Add-ons and integrations

Scaling

Autopilot automatically scales the cluster’s resources based on your Pod specifications, so that you can focus on your Pods. To automatically increase or decrease the number of Pods, you can implement Horizontal pod autoscaling using the standard Kubernetes CPU or memory metrics, or using custom metrics through Cloud Monitoring.

Allowable resource ranges

The following table lists the allowable resource ranges for Autopilot Pods. All values apply to the sum of all container resource requests in the Pod, unless noted. Pod vCPU are available in increments of 0.25 vCPU. In addition to the minimum values, the CPU:memory ratio must be in the range of 1 vCPU:1 GiB to 1 vCPU:6.5 GiB. Resources outside of the allowable ratio ranges will be scaled up. For more information, see resource ranges and ratios management and resource limitation examples.

ResourceMinimum resourcesMaximum resources
Normal PodsDaemonSet PodsNormal and DaemonSet Pods
CPU250 mCPU10 mCPU28 vCPU2
Memory512 MiB10 MiB80 GiB2
Ephemeral storage10 MiB (per container)10 MiB (per container)10 GiB

2The maximum CPU and memory limits for normal Pods are further reduced by the sum total of the resource requests of all DaemonSet Pods.

Default container resource requests

Autopilot relies on what you specify in your deployment configuration to provision resources. If you do not specify resource requests for any container in the Pod, Autopilot applies default values. These defaults are designed to give the containers in your Pods an average amount of resources, which are suitable for many smaller workloads.Important: Google recommends that you explicitly set your resource requests for each container to meet your application requirements, as these default values might not be sufficient, or optimal.

Autopilot applies these default values to resources that are not defined in the Pod specification.

ResourceContainers in normal PodsContainers in DaemonSets
CPU500 mCPU50 mCPU
Memory2 GiB100 MiB
Ephemeral storage1 GiB100 MiB

For more information about Autopilot cluster limits, see Quotas and limits.

Workload limitations and restrictions

Autopilot supports most workloads that run your applications. In order for GKE to offer management of the nodes and provide you with a more streamlined operational experience, there are a few restrictions and limitations when compared to GKE Standard. Some of these limitations are security best practices, while others allow Autopilot clusters to be safely managed. Workload limitations apply to all Pods, including those launched by Deployments, DaemonSets, ReplicaSets, ReplicationControllers, StatefulSets, Jobs, and CronJobs.

Host options restrictions

HostPort and hostNetwork are not permitted because node management is handled by GKE. Using hostPath volumes in write mode is prohibited, while using hostPath volumes in read mode is allowed only for /var/log/ path prefixes. Using host namespaces in workloads is prohibited.

Linux workload limitations

Autopilot supports only the following Linux capabilities for workloads:

"SETPCAP", "MKNOD", "AUDIT_WRITE", "CHOWN", "DAC_OVERRIDE", "FOWNER",
"FSETID", "KILL", "SETGID", "SETUID", "NET_BIND_SERVICE", "SYS_CHROOT", "SETFCAP"

In GKE version 1.21 and later, the "SYS_PTRACE" capability is also supported for workloads.

Node selectors and node affinity

Zonal affinity topologies are supported. Node affinity and node selectors are limited for use only with the following keys: topology.kubernetes.io/regiontopology.kubernetes.io/zonefailure-domain.beta.kubernetes.io/regionfailure-domain.beta.kubernetes.io/zonecloud.google.com/gke-os-distributionkubernetes.io/os, and kubernetes.io/arch. Not all values of OS and arch are supported in Autopilot.

Node selectors and node affinities also support the cloud.google.com/gke-spot key to automatically provision Spot Pods in clusters running GKE version 1.21.4 and later.

No Container Threat Detection

Autopilot does not support Container Threat Detection.

No privileged Pods

Privileged mode for containers in workloads is mainly used to make changes to nodes, like changing kubelet or networking settings. With Autopilot clusters, node changes aren’t allowed, so these types of Pods are also not allowed. This restriction might impact some admin workloads.

Pod affinity and anti-affinity

Although GKE manages your nodes for you in Autopilot, you retain the ability to schedule your Pods. Autopilot supports Pod affinity , so that you can co-locate Pods together on a single node for network efficiency. For example, you can use Pod affinity to deploy frontend Pods on nodes with backend Pods. Pod affinity is limited for use only with the following keys: topology.kubernetes.io/regiontopology.kubernetes.io/zonefailure-domain.beta.kubernetes.io/region, and failure-domain.beta.kubernetes.io/zone.

Autopilot also supports anti-affinity, so that you can spread Pods across nodes to avoid single points of failure. For example, you can use Pod anti-affinity to prevent frontend Pods from co-locating with backend Pods.

Defaults and resource limitations when using Pod anti-affinity

Autopilot supports Pod anti-affinity, so that you can prevent two Pods from co-locating on the same node. When using anti-affinity, Autopilot must allocate additional compute resources to ensure proper Pod separation, as defined by the PodSpec. When using Pod anti-affinity, the defaults and minimum resource limits increase. For all containers listed in the PodSpec:

ResourceDefault value
CPU0.75 vCPU
Memory2 GiB
Ephemeral Storage1 GiB

When using Pod anti-affinity, the same resource limitation rules and logic apply, but with higher vCPU increments. Pod vCPU are offered in a minimum of 0.5 vCPU and increments of 0.5 vCPU (rounded up to the nearest increment). For example, if you request 0.66 vCPU total (among all your containers using anti-affinity), your PodSpec is modified during admission and set to 1 vCPU. Your Pod has full access to the higher resource, with the extra resource divided among the resource requests of all containers.

Tolerations supported only for workload separation

Tolerations are supported only for workload separation. Taints are automatically added by node auto-provisioning as needed.

Resource ranges and ratio management

  • Pod vCPU increments: Pod vCPU are available in increments of 0.25 vCPU (rounded up). For example, if you request 0.66 vCPU total (among all your containers), your PodSpec is modified during admission and set to 0.75. Your Pod has full access to the higher resource, with the extra resource divided among the resource requests of all containers. The minimum value is 250 milliCPU (mCPU). DaemonSet vCPU are offered in increments of 10 mCPU (rounded up to the nearest increment).
  • Memory to CPU ratio range: The ratio of memory (in GiB) to vCPU must be in the range 1 to 6.5 vCPU. For example, you can have a Pod with 1 vCPU and 1 GiB of memory, or 1 vCPU and 6.5 GiB memory, but not 1 vCPU and 7 GiB memory. In order to deliver the resources request, GKE scales up whichever resource is too low. For example, if you request 1 vCPU and 7 GiB memory, your PodSpec is modified to 1.25 vCPU and 7 GiB memory on admission. Similarly, if you request 1 vCPU and 800 MiB memory, your PodSpec is modified to 1 vCPU and 1 GiB RAM, with the additional resource divided among the containers.

The CPU and memory increment and ratio requirements and potential scale-up of resource requests is calculated after the defaults are applied to containers with missing resource requests.

Containers with no resource requests will default to the standard minimums of 500 mCPU and 1 GiB memory. For CPU and memory, when GKE scales a resource request up (for example, to meet the minimum requirement, or the ratio requirement), the additional resource is allocated evenly between containers. Rounded up values are distributed proportionally across containers. For example, a container that has twice as much memory than the other containers will receive twice as much additional memory.

  • Ephemeral storage: The available range is between 10 MiB and 10 GiB. This impacts the container writable layer and emptyDir mounts.

Ephemeral storage has a minimum request per container, so if the ephemeral storage requests for a container are less than the minimum, Autopilot increases the request to the minimum. Ephemeral storage does not have a minimum request per Pod. Ephemeral storage has a maximum request per Pod which is cumulative across all containers. If the cumulative value is more than the maximum, Autopilot scales the request back to the maximum while also ensuring that the ratio of requests between containers remains the same.

Resource limitation examples

Example 1: For a single container with < 250 mCPU minimum:

Container numberOriginal resource requestsModified requests
1CPU: 180 mCPU
Memory: 0.5 GiB
Ephemeral storage: 10 MiB
CPU: 250 mCPU
Memory: 0.5 GiB
Ephemeral storage: 10 MiB
Total Pod resourcesCPU: 250 mCPU
Memory: 0.5 GiB
Ephemeral storage: 10 MiB

Example 2: For multiple containers with a total of < 250 mCPU minimum, Autopilot distributes the remainder of the resources (up to 250 vCPU) evenly between all containers in the Pod.

Container numberOriginal resource requestsModified requests
1CPU: 70 mCPU
Memory: 0.5 GiB
Ephemeral storage: 10 MiB
CPU: 84 mCPU
Memory: 0.5 GiB
Ephemeral storage: 10 MiB
2CPU: 70 mCPU
Memory: 0.5 GiB
Ephemeral storage: 10 MiB
CPU: 83 mCPU
Memory: 0.5 GiB
Ephemeral storage: 10 MiB
3CPU: 70 mCPU
Memory: 0.5 GiB
Ephemeral storage: 10 MiB
CPU: 83 mCPU
Memory: 0.5 GiB
Ephemeral storage: 10 MiB
Total Pod resourcesCPU: 250 mCPU
Memory: 1.5 GiB
Ephemeral storage: 30 MiB

Example 3: For multiple containers with total resources >= 250 mCPU, the CPU is rounded to multiples of 250 mCPU and the extra CPU is spread across all containers in the ratio of their original requests. In this example, the original cumulative CPU is 320 mCPU and is modified to a total of 500 mCPU. The extra 180 mCPU is spread across the containers:

Container numberOriginal resource requestsModified requests
1CPU: 170 mCPU
Memory: 0.5 GiB
Ephemeral storage: 10 MiB
CPU: 266 mCPU
Memory: 0.5 GiB
Ephemeral storage: 10 MiB
2CPU: 80 mCPU
Memory: 0.5 GiB
Ephemeral storage: 10 MiB
CPU: 125 mCPU
Memory: 0.5 GiB
Ephemeral storage: 10 MiB
3CPU: 70 mCPU
Memory: 0.5 GiB
Ephemeral storage: 10 MiB
CPU: 109 mCPU
Memory: 0.5 GiB
Ephemeral storage: 10 MiB
4Init container, resources not definedWill receive Pod resources
Total Pod resourcesCPU: 500 mCPU
Memory: 1.5 GiB
Ephemeral storage: 30 MiB

Example 4: For a single container where the CPU is too low for the amount of memory (1 vCPU:6.5 GiB maximum). The maximum allowed ratio for CPU to memory is 1:6.5. If the ratio is higher than that, the CPU request is increased and then rounded up if necessary:

Container numberOriginal resource requestsModified requests
1CPU: 250 mCPU
Memory: 4 GiB
Ephemeral storage: 10 MiB
CPU: 750 mCPU
Memory: 4 GiB
Ephemeral storage: 10 MiB
Total Pod resourcesCPU: 750 mCPU
Memory: 4 GiB
Ephemeral storage: 10 MiB

Example 5: For a single container where the memory is too low for the amount of CPU (1 vCPU:1 GiB minimum). The minimum allowed ratio for CPU to memory is 1:1. If the ratio is lower than that, the memory request is increased.

Container numberOriginal resource requestsModified requests
1CPU: 4 vCPU
Memory: 1 GiB
Ephemeral storage: 10 MiB
CPU: 4 vCPU
Memory: 4 GiB
Ephemeral storage: 10 MiB
Total Pod resourcesCPU: 4 mCPU
Memory: 4 GiB
Ephemeral storage: 10 MiB

Example 6: For a single container with < 250 mCPU minimum, where after adjusting, the CPU is too low for the amount of memory (1 vCPU:6.5 GiB maximum).

Container numberOriginal resource requestsModified requests
1CPU: 100 mCPU
Memory: 50 MiB
Ephemeral storage: 10 MiB
CPU: 250 mCPU
Memory: 256 MiB
Ephemeral storage: 10 MiB
Total Pod resourcesCPU: 250 mCPU
Memory: 256 MiB
Ephemeral storage: 10 MiB

Example 7: For a single container with ephemeral storage requests > 10 GiB, the maximum allowed ephemeral storage request is 10 GiB. If the request is greater than the maximum value, the request is downscaled to 10 GiB.

Container numberOriginal resource requestsModified requests
1CPU: 250 mCPU
Memory: 256 MiB
Ephemeral storage: 11 GiB
CPU: 250 mCPU
Memory: 256 MiB
Ephemeral storage: 10 GiB
Total Pod resourcesCPU: 250 mCPU
Memory: 256 MiB
Ephemeral storage: 10 GiB

Example 8: For multiple containers with ephemeral storage requests > 10 GiB, all container ephemeral storage requests are downscaled to make the final cumulative storage request of 10 GiB.

Container numberOriginal resource requestsModified requests
1CPU: 250 mCPU
Memory: 256 MiB
Ephemeral storage: 5 GiB
CPU: 250 mCPU
Memory: 256 MiB
Ephemeral storage: 2.94 GiB
2CPU: 250 mCPU
Memory: 256 MiB
Ephemeral storage: 6 GiB
CPU: 250 mCPU
Memory: 256 MiB
Ephemeral storage: 3.53 GiB
3CPU: 250 mCPU
Memory: 256 MiB
Ephemeral storage: 6 GiB
CPU: 250 mCPU
Memory: 256 MiB
Ephemeral storage: 3.53 GiB
Total Pod resourcesCPU: 750 mCPU
Memory: 768 MiB
Ephemeral storage: 10 GiB

Security limitations

Container isolation

Autopilot enforces a hardened configuration for your Pods that provides enhanced security isolation and helps limit the impact of container escape vulnerabilities on your cluster:

  • The container runtime default seccomp profile is applied, by default, to all Pods in your cluster.
  • The CAP_NET_RAW container permission is dropped for all containers. The CAP_NET_RAW permission is not typically used and was the subject of multiple container escape vulnerabilities. The lack of CAP_NET_RAW might cause the use of ping to fail inside your container.
  • Workload Identity is enforced and prevents Pod access to the underlying Compute Engine service account and other sensitive node metadata.
  • Services with spec.ExternalIPs set are blocked to protect against CVE-2020-8554. These services are rarely used.
  • The following StorageTypes are allowed. Other StorageTypes are blocked because they require privileges over the node:"configMap", "csi", "downwardAPI", "emptyDir", "gcePersistentDisk", "hostPath", "nfs", "persistentVolumeClaim", "projected", "secret"

Pod security policies

Autopilot enforces settings that provide enhanced isolation for your containers. Kubernetes PodSecurityPolicy is not supported on Autopilot clusters. In GKE versions older than 1.21, OPA Gatekeeper and Policy Controller are also not supported.

Security boundaries in Autopilot

At the Kubernetes layer, the GKE Autopilot mode provides the Kubernetes API but removes permissions to use some highly privileged Kubernetes primitives, like privileged Pods, with the goal to limit the ability to access, modify, or directly control the node virtual machine (VM).

These restrictions are put in place for GKE Autopilot mode to limit workloads from having low-level access to the node VM, in order to allow Google Cloud to offer full management of nodes, and a Pod-level SLA.Important: The security boundary for GKE nodes is the single-tenant virtual machine, and as such the ability to access the node VM from pods is not considered a security boundary for Autopilot. Use of any node-level access is inconsistent with the features of GKE Autopilot, is not currently supported, and may be removed without notice. If you require node VM level access, consider using GKE Standard.

Our intent is to prevent unintended access to the node virtual machine. We accept submissions to that effect through the Google Vulnerability Reward Program (VRP) and will reward reports at the discretion of the Google VRP reward panel.

By design, privileged users, like cluster administrators, have full control of any GKE cluster. As a security best practice, we recommend that you avoid granting powerful GKE/Kubernetes privileges widely and instead use namespace admin delegation wherever possible as described in our multi-tenancy guidance.

Workloads on Autopilot continue to enjoy the same security as GKE Standard mode, where single-tenant VMs are provisioned in the user’s project for their exclusive use. And, like Standard, on each individual VM, Autopilot workloads within a cluster might run together on a VM with a kernel that is security-hardened, but shared.

Since the shared kernel represents a single security boundary, GKE recommends that if you require strong isolation, such as high-risk or untrusted workloads, run your workloads on GKE Standard clusters using GKE Sandbox to provide multi-layer security protection.

Other limitations

Certificate signing requests

You cannot create certificate signing requests within Autopilot.

External monitoring tools

Most external monitoring tools require access that is restricted. Solutions from several Google Cloud partners are available for use on Autopilot, however not all are supported, and custom monitoring tools cannot be installed on Autopilot clusters.

External services

External IP Services are not permitted on Autopilot clusters. To give a Service an external IP, you can use a LoadBalancer type of Service or use an Ingress to add the Service to an external IP shared among several services.

Init containers

Init containers run in serial before the application containers start. By default, GKE allocates the full resources available to the Pod to each init container.

Unlike your other containers, GKE recommends leaving the resource requests unspecified for init containers, so that the containers have the full resources. If you set lower resources, your init container is constrained unnecessarily, and if you set higher resources, then you might increase your bill for the lifetime of the Pod.

Managed namespaces

The kube-system namespace is managed, meaning that all resources in this namespace cannot be altered and new resources cannot be created.

No changes to nodes

Since GKE manages the nodes for you for Autopilot clusters, you cannot alter the nodes.

No conversion

Converting Standard clusters to Autopilot mode and converting Autopilot clusters to Standard mode is not supported.

No direct external inbound connections for private clusters

Autopilot clusters with private nodes do not have external IPs and cannot accept inbound connections directly. If you deploy services on a NodePort, you cannot access those services from outside the VPC, such as from the internet. To expose applications externally in Autopilot clusters, use Services. For more information, see Exposing applications using services.

No Pod bursting

For Standard clusters, Pods can be configured to burst into unused capacity on the node. For Autopilot clusters, since all Pods have limits set on requests, resource bursting is not possible. It is important to ensure that your Pod specification defines adequate resources for the resource requests, and does not rely on bursting.

No SSH

Since you’re no longer provisioning or managing the nodes in Autopilot, there’s no SSH access. GKE handles all operational aspects of the nodes, including node health and all Kubernetes components running on the nodes.

Resource limits

In an Autopilot cluster, each Pod is treated as a Guaranteed QoS Class Pod, with limits that are equal to requests. Autopilot automatically sets resource limits equal to requests if you do not have resource limits specified. If you do specify resource limits, your limits will be overridden and set to be equal to the requests.

Webhooks limitations

In GKE version 1.21 and later, you can also create mutating dynamic admission webhooks. However, Autopilot modifies mutating webhooks objects to add a namespace selector which excludes the resources in managed namespaces (e.g. kube-system) from being intercepted. Additionally, webhooks which specify one or more of following resources (and any of their sub-resources) in the rules, will be rejected:

- group: ""
  resource: nodes
- group: ""
  resource: persistentvolumes
- group: certificates.k8s.io
  resource: certificatesigningrequests
- group: authentication.k8s.io
  resource: tokenreviews

Troubleshooting

Cannot create a cluster: 0 nodes registered

When you create an Autopilot cluster, it fails with the following error:

All cluster resources were brought up, but: only 0 nodes out of 2 have registered.

To resolve the issue, ensure that the default Compute Engine service account is not disabled. Run the following command to check:

gcloud iam service-accounts describe SERVICE_ACCOUNT

Replace SERVICE_ACCOUNT with the numeric service account ID or service account email address (like 123456789876543212345 or [email protected]).

Nodes fail to scale up

After creating an Autopilot cluster, the logs show the following message:

    "napFailureReasons": [
            {
              "messageId": "no.scale.up.nap.pod.zonal.resources.exceeded",
              ...

This error refers to a noScaleUp event, where node auto-provisioning did not provision any node group for the Pod in the zone because doing so would violate resource limits.

If you encounter this error, confirm the following:

Pricing

One autopilot cluster or zonal cluster per billing account is free.

Cluster management fee of $0.10 per cluster/hour apply, except for Anthos clusters. User pods in autopilot clusters are billed per second for CPU cores, memory, and ephemeral storage, until a pod is deleted. Worker nodes in standard clusters accrue compute costs, until a cluster is deleted.

Take the next step

Start building on Google Cloud with $500 in free credits and 20+ always free products.