A RetroSearch Logo

Home - News ( United States | United Kingdom | Italy | Germany ) - Football scores

Search Query:

Showing content from https://cloud.google.com/kubernetes-engine/docs/concepts/regional-clusters below:

Regional clusters | Google Kubernetes Engine (GKE)

Regional clusters

Stay organized with collections Save and categorize content based on your preferences.

This page explains how regional clusters work in Google Kubernetes Engine (GKE).

Regional clusters increase the availability of a cluster by replicating the control plane across multiple zones in a region.

This configuration provides the following benefits:

Also, by default, regional clusters are created as multi-zonal clusters, so worker nodes are distributed across multiple zones in a region. This increases the availability of your workload, if you run sufficient replicas of the workload.

GKE Autopilot clusters are always regional. If you use GKE Standard, you can choose to create regional or zonal clusters. To learn about the different cluster availability types, see Cluster availability.

In regional clusters, including Autopilot clusters, the control plane is replicated across multiple zones of a region. GKE automatically replicates nodes across zones in the same region. In Standard clusters and node pools, you can optionally manually specify the zone(s) in which the nodes run. All zones must be within the same region as the control plane.

Note: Use regional clusters to run your production workloads, as they generally offer higher availability than zonal clusters. For more information about region-specific considerations, see Geography and regions.

After creating a regional cluster, you cannot change it to a zonal cluster.

How regional clusters work

Regional clusters replicate the cluster's control plane and nodes across multiple zones within a single region. For example, using the default configuration, a regional cluster in the us-east1 region creates multiple replicas of the control plane in different us-east1 zones and provisions nodes in three us-east1 zones: us-east1-b, us-east1-c, and us-east1-d. In the event of an infrastructure outage, Autopilot workloads continue to run and GKE automatically rebalances nodes. If you use Standard clusters, you must rebalance nodes manually or by using the cluster autoscaler.

Limitations Pricing

All Autopilot clusters are regional, and are subject to the Autopilot pricing model.

In Standard mode, regional clusters require more of your project's regional quotas than a similar zonal or multi-zonal cluster. Ensure that you understand your quotas and Standard pricing before using regional clusters. If you encounter an Insufficient regional quota to satisfy request for resource error, your request exceeds your available quota in the current region.

Also, you are charged for node-to-node traffic across zones. For example, if a workload running in one zone needs to communicate with a workload in a different zone, the cross-zone traffic incurs cost. For more information, see Egress between zones in the same region (per GB) in the Compute Engine pricing page.

Persistent storage in regional clusters

Zonal persistent disks are zonal resources and regional persistent disks are multi-zonal resources. When adding persistent storage unless a zone is specified, GKE assigns the disk to a single, random zone. To learn how to control the zones, see Zones in persistent disks.

Autoscaling regional clusters

Keep the following considerations in mind when using the cluster autoscaler to automatically scale node pools in regional Standard mode clusters.

You can also learn more about Autoscaling limits for regional clusters or about how Cluster Autoscaler balances across zones.

These considerations only apply to Standard mode clusters with the cluster autoscaler.

Overprovisioning scaling limits

To maintain capacity in the unlikely event of zonal failure, you can allow GKE to overprovision your scaling limits, to ensure a minimum level of availability even when some zones are unavailable.

For example, if you overprovision a three-zone cluster to 150% (50% excess capacity), you can ensure that 100% of traffic is routed to available zones if one-third of the cluster's capacity is lost. In the preceding example, you would accomplish this by specifying a maximum of six nodes per zone rather than four. If one zone fails, the cluster scales to 12 nodes in the remaining zones.

Similarly, if you overprovision a two-zone cluster to 200%, you can ensure that 100% of traffic is rerouted if half of the cluster's capacity is lost.

You can learn more about the cluster autoscaler or read the FAQ for autoscaling in the Kubernetes documentation.

What's next

Except as otherwise noted, the content of this page is licensed under the Creative Commons Attribution 4.0 License, and code samples are licensed under the Apache 2.0 License. For details, see the Google Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.

Last updated 2025-07-02 UTC.

[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-07-02 UTC."],[],[]]


RetroSearch is an open source project built by @garambo | Open a GitHub Issue

Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo

HTML: 3.2 | Encoding: UTF-8 | Version: 0.7.4