GKE logo

Google Kubernetes Engine (GKE)

The most scalable and fully automated Kubernetes service

Put your containers on autopilot and securely run your enterprise workloads at scale—with little to no Kubernetes expertise required.

Get one Zonal or Autopilot cluster free per month. New customers get $300 in free credits to try GKE and other Google Cloud products.

Features

Support for 65,000-node clusters for next gen AI

In anticipation of even larger models, we are introducing support for 65,000-node clusters. To develop cutting-edge AI models, teams need to be able to allocate computing resources across diverse workloads. This includes not only model training but also serving, inference, conducting ad hoc research, and managing auxiliary tasks. Centralizing computing power within the smallest number of clusters provides the flexibility to quickly adapt to changes in demand from inference serving, research, and training workloads.




Increased velocity, reduced risk, and lower TCO

With the new premium GKE Enterprise edition, platform teams benefit from increased velocity by configuring and observing multiple clusters from one place, defining configuration for teams rather than clusters, and providing self-service options for developers for deployment and management of apps. You can reduce risk using advanced security and GitOps-based configuration management. Lower total cost of ownership (TCO) with a fully integrated and managed solution—adding up to a 196% ROI in three years.

Flexible editions

GKE Standard edition provides fully automated cluster life cycle management, pod and cluster autoscaling, cost visibility, and automated infrastructure cost optimization. It includes all the existing benefits of GKE and offers both the Autopilot and Standard operation modes. The new premium GKE Enterprise edition offers all of the above, plus management, governance, security, and configuration for multiple teams and clusters—all with a unified console experience and integrated service mesh.

Serverless Kubernetes experience using Autopilot

GKE Autopilot is a hands-off operations mode that manages your cluster’s underlying compute (without you needing to configure or monitor)—while still delivering a complete Kubernetes experience. And with per-pod billing, Autopilot ensures you pay only for your running pods, not system components, operating system overhead, or unallocated capacity for up to 85% savings from resource and operational efficiency. Both Autopilot and Standard operations mode are available as part of the GKE Enterprise edition.

Automated security and compliance monitoring

GKE threat detection is powered by Security Command Center (SCC), and surfaces threats affecting your GKE clusters in near real-time by continuously monitoring GKE audit logs.


GKE compliance provides streamlined real-time insights, automated reports, and the freedom to innovate securely on Google Cloud.

Pod and cluster autoscaling

GKE implements the full Kubernetes API, four-way autoscaling, release channels, and multi-cluster support. Horizontal pod autoscaling can be based on CPU utilization or custom metrics. Cluster autoscaling works on a per-node-pool basis, and vertical pod autoscaling continuously analyzes the CPU and memory usage of pods, automatically adjusting CPU and memory requests.

Container-native networking and security

Privately networked clusters in GKE can be restricted to a private endpoint or a public endpoint that only certain address ranges can access. GKE Sandbox for the Standard mode of operation provides a second layer of defense between containerized workloads on GKE for enhanced workload security. GKE clusters inherently support Kubernetes Network Policy to restrict traffic with pod-level firewall rules.

Prebuilt Kubernetes applications and templates

Get access to enterprise-ready containerized solutions with prebuilt deployment templates, featuring portability, simplified licensing, and consolidated billing. These are not just container images, but open source, Google-built, and commercial applications that increase developer productivity. Click to deploy on-premises or in third-party clouds from Google Cloud Marketplace.

GPU and TPU support

GKE supports GPUs and TPUs and makes it easy to run ML, GPGPU, HPC, and other workloads that benefit from specialized hardware accelerators.

Multi-team management using fleet team scopes

Use fleets to organize clusters and workloads, and assign resources to multiple teams easily to improve velocity and delegate ownership. Team scopes let you define subsets of fleet resources on a per-team basis, with each scope associated with one or more fleet member clusters.

Multi-cluster management using fleets

You might choose multiple clusters to separate services across environments, tiers, locales, teams, or infrastructure providers. Fleets and the Google Cloud components and features that support them strive to make managing multiple clusters as easy as possible.

Backup for GKE

Backup for GKE is an easy way for customers running stateful workloads on GKE to protect, manage, and restore their containerized applications and data.

Multi-cloud support with workload portability

GKE runs Certified Kubernetes, enabling workload portability to other Kubernetes platforms across clouds and on-premises. You can also run your apps anywhere with consistency using GKE on Google Cloud, GKE on AWS, or GKE on Azure.

Hybrid support

Take advantage of Kubernetes and cloud technology in your own data center through Google Distributed Cloud. Get the GKE experience with quick, managed, and simple installs as well as upgrades validated by Google.

Managed service mesh

Manage, observe, and secure your services with Google’s implementation of the powerful Istio open source project. Simplify traffic management and monitoring with a fully managed service mesh.

Managed GitOps

Create and enforce consistent configurations and security policies across clusters, fleets, and teams with managed GitOps config deployment.

Identity and access management

Control access in the cluster with your Google accounts and role permissions.

Hybrid networking

Reserve an IP address range for your cluster, allowing your cluster IPs to coexist with private network IPs using Google Cloud VPN.

Security and compliance

GKE is backed by a Google security team of over 750 experts and is both HIPAA and PCI DSS compliant.

Integrated logging and monitoring

Enable Cloud Logging and Cloud Monitoring with simple checkbox configurations, making it easy to gain insight into how your application is running.

Cluster options

Choose clusters tailored to the availability, version stability, isolation, and pod traffic requirements of your workloads.

Auto scale

Automatically scale your application deployment up and down based on resource utilization (CPU, memory).

Auto upgrade

Automatically keep your cluster up to date with the latest release version of Kubernetes.

Auto repair

When auto repair is enabled, if a node fails a health check, GKE initiates a repair process for that node.

Resource limits

Kubernetes allows you to specify how much CPU and memory (RAM) each container needs, which is used to better organize workloads within your cluster.

Container isolation

Use GKE Sandbox for a second layer of defense between containerized workloads on GKE for enhanced workload security.

Stateful application support

GKE isn't just for 12-factor apps. You can attach persistent storage to containers, and even host complete databases.

Docker image support

GKE supports the common Docker container format.

OS built for containers

GKE runs on Container-Optimized OS, a hardened OS built and managed by Google.

Private container registry

Integrating with Google Container Registry makes it easy to store and access your private Docker images.

Fast, consistent builds

Use Cloud Build to reliably deploy your containers on GKE without needing to set up authentication.

Built-in dashboard

Google Cloud console offers useful dashboards for your project's clusters and their resources. You can use these dashboards to view, inspect, manage, and delete resources in your clusters.

Spot VMs

Affordable compute instances suitable for batch jobs and fault-tolerant workloads. Spot VMs provide significant savings of up to 91% while still getting the same performance and capabilities as regular VMs.

Persistent disks support

Durable, high-performance block storage for container instances. Data is stored redundantly for integrity, flexibility to resize storage without interruption, and automatic encryption. You can create persistent disks in HDD or SSD formats. You can also take snapshots of your persistent disk and create new persistent disks from that snapshot.

Local SSD support

GKE offers always encrypted, local, solid-state drive (SSD) block storage. Local SSDs are physically attached to the server that hosts the virtual machine instance for very high input/output operations per second (IOPS) and very low latency compared to persistent disks.

Global load balancing

Global load-balancing technology helps you distribute incoming requests across pools of instances across multiple regions, so you can achieve maximum performance, throughput, and availability at low cost.

Linux and Windows support

Fully supported for both Linux and Windows workloads, GKE can run both Windows Server and Linux nodes.

Serverless containers

Run stateless serverless containers abstracting away all infrastructure management and automatically scale them with Cloud Run.

Usage metering

Fine-grained visibility to your Kubernetes clusters. See your GKE clusters' resource usage broken down by namespaces and labels, and attribute it to meaningful entities.

Release channels

Release channels provide more control over which automatic updates a given cluster receives, based on the stability requirements of the cluster and its workloads. You can choose rapid, regular, or stable. Each has a different release cadence and targets different types of workloads.

Software supply chain security

Verify, enforce, and improve security of infrastructure components and packages used for container images with Artifact Analysis.

Per-second billing

Google bills in second-level increments. You pay only for the compute time that you use.

How It Works

A GKE cluster has a control plane and machines called nodes. Nodes run the services supporting the containers that make up your workload. The control plane decides what runs on those nodes, including scheduling and scaling. Autopilot mode manages this complexity; you simply deploy and run your apps.

GKE technical diagram
Google Kubernetes Engine in a minute (1:21)

Common Uses

Manage multi-cluster infrastructure

Simplify multi-cluster deployments with fleets

Use fleets to simplify how you manage multi-cluster deployments—such as separating production from non-production environments, or separating services across tiers, locations, or teams. Fleets let you group and normalize Kubernetes clusters, making it easier to administer infrastructure and adopt Google best practices.

Learn about fleet management
Multi-cluster infrastructure

    Simplify multi-cluster deployments with fleets

    Use fleets to simplify how you manage multi-cluster deployments—such as separating production from non-production environments, or separating services across tiers, locations, or teams. Fleets let you group and normalize Kubernetes clusters, making it easier to administer infrastructure and adopt Google best practices.

    Learn about fleet management
    Multi-cluster infrastructure

      Find the right partner to manage multi-cluster infra

      Securely manage multi-cluster infrastructure and workloads with the help of Enterprise edition launch partners.

      Find a GKE partner
      Launch partners logos

        Securely run optimized AI workloads

        Run optimized AI workloads with platform orchestration

        A robust AI/ML platform considers the following layers: (i) Infrastructure orchestration that support GPUs for training and serving workloads at scale, (ii) Flexible integration with distributed computing and data processing frameworks, and (iii) Support for multiple teams on the same infrastructure to maximize utilization of resources.

        Learn more about AI/ML orchestration on GKE
        Frameworks

        Run optimized AI workloads with platform orchestration

        A robust AI/ML platform considers the following layers: (i) Infrastructure orchestration that support GPUs for training and serving workloads at scale, (ii) Flexible integration with distributed computing and data processing frameworks, and (iii) Support for multiple teams on the same infrastructure to maximize utilization of resources.

        Learn more about AI/ML orchestration on GKE
        Frameworks

        GKE shared-GPU helps to search for neutrinos

        Hear from the San Diego Supercomputer Center (SDSC) and University of Wisconsin-Madison about how GPU sharing in Google Kubernetes Engines is helping them detect neutrinos at the South Pole with the gigaton-scale IceCube Neutrino Observatory.
        Read to learn more
        South Pole

        Continuous integration and delivery

        Create a continuous delivery pipeline

         This hands-on lab shows you how to create a continuous delivery pipeline using Google Kubernetes Engine, Google Cloud Source Repositories, Google Cloud Container Builder, and Spinnaker. After you create a sample application, you configure these services to automatically build, test, and deploy it. 

        Start hands-on lab
        Diagram

        Create a continuous delivery pipeline

         This hands-on lab shows you how to create a continuous delivery pipeline using Google Kubernetes Engine, Google Cloud Source Repositories, Google Cloud Container Builder, and Spinnaker. After you create a sample application, you configure these services to automatically build, test, and deploy it. 

        Start hands-on lab
        Diagram

        Deploying and running applications

        Deploy a containerized web application

        Create a containerized web app, test it locally, and then deploy to a Google Kubernetes Engine (GKE) cluster—all directly in the Cloud Shell Editor. By the end of this short tutorial, you'll understand how to build, edit, and debug a Kubernetes app.

        Start tutorial

        Deploy a containerized web application

        Create a containerized web app, test it locally, and then deploy to a Google Kubernetes Engine (GKE) cluster—all directly in the Cloud Shell Editor. By the end of this short tutorial, you'll understand how to build, edit, and debug a Kubernetes app.

        Start tutorial

        Find the right partner to deploy and run

        Deploy and run on GKE with the help of our trusted partners, including WALT Labs, Zencore, FTG, and more.

        Find a GKE partner

          Current deploys and runs on GKE

          Current, a leading challenger bank based in New York City, now hosts most of its apps in Docker containers, including its business-critical GraphQL API, using GKE to automate cluster deployment and management of containerized apps.
          Read how Current deployed apps with GKE
          Current (the company) Visa credit cards

          Migrate workloads

          Migrating a two-tier application to GKE

          Use Migrate to Containers to move and convert workloads directly into containers in GKE. Migrate a two-tiered LAMP stack application, with both app and database VMs, from VMware to GKE.

          Reference diagram

          Migrating a two-tier application to GKE

          Use Migrate to Containers to move and convert workloads directly into containers in GKE. Migrate a two-tiered LAMP stack application, with both app and database VMs, from VMware to GKE.

          Reference diagram

          Migration partners and services

          Work with a trusted partner to get Google Kubernetes Engine on-prem and bring Kubernetes' world-class management to private infrastructure. Or tap into migration services from the Google Cloud Marketplace.

          Find a migration partner

          Pricing

          How GKE pricing worksAfter free credits are used, total cost is based on edition, cluster operation mode, cluster management fees, and applicable inbound data transfer fees.
          ServiceDescriptionPrice (USD)

          Free tier

          The GKE free tier provides $74.40 in monthly credits per billing account that are applied to zonal and Autopilot clusters.

          Free

          Kubernetes

          Enterprise edition

          Includes standard edition features and multi-team, multi-cluster, self-service operations, advanced security, service mesh, configuration, and a unified console experience.

          $0.0083

          Per vCPU per hour

          Standard edition

          Includes fully automated cluster life cycle management, pod and cluster autoscaling, cost visibility, and automated infrastructure cost optimization.

          $0.10

          Per cluster per hour

          Compute

          Autopilot mode: CPU, memory, and compute resources that are provisioned for your Pods.

          Standard mode: You are billed for each instance according to Compute Engine's pricing.

          Refer to Compute Engine pricing

          Learn more about GKE pricing. View all pricing details

          How GKE pricing works

          After free credits are used, total cost is based on edition, cluster operation mode, cluster management fees, and applicable inbound data transfer fees.

          Free tier

          Description

          The GKE free tier provides $74.40 in monthly credits per billing account that are applied to zonal and Autopilot clusters.

          Price (USD)

          Free

          Kubernetes

          Description

          Enterprise edition

          Includes standard edition features and multi-team, multi-cluster, self-service operations, advanced security, service mesh, configuration, and a unified console experience.

          Price (USD)

          $0.0083

          Per vCPU per hour

          Standard edition

          Includes fully automated cluster life cycle management, pod and cluster autoscaling, cost visibility, and automated infrastructure cost optimization.

          Description

          $0.10

          Per cluster per hour

          Compute

          Description

          Autopilot mode: CPU, memory, and compute resources that are provisioned for your Pods.

          Standard mode: You are billed for each instance according to Compute Engine's pricing.

          Price (USD)

          Refer to Compute Engine pricing

          Learn more about GKE pricing. View all pricing details

          Pricing calculator

          Estimate your monthly GKE costs, including region specific pricing and fees.

          Custom quote

          Connect with our sales team to get a custom quote for your organization.

          Start your proof of concept

          New customers get $300 in free credits to try GKE and other Google Cloud products

          Have a large project?

          Deploy an app to a GKE cluster

          Click to deploy Kubernetes applications

          Get expert help evaluating and implementing GKE

          Google Cloud
          • ‪English‬
          • ‪Deutsch‬
          • ‪Español‬
          • ‪Español (Latinoamérica)‬
          • ‪Français‬
          • ‪Indonesia‬
          • ‪Italiano‬
          • ‪Português (Brasil)‬
          • ‪简体中文‬
          • ‪繁體中文‬
          • ‪日本語‬
          • ‪한국어‬
          Console
          Google Cloud
            翻译: