<?xml version="1.0" encoding="UTF-8"?>
<!-- AUTOGENERATED FILE. DO NOT EDIT. -->
<feed xmlns="http://www.w3.org/2005/Atom">
  <id>tag:google.com,2016:gke-new-features-release-notes</id>
  <title>Google Kubernetes Engine New Features - Release notes</title>
  <link rel="self" href="https://docs.cloud.google.com/feeds/gke-new-features-release-notes.xml"/>
  <author>
    <name>Google Cloud Platform</name>
  </author>
  <updated>2026-04-08T00:00:00-07:00</updated>

  <entry>
    <title>April 08, 2026</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#April_08_2026</id>
    <updated>2026-04-08T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#April_08_2026"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p><a href="https://github.com/kubernetes-sigs/gateway-api/releases/tag/v1.5.0">Gateway API v1.5</a>
is supported in GKE version 1.35.2-gke.1842000 and later.
The GKE Gateway controller passes core conformance tests for
this version of the Gateway API.</p>
<h3>Feature</h3>
<p>GKE managed DRANET is now Generally Available (GA)
for GKE version 1.35.2-gke.1842000 or later.</p>
<p>GKE DRANET is a managed feature that implements the
Kubernetes Dynamic Resource Allocation (DRA) API for high-performance
networking. The GA release expands support beyond the preview phase to
include the following hardware:</p>
<ul>
<li>NVIDIA GPU Instances: Support for instances starting from A3 Ultra, including A4, A4X, and A4X Max.</li>
<li>Cloud TPU Instances: Support for TPU v6e and TPU v7x.</li>
</ul>
<p>For more information, see
<a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/allocate-network-resources-dra">Allocate network resources by using GKE managed DRANET</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>March 25, 2026</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#March_25_2026</id>
    <updated>2026-03-25T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#March_25_2026"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>To provide more controls over the control plane version upgrade, you can now do
the following:</p>
<ul>
<li>Configure a frequency of disruption from auto-upgrades by using the cluster
disruption budget. For more information, see
<a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/cluster-disruption-budget">Control the frequency of disruption from auto-upgrades</a>.</li>
<li>Continue using an existing control plane patch for a longer period, which
facilitates large-scale upgrade and downgrade operations. For more
information, see
<a href="https://docs.cloud.google.com/kubernetes-engine/versioning#patch-version-support">Patch version support</a>.</li>
</ul>
]]>
    </content>
  </entry>

  <entry>
    <title>March 13, 2026</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#March_13_2026</id>
    <updated>2026-03-13T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#March_13_2026"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>In GKE version 1.35 and later, all organization and cluster
 administrators can granularly control which privileged Autopilot
 partner workloads can run in GKE clusters. Additionally,
 approved customers can authorize and run their own privileged workloads in
 Autopilot mode by using custom allowlists.</p>
<p>For more information, see
 <a href="https://docs.cloud.google.com/kubernetes-engine/docs/concepts/about-autopilot-privileged-workloads">About Autopilot privileged  workloads</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>March 10, 2026</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#March_10_2026</id>
    <updated>2026-03-10T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#March_10_2026"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Managed OpenTelemetry for GKE is available in Preview for
clusters running version 1.34.1-gke.2178000 or later. Managed OpenTelemetry for
GKE provides a fully managed and simplified experience for
collecting OpenTelemetry Protocol (OTLP) traces, metrics, and logs on
GKE. This feature includes the following characteristics:</p>
<ul>
<li><strong>Managed collection:</strong> an in-cluster OTLP endpoint that automatically routes
telemetry to the Cloud Telemetry API.</li>
<li><strong>Automatic configuration:</strong> a new Instrumentation custom resource that
automatically injects environment variables into your workloads to simplify
OTLP ingestion.</li>
</ul>
<p>For more information, see
<a href="https://docs.cloud.google.com/kubernetes-engine/docs/concepts/managed-otel-gke">Managed OpenTelemetry for GKE</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>March 05, 2026</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#March_05_2026</id>
    <updated>2026-03-05T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#March_05_2026"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>GKE Inference Quickstart (GIQ) now offers
recommendations for distributed AI inference. This enables you to deploy
optimized, full configurations for advanced models, such as the Qwen and gpt-oss
model families, on NVIDIA GPUs and Cloud TPUs.</p>
<p>This release introduces GKE Inference Gateway by integrating
llm-d inference scheduling. You can select optimized configurations for
workloads like Advanced Customer Support, Code Completion, and Deep Research.
This tunes your infrastructure to meet the specific latency and throughput
requirements of these applications.</p>
<p>For more information, see <a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/machine-learning/inference/inference-quickstart">Analyze model serving performance and costs with Inference Quickstart</a>.</p>
<h3>Feature</h3>
<p>You can use automated disk type selection for Hyperdisk volumes on
GKE. This feature allows GKE to automatically
select the most appropriate disk type based on the machine type of the node
where your workload is scheduled.</p>
<p>With this feature, you can create a single StorageClass that supports clusters
with mixed VM generations. For example, GKE can provision
Hyperdisk on compatible instances (such as C3 or C4) while automatically falling
back to Persistent Disk on other generations.</p>
<p>For more information, see <a href="https://docs.cloud.google.com/kubernetes-engine/docs/concepts/hyperdisk#automated_disk_type_selection">Automated disk type selection</a>.</p>
<h3>Feature</h3>
<p>The <a href="https://docs.cloud.google.com/compute/docs/compute-optimized-machines#h4d_series">H4D machine series</a>,
designed for high performance computing (HPC) workloads, is generally available
for GKE clusters. Based on 5th generation AMD EPYC Turin with
Cloud RDMA 200 Gbps networking, H4D VMs offer 192 cores (SMT disabled), up to
1,488 GB of memory, and 3,750 GiB of Local SSD. H4D is optimized for
tightly-coupled applications that scale across multiple nodes and offers
RDMA-enabled 200 Gbps networking.</p>
<p>You can use H4D with GKE clusters in Standard, or with
the Performance compute class in Autopilot. For more information, see
<a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/run-hpc-workloads">Run high performance computing (HPC) workloads with H4D</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>February 24, 2026</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#February_24_2026</id>
    <updated>2026-02-24T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#February_24_2026"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>The release note for <a href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#November_11_2025">November 11, 2025</a> has been updated to correct the version requirements for using N4D machine types. Cluster autoscaler was incorrectly included in the list of features requiring GKE version 1.34.1-gke.2037000 or later. You can use any available GKE version to use N4D and Cluster autoscaler.</p>
<h3>Feature</h3>
<p>You can create a bare metal instance from the <a href="https://docs.cloud.google.com/compute/docs/general-purpose-machines#c4a_series">C4A machine series</a> with the <code>c4a-highmem-96-metal</code> machine type. This machine type is available in Public Preview for Standard clusters running GKE version 1.35.0-gke.2232000 or later. You can select this machine type by using the <code>--machine-type</code> flag when creating a cluster or node pool. For more information about the requirements and limitations of this machine type, see the <a href="https://docs.cloud.google.com/kubernetes-engine/docs/concepts/arm-on-gke#arm-requirements-limitations">Requirements and limitations</a> section of the "Arm workloads on GKE" document.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>February 13, 2026</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#February_13_2026</id>
    <updated>2026-02-13T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#February_13_2026"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>You can now determine the status and health of a TPU slice and partition by monitoring these new beta system metrics:</p>
<ul>
<li><code>kubernetes.io/accelerator/slice/state</code>: Indicates the current status of the slice.</li>
<li><code>kubernetes.io/accelerator/partition/state</code>: Indicates the health of the partition.</li>
</ul>
<p>For more information, see the <a href="https://docs.cloud.google.com/monitoring/api/metrics_kubernetes#kubernetes-kubernetes">GKE system metrics</a> documentation.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>February 05, 2026</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#February_05_2026</id>
    <updated>2026-02-05T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#February_05_2026"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Image streaming is now available in the <code>asia-southeast3</code> region. For more information,
see the <a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/image-streaming">Image streaming</a> documentation.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>February 03, 2026</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#February_03_2026</id>
    <updated>2026-02-03T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#February_03_2026"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Image streaming and secondary boot disks are now generally available (GA) for
nodes using the Ubuntu with containerd (<code>UBUNTU_CONTAINERD</code>) image type. These
features improve workload startup performance on GKE Standard and Autopilot clusters
through image data streaming and preloaded disk data. To use these features on
Ubuntu nodes, your cluster must be running GKE version 1.35.0-gke.1403000 or later.</p>
<p>For more information, see the documentation for
<a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/image-streaming">Image Streaming</a>
and <a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/data-container-image-preloading">Using Secondary Boot Disks</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>January 27, 2026</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#January_27_2026</id>
    <updated>2026-01-27T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#January_27_2026"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Stream Control Transmission Protocol (SCTP) support on GKE Dataplane V2 is now
generally available (GA). You can now deploy workloads that use SCTP on
GKE Standard clusters. This feature enables direct SCTP
communication for Pod-to-Pod and Pod-to-Service traffic.</p>
<p>SCTP support requires clusters to use GKE Dataplane V2 and Ubuntu node images.
This feature is available in GKE version 1.32.2-gke.1297000 or
later.</p>
<p>For more information, see <a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/deploy-workloads-with-sctp">Deploy workloads with
SCTP</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>January 26, 2026</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#January_26_2026</id>
    <updated>2026-01-26T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#January_26_2026"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>The <a href="https://docs.cloud.google.com/compute/docs/general-purpose-machines#n4a_series">N4A machine
series</a>
is generally available for GKE clusters in Autopilot and
Standard modes. For more information, see <a href="https://docs.cloud.google.com/kubernetes-engine/docs/concepts/arm-on-gke">Arm workloads on
GKE</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>January 21, 2026</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#January_21_2026</id>
    <updated>2026-01-21T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#January_21_2026"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>You can now determine which Kubernetes JobSets are scheduled on which
GKE node pools and nodes by monitoring the new generally
available system metrics:</p>
<ul>
<li><code>kubernetes.io/jobset/assigned_node_pools</code>: GKE node pools
where a Kubernetes JobSet has scheduled Pods.</li>
<li><code>kubernetes.io/jobset/assigned_nodes</code>: GKE nodes where a
Kubernetes JobSet has scheduled Pods.</li>
<li><code>kubernetes.io/node_pool/assigned_jobsets</code>: Kubernetes JobSets that have
scheduled Pods on a GKE node pool.</li>
<li><code>kubernetes.io/node/assigned_jobsets</code>: Kubernetes JobSets that have
scheduled Pods on a GKE node.</li>
</ul>
]]>
    </content>
  </entry>

  <entry>
    <title>January 20, 2026</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#January_20_2026</id>
    <updated>2026-01-20T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#January_20_2026"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>The <code>asia-southeast3</code> region in Bangkok, Thailand is available. For more
information, see the
<a href="https://cloud.google.com/about/locations/">Global Locations</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>January 07, 2026</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#January_07_2026</id>
    <updated>2026-01-07T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#January_07_2026"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>NodeLocal DNSCache is enabled by default on new Standard
GKE clusters which are created running version 1.34.1-gke.3720000
or later. NodeLocal DNSCache is a GKE add-on that improves DNS
performance by running a DNS cache directly on each cluster node as a DaemonSet.
To learn more, see <a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/nodelocal-dns-cache">Set up NodeLocal
DNSCache</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>December 29, 2025</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#December_29_2025</id>
    <updated>2025-12-29T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#December_29_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<h4 id="new_features_in_135">New features in 1.35</h4>
<ul>
<li><strong>In-place Pod Resize:</strong> <a href="https://kubernetes.io/docs/tasks/configure-pod-container/resize-container-resources/">In-place Pod Resize</a> is now GA. This feature allows Pod CPU and memory requests and limits to be modified in-place without Pod or container restart.</li>
<li><strong>Writable cgroups:</strong> GKE <a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/writable-cgroups">Writable cgroups</a> for containers is now GA. This feature allows workloads to manage resources for child processes using the Linux cgroups API, improving reliability for applications like <a href="https://www.ray.io/">Ray</a>.</li>
</ul>
]]>
    </content>
  </entry>

  <entry>
    <title>December 19, 2025</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#December_19_2025</id>
    <updated>2025-12-19T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#December_19_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Rollout sequencing with custom stages is now available in Preview. This feature
offers granular control over upgrading groups of clusters within a fleet,
allowing you to progressively roll out GKE versions across environments. For
more information see
<a href="https://docs.cloud.google.com/kubernetes-engine/docs/concepts/rollout-sequencing-custom-stages/about-rollout-sequencing">About rollout sequencing with custom stages</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>December 15, 2025</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#December_15_2025</id>
    <updated>2025-12-15T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#December_15_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>GKE Autopilot now supports N4A machine types in
Public Preview, available on clusters running
version 1.34.1-gke.3403001 or later.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>December 10, 2025</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#December_10_2025</id>
    <updated>2025-12-10T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#December_10_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>In GKE version 1.34.1-gke.2541000 and later, you can specify
secure tags for firewalls in the
<a href="https://docs.cloud.google.com/kubernetes-engine/docs/reference/crds/computeclass#resourceManagerTags"><code>spec.nodePoolConfig.resourceManagerTags</code> field</a>
in ComputeClasses. GKE adds those secure tags to the nodes that
GKE creates for that ComputeClass, so that you can target
nodes by using these tags in firewall policies. For more information, see
<a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/tags-firewall-policies">Selectively enforce firewall policies in GKE</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>December 03, 2025</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#December_03_2025</id>
    <updated>2025-12-03T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#December_03_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>GKE Inference Gateway is generally available (GA) and ready for production
workloads. This release introduces major performance, security, and usability
enhancements since the Public Preview.</p>
<ul>
<li><strong>Stable v1 API</strong>: The API has graduated to v1. The <code>InferenceModel</code> resource
is replaced by the <code>InferenceObjective</code> resource for a clearer definition of
serving goals. A zero-downtime migration path is available.</li>
<li><strong>Prefix-Aware Routing</strong>: A new, intelligent routing feature inspects request
context and routes requests with shared prefixes (like in conversational AI)
to the same model replica. This can maximize KV cache hits and improve
Time-to-First-Token (TTFT) latency by up to 96%.</li>
<li><strong>API Key Authentication</strong>: Secure your endpoints by enforcing API key
validation through a new integration with Apigee.</li>
<li><strong>Body-Based Routing</strong>: The gateway can route requests using the model
field directly from the HTTP request body, which enables native
compatibility with the OpenAI API specification.</li>
</ul>
<p>For more information see
<a href="https://docs.cloud.google.com/kubernetes-engine/docs/concepts/about-gke-inference-gateway">About GKE Inference Gateway</a>
and <a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/deploy-gke-inference-gateway">Deploy GKE Inference Gateway</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>November 27, 2025</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#November_27_2025</id>
    <updated>2025-11-27T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#November_27_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>TPU7x (Ironwood), Google's seventh-generation TPU for large-scale AI workloads,
is available in <a href="https://docs.cloud.google.com/products#product-launch-stages">Preview</a> in GKE
Standard clusters that run version 1.34.0-gke.2201000 and later, and in
Autopilot clusters that run version 1.34.1-gke.3084001 and later.
TPU7x offers a significant performance increase compared to previous
generations, with 2307 TFLOPs of BF16 performance and 192 GB of
high-bandwidth memory (HBM) per chip. For more information, see
<a href="https://docs.cloud.google.com/kubernetes-engine/docs/concepts/tpus#ironwood-benefits">Get started with Ironwood (TPU7x)</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>November 24, 2025</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#November_24_2025</id>
    <updated>2025-11-24T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#November_24_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Fast-starting nodes are now generally available. GKE provisions
fast-starting nodes on a best-effort basis in Autopilot when workloads
use compatible configurations. For more information, see
<a href="https://docs.cloud.google.com/kubernetes-engine/docs/concepts/fast-starting-nodes">About quicker workload startup with fast-starting nodes</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>November 17, 2025</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#November_17_2025</id>
    <updated>2025-11-17T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#November_17_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>NVIDIA recommends that Kubernetes clusters enable Coherent Driver-Based Memory
Management (CDMM) to resolve memory over-reporting. CDMM is enabled by default
on A4X nodes running the R580 GPU driver in GKE clusters with the
following versions:</p>
<ul>
<li><strong>1.33 or later</strong>: 1.33.4-gke.1036000 or later</li>
<li><strong>1.32</strong>: 1.32.8-gke.1108000 or later</li>
</ul>
<p>CDMM allows GPU memory to be managed through the driver instead of the operating
system (OS), avoiding OS onlining of GPU memory, and exposing the GPU memory as
a Non-Uniform Memory Access (NUMA) node to the OS.</p>
<p>For more information about CDMM, see
<a href="https://docs.nvidia.com/datacenter/tesla/tesla-release-notes-580-65-06/index.html#hardware-software-support">Hardware and Software Support</a>.
To create GKE clusters with A4X, see the following documents:</p>
<ul>
<li><a href="https://docs.cloud.google.com/ai-hypercomputer/docs/create/gke-ai-hypercompute">Create an AI-optimized GKE cluster with default configuration</a></li>
<li><a href="https://docs.cloud.google.com/ai-hypercomputer/docs/create/gke-ai-hypercompute-custom-a4x">Create a custom AI-optimized GKE cluster which uses A4X</a></li>
</ul>
]]>
    </content>
  </entry>

  <entry>
    <title>November 11, 2025</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#November_11_2025</id>
    <updated>2025-11-11T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#November_11_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>This note was updated on February 24, 2026. Cluster autoscaler was incorrectly included in the list of features which required 1.34 support with N4D. The correct features are now listed as follows.</p>
<p>The N4D machine family is now Generally Available (GA) for
Standard and Autopilot mode. N4D instances are powered by the
fifth generation AMD EPYC SP5 processors (Turin). The N4D machine series is
available as follows:</p>
<ul>
<li><strong>Compute classes, node pool auto-creation, and Autopilot mode</strong>:
GKE version 1.34.1-gke.2037000 and later.</li>
<li><strong>Manually created node pools in Standard mode</strong>: all available
GKE versions.</li>
</ul>
<p>For more information, see
<a href="https://docs.cloud.google.com/compute/docs/general-purpose-machines#n4d_series">N4D machine series</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>November 07, 2025</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#November_07_2025</id>
    <updated>2025-11-07T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#November_07_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<aside class="special"><strong>Important:</strong><span> This note is incorrect. For the correct note, see the entry for
<a href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#April_08_2026">April 8, 2026</a>.</span></aside>
<p>In GKE version 1.34.1-gke.2037001 and later, the
GKE logging agent in your clusters can process logs up to two
times faster per node than in version 1.33 and earlier. The logging agent also
uses less node resources, which improves efficiency especially if you use
<a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/adjust-log-throughput#enable">high-throughput logging</a>.
These improvements to the logging agent are automatically enabled in version
1.34.1-gke.2037001 and later.</p>
<h3>Feature</h3>
<p>In version 1.34.1-gke.1829001 and later, GKE can
<a href="https://docs.cloud.google.com/kubernetes-engine/docs/concepts/node-auto-provisioning">auto-create</a> multiple
node pools concurrently to improve the speed with which multiple new node pools
become ready.</p>
<h3>Feature</h3>
<p>In GKE version 1.35 and later, GKE rejects
anonymous requests to cluster endpoints (except for the <code>livez</code>, <code>/healthz</code>, and
<code>/readyz</code> health check endpoints) by default for all new Autopilot or
Standard clusters. Existing clusters aren't affected by this change. To
allow anonymous requests to cluster endpoints, explicitly specify a value of
<code>ENABLED</code> in the <code>--anonymous-authentication-config</code> flag or the
<code>AnonymousAuthenticationConfig.mode</code> API field. For more information, see
<a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/hardening-your-cluster#restrict-anon-access">Restrict anonymous access to cluster endpoints</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>October 31, 2025</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#October_31_2025</id>
    <updated>2025-10-31T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#October_31_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>The Multi-Cluster Services (MCS) feature has been updated with a finalizer to
more effectively prevent potential resource leaks and ensure a full cleanup
during the feature's disablement process. As a result of this improvement, the
disablement procedure has been updated. For more details on how to disable MCS, see <a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/multi-cluster-services#disabling_mcs">Disabling
MCS</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>October 28, 2025</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#October_28_2025</id>
    <updated>2025-10-28T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#October_28_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Autoscaled blue-green upgrades are a type of node upgrade strategy that
maximizes the amount of time before disruption-intolerant workloads are evicted,
while minimizing cost. This feature is available in Preview for
GKE Standard node pools. For more information, see
<a href="https://docs.cloud.google.com/kubernetes-engine/docs/concepts/node-pool-upgrade-strategies#autoscaled-blue-green-upgrade-strategy">Autoscaled blue-green upgrades</a>.</p>
<h3>Feature</h3>
<p>You can use the G4 VM, powered by NVIDIA's RTX PRO 6000 GPUs, with
GKE Autopilot in version 1.34.1-gke.1829001 or later. To
get started, see <a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/autopilot-gpus">Deploy GPU workloads in
Autopilot</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>October 21, 2025</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#October_21_2025</id>
    <updated>2025-10-21T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#October_21_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>The G4 VM, powered by NVIDIA's RTX PRO 6000 Blackwell Server Edition GPUs with
the AMD EPYC Turin CPU platform, is generally available on GKE.
G4 instances have up to 384 vCPUs, 1,440 GB of memory, 12 TiB of Titanium SSD
disks attached, and up to 400 Gbps of standard network performance. The G4 VM
offers a leap in performance with up to 9 times the throughput of G2 instances
for workloads such as AI development, and graphics rendering. G4 VMs are
currently available with 1, 2, 4, or 8 GPUs.</p>
<ul>
<li>For GKE Standard, use GKE version
1.34.0-gke.1662000 or later. To get started, see <a href="https://cloud.google.com/kubernetes-engine/docs/how-to/gpus">Run GPUs in
GKE Standard node
pools</a>.</li>
</ul>
]]>
    </content>
  </entry>

  <entry>
    <title>October 09, 2025</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#October_09_2025</id>
    <updated>2025-10-09T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#October_09_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>The following networking features are available:</p>
<ul>
<li><p>In GKE version 1.33.4-gke.1055000 or later, you can control
how external traffic reaches your Services on GKE clusters by
using Network Service Tiers. You can configure the network tier to use either
Standard Tier or Premium Tier when you create or update clusters or when you
update LoadBalancer Services. For more information, see <a href="https://cloud.google.com/kubernetes-engine/docs/how-to/network-tiers">Configure external
traffic with Network Service Tiers</a>.</p></li>
<li><p>Starting with GKE versions 1.33 and later, you can enable
automatic IP address management (auto IPAM) on GKE clusters. Auto
IPAM dynamically adds or removes additional IP address ranges for nodes and Pods
as the cluster scales up or down. This feature eliminates the need for large,
potentially wasteful, upfront IP reservations and manual intervention during
cluster scaling. For more information, see <a href="https://cloud.google.com/kubernetes-engine/docs/how-to/enable-auto-ipam">Use auto IP address
management</a>.</p></li>
<li><p>In GKE version 1.30.3-gke.1211000 and later, you can assign
additional subnets to a VPC-native cluster. Additional subnets
assigned to a cluster let you create new node pools where IPv4 addresses for
both nodes and Pods come from the additional subnet ranges. This enhancement
removes single-subnet limitations, increases scalability, and enhances the
flexibility of your GKE clusters. For more information, see <a href="https://cloud.google.com/kubernetes-engine/docs/how-to/multi-subnet-cluster">Add subnets to
clusters</a>.</p></li>
</ul>
<h3>Feature</h3>
<p>For AI models deployed on a GKE cluster, you can view details
about these deployments in the Google Cloud console. The pages include deployment
details, logs, and <a href="https://cloud.google.com/kubernetes-engine/docs/how-to/configure-automatic-application-monitoring#aiml-dashboard">observability
dashboards</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>October 07, 2025</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#October_07_2025</id>
    <updated>2025-10-07T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#October_07_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Starting with GKE version 1.33.2-gke.1240000 and later, you can specify the
network tier (Standard or Premium) for ephemeral IP addresses used by
the <code>gke-l7-regional-external-managed-mc</code> GatewayClass. For more information,
see <a href="https://cloud.google.com/kubernetes-engine/docs/how-to/deploying-gateways#configure-network-tier">Configure Network
Tier</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>October 01, 2025</title>
    <id>tag:google.com,2016:gke-new-features-release-notes#October_01_2025</id>
    <updated>2025-10-01T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/kubernetes-engine/docs/release-notes#October_01_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>The GKE cluster autoscaler now allows for a significantly longer node drain time. From GKE version 1.32.7-gke.1079000 and later, the graceful node drain timeout has been increased from 10 minutes to 1 hour. For more information, see <a href="https://docs.cloud.google.com/kubernetes-engine/docs/concepts/cluster-autoscaler#how_cluster_autoscaler_works">How cluster autoscaler works</a>.</p>
<h3>Feature</h3>
<p>The <a href="https://cloud.google.com/kubernetes-engine/docs/concepts/verticalpodautoscaler#inplaceorrecreate_mode"><code>InPlaceOrRecreate</code></a> mode for Vertical Pod Autoscaler (VPA) is now available for Public Preview in GKE.</p>
<p>This mode uses <a href="https://github.com/kubernetes/autoscaler/tree/master/vertical-pod-autoscaler/enhancements/4016-in-place-updates-support">In-Place Pod Resize (IPPR/IPPU)</a>, which allows VPA to automatically adjust workload resources, without requiring Pod recreation. This seamless rightsizing capability helps ensure better service continuity and helps minimize costs by optimizing resource allocation, particularly during idle periods.</p>
<p>VPA is enabled by default in Autopilot clusters. For Standard clusters, you must first enable VPA. For more information on configuring a VPA object, see
<a href="https://cloud.google.com/kubernetes-engine/docs/how-to/vertical-pod-autoscaling#gcloud">Set Pod resource requests automatically</a>.</p>
]]>
    </content>
  </entry>

</feed>
