<?xml version="1.0" encoding="utf-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:media="http://search.yahoo.com/mrss/"><channel><title>Storage &amp; Data Transfer</title><link>https://cloud.google.com/blog/products/storage-data-transfer/</link><description>Storage &amp; Data Transfer</description><atom:link href="https://cloudblog.withgoogle.com/blog/products/storage-data-transfer/rss/" rel="self"></atom:link><language>en</language><lastBuildDate>Wed, 22 Apr 2026 12:00:24 +0000</lastBuildDate><image><url>https://cloud.google.com/blog/products/storage-data-transfer/static/blog/images/google.a51985becaa6.png</url><title>Storage &amp; Data Transfer</title><link>https://cloud.google.com/blog/products/storage-data-transfer/</link></image><item><title>Cross-cloud infrastructure innovation for the agentic enterprise</title><link>https://cloud.google.com/blog/products/compute/cross-cloud-infrastructure-at-next26/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The era of agentic AI is accelerating from human- to machine-speed operations, while also creating profound stress on legacy technology infrastructure. This new reality pushes foundational systems to their limits: agents generate thousands of internal messages and complex queries, spawning more agents, all of which can rapidly overwhelm traditional networks and databases, and expose new security vulnerabilities.&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Unlocking AI's full potential in the era of agents requires a secure, adaptive foundation. We call it &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;cross-cloud infrastructure for the agentic enterprise&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; – and at Google Cloud Next ‘26, we’re launching a powerful set of new innovations across four areas:&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;What’s new:&lt;/strong&gt;&lt;/p&gt;
&lt;ol&gt;
&lt;li role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Fluid compute: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Google Compute Engine and Kubernetes services work together to enable cost-effective, high-speed AI agents and enterprise workloads with new compute and orchestration capabilities. &lt;/span&gt;&lt;/li&gt;
&lt;li role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Secure cross-cloud connectivity: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Agent Gateway, Cloud Armor, and other tools deliver a secure, governed, and simplified networking foundation for AI agents, including observability of agentic traffic across clouds.&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;/li&gt;
&lt;li role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Unified data layer: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Smart Storage, Knowledge Catalog, and other innovations transform passive data archives into dynamic reasoning engines, giving AI agents the context they need to execute.&lt;/span&gt;&lt;/li&gt;
&lt;li role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Digital sovereignty: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Confidential External Key Management and new features in Google Distributed Cloud bring Google’s leading models and AI enablers wherever your data lives.&lt;/span&gt;&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Let’s take a closer look at all the news for each of these four areas.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Fluid compute&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Agentic workloads are dynamic and unpredictable, impacting both traditional enterprise applications and the AI agents themselves. Fluid compute is enabled by Google Compute Engine and Google Kubernetes services working together to dynamically adapt and shift weight in real-time, enabling cost-effective, high-speed AI agents and operational enterprise workloads for all customers. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;While our &lt;/span&gt;&lt;a href="https://cloud.google.com/blog/products/compute/ai-infrastructure-at-next26"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;AI Hypercomputer delivers raw power for large-scale AI model training&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, fluid compute addresses the needs of operational workloads and agents. As agents move toward reasoning and reinforcement learning, CPUs are reclaiming a central role, excelling at the "branchy" logic, complex control flows, and secure execution sandboxes (like those for agentic orchestration, RL, SLM inference, and RAG) that agent workflows demand. CPUs also provide the critical isolation needed for secure agent execution, complementing the parallel processing strength of GPUs and TPUs used in training.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;We are introducing new CPU families, GKE capabilities, and Hyperdisk block storage capabilities to run traditional workloads and AI agents securely at scale, including:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Google C4N Series&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;: These VMs help ensure your enterprise workloads don't slow down under the demands of agentic AI by processing up to 95 million packets per second, up to 40% faster than other leading hyperscalers.&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;This eliminates I/O bottlenecks for demanding workloads like security appliances, streaming media, and open source databases, even when utilizing smaller instance sizes.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Google M4N Series with Hyperdisk Extreme&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;: M4N removes data pipeline bottlenecks and eliminates overprovisioning to deliver industry leading per-core IOPS and throughput required to handle massive data I/O from agents, analytics, and mission-critical databases. M4N provides 26.57 GB of RAM per vCPU, allowing you to scale mission-critical workloads cost-effectively on fewer cores. For example, M4N with Hyperdisk Extreme reduces Oracle workload total cost of ownership by over 20% compared to leading hyperscale clouds.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;GKE Agent Sandbox: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;This solution secures agents with trusted gVisor isolation and handles demand spikes, launching up to 300 sandboxes per second, per cluster. Backed by the only managed sandbox technology available among leading hyperscale clouds, it achieves up to 30% better price-performance than competitors when running AI agents on GKE Agent Sandbox with Google Axion N4A. &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p style="padding-left: 40px;"&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;“Wayfair's AI strategy is built on years of systematic infrastructure modernization on Google Cloud — migrating our core eCommerce engine and databases off legacy systems, decomposing monolithic services into cloud-native architecture, and unifying our data and analytics platform. That foundation is what makes everything else possible. Today, Gemini Enterprise Agent Platform is powering everything from catalog enrichment to generative shopping experiences that help customers create a home that's just right for them — and it's the same foundation preparing us for the agentic era, where AI doesn't just assist but actively drives discovery, personalization, and commerce across every customer touchpoint and across our business.”&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; - Fiona Tan, Chief Technology Officer, Wayfair&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Explore all our latest compute innovations in &lt;/span&gt;&lt;a href="https://cloud.google.com/blog/products/compute/whats-new-in-compute-at-next26"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;this blog&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. &lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Secure cross-cloud connectivity &lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Agentic AI replaces predictable human requests with autonomous “reasoning loops,” in which agents call other agents that, in turn, call LLMs, triggering massive, sudden surges in compute and machine-to-machine traffic. This shift creates unique challenges for network predictability and security of non-human identities. Optimized for agentic AI, our Cross-Cloud Network moves data across diverse environments, connecting employees, customers, and agents with visibility and security. New in Cross-Cloud Network are:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Agent Gateway:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Governs and orchestrates your enterprise agentic traffic as the “air traffic controller” in &lt;/span&gt;&lt;a href="https://cloud.google.com/blog/products/ai-machine-learning/introducing-gemini-enterprise-agent-platform"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Gemini Enterprise Agent Platform&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. It natively understands agent protocols like MCP and A2A to inspect and govern every agent interaction. By integrating with Google and third-party identity and AI safety services, it enables deep inspection to verify access, block attacks, and protect sensitive data, maintaining compliance across your core business.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Cloud Network Insights&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;: Delivers broad visibility across your hybrid and multi-cloud infrastructure to drive faster troubleshooting and network resolutions. Continuously monitor your end-to-end agent, network and web performance across Google Cloud, AWS, Azure, data centers, internet applications, and agentic workloads. Using synthetic traffic analytics, Cloud Network Insights provides hop-by-hop network path visibility to help you pinpoint the source of degradations and is coupled with AI-powered insights from Gemini Cloud Assist to deliver more autonomous operations.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Enhanced Cloud Next Generation Firewall (NGFW) and Cloud Armor&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;: Provides machine-speed, AI-powered protection to combat the rapid explosion of AI-generated polymorphic malware and zero-day exploits. Cloud NGFW advanced malware sandbox delivers real-time inline prevention of AI-generated threats, while Cloud Armor managed rules provides automated protection against both known and unknown Common Vulnerabilities and Exposures (CVEs). Together with Model Armor, these services analyze the intent and content of AI agent communications.  &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Discover more about how we &lt;/span&gt;&lt;a href="https://cloud.google.com/blog/products/networking/whats-new-in-cloud-networking-at-next26"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;optimized networking for AI&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; in and outside of the data center. &lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Unified data layer&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;AI agents are only as powerful as the data they can access and the context they’re given. More applications and platforms are using structured and unstructured data, but it can be difficult to catalog, find, and act on that data at scale, leading to less effective agent interactions. To close the gap, your agents need all of your data brought together into a cohesive, queryable knowledge engine, or unified data layer. This way, your agents can identify and access accurate sources. At Next ‘26, we’re enhancing the unified data layer with:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Smart Storage&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;: This solution transforms dark data into a powerful knowledge asset for AI agents and training by embedding new semantic intelligence directly into your data objects. With new Google Cloud Storage capabilities like automated annotation, entity extraction, and semantic search, your agents can instantly find and use the specific data they need — whether it's hidden in spreadsheets, PDFs, or other unstructured formats across your entire organization. This significantly speeds up the development and deployment of your AI solutions. Learn more about &lt;/span&gt;&lt;a href="https://cloud.google.com/blog/products/storage-data-transfer/next26-storage-announcements"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;storage innovations to accelerate your AI workloads&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Knowledge Catalog&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;: Knowledge Catalog maps business meaning across your entire data estate, providing a grounded source of truth so agents can deliver the most accurate results. This foundation enables AI training and inferencing and doesn’t require you to migrate your data; your agents interact with it directly, wherever it lives, with full context and governance, making modernization easier.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Part of our &lt;/span&gt;&lt;a href="https://cloud.google.com/transform/shift-system-of-action-architecting-the-agentic-data-cloud-AI"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Agentic Data Cloud&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, Smart Storage and Knowledge Catalog can take your data from a passive archive into a dynamic reasoning engine.&lt;/span&gt;&lt;/p&gt;
&lt;p style="padding-left: 40px;"&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;“AI is critical to making our customers’ smart home and security solutions more intelligent and convenient. By leveraging Google Cloud’s Smart Storage, we auto-annotate rich metadata delivered in BigQuery. We’ve scaled and accelerated our data discovery and curation efforts, speeding up our AI development process from months to weeks, continuously delivering innovations that build trust and enhance the overall home experience.”&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; - Brandon Bunker, VP of Product, AI, Vivint&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Digital sovereignty&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;In the agentic era, digital sovereignty is a fundamental requirement for public sector and enterprise customers looking to accelerate innovation — without sacrificing control. There’s no one-size-fits-all solution, which is why we’ve designed a comprehensive set of offerings to meet different sovereign AI needs anywhere: public cloud, on-premises, or hybrid. New capabilities in our sovereign AI portfolio include: &lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Confidential External Key Management:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Organizations can use Confidential External Key Management to maintain complete possession, custody, and control of your encryption keys and the policies that govern them. Confidential External Key Management leverages &lt;/span&gt;&lt;a href="https://cloud.google.com/security/products/confidential-computing"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Confidential Compute&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; to host the key management endpoint in a tamper-proof environment within Google Cloud. You are in control and determine where your keys are stored, who can access them, and under what circumstances. Even highly privileged Google administrators cannot access your keys without authorization, which you can revoke at any time. Your data, your control.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Gemini on Google Distributed Cloud: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;With Gemini on GDC, companies can securely deploy Gemini in sensitive environments, while meeting data sovereignty needs. Your choice of deployment models includes managed software on your connected hardware or a fully disconnected, air-gapped solution. You can now scale with Google's leading AI capabilities even in the most restricted, high-security environments — from powerful Gemini models to advanced coding, search, and other agentic capabilities.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;In addition, Google Distributed Cloud supports an end-to-end AI stack, combining our latest-generation AI infrastructure with Gemini models to accelerate and enhance all your sovereign AI workloads. This stack includes:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;NVIDIA Blackwell GPUs:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; NVIDIA Blackwell (NVIDIA HGX B200) and NVIDIA Blackwell Ultra platforms (NVIDIA HGX B300) GPUs accelerate AI performance, leveraging fifth-gen NVIDIA NVLink to deliver data-center scale bandwidth directly to your environment.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;New VM families:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; New A4 family offerings provide the ability to handle the most demanding inference tasks, delivering a 2.25x increase in peak compute. Memory-Optimized M2 and M3&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;brings the&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;high memory-to-vCPU ratios needed for massive ERP and data analytics workloads on-premises.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Enhanced storage: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Eliminate storage bottlenecks with 6x storage capacity per zone and a 10x performance boost, giving you the ability to do AI reasoning on-premises. Now, your data infrastructure moves at the speed of AI reasoning.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p style="padding-left: 40px;"&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;"&lt;/span&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;Our customers demand high-performance, private AI inference without the risks of multi-tenancy. Google Distributed Cloud allows us to provide dedicated, low-latency environments that meet strict sensitive data requirements. With the ability to run Gemini on B200s and B300s, we can significantly increase inference speeds and provide the token throughput our clients need to scale."&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;- &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;Dave Driggers, CEO &amp;amp; Co-founder, Cirrascale Cloud Services&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Transforming vision into reality &lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;When these product areas converge, your infrastructure evolves into a high-performing, secure, adaptive foundation for the agentic era. We're not just offering tools; we're providing the architectural blueprint to enable enterprises and the public sector to rapidly embrace the full power of AI and agents with confidence.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;To learn more about key industry trends for AI Infrastructure, read our &lt;/span&gt;&lt;a href="https://cloud.google.com/resources/content/state-of-infrastructure-in-the-agentic-ai-era?utm_source=cgc-blog&amp;amp;utm_medium=blog&amp;amp;utm_campaign=FY26-Q1-GLOBAL-STO121-website-dl-State-AI-Infra-172614&amp;amp;utm_content=state-of-infra-agentic-ai-era-report&amp;amp;utm_term=state-of-infra-agentic-ai-era-report"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;State of Infrastructure in the Agentic AI Era report&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. &lt;/span&gt;&lt;/p&gt;&lt;/div&gt;</description><pubDate>Wed, 22 Apr 2026 12:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/compute/cross-cloud-infrastructure-at-next26/</guid><category>Networking</category><category>Storage &amp; Data Transfer</category><category>Infrastructure</category><category>Google Cloud Next</category><category>Compute</category><media:content height="540" url="https://storage.googleapis.com/gweb-cloudblog-publish/images/GCN26_102_BlogHeader_2436x1200_Opt_4_Light.max-600x600.jpg" width="540"></media:content><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>Cross-cloud infrastructure innovation for the agentic enterprise</title><description></description><image>https://storage.googleapis.com/gweb-cloudblog-publish/images/GCN26_102_BlogHeader_2436x1200_Opt_4_Light.max-600x600.jpg</image><site_name>Google</site_name><url>https://cloud.google.com/blog/products/compute/cross-cloud-infrastructure-at-next26/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Nirav Mehta</name><title>VP, Product Management, Compute Platforms</title><department></department><company></company></author><author xmlns:author="http://www.w3.org/2005/Atom"><name>Muninder Sambi</name><title>VP, Google Distributed Cloud</title><department></department><company></company></author></item><item><title>Storage innovations to accelerate your AI workloads at Next ‘26</title><link>https://cloud.google.com/blog/products/storage-data-transfer/next26-storage-announcements/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;At Google Cloud Next, we are announcing innovations across every layer of our storage stacks — performance, intelligence, and management — to ensure your data is as fast and as useful as the AI models, apps and agents you are building.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Why it matters:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Storage is no longer just a place to keep data. When training AI models, storage is the engine that feeds data-hungry accelerators. During AI inference, it’s the access layer that makes it responsive, acting as the source for the context that AI agents need to be effective. When storage performance falls short, accelerators sit idle, agents respond slowly, and data remains invisible to AI models. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;But storage performance is only half the battle; you also need storage that’s &lt;/span&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;smart&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;. With the help of Google’s AI models integrated directly into the storage layer, you’re no longer just storing bits, but data that has full context about its content. In this new era of smart storage, raw data becomes a valuable asset that’s ready to use by a variety of downstream AI and enterprise applications.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;What’s new:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;High-performance storage infrastructure: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;New Rapid family of features in Cloud Storage for high-performance object storage; delivering 10x performance enhancements plus a new cost-effective Dynamic tier for Google Cloud Managed Lustre.&lt;/span&gt;&lt;/li&gt;
&lt;li role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Smart Storage: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Unlocking unstructured data with automated metadata annotation, and AI agent connectivity via MCP.&lt;/span&gt;&lt;/li&gt;
&lt;li role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Storage Intelligence: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Streamlined data management through zero-configuration dashboards, aggregated activity views, and enhanced batch operations.&lt;/span&gt;&lt;/li&gt;
&lt;li role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Enhanced ecosystem:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Expanded capabilities across Google Cloud NetApp Volumes, Filestore for GKE, and our backup and data protection portfolio.&lt;/span&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Let’s take a deeper look at the storage enhancements we are unveiling this week.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Storage infrastructure that keeps up with AI&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;As AI models scale, getting data from the storage to the compute layer fast enough can be a bottleneck. New storage capabilities move performance directly into the storage layer, reducing total cost of ownership (TCO) and keeping accelerators fully utilized.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Cloud Storage Rapid&lt;br/&gt;&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Cloud-based object storage like our Cloud Storage is scalable and cost-effective, but bottlenecks can stall AI jobs and waste expensive compute cycles. Every time a training cluster waits on a read or a checkpoint write stalls, you're paying for accelerators that aren't doing useful work.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;a href="https://docs.cloud.google.com/storage/docs/rapid/high-performance-storage"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Cloud Storage Rapid&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; marks a fundamental shift in designing AI infrastructure: you no longer have to choose between the reliability of object storage and the high performance of a specialized AI storage system. Cloud Storage Rapid lets you leverage the industry-leading durability, massive distributed scale, and cost-effective auto-tiering of object storage, while simultaneously achieving extreme throughput, frequent I/Os, and ultra-low latency. With native integrations into PyTorch and JAX, Cloud Storage Rapid is optimized out-of-box for the most popular AI/ML ecosystem frameworks, so that your data preparation, training, and inference workloads run on a high-performance and reliable foundation.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The Cloud Storage Rapid family consists of two offerings: &lt;/span&gt;&lt;a href="https://docs.cloud.google.com/storage/docs/rapid/rapid-bucket"&gt;&lt;strong style="text-decoration: underline; vertical-align: baseline;"&gt;Rapid Bucket&lt;/strong&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; and &lt;/span&gt;&lt;a href="https://docs.cloud.google.com/storage/docs/anywhere-cache"&gt;&lt;strong style="text-decoration: underline; vertical-align: baseline;"&gt;Rapid Cache&lt;/strong&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href="https://docs.cloud.google.com/storage/docs/rapid/rapid-bucket"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Rapid Bucket&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;,&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;now generally available, &lt;/span&gt;&lt;a href="https://cloud.google.com/blog/products/storage-data-transfer/how-the-colossus-stateful-protocol-benefits-rapid-storage"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;leverages Colossus&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, the Google distributed storage system that powers Gemini and YouTube, to deliver more than &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;15 TB/s of bandwidth, 20 million requests per second, and sub-millisecond latency&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; in a single zonal bucket. With access via high-performance gRPC and S3-compatible APIs, Rapid Bucket increases accelerator utilization for multi-modal training with &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;50% reduced GPU blocked time &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;and 2.5x faster data loading. Checkpoint restores are 5x faster and checkpoint writes are 3.2x faster compared to traditional object storage, reducing workload interruptions and wasted GPU time.&lt;/span&gt;&lt;/li&gt;
&lt;/ul&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/1_x1nf9ws.max-1000x1000.png"
        
          alt="1"&gt;
        
        &lt;/a&gt;
      
        &lt;figcaption class="article-image__caption "&gt;&lt;p data-block-key="opmmh"&gt;Checkpoint writes are 3.2x faster and restores are 5x faster with Rapid Bucket&lt;/p&gt;&lt;/figcaption&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;a href="https://docs.cloud.google.com/storage/docs/anywhere-cache"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Rapid Cache&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, formerly Anywhere Cache, accelerates bandwidth for bursty workloads like model loading for inference, delivering an aggregate read throughput of 2.5 TB/s for existing buckets, with no code changes. The new &lt;/span&gt;&lt;a href="https://docs.cloud.google.com/storage/docs/rapid/rapid-cache#ingest-on-write"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;ingest-on-write&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; feature provides up to &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;2.2x faster checkpoint restores&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;, allowing training clusters to recover faster from interruptions. &lt;br/&gt;&lt;br/&gt;&lt;/span&gt;&lt;span style="font-family: -apple-system, BlinkMacSystemFont, 'Segoe UI', Roboto, Oxygen, Ubuntu, Cantarell, 'Open Sans', 'Helvetica Neue', sans-serif;"&gt;Rapid Cache’s combination of simplicity and performance has resulted in strong adoption, including from cutting-edge AI/ML customers like Thinking Machines Lab.&lt;br/&gt;&lt;br/&gt;&lt;/span&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;“Rapid Cache has become a core foundation of our AI/ML data infrastructure, supporting our critical workflows, from data prep and pretraining to training and model loading. By acting as a crucial bandwidth shield and booster, it enables us to scale our data-intensive workloads across our entire fleet without compromise, providing us with the on-demand high bandwidth and consistent stability that we need to innovate at speed.” &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;- &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;James Sun, Member of Technical Staff, &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;Thinking Machines Lab&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Google Cloud Managed Lustre&lt;br/&gt;&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;The Lustre parallel file system is the industry standard for organizations whose AI training and inference workloads require high throughput and sub-millisecond latency, and is trusted by AI labs and HPC centers worldwide to feed thousands of accelerators simultaneously and keep them saturated under pressure. &lt;/span&gt;&lt;a href="https://docs.cloud.google.com/managed-lustre/docs/overview"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Google Cloud Managed Lustre&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; brings that capability as a fully managed service, and with today's announcements, it is the most performant managed Lustre offering available in any cloud.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Managed Lustre now delivers up to &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;10 TB/s of throughput&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; — a 10x increase since last year and 4–20x higher than managed Lustre offerings from other hyperscalers for a single instance. Powered by C4NX VMs and Hyperdisk Exapools, Managed Lustre writes and restores checkpoints 2.6x faster when compared to other Google Cloud storage solutions.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The new &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;Dynamic tier&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; ($0.06/GB-month) delivers the low-latency performance required for intense AI workloads like training and checkpointing. By serving data from persistent disk rather than relying on object-based caching, we eliminate a performance cliff — helping ensure your data remains responsive and your accelerators stay productive. A single SKU provides simple, predictable billing without the hidden complexity of traditional data tiering.&lt;/span&gt;&lt;/p&gt;
&lt;p style="padding-left: 40px;"&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;“By integrating Managed Lustre we eliminated the typical onboarding bottlenecks, allowing us to hit the ground running with the inferencing workload. This high-throughput, low-latency storage keeps our B200 GPUs fully saturated, driving a substantial performance gain in LLM inference over the H200. For our customers, this translates directly into faster, more responsive AI agents that can handle complex reasoning at a fraction of the previous latency.”&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; - Lavnaya Karanam, Software Engineering PMTS, Salesforce&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Smart Storage: Context for the AI era&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The beauty of an object storage system like Cloud Storage has long been its simplicity: the system knows the object’s name, its size, and when it was created. But if you want to understand the object’s content — what entities it references, whether it contains sensitive PII, or whether it’s relevant to a pending query — you need to use custom pipelines, separate databases, and bespoke enrichment systems.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;AI has changed the equation. To fine-tune a model, you need to select the right objects from the get-go, from a corpus of millions. Building an agent requires retrieving the right context for each decision. To meet a compliance obligation, you need to know what every file contains up front, before it becomes a liability. In each case, the bottleneck isn’t compute or model quality — it’s the inability to describe, find, and act on objects at scale. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;To bridge that gap between stored and usable data, last year we introduced &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;Smart Storage,&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; a set of capabilities built directly into Cloud Storage that makes every object self-describing. New Smart Storage capabilities include:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Automated annotations&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;, which eliminates the need to build and maintain custom annotation pipelines. With Smart Storage enabled, Cloud Storage can now automatically generate context — including image annotations — so your data is discoverable and usable from the moment it lands. You pay to annotate the data once at write time, and every downstream system can use those annotations immediately for the life of the object.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Cloud Storage MCP server &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;lets you read, write, and analyze Cloud Storage data using the standard MCP protocol. &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Smart Storage enables these capabilities, and others, thanks to its &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;object context,&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; now generally available. This metadata substrate adds structured, mutable, IAM-governed context to every object. Customers write their own tags and classifications; Google's annotation pipelines automatically attach labels, extracted entities, and compliance signals.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/2_EJZvHi4.max-1000x1000.png"
        
          alt="2"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;With Smart Storage, ML teams can select training datasets from semantic criteria without building retrieval pipelines. AI agents can ground their reasoning in enterprise data without a separate retrieval layer. &lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Storage Intelligence: Data management at AI scale&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;As data estates grow to hundreds of petabytes, storage costs can spike without warning, and security blind spots can multiply across billions of objects. To manage this, teams have to stitch together multiple tools just to answer basic questions about their own data.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Last year we launched &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;Storage Intelligence&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; to give enterprises a unified management experience built directly into Cloud Storage. Today, 70% of our largest customers use Storage Intelligence, each of whom manage over 50 billion objects.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Storage Intelligence provides a single view across your entire project or organization, with unique capabilities like bucket relocations across regions. Today, we're making it significantly more powerful with:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;New &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;zero-configuration dashboards&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; instantly surface cost anomalies and integrate Security Command Center’s Data Security Posture Management (DSPM) data governance feature, to detect critical security vulnerabilities across Cloud Storage — no setup required. &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;New &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;object events and bucket activity&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; tables in Insights Datasets now drive deeper cost analysis and accelerate operational tasks. You can use these insights to perform a wide range of analyses, from optimizing bucket placement based on egress patterns to quickly troubleshooting 429 errors by finding the impacted objects.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Enhanced batch operations&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; make it even simpler to act on billions of objects with new change ACL and storage class operations, and support for multi-bucket operations.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Enhancing the storage ecosystem&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Beyond our core storage offerings, we are streamlining how enterprises migrate to and protect data in the cloud.&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Google Cloud NetApp Volumes:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; With the launch of Flex Unified, NetApp Volumes now provides a unified enterprise storage platform that bridges the data center and the cloud, provisioning both block (iSCSI, NVMe/TCP) and file (NFS/SMB) on the same storage pool. New &lt;/span&gt;&lt;a href="https://docs.cloud.google.com/netapp/volumes/docs/ontap/overview"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;ONTAP-mode&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; lets you bring your existing automation (Terraform, Ansible) and ONTAP APIs directly to NetApp Volumes.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Filestore for GKE:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Developers building AI workloads on Google Kubernetes Engine (GKE) can start small, with shares as small as 100 GiB, and scale capacity and IOPS independently. At the same time, tighter integration to the &lt;/span&gt;&lt;a href="https://cloud.google.com/blog/products/storage-data-transfer/a-peek-behind-colossus-googles-file-system"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Colossus&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; distributed file system provides more scale and enterprise capabilities.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Data protection:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Google Cloud Backup and DR now features agentic AI capabilities that can autonomously audit your backup estate and remediate coverage gaps, with new GA integrations for AlloyDB and Filestore.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Where to start&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;As you navigate today’s generational AI shift, you need a storage foundation to support ever-larger, more intelligent, and autonomous models. With new high-performance and intelligent storage layers, plus enhanced storage management tools and a deeper data protection bench, Google Cloud’s storage platforms lets you understand and use your enterprise data in ways that weren’t previously possible, allowing you to:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Reduce the AI data bottleneck: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Saturate compute and accelerate ROI. Keep your expensive GPUs and TPUs fully productive with high-throughput storage that delivers the extreme performance required for large-scale training and inference.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Build agent-ready data foundations:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Shift from building custom pipelines to an active knowledge base where self-describing objects let AI agents instantly reason over data without manual prep.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Minimize blind spots across exabytes:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Replace fragmented management tools with zero-configuration dashboards and datasets to instantly surface cost anomalies and security risks across billions of objects.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Embrace the storage ecosystem:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Streamline migration and protection. Bridge your data center to the cloud, scale containerized apps, and automate data resilience with agentic AI.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Visit the &lt;/span&gt;&lt;a href="https://console.cloud.google.com/storage"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Google Cloud Storage console&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; to explore these new features, read more about &lt;/span&gt;&lt;a href="https://docs.cloud.google.com/storage/docs/rapid/high-performance-storage"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Cloud Storage Rapid&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, or explore our &lt;/span&gt;&lt;a href="https://www.googlecloudevents.com/next-vegas/session-library?session_id=3913124&amp;amp;name=google-cloud-storage-products-the-ai-ready-foundation-for-your-data" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Next '26 storage sessions&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;</description><pubDate>Wed, 22 Apr 2026 12:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/storage-data-transfer/next26-storage-announcements/</guid><category>Google Cloud Next</category><category>Storage &amp; Data Transfer</category><media:content height="540" url="https://storage.googleapis.com/gweb-cloudblog-publish/images/GCN26_102_BlogHeader_2436x1200_Opt_10_Dark.max-600x600.jpg" width="540"></media:content><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>Storage innovations to accelerate your AI workloads at Next ‘26</title><description></description><image>https://storage.googleapis.com/gweb-cloudblog-publish/images/GCN26_102_BlogHeader_2436x1200_Opt_10_Dark.max-600x600.jpg</image><site_name>Google</site_name><url>https://cloud.google.com/blog/products/storage-data-transfer/next26-storage-announcements/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Sameet Agarwal</name><title>VP/GM, Storage, Google Cloud</title><department></department><company></company></author><author xmlns:author="http://www.w3.org/2005/Atom"><name>Asad Khan</name><title>Sr. Director of Product Management, Google Cloud</title><department></department><company></company></author></item><item><title>New GKE Cloud Storage FUSE Profiles take the guesswork out of configuring AI storage</title><link>https://cloud.google.com/blog/products/containers-kubernetes/optimize-aiml-workloads-with-gke-cloud-storage-fuse-profiles/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;In the world of AI/ML, data is the fuel that drives training and inference workloads. For Google Kubernetes Engine (GKE) users, Cloud Storage FUSE provides high-performance, scalable access to data stored in Google Cloud Storage. However, we learned from customers that getting the maximum performance out of Cloud Storage FUSE can be complex.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Today, we are excited to introduce GKE Cloud Storage FUSE Profiles, a new feature designed to automate performance tuning and accelerate data access for your AI/ML workloads (training, checkpointing, or inference) with minimal operational overhead. With these profiles, tuned for your specific workload needs, you can enjoy high performance of Cloud Storage FUSE out of the box.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Before &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;(manual tuning)&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;#x27;apiVersion: v1\r\nkind: PersistentVolume\r\nmetadata:\r\n  name: serving-bucket-pv\r\nspec:\r\n  accessModes:\r\n  - ReadWriteMany\r\n  capacity:\r\n    storage: 64Gi\r\n  persistentVolumeReclaimPolicy: Retain\r\n  storageClassName: &amp;quot;&amp;quot;\r\n  claimRef:\r\n    name: serving-bucket-pvc\r\n  mountOptions:\r\n    - implicit-dirs\r\n    - metadata-cache:ttl-secs:-1\r\n    - metadata-cache:stat-cache-max-size-mb:-1\r\n    - metadata-cache:type-cache-max-size-mb:-1\r\n    - file-cache:max-size-mb:-1\r\n    - file-cache:cache-file-for-range-read:true\r\n    - file-system:kernel-list-cache-ttl-secs:-1\r\n    - file-cache:enable-parallel-downloads:true\r\n    - read_ahead_kb=1024\r\n  csi:\r\n    driver: gcsfuse.csi.storage.gke.io\r\n    volumeHandle: BUCKET_NAME\r\n    volumeAttributes:\r\n      skipCSIBucketAccessCheck: &amp;quot;true&amp;quot;\r\n      gcsfuseMetadataPrefetchOnMount: &amp;quot;true&amp;quot;\r\n---\r\napiVersion: v1\r\nkind: PersistentVolumeClaim\r\nmetadata:\r\n  name: serving-bucket-pvc\r\nspec:\r\n  accessModes:\r\n  - ReadWriteMany\r\n  resources:\r\n    requests:\r\n      storage: 64Gi\r\n  volumeName: serving-bucket-pv\r\n  storageClassName: &amp;quot;&amp;quot;\r\n–--\r\napiVersion: v1\r\nkind: Pod\r\nmetadata:\r\n  name: gcs-fuse-csi-example-pod\r\n  annotations:\r\n    gke-gcsfuse/volumes: &amp;quot;true&amp;quot;\r\nspec:\r\n  containers:\r\n    # Your workload container spec\r\n    ...\r\n    volumeMounts:\r\n    - name: serving-bucket-vol\r\n      mountPath: /serving-data\r\n      readOnly: true\r\n  serviceAccountName: KSA_NAME \r\n  volumes:\r\n    - name: gke-gcsfuse-cache # gcsfuse file cache backed by RAM Disk\r\n      emptyDir:\r\n        medium: Memory \r\n  - name: serving-bucket-vol\r\n    persistentVolumeClaim:\r\n      claimName: serving-bucket-pvc&amp;#x27;), (&amp;#x27;language&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8eb14a60&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;After &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;(Cloud Storage FUSE mount options, CSI configs, and file cache medium automatically configured!)&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;#x27;apiVersion: v1\r\nkind: PersistentVolume\r\nmetadata:\r\n  name: serving-bucket-pv\r\nspec:\r\n  accessModes:\r\n  - ReadWriteMany\r\n  capacity:\r\n    storage: 64Gi\r\n  persistentVolumeReclaimPolicy: Retain\r\n  storageClassName: gcsfusecsi-serving\r\n  claimRef:\r\n    name: serving-bucket-pvc\r\n  csi:\r\n    driver: gcsfuse.csi.storage.gke.io\r\n    volumeHandle: BUCKET_NAME\r\n---\r\napiVersion: v1\r\nkind: PersistentVolumeClaim\r\nmetadata:\r\n  name: serving-bucket-pvc\r\nspec:\r\n  accessModes:\r\n  - ReadWriteMany\r\n  resources:\r\n    requests:\r\n      storage: 64Gi\r\n  volumeName: serving-bucket-pv\r\n  storageClassName: gcsfusecsi-serving\r\n–--\r\napiVersion: v1\r\nkind: Pod\r\nmetadata:\r\n  name: gcs-fuse-csi-example-pod\r\n  annotations:\r\n    gke-gcsfuse/volumes: &amp;quot;true&amp;quot;\r\nspec:\r\n  containers:\r\n    # Your workload container spec\r\n    ...\r\n    volumeMounts:\r\n    - name: serving-bucket-vol\r\n      mountPath: /serving-data\r\n      readOnly: true\r\n  serviceAccountName: KSA_NAME \r\n  volumes: \r\n  - name: serving-bucket-vol\r\n    persistentVolumeClaim:\r\n      claimName: serving-bucket-pvc&amp;#x27;), (&amp;#x27;language&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8eb14340&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;The trouble with optimizing Cloud Storage FUSE&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Optimizing Cloud Storage FUSE for high-performance workloads is a multi-dimensional problem. Historically, users had to navigate &lt;/span&gt;&lt;a href="https://docs.cloud.google.com/storage/docs/cloud-storage-fuse/performance"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;manual configuration guides&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; that could span dozens of pages. And as AI/ML has evolved, Cloud Storage FUSE’s capabilities have also increased, with new mount options available to accelerate your workloads. The "right" settings were never static; they depended heavily on a variety of dynamic factors:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Bucket characteristics&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;: The total size of your dataset and the number of objects significantly impact metadata and file cache requirements.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Infrastructure variability:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Optimal configurations change based on whether you are using GPUs, TPUs, or general-purpose compute.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Node resources: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Available RAM and Local SSD capacity determine how much data can be cached locally to minimize expensive round-trips to Cloud Storage.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Workload patterns: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;A training workload (high-throughput reads of large datasets) requires different tuning than a checkpointing workload (bursty, high-throughput writes) or a serving workload (latency-sensitive model loading).&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;In fact, many customers leave available performance on the table or face reliability issues (e.g., Pod Out-of-Memory kills) due to unoptimized or misconfigured Cloud Storage FUSE settings.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Introducing Cloud Storage FUSE Profiles for GKE&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;GKE Cloud Storage FUSE Profiles simplify this complexity with pre-defined, dynamically managed StorageClasses tailored for specific AI/ML patterns. Instead of manually adjusting dozens of mount options, you simply select a profile that matches your workload type.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;These profiles operate on a layered model. They take the base best practices from Cloud Storage FUSE and add a GKE-specific intelligence layer. When you deploy a Pod using a profile, GKE automatically:&lt;/span&gt;&lt;/p&gt;
&lt;ol&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Scans your bucket (or a specific directory) to understand its size and object count.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Analyzes the target node to check for available RAM, Local SSD, and accelerator types.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Calculates optimal cache sizes and selects the best backing medium (RAM or Local SSD) automatically.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;We are launching with three primary profiles:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li role="presentation"&gt;&lt;code style="vertical-align: baseline;"&gt;gcsfusecsi-training&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt;: Optimized for high-throughput reads to keep GPUs and TPUs fed with data.&lt;/span&gt;&lt;/li&gt;
&lt;li role="presentation"&gt;&lt;code style="vertical-align: baseline;"&gt;gcsfusecsi-serving&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt;: Optimized for model loading and inference, with automated &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/anywhere-cache"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Rapid Cache&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; integration.&lt;/span&gt;&lt;/li&gt;
&lt;li role="presentation"&gt;&lt;code style="vertical-align: baseline;"&gt;gcsfusecsi-checkpointing&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt;: Optimized for fast, reliable writes of large multi-gigabyte checkpoint files.&lt;/span&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Using GKE Cloud Storage FUSE Profiles delivers several benefits:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Simplified tuning:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Replace complex, error-prone manual configurations with three simple, purpose-built StorageClasses.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Dynamic, resource-aware optimization:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; The CSI driver automatically adjusts cache sizes based on real-time environment signals, so that you can maximize performance without risking node stability.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Accelerated read performance:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; The serving profile automatically triggers Rapid Cache, placing your data closer to your compute for faster cold-start model loading.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li&gt;&lt;strong style="vertical-align: baseline;"&gt;Granular performance insights:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Gain visibility into automated tuning decisions through structured logs that detail exactly why specific cache sizes and mediums were selected for your Pod.&lt;/span&gt;&lt;/li&gt;
&lt;/ul&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/image1_4Ng3Hpa.max-1000x1000.png"
        
          alt="image1"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Using GKE Cloud Storage FUSE Profiles inference profile, we were able to reduce model loading time for a Qwen3-235B-A22B workload on TPUs (480GB) from 39 hours to just 14 minutes, helping customers achieve the maximum benefit of Cloud Storage FUSE GCSFuse out-of-the-box.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;How to use Cloud Storage FUSE Profiles on GKE&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;To get started, ensure your cluster is running GKE version 1.35.1-gke.1616000 or later with the Cloud Storage FUSE CSI driver enabled.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;1. Identify the StorageClass&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;GKE comes pre-installed with the profile-based StorageClasses. You can verify them with:&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;#x27;kubectl get sc -l gke-gcsfuse/profile=true&amp;#x27;), (&amp;#x27;language&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8eb14e20&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;2. Create your PV and PVC&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;When creating your PersistentVolume, point it to your Cloud Storage bucket. GKE automatically initiates a bucket scan to determine the optimal configuration.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;#x27;apiVersion: v1\r\nkind: PersistentVolume\r\nmetadata:\r\n  name: gcs-pv\r\nspec:\r\n  accessModes:\r\n    - ReadWriteMany\r\n  capacity:\r\n    storage: 5Gi\r\n  persistentVolumeReclaimPolicy: Retain  \r\n  storageClassName: gcsfusecsi-training\r\n  mountOptions:\r\n    - only-dir=my-ml-dataset-subdirectory # Optional\r\n  csi:\r\n    driver: gcsfuse.csi.storage.gke.io\r\n    volumeHandle: my-ml-dataset-bucket\r\n---\r\napiVersion: v1\r\nkind: PersistentVolumeClaim\r\nmetadata:\r\n  name: gcs-pvc\r\nspec:\r\n  accessModes:\r\n    - ReadWriteMany\r\n  resources:\r\n    requests:\r\n      storage: 5Gi\r\n  storageClassName: gcsfusecsi-training\r\n  volumeName: gcs-pv&amp;#x27;), (&amp;#x27;language&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8eb14610&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;3. Create your Deployment&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Once your Persistent Volume Claim (PVC) is bound, simply consume it in your Deployment as you would any other volume. GKE mounts the volume with the precise settings your hardware and dataset require.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;#x27;apiVersion: apps/v1\r\nkind: Deployment\r\nmetadata:\r\n  name: my-deployment\r\nspec:\r\n  replicas: 3\r\n  selector:\r\n    matchLabels:\r\n      app: my-app\r\n  template:\r\n    metadata:\r\n      labels:\r\n        app: my-app\r\n      annotations:\r\n        gke-gcsfuse/volumes: &amp;quot;true&amp;quot;\r\n    spec:\r\n      serviceAccountName: my-ksa\r\n      containers:\r\n      - name: my-container\r\n        image: busybox\r\n        volumeMounts:\r\n        - name: my-gcs-volume\r\n          mountPath: &amp;quot;/data&amp;quot;\r\n      volumes:\r\n      - name: my-gcs-volume\r\n        persistentVolumeClaim:\r\n          claimName: gcs-pvc&amp;#x27;), (&amp;#x27;language&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8eb14b50&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;After it's deployed, the CSI driver automatically calculates optimal cache sizes and mount options based on your node's resources, such as GPUs or TPUs, memory, Local SSD, the bucket or sub-directory size, and the sidecar resource limits.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Get started today&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;GKE Cloud Storage FUSE Profiles remove the guesswork from configuring your cloud storage for high performance. By moving from manual "knob-turning" to automated, workload-aware profiles, you can spend less time debugging storage throughput and more time building the next generation of AI.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Ready to get started? GKE Cloud Storage FUSE Profiles are generally available in version 1.35.1-gke.1616000. Explore the &lt;/span&gt;&lt;a href="https://docs.cloud.google.com/kubernetes-engine/docs/how-to/persistent-volumes/gcsfuse-profiles"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;official documentation&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; to configure Cloud Storage FUSE profiles in GKE for your AI/ML workloads!&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;</description><pubDate>Wed, 08 Apr 2026 16:30:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/containers-kubernetes/optimize-aiml-workloads-with-gke-cloud-storage-fuse-profiles/</guid><category>AI &amp; Machine Learning</category><category>GKE</category><category>Storage &amp; Data Transfer</category><category>Containers &amp; Kubernetes</category><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>New GKE Cloud Storage FUSE Profiles take the guesswork out of configuring AI storage</title><description></description><site_name>Google</site_name><url>https://cloud.google.com/blog/products/containers-kubernetes/optimize-aiml-workloads-with-gke-cloud-storage-fuse-profiles/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Nishtha Jain</name><title>Engineering Manager</title><department></department><company></company></author><author xmlns:author="http://www.w3.org/2005/Atom"><name>Uriel Guzmán-Mendoza</name><title>Software Engineer</title><department></department><company></company></author></item><item><title>Accelerate model downloads on GKE with NVIDIA Run:ai Model Streamer</title><link>https://cloud.google.com/blog/products/containers-kubernetes/nvidia-runai-model-streamer-supports-cloud-storage/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;As large language models (LLMs) continue to grow in size and complexity, the time it takes to load them from storage to accelerator memory for inference can become a significant bottleneck. This "cold start" problem isn't just a minor delay — it's a critical barrier to building resilient, scalable, and cost-effective AI services. Every minute spent loading a model is a minute a GPU is sitting idle, a minute your service is delayed from scaling to meet demand, and a minute a user request is waiting.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Google Cloud and NVIDIA are committed to removing these barriers. We’re excited to highlight a powerful, open-source collaboration that helps AI developers do just that: the NVIDIA Run:ai Model Streamer now comes with native &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/introduction"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Google Cloud Storage&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; support, supercharging vLLM inference workloads on Google Kubernetes Engine (GKE). Accessing data for AI/ML from Cloud Storage on GKE has never been faster!&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/image1_uEwzVCo.max-1000x1000.png"
        
          alt="image1"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The chart above shows how quickly the model streamer can fetch a 141GB Llama 3.3-7 70B model from Cloud Storage as compared to the default vLLM model loader (lower is better). &lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Boost resilience and scalability with fewer cold starts&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;For an inference server running on Kubernetes, a "cold start" involves several steps: pulling the container image, starting the process, and — most time-consuming of all — loading the model weights into GPU memory. For large models, this loading phase can take many minutes, with painful consequences such as slow auto-scaling and idling GPUs as they wait for the workload to start up. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;By streaming the model into GPU memory, the model streamer slashes potentially the most time-consuming part of the startup process. Instead of waiting for an entire model to be downloaded before loading, the streamer fetches model tensors directly from object storage and streams them concurrently to GPU memory. This dramatically reduces model loading times from minutes to seconds.  &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;For workloads that rely on model parallelism— where a single model is partitioned and executed across multiple GPUs— the model streamer goes a step further. Its distributed streaming capability is optimized to take full advantage of &lt;/span&gt;&lt;a href="https://www.nvidia.com/en-us/data-center/nvlink/" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;NVIDIA NVLink&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, using high-bandwidth GPU-to-GPU communication to coordinate loading across multiple processes. Reading the weights from storage is divided efficiently and evenly across all participating processes, with each one fetching a portion of the model weights from storage and then sharing its segment with the others over NVLink. This allows even multi-GPU deployments to benefit from faster startups and fewer cold-start bottlenecks.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Performance and simplicity&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The latest updates to the Model Streamer introduce first-class support for Cloud Storage, creating an integrated and high-performance experience for Google Cloud users. This integration is designed to be simple, fast, and secure, especially for workloads running on GKE.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;For users of popular inference servers like &lt;/span&gt;&lt;a href="https://docs.vllm.ai/en/stable/models/extensions/runai_model_streamer.html" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;vLLM&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, enabling the streamer is as simple as adding a single flag to your vLLM command line:&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;code style="vertical-align: baseline;"&gt; &lt;/code&gt;&lt;code style="vertical-align: baseline;"&gt;--load-format=runai_streamer&lt;/code&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Here’s how easy it is to launch a model stored in a Cloud Storage bucket with vLLM:&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;#x27;vllm serve gs://your-gcs-bucket/path/to/your/model \r\n--load-format=runai_streamer&amp;#x27;), (&amp;#x27;language&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8eb61970&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The NVIDIA Run:ai Model Streamer is a key component for Vertex AI Model Garden's large model deployments. With container image streaming and model weight streaming, we have been able to significantly improve the first deployment and autoscaling experience for our users, and the efficiency of NVIDIA GPUs.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;When running on GKE, the Model Streamer can automatically use the cluster's &lt;/span&gt;&lt;a href="https://cloud.google.com/kubernetes-engine/docs/how-to/workload-identity"&gt;&lt;strong style="text-decoration: underline; vertical-align: baseline;"&gt;Workload Identity&lt;/strong&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. This means you no longer need to manually manage and mount service account keys, simplifying your deployment manifests and enhancing your security posture. The following deployment manifest shows how to launch a container serving Llama3 70B on GKE. We have added the model loader &lt;/span&gt;&lt;a href="https://docs.vllm.ai/en/stable/models/extensions/runai_model_streamer/#tunable-parameters" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;distributed&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; option to accelerate loads when model parallelism &amp;gt; 1:&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;#x27;apiVersion: apps/v1\r\nkind: Deployment\r\n…\r\n   spec:\r\n     serviceAccountName: gcs-access\r\n     containers:\r\n       - args:\r\n           - --model=gs://your-gcs-bucket/path/to/your/model \r\n           - --load-format=runai_streamer\r\n \t\t- --model-loader-extra-config={&amp;quot;distributed&amp;quot;:true}\r\n\t\t…\r\n         command:\r\n           - python3\r\n           - -m\r\n           - vllm.entrypoints.openai.api_server\r\n         image: vllm/vllm-openai:latest\r\n         ….&amp;#x27;), (&amp;#x27;language&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8eb61dc0&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;That’s it! The streamer handles the rest, auto-tuning streaming concurrency to match your VM’s performance. For more details, see the documentation on &lt;/span&gt;&lt;a href="https://cloud.google.com/kubernetes-engine/docs/how-to/persistent-volumes/run-ai-model-streamer"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;optimizing vLLM model loading on GKE&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Combining NVIDIA Run:ai Model Streamer with Cloud Storage Anywhere Cache&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;a href="https://cloud.google.com/storage/docs/anywhere-cache"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Anywhere Cache&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; provides zonally co-located SSD-backed caching for data stored in a regional or multi-regional Cloud Storage bucket. Reducing latency by up to 70% and providing up to 2.5 TB/s of read throughput, Anywhere Cache is a great solution for scale-out inference workloads where the same model is downloaded multiple times across a series of nodes. Together, Anywhere Cache server-side acceleration, along with the NVIDIA Run:ai Model Streamer’s client-side acceleration, create an easy-to-manage, extremely performant model-loading system.  &lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Get started today&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The NVIDIA Run:ai Model Streamer is evolving into a critical piece of the AI infrastructure puzzle, enabling teams to build faster, more resilient, and more flexible MLOps pipelines on GKE. &lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;To learn more about how to use the model streamer on GKE see our &lt;/span&gt;&lt;a href="https://cloud.google.com/kubernetes-engine/docs/how-to/persistent-volumes/run-ai-model-streamer"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;GKE NVIDIA Run:ai Guide&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;For detailed instructions on using the streamer with vLLM, see the&lt;/span&gt;&lt;a href="https://docs.vllm.ai/en/stable/models/extensions/runai_model_streamer.html" rel="noopener" target="_blank"&gt;&lt;span style="vertical-align: baseline;"&gt; &lt;/span&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;official vLLM documentation&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li&gt;&lt;span style="vertical-align: baseline;"&gt;To learn more and contribute to the model streamers ongoing development check out the &lt;/span&gt;&lt;a href="https://github.com/run-ai/runai-model-streamer" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;NVIDIA Run:ai Model Streamer project on GitHub&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/li&gt;
&lt;/ul&gt;&lt;/div&gt;</description><pubDate>Thu, 04 Dec 2025 17:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/containers-kubernetes/nvidia-runai-model-streamer-supports-cloud-storage/</guid><category>AI &amp; Machine Learning</category><category>GKE</category><category>Storage &amp; Data Transfer</category><category>Containers &amp; Kubernetes</category><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>Accelerate model downloads on GKE with NVIDIA Run:ai Model Streamer</title><description></description><site_name>Google</site_name><url>https://cloud.google.com/blog/products/containers-kubernetes/nvidia-runai-model-streamer-supports-cloud-storage/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Peter Schuurman</name><title>Software Engineer, Google</title><department></department><company></company></author><author xmlns:author="http://www.w3.org/2005/Atom"><name>Brian Kaufman</name><title>Senior Product Manager, Google</title><department></department><company></company></author></item><item><title>Reducing TCO for AI inferencing with external KV Cache on Managed Lustre</title><link>https://cloud.google.com/blog/products/storage-data-transfer/choosing-google-cloud-managed-lustre-for-your-external-kv-cache/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The demand for AI inference infrastructure is accelerating, with market spend expected to soon surpass investment in training the models themselves. This growth is driven by the demand for richer experiences, particularly through support for larger context windows and the rise of agentic AI. As organizations aim to improve user experience while optimizing costs, efficient management of inference resources is paramount.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;According to an internal experimental study of large model inferencing, external key-value caches — KV Cache or, “attention caches” — on high-performance storage like &lt;/span&gt;&lt;a href="https://cloud.google.com/products/managed-lustre?e=48754805&amp;amp;hl=en"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Google Cloud Managed Lustre&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, can reduce total cost of ownership (TCO) by up to 35%, allowing organizations to serve the same workload with ~40% fewer GPUs by offloading prefill compute to I/O&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;. In this blog, we explore the core challenges of managing long-context AI inference and detail how Google Cloud Managed Lustre provides the high-performance external storage solution required to achieve these significant cost and efficiency benefits.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;About KV Cache&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;During the inference phase, a KV Cache&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;is a critical optimization technique for the efficient operation of Transformer-based large language models (LLMs).&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The key innovation of the Transformer was the complete elimination of sequential processing (recurrence), which was achieved by introducing the self-attention mechanism to allow every element in a sequence to instantaneously and dynamically compare itself to and assess the relevance of every other element (a global, all-at-once evaluation). Within this self-attention mechanism, the model computes Key (K) and Value (V) vectors of all preceding tokens in the sequence. To generate the next token during the inference phase, the model needs the K and V vectors of &lt;/span&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;all&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; the previous tokens.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;This is where the&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;KV Cache&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;comes into play. The KV Cache stores these K and V vectors after the initial context processing (known as the "prefill" stage), thereby avoiding the redundant, costly re-computation of the context sequence when generating subsequent tokens. By eliminating this re-computation, the KV Cache vastly speeds up the overall inference process. While smaller caches can fit in high-bandwidth memory (HBM) or host DRAM — up to a few TBs of memory may be available in a single multi-accelerator server — managing a KV Cache for contexts across multiple concurrent users that exceed the memory capacity often requires external or hierarchical storage solutions.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;These large contexts can make the "prefill" computation — the calculation that an AI model performs when processing a large context window — very expensive:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;For a large context of 100K or more tokens, the prefill computation may cause the time to first token (TTFT) to increase to tens of seconds.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Prefill computation requires a high number of floating-point operations (FLOPs). KV Cache reuse saves these costs and makes additional resources available on the accelerator.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The growth of agentic AI is likely to make the challenge of managing a long context even greater. Unlike a simple chatbot, agentic AI is built for action. It moves beyond conversation to solve problems proactively, completing tasks on your behalf. To do this, it actively gathers context from a wide range of digital sources. Agentic AI may, for example: check live flight data, pull a customer's history from a database, research topics on the web, and/or keep organized notes in its own files. Agentic AI thereby builds a rich understanding of its environment, but often increases context lengths and their associated KV Cache size.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The key to managing performance costs at scale is to ensure that the accelerator is utilized as fully as possible. High-performance, scale-out storage provides the required greater throughput per accelerator and therefore translates into lighter resource requirements.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;External KV Cache on Google Cloud Managed Lustre&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;We believe that &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;Google Cloud Managed Lustre&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; should be your primary storage solution for external KV Cache. On GPUs, Lustre is assisted by locally attached SSDs. And on TPUs, where local SSDs are not available, Lustre’s role is even more central.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;A recent LMCache blog post by Google’s Danna Wang, “&lt;/span&gt;&lt;a href="https://blog.lmcache.ai/2025-10-07-LMCache-on-GKE/" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;LMCache on Google Kubernetes Engine: Boosting LLM Inference Performance with KV Cache on Tiered Storage&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;,” demonstrates the foundational value of host-level offloading. Our Managed Lustre strategy is the next evolution of this host-offloading concept. While Local SSDs and CPU RAM are effective node-local tiers, they are fixed in size and cannot be shared. Managed Lustre provides a parallel file system to act as the massive, high-throughput external storage, making it a great solution for large-scale, multi-node, and multi-tenant AI inference workloads where the cache exceeds the capacity of the host machine.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Here’s an example of how the performance gains of Managed Lustre can reduce your TCO:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;In an experiment with a 50K token context and a high cache hit rate (about 75%), using&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Managed Lustre improved total inference throughput by 75% and reduced MTTF by more than 40% compared to using KV Cache in host memory alone (further detail below).&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;TCO analysis yielded a 35% savings from using an external attention/KV Cache for a workload processing 1 million Tokens per Second (TPS) and leveraging A3-Ultra VMs and Managed Lustre, when compared to a workload leveraging no external storage.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Our experiment demonstrated that with configuration tuning and an improvement in KV Cache software to adopt more I/O parallelism, Managed Lustre can substantially improve inference performance.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Total Cost of Ownership: Analysis&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;When evaluating a KV Cache solution, it's critical to consider the TCO, which includes not just compute and storage costs but also operational expenses and potential savings. Our analysis shows that a high performance storage-backed KV Cache, like one built on Managed Lustre, provides a compelling TCO advantage compared to purely memory-based solutions.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Cost savings&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;After taking incremental storage costs into account, we project that the TCO for a file-system-backed KV Cache solution, processing 1m TPS, is 35% lower compared to a memory-only solution. This makes it a more scalable and economically viable option for large-scale AI inference deployments.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The primary TCO benefit comes from a more efficient utilization of expensive compute resources. By offloading KV Cache to a high-performance storage solution, you can achieve a higher inference throughput per accelerator. This means that fewer accelerators are needed for the same workload: You can handle a specific number of queries per second with ~40% fewer accelerators, resulting in direct cost savings.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;TCO model assumptions&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The TCO calculation includes several key components:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Storage costs (list price):&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; These are the costs of Managed Lustre. Testing used the 1000 MB/s per TiB Performance Tier. The TCO model includes sufficient Lustre capacity (73 A3-Ultra machines, with 18 TiB Lustre capacity per machine) to hit the 1m TPS target rate.&lt;/span&gt;&lt;/li&gt;
&lt;li role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Compute costs (list price):&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; A3-Ultra VMs each with 8x H200s GPUs and 8x 141 GB HBM (spot prices will be lower).&lt;/span&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Performance benchmarks&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Our experiments demonstrated Google Cloud Managed Lustre’s ability to deliver the high-performance I/O necessary with a state-of-the-art LLM. These experiments serve Deepseek-R1 on a Google Cloud A3-Ultra machine (8x H200s; 8x 141GB HBMs). The experiments ran a synthetic serving workload with a 50K token context and a high cache-hit rate (about 75% hit rate) with a total KV Cache size of about 3.4TiB. The memory-only baseline uses 1 TiB host memory for KV Cache. We experimented with two variants of Managed Lustre at high and low I/O parallelism. For high I/O parallelism, we utilized 32 I/O worker threads to read KV Cache data from Lustre in parallel.&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Lustre improved total inference throughput by 75% and reduced the mean time to first token by greater than 40% compared to using KV Cache in host memory alone.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/1_WcX8I6F.max-1000x1000.png"
        
          alt="1"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/2_vxy9BwW.max-1000x1000.png"
        
          alt="2"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Ready to optimize your inference workloads?&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;To get started with an external KV Cache solution that solves the capacity limits of long context windows and delivers significant performance gains on your large-scale LLMs, follow these steps:&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;1. Provision your infrastructure; create a Managed Lustre instance:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Provision your Lustre file system in the same region and zone as your target accelerators (GPUs or TPUs) for optimal low-latency access.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Deploy your inference engine: Deploy your LLM using a high-performance inference server like vLLM or a similar framework that supports an external KV Cache or paged-attention architecture.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;2. Configure for performance&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Once you’ve mounted Managed Lustre, you must configure your inference engine software to leverage the high-performance storage:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Implement direct I/O:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Configure your application to access Managed Lustre using the &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;o_direct&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; flag. This bypasses the general-purpose file system cache, allowing the inference engine to manage the critical host memory more effectively.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Tune I/O parallelism:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Depending on your inference KV Cache software, its out-of-the-box storage I/O parallelism may not be ideal. You may need to tune the KV Cache software to read KV chunk files with enhanced parallelism to maximize performance.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;To take the next step, read the documentation about how to &lt;/span&gt;&lt;a href="https://cloud.google.com/managed-lustre/docs/overview"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;get started with Managed Lustre&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;</description><pubDate>Fri, 31 Oct 2025 16:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/storage-data-transfer/choosing-google-cloud-managed-lustre-for-your-external-kv-cache/</guid><category>AI &amp; Machine Learning</category><category>Storage &amp; Data Transfer</category><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>Reducing TCO for AI inferencing with external KV Cache on Managed Lustre</title><description></description><site_name>Google</site_name><url>https://cloud.google.com/blog/products/storage-data-transfer/choosing-google-cloud-managed-lustre-for-your-external-kv-cache/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Kai Shen</name><title>Distinguished Software Engineer</title><department></department><company></company></author><author xmlns:author="http://www.w3.org/2005/Atom"><name>Barak Epstein</name><title>Sr Product Manager</title><department></department><company></company></author></item><item><title>From dark data to bright insights: The dawn of smart storage</title><link>https://cloud.google.com/blog/products/storage-data-transfer/make-your-unstructured-data-smart-with-cloud-storage/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Organizations interested in AI today have access to amazing computational power with Tensor Processing Units (TPUs) and Graphical Processing Units (GPUs), while foundational models like Gemini are redefining what's possible. Yet for many enterprises a critical obstacle to AI is the data itself, specifically unstructured data. According to Enterprise Strategy Group, for most organizations, 61% of their total data is unstructured, the vast majority of which sits unanalyzed and unlabeled in archives, so-called "dark data." But with the help of AI, this untapped resource is an opportunity to unlock a veritable treasure trove of insights. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;At the same time, when it comes to unstructured data, traditional tools only scratch the surface,&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; and subject matter experts must build massive, manual preprocessing pipelines and define the data’s semantic meaning. &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;This prevents any real analysis at scale, preventing companies from using even a fraction of what they store.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Now imagine a world where your unstructured data isn't just stored, but &lt;/span&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;understood&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;. A world where you can ask complex questions of data such as images, videos, and documents, and get interesting answers in return. This isn't just a futuristic vision — the era of smart storage is upon us. Today we are announcing new auto annotate and object contexts features that use AI to generate metadata and insights on your data, so you can then use your dark data for discovery, curation, and governance at scale. Better yet, the new features relieve you from having to build and manage your own object-analysis data pipelines&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; &lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Leveraging AI to transform dark data&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Now, as unstructured data lands in Google Cloud, it's no longer treated as a passive object. Instead, a data pipeline leverages AI to automatically process and understand the data, surfacing key insights and connections. Two new features are integral to this vision: &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;auto annotate&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;, which enriches your data by automatically generating metadata using Google’s pretrained AI models,&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;and&lt;/span&gt; &lt;a href="https://cloud.google.com/storage/docs/object-contexts"&gt;&lt;strong style="text-decoration: underline; vertical-align: baseline;"&gt;object contexts&lt;/strong&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, which lets you attach custom, actionable tags to your data. Together, these two features can help transform passive data into active assets, unlocking use cases such as rapid data discovery for AI model training, streamlined data curation to reduce model bias, enhanced data governance to protect sensitive information, and the ability to build powerful, stateful workflows directly on your storage.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Making your data smart&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Auto annotate,&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;currently in a limited experimental release, automatically generates rich metadata (“annotations”) about objects stored in Cloud Storage buckets by applying Google's advanced AI models, starting with image objects. Getting started is simple: enable auto annotate for your selected buckets or an entire project, pick one or more available models, and your entire image library will be annotated. Furthermore, new images are automatically annotated as they are uploaded. An annotation’s lifecycle is always tied to its object’s, simplifying management and helping to ensure consistency. &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;Importantly, auto annotate operates under your control, only accessing object content to which you have explicitly granted permissions. Then, you can query the annotations, which are available as object contexts, through Cloud Storage API calls and &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/insights/datasets"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Storage Insights datasets&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; The initial release uses pretrained models for generating annotations: object detection with confidence scores, image labeling, and objectionable content detection.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/1_-_Auto_annotations_car.max-1000x1000.png"
        
          alt="1 - Auto annotations car"&gt;
        
        &lt;/a&gt;
      
        &lt;figcaption class="article-image__caption "&gt;&lt;p data-block-key="bzpog"&gt;a sample of generated annotations for an object&lt;/p&gt;&lt;/figcaption&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Then, with object contexts&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;, you can attach custom key-value pair metadata&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; directly to objects in Cloud Storage, including information generated by the new auto annotate feature. Currently in preview, object contexts are natively integrated with Cloud Storage APIs for listing and batch operations, as well as Storage Insights datasets for analysis in BigQuery. Each context includes object creation and modification timestamps, providing valuable lineage information. You &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;can use Identity and Access Management (IAM) permissions to control who can add, change, or remove object contexts. &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;W&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;hen migrating data from Amazon S3 using Cloud Storage APIs, existing S3 Object Tags are automatically converted into contexts&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;In short, object contexts provide a flexible and native way to add context to enrich your data. Combined with a smart storage feature like auto annotations, object contexts convert data into information, letting you build sophisticated data management workflows directly within Cloud Storage.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Now, let’s take a deeper look at some of the new use cases these smart storage features deliver.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;1. Data discovery &lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;One of the most significant challenges in building new AI applications is data discovery — how to find the most relevant data across an enterprise's vast and often siloed data stores. Locating specific images or information within petabytes of unstructured data can feel impossible. &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;Auto annotate automatically generates rich, descriptive annotations for your data in Cloud Storage.&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; Annotations, including labels and detected objects, are available within object contexts and fully indexed in BigQuery. After generating embeddings for them, you can then use BigQuery to run a semantic search for these annotations, effectively solving the "needle in a haystack" problem. &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;For example, a large retailer with millions of product images can use auto annotate and BigQuery to quickly find 'red dresses' or 'leather sofas', accelerating catalog management and marketing efforts.&lt;/span&gt;&lt;/p&gt;
&lt;h3 role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;2. Data curation for AI&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Building effective AI models requires carefully curated datasets. Sifting through data to ensure it is widely representative (e.g., "does this dataset have cars in multiple colors?") to reduce model bias, or to select specific training examples (e.g., “Find images with red cars”), is both time-consuming and error-prone. Auto annotate can identify attributes like colors and object types, to automate selecting balanced datasets. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;For instance, an autonomous vehicle company training models could use petabytes of on-road camera data to recognize traffic signs, using auto annotate to identify and extract images that contain the word ‘Stop’ or 'Pedestrian Crossing'.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/docs.google.com_presentation_d_1WSXexxs8Eb.max-1000x1000.png"
        
          alt="2 - Contexts car"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Vivint, a smart home and security company, has been using auto annotate to find and understand their data.&lt;/span&gt;&lt;/p&gt;
&lt;p style="padding-left: 40px;"&gt;&lt;span style="vertical-align: baseline;"&gt;“&lt;/span&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;Our customers trust us to help make their homes and lives safer, smarter, and more convenient, and AI is at the heart of our product and customer experience innovations. Cloud Storage auto annotate’s rich metadata delivered in BigQuery helps us scale our data discovery and curation efforts, speeding up our AI development process from 6 months to as little as 1 month by finding the needle-in-a-haystack data essential to improve our models.” &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;- Brandon Bunker, VP of Product, AI, Vivint&lt;/span&gt;&lt;/p&gt;
&lt;h3 role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;3. Governing unstructured data at scale&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Unstructured data is constantly growing, and manually managing and governing that data to identify sensitive information, detect policy violations, or categorize it for lifecycle management is a challenge. Auto annotate and object contexts help solve these data governance and compliance challenges. &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;For example, a retail customer can use auto annotate to identify and flag images containing visible customer personally identifiable information (PII) such as shipping labels or order forms.&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;This information, stored in object context, can then trigger automated governance actions such as moving flagged objects to a restricted bucket or initiating a review process.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;BigID, a partner building solutions on Cloud Storage, reports that using object contexts is helping them manage their customers’ risk&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;:&lt;/span&gt;&lt;/p&gt;
&lt;p style="padding-left: 40px;"&gt;&lt;span style="vertical-align: baseline;"&gt;“&lt;/span&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;Object contexts gives us a way to take the outputs of BigID's industry-leading data classification solutions and apply labels to Cloud Storage objects. Object contexts will allow BigID labels to shed light onto data in Cloud Storage: identifying objects which contain sensitive information and helping them understand and manage their risk across AI, security, and privacy." &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;- Marc Hebrard, Principal Technical Architect, BigID&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; &lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;The future is bright for your data&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;At Google Cloud, we’re committed to building a future where your data is not just a passive asset but an active catalyst for innovation. &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;Don't keep your valuable data in the dark. Bring your data to Cloud Storage and enable auto annotation and object contexts to unlock its full potential with Gemini, Vertex AI, and BigQuery.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;You can start using object contexts today, and &lt;/span&gt;&lt;a href="mailto:storage-ai@google.com"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;reach out to us&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; for an early look at auto annotate. Once you have access, s&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;imply enable auto annotate for selected buckets or on an entire project, pick one or more available models, and your entire image library will be annotated. Y&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;ou can then query the annotations that are available as object contexts through Cloud Storage API calls and Storage Insights datasets.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;To learn more, read about our end-to-end vision in a showcase paper with Enterprise Strategy Group: &lt;/span&gt;&lt;a href="https://services.google.com/fh/files/misc/google_cloud_smart_storage_esg.pdf" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Illuminating Dark Data With Smart Storage from Google Cloud&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. &lt;/span&gt;&lt;/p&gt;&lt;/div&gt;</description><pubDate>Tue, 14 Oct 2025 16:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/storage-data-transfer/make-your-unstructured-data-smart-with-cloud-storage/</guid><category>Storage &amp; Data Transfer</category><media:content height="540" url="https://storage.googleapis.com/gweb-cloudblog-publish/images/smart_storage.max-600x600.jpg" width="540"></media:content><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>From dark data to bright insights: The dawn of smart storage</title><description></description><image>https://storage.googleapis.com/gweb-cloudblog-publish/images/smart_storage.max-600x600.jpg</image><site_name>Google</site_name><url>https://cloud.google.com/blog/products/storage-data-transfer/make-your-unstructured-data-smart-with-cloud-storage/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Asad Khan</name><title>Sr. Director of Product Management, Google Cloud</title><department></department><company></company></author><author xmlns:author="http://www.w3.org/2005/Atom"><name>Manjul Sahay</name><title>Group Product Manager, Google Cloud Storage</title><department></department><company></company></author></item><item><title>Power your enterprise applications in the cloud with unified block and file storage</title><link>https://cloud.google.com/blog/products/storage-data-transfer/announcing-enhancements-to-google-cloud-netapp-volumes/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Migrating enterprise applications to the cloud requires a storage foundation that can handle everything from high-performance block workloads to globally distributed file access. To solve these challenges, we’re thrilled to announce two new capabilities for Google Cloud NetApp Volumes: unified iSCSI block and file storage to enable your storage area network (SAN) migrations, and NetApp FlexCache to accelerate your hybrid cloud workloads. These features, along with a new integration for agents built with Gemini Enterprise, can help you modernize even your most demanding applications.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Run your most demanding SAN workloads on Google Cloud&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;For decades, enterprises have relied on NetApp for both network attached storage (NAS) and SAN workloads on-premises. We’re now bringing that same trusted technology to a fully managed cloud service, allowing you to migrate latency-sensitive applications to Google Cloud without changing their underlying architecture.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Our unified service is engineered for enterprise-grade performance, with features including:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Low latency engineered for your most demanding applications&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Throughput that can burst up to 5 GiB/s with up to 160K random IOPS per volume&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Independent scaling of capacity, throughput, and IOPS to control costs&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Integrated data protection with &lt;/span&gt;&lt;a href="https://cloud.google.com/netapp/volumes/docs/configure-and-use/volume-snapshots/overview"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;NetApp Snapshots&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; for rapid recovery and ransomware defense&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;iSCSI block protocol support is available now via private preview for interested customers.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Accelerate your hybrid cloud with NetApp FlexCache&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;For organizations with distributed teams and a hybrid cloud strategy, providing fast access to shared datasets is critical. &lt;/span&gt;&lt;a href="https://cloud.google.com/netapp/volumes/docs/configure-and-use/volumes/cache-ontap-volumes/overview"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;NetApp FlexCache&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, a new capability for Google Cloud NetApp Volumes, provides high-performance, local read caches of remote volumes. This helps distributed teams access shared datasets as if they were local, and supports compute bursting for workloads that need low-latency data access, improving productivity and collaboration across your entire organization. FlexCache is available now in preview via an allowlist.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-aside"&gt;&lt;dl&gt;
    &lt;dt&gt;aside_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;title&amp;#x27;, &amp;#x27;Try Google Cloud for free&amp;#x27;), (&amp;#x27;body&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8ef9bca0&amp;gt;), (&amp;#x27;btn_text&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;href&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;image&amp;#x27;, None)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Bring your enterprise data to Gemini Enterprise&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;We’re also announcing that Google Cloud NetApp Volumes now serves as a data store for Gemini Enterprise. This integration unlocks new possibilities for retrieval-augmented generation (RAG), allowing you to ground your AI models on your own secure, factual, enterprise-grade data. Your data remains securely governed in NetApp Volumes and is quickly available for search and inference workflows, without the need for complex ETL or manual integrations.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Additional enhancements for your cloud environment&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Google Cloud NetApp Volumes has several other new capabilities to help you modernize your data estate:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;a href="https://cloud.google.com/netapp/volumes/docs/migrate/ontap/overview"&gt;&lt;strong style="text-decoration: underline; vertical-align: baseline;"&gt;NetApp SnapMirror&lt;/strong&gt;&lt;/a&gt;&lt;strong style="vertical-align: baseline;"&gt;:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; You can now quickly replicate mission-critical data between on-prem NetApp systems and Google Cloud, providing a zero recovery point objective (RPO) and near-zero recovery time objective (RTO).&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;High-performance for large volumes: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;For applications with massive datasets such as HPC, AI, and EDA, we now offer large-capacity volumes that scale from 15TiB to 3PiB, with over 21GiB/s of throughput per volume.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;a href="https://cloud.google.com/netapp/volumes/docs/configure-and-use/volumes/manage-auto-tiering"&gt;&lt;strong style="text-decoration: underline; vertical-align: baseline;"&gt;Auto-tiering&lt;/strong&gt;&lt;/a&gt;&lt;strong style="vertical-align: baseline;"&gt;:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; To help you manage costs, built-in &lt;/span&gt;&lt;a href="https://cloud.google.com/netapp/volumes/docs/configure-and-use/volumes/manage-auto-tiering"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;auto-tiering&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; dynamically moves infrequently accessed data to lower-cost storage, with cold data priced at just $0.03/GiB for the Flex service level. As a turnkey, integrated feature, auto-tiering is transparent to any application built on Google Cloud NetApp Volumes, and can support a tiering threshold of anywhere from 2-183 days, with dynamically adjustable policy support.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Get started&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Whether you’re migrating your enterprise SAN data, powering AI with Gemini Enterprise, or running high-throughput EDA workloads, Google Cloud NetApp Volumes can help you modernize your data estate. To learn more and get started, &lt;/span&gt;&lt;a href="https://cloud.google.com/netapp-volumes?e=48754805"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;explore the product documentation&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;</description><pubDate>Tue, 14 Oct 2025 16:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/storage-data-transfer/announcing-enhancements-to-google-cloud-netapp-volumes/</guid><category>Partners</category><category>Storage &amp; Data Transfer</category><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>Power your enterprise applications in the cloud with unified block and file storage</title><description></description><site_name>Google</site_name><url>https://cloud.google.com/blog/products/storage-data-transfer/announcing-enhancements-to-google-cloud-netapp-volumes/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Asad Khan</name><title>Sr. Director of Product Management, Google Cloud</title><department></department><company></company></author><author xmlns:author="http://www.w3.org/2005/Atom"><name>Brendan Power</name><title>Group Product Manager, Google Storage</title><department></department><company></company></author></item><item><title>The future of media sanitization at Google</title><link>https://cloud.google.com/blog/products/identity-security/the-future-of-media-sanitization-at-google/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;At Google, protecting your data is our most important responsibility, and we are committed to keeping your data safe. To further this commitment, we are proud to announce that starting in November 2025, we will start transitioning our approach to media sanitization to fully rely on a robust and layered encryption strategy.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;This marks a move away from the "brute force disk erase" process we have used for nearly two decades. While overwriting data has been an effective method, the storage technology landscape has changed dramatically. This process is no longer sustainable due to the size and technological complexity of today's modern media.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;A smarter approach: Cryptographic erasure&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;To address these challenges, we are embracing a more modern and efficient method of media sanitization: cryptographic erasure.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;By default, all user data in Google's services is protected by multiple layers of encryption. Cryptographic erasure leverages this encryption to sanitize media. Instead of overwriting the entire drive, we securely delete the cryptographic keys that are used to encrypt the data. Once the keys are gone, the data is rendered unreadable and unrecoverable.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;This method is not only faster but also aligns with industry best practices. The National Institute of Standards and Technology (NIST) recognizes cryptographic erasure as a valid sanitization technique in its special publication 800-88. We are committed to meeting and exceeding these standards to ensure the security of your data.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Enhancing security through innovation&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;We implement cryptographic erasure with multiple layers of security, employing a defense in depth strategy. Our trust-but-verify model uses independent verification mechanisms to ensure permanent deletion of media encryption keys.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;We also protect secrets involved in this process, like storage device keys, with industry-leading measures. Multiple key rotations enhance the security of customer data through independent layers of trusted encryption.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Sustainability and the circular economy&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Our previous method of media erasure had an environmental cost. Any storage device that failed our rigorous verification process was physically destroyed. This resulted in the destruction of a significant number of devices each year. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Cryptographic erasure allows us to move towards a more sustainable, circular economy. By eliminating the need to physically destroy drives, we can reuse more of our hardware. This also allows us to recover valuable rare earth materials, such as neodymium magnets, from end-of-life media. This innovative magnet recovery process is a major accomplishment in sustainable manufacturing, showcasing our commitment to responsible growth.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Our path forward&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;We have consistently been strong advocates for doing what is truly right for our users, the broader industry, and the world at large. &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;This transition to cryptographic erasure is a direct reflection of that commitment.&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; It allows us to enhance security, align with the highest industry standards, and build a more sustainable future for our infrastructure. We believe this is the right path forward for our users, the industry, and the environment.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;For more information about encryption at rest, including encryption key management, see our &lt;/span&gt;&lt;a href="https://cloud.google.com/docs/security/encryption/default-encryption#googles_default_encryption"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;default encryption at rest&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; security whitepaper.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;</description><pubDate>Mon, 13 Oct 2025 16:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/identity-security/the-future-of-media-sanitization-at-google/</guid><category>Infrastructure</category><category>Storage &amp; Data Transfer</category><category>Security &amp; Identity</category><media:content height="540" url="https://storage.googleapis.com/gweb-cloudblog-publish/images/The_future_of_media_sanitization_at_Google.max-600x600.jpg" width="540"></media:content><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>The future of media sanitization at Google</title><description></description><image>https://storage.googleapis.com/gweb-cloudblog-publish/images/The_future_of_media_sanitization_at_Google.max-600x600.jpg</image><site_name>Google</site_name><url>https://cloud.google.com/blog/products/identity-security/the-future-of-media-sanitization-at-google/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Paul B. Pescitelli</name><title>Director, Data Center Information Security</title><department></department><company></company></author></item><item><title>11 ways to reduce your Google Cloud compute costs today</title><link>https://cloud.google.com/blog/products/compute/cost-saving-strategies-when-migrating-to-google-cloud-compute/</link><description>&lt;div class="block-paragraph"&gt;&lt;p data-block-key="t3t6l"&gt;As the saying goes, "a penny saved is a penny earned," and this couldn't be more true when it comes to cloud infrastructure. In today's competitive business landscape, you need to maintain the performance to meet your business needs. Luckily, Google Cloud’s &lt;a href="https://cloud.google.com/products/compute"&gt;Compute Engine&lt;/a&gt; and block storage services offer numerous opportunities to reduce costs without sacrificing performance, especially in the context of your migration and modernization initiatives.&lt;/p&gt;&lt;p data-block-key="fodf8"&gt;In this article, we'll explore &lt;b&gt;11 key ways&lt;/b&gt; to optimize your infrastructure spending on Google Cloud, from simple adjustments to strategic decisions that can result in significant long-term savings.&lt;/p&gt;&lt;h3 data-block-key="58qqa"&gt;&lt;b&gt;1. Choose the right VM instances&lt;/b&gt;&lt;/h3&gt;&lt;p data-block-key="bhado"&gt;One of the most effective ways to reduce Compute Engine costs is to ensure that you’ve properly selected and right-sized your virtual machines (VMs) for their workloads to support your migration and modernization efforts. Whether you're new to Google Cloud or already using Compute Engine, adopting the latest-generation VMs — such as &lt;a href="https://cloud.google.com/compute/docs/general-purpose-machines#n4_series"&gt;N4&lt;/a&gt;, &lt;a href="https://cloud.google.com/compute/docs/general-purpose-machines#c4_series"&gt;C4&lt;/a&gt;, &lt;a href="https://cloud.google.com/compute/docs/general-purpose-machines#c4d_series"&gt;C4D&lt;/a&gt;, and &lt;a href="https://cloud.google.com/compute/docs/general-purpose-machines#c4a_series"&gt;C4A&lt;/a&gt; — can deliver substantial savings and improved price-performance.&lt;/p&gt;&lt;p data-block-key="8rrjo"&gt;Powered by Google Cloud’s &lt;a href="https://cloud.google.com/titanium?e=48754805&amp;amp;hl=en"&gt;Titanium&lt;/a&gt; architecture, our latest-generation VMs offer faster CPUs, higher memory bandwidth, and more efficient virtualization than their predecessors, so you can handle the same workloads with fewer resources. For existing customers, migrating from older VM generations to the newest VMs can significantly lower total costs while helping you exceed current performance levels. Organizations that have made the switch often report 20–40% better performance along with meaningful reductions in cloud compute spend. For example, &lt;a href="https://www.elastic.co/blog/elasticsearch-runs-faster-google-axion-processors" target="_blank"&gt;Elastic&lt;/a&gt; leveraged the general-purpose C4A machine series based on &lt;a href="https://cloud.google.com/blog/products/compute/introducing-googles-new-arm-based-cpu?e=48754805"&gt;Google Cloud's Arm-based Axion CPUs&lt;/a&gt;, to achieve a compelling efficiency and performance uplift for their workloads.&lt;/p&gt;&lt;p data-block-key="febac"&gt;Beyond &lt;a href="https://cloud.google.com/compute/docs/general-purpose-machines"&gt;general-purpose VMs&lt;/a&gt;, we also offer specialized machine types to address unique customer requirements. Compute-optimized HPC VMs like &lt;a href="https://cloud.google.com/blog/products/compute/new-h4d-vms-optimized-for-hpc?e=48754805"&gt;H4D&lt;/a&gt; are designed for high-performance computing and data analytics, offering extreme performance for demanding workloads. &lt;a href="https://cloud.google.com/compute/docs/memory-optimized-machines#m4_series"&gt;M4&lt;/a&gt; and &lt;a href="https://cloud.google.com/compute/docs/memory-optimized-machines#x4_series"&gt;X4&lt;/a&gt; instances cater to memory-intensive applications, while &lt;a href="https://cloud.google.com/compute/docs/storage-optimized-machines#z3_series"&gt;Z3&lt;/a&gt; instances are ideal for storage-intensive workloads. Furthermore, if you need complete control over your hardware environment and maximum performance isolation, we offer &lt;a href="https://cloud.google.com/compute/docs/instances/bare-metal-instances#:~:text=Bare%20metal%20instances%20provide%20direct,same%20way%20as%20VM%20instances."&gt;bare metal instances&lt;/a&gt;.&lt;/p&gt;&lt;p data-block-key="eqplt"&gt;These options help ensure that even the most specialized and performance-sensitive workloads can find an optimal and cost-effective home within the Compute Engine portfolio.&lt;/p&gt;&lt;h3 data-block-key="5i8hm"&gt;&lt;b&gt;2. Optimize your block storage selections&lt;/b&gt;&lt;/h3&gt;&lt;p data-block-key="bsaae"&gt;The best way to lower your block storage TCO, while ensuring your workloads remain successful, is to drive high resource efficiency. &lt;a href="https://cloud.google.com/compute/docs/disks/hyperdisks"&gt;Hyperdisk&lt;/a&gt; makes it simple to drive high performance and high efficiency by enabling you to optimize your block storage to your workload and through Storage Pools. We’ll discuss each of these capabilities, and how you can use them to lower your block storage TCO below.&lt;/p&gt;&lt;p data-block-key="6kmjp"&gt;Workload Optimized: With Hyperdisk, you can independently tune capacity and performance to match your block storage resources to your workload. Hyperdisk enables you to independently provision performance and capacity at the volume level. You can leverage this capability to purchase just the capacity and performance you need, no more and no less. You can also take advantage of Hyperdisk Balanced’s “baseline” performance (i.e. included free with every volume), you can serve the vast majority of your VMs without purchasing any extra performance.&lt;/p&gt;&lt;p data-block-key="at87k"&gt;Storage Pools: Hyperdisk is the only hyperscale cloud block storage to offer thin-provisioned performance and capacity. With Hyperdisk Storage Pools, you can provision the aggregate performance and capacity your workload requires, while still provisioning the volume level capacity performance your workloads request (also known as &lt;a href="https://en.wikipedia.org/wiki/Thin_provisioning" target="_blank"&gt;thin-provisioning&lt;/a&gt;). This allows you to pay for the resources you need, not the sum of the volumes you’ve provisioned. As a result, you can &lt;a href="https://cloud.google.com/blog/products/storage-data-transfer/hyperdisk-storage-pools-is-now-generally-available#:~:text=Infrastructure%20Manager%2C%20REWE-,Get%20started,use%20and%20manage%20your%20pools."&gt;lower your overall block storage TCO by as much as 50%.&lt;/a&gt;&lt;/p&gt;&lt;p data-block-key="929m4"&gt;For more information on how to select the right block storage for your workload and to see how customers have benefitted from Hyperdisk, read this &lt;a href="https://cloud.google.com/blog/products/storage-data-transfer/how-to-choose-the-right-hyperdisk-block-storage-for-your-use-case?e=48754805"&gt;blog&lt;/a&gt;.&lt;/p&gt;&lt;h3 data-block-key="e9sir"&gt;&lt;b&gt;3. Consider custom compute classes&lt;/b&gt;&lt;/h3&gt;&lt;p data-block-key="616a2"&gt;To get the most out of our latest-generation VMs, Google Kubernetes Engine (GKE) &lt;a href="https://cloud.google.com/kubernetes-engine/docs/concepts/about-custom-compute-classes"&gt;&lt;b&gt;custom compute classes&lt;/b&gt;&lt;/a&gt; (CCC) offer an advanced way to optimize compute choices and provide high availability. Instead of being limited to a single machine type for your workloads, you can define a prioritized list of VM instance types. This allows you to set the newest, most price-performant VMs — including our latest-generation VMs — as your top priority. GKE custom compute classes provide the capability to automatically and seamlessly spin up instances based on your specified priority list. This feature helps you maximize the availability of your compute capacity while still aiming for the most cost-effective options, so your workloads can scale reliably without manual intervention.&lt;/p&gt;&lt;p data-block-key="b9b8c"&gt;Here are some specific use cases for how custom compute classes can help you optimize costs:&lt;/p&gt;&lt;ul&gt;&lt;li data-block-key="b747a"&gt;&lt;b&gt;Autoscaling cost-performant fallbacks:&lt;/b&gt; When demand peaks, you might be tempted to autoscale using a highly available but less cost-efficient VM type. CCC allows you to take a tiered approach. You can set up several cost-efficient fallback alternatives, so that as demand increases, GKE first attempts to use the most cost-effective options, and progressively moves to the other choices in your list when necessary to meet demand.&lt;/li&gt;&lt;li data-block-key="smd6"&gt;&lt;b&gt;AI/ML inference:&lt;/b&gt; Running AI/ML inference workloads often involves significant compute resources. Instead of maintaining a large, static reservation that might sit idle during off-peak times, CCC lets you provision a minimal base reservation and leverage more cost-effective capacity types, such as &lt;a href="https://docs.google.com/document/d/1KLJ97-xgtX9pDaodkMsXN18xJOYiBFZUSqkdYb--_44/edit?tab=t.0" target="_blank"&gt;Spot VMs&lt;/a&gt;, to handle peak inference demand — all orchestrated through your CCC configuration.&lt;/li&gt;&lt;li data-block-key="4pm95"&gt;&lt;b&gt;Adopting new VM generations:&lt;/b&gt; Combine the power of GKE custom compute classes with &lt;a href="https://cloud.google.com/compute/docs/instances/committed-use-discounts-overview#spend_based"&gt;Compute Flexible committed use discounts&lt;/a&gt; (Flex CUDs) to de-risk the adoption of new, cost-efficient VM series like N4 and C4. With CCC, you can define fallback options, providing workload resilience, while Flex CUDs offer financial adaptability, as the discounts apply across your total eligible compute spend, regardless of the specific VM series you use. This dual approach is a safe, cost-effective strategy for leveraging the latest hardware without disruption. For more information, read this &lt;a href="https://cloud.google.com/blog/products/compute/adopt-new-vm-series-with-gke-compute-classes-flexible-cuds/?e=48754805"&gt;blog&lt;/a&gt;.&lt;/li&gt;&lt;li data-block-key="5fieq"&gt;&lt;b&gt;Using flexible Spot VMs:&lt;/b&gt; Spot VMs offer significant savings but can be preempted. Being constrained to a single Spot VM shape increases the risk that capacity will not be available. With CCC, you can define multiple fallback Spot VM types. This "spot surfing" capability allows the application to remain on cost-efficient Spot capacity by automatically pivoting to alternative Spot instance types if the primary choice is unavailable.&lt;/li&gt;&lt;/ul&gt;&lt;p data-block-key="9pj6j"&gt;In short, by leveraging GKE CCC, you can artfully mix and match various VM types and consumption models, including On-Demand, Spot, DWS FlexStart, and instances covered by CUDs, to build a resilient and highly cost-optimized infrastructure that adapts to the unique needs and patterns of your workloads.&lt;/p&gt;&lt;h3 data-block-key="4f7sf"&gt;&lt;b&gt;4. Leverage custom machine types (CMT)&lt;/b&gt;&lt;/h3&gt;&lt;p data-block-key="44on9"&gt;&lt;a href="https://cloud.google.com/compute/docs/instances/creating-instance-with-custom-machine-type"&gt;Custom machine types&lt;/a&gt;, available on N4 VMs, allow you to precisely configure virtual machines to your exact specifications. Rather than selecting from predefined machine types that might include excess capacity, you can tailor the CPU-to-memory ratio specifically for your workloads, so you only pay for resources you actually use. This targeted approach minimizes waste and can significantly reduce your cloud spend, especially when migrating from on-premises to Google Cloud or from other cloud providers.&lt;/p&gt;&lt;p data-block-key="ditbj"&gt;This flexibility becomes particularly valuable if your applications have unique resource profiles that don't align well with our standard offerings. Custom machine types let you create the perfect environment for your needs. By avoiding the compromise of over-provisioning certain resources while potentially constraining others, you can achieve both better performance and more efficient spending across your Compute Engine deployment.&lt;/p&gt;&lt;p data-block-key="16cdr"&gt;As an example, take a memory-intensive workload that runs best with 16 vCPU, and 70 GB memory. Normally, you would need to pick a VM with 128 GB memory with our standard shapes, or in other cloud contexts, resulting in higher costs to run your workload due to the extra provisioned resources. Instead, with custom machine types, you can easily launch a VM with 16 vCPU and 70 GB memory, resulting in an 18% cost savings vs standard N4-highmem-16 VMs.&lt;/p&gt;&lt;h3 data-block-key="ei6g2"&gt;&lt;b&gt;5. Make the most of committed use discounts&lt;/b&gt;&lt;/h3&gt;&lt;p data-block-key="3vd4p"&gt;CUDs are a strategic cost-saving opportunity for organizations with steady, predictable computing needs. By committing to resource usage over one- or three-year periods, you can reduce cloud costs by up to 70% compared to on-demand pricing. This approach not only helps ensure budget predictability but also converts fixed infrastructure spending into a financial advantage, making it ideal for stable workloads that support core business functions.&lt;/p&gt;&lt;p data-block-key="bnjpk"&gt;Google Cloud offers flexible CUD structures to align with various operational models. Resource-based commitments target specific machine types and regions, flexible commitments apply discounts across projects, regions, and machine series — great for dynamic environments. By analyzing historical usage and forecasting future needs, you can identify workloads suited for these discounts, reinvesting the savings into innovation and scaling initiatives.&lt;/p&gt;&lt;h3 data-block-key="einu0"&gt;&lt;b&gt;6. Manage unused disk space&lt;/b&gt;&lt;/h3&gt;&lt;p data-block-key="5n36i"&gt;You pay for the total provisioned disk space, regardless of how much you actually use. Many organizations tend to over-provision storage "just in case," which often leads to unnecessary and costly waste. For instance, if you provision a 100GB disk but only use 20GB, you're still paying for the entire 100GB. Being intentional and precise with your storage allocations — rather than rounding up to common sizes — can lead to significant cost savings.&lt;/p&gt;&lt;p data-block-key="d58ss"&gt;To optimize spending, it's important to adopt a few best practices. Using &lt;a href="https://cloud.google.com/stackdriver/docs/solutions/agents/ops-agent"&gt;Ops Agent&lt;/a&gt;, regularly audit disk usage across your infrastructure to identify and eliminate inefficiencies. Resize disks to align with actual consumption, allowing a reasonable buffer for growth. Implement automated alerts in &lt;a href="https://cloud.google.com/monitoring?e=48754805&amp;amp;hl=en"&gt;Cloud Monitoring&lt;/a&gt; to detect underutilized disks and take corrective action. For stateless applications, consider using smaller boot disk images to minimize overhead and reduce costs even further.&lt;/p&gt;&lt;p data-block-key="4hu1t"&gt;In addition, consider the following optimization strategies to further reduce costs and improve efficiency:&lt;/p&gt;&lt;ul&gt;&lt;li data-block-key="as6ik"&gt;Use Google Cloud’s monitoring tools to track CPU, memory, and disk usage over time.&lt;/li&gt;&lt;li data-block-key="3uj3j"&gt;Establish a regular review cycle to identify and right-size over-provisioned resources.&lt;/li&gt;&lt;li data-block-key="7jobf"&gt;Test workloads across different VM configurations to find the optimal balance between cost and performance.&lt;/li&gt;&lt;/ul&gt;&lt;h3 data-block-key="a49bh"&gt;&lt;b&gt;7. Use Spot VMs&lt;/b&gt;&lt;/h3&gt;&lt;p data-block-key="52fp7"&gt;&lt;a href="https://cloud.google.com/compute/docs/instances/spot"&gt;Spot VMs&lt;/a&gt; provide the same machine types and configuration​​ options as standard virtual machines but at a significantly reduced cost — typically offering a 60% to 91% discount. This cost efficiency comes with the tradeoff of potential preemption at short notice, making them most suitable for workloads that are fault-tolerant and can recover quickly from unexpected interruptions. Spot VMs are designed to take advantage of unused compute capacity, allowing you to optimize your cloud spending without compromising access to high-performance resources.&lt;/p&gt;&lt;p data-block-key="8abo0"&gt;Strong use cases for Spot VMs include batch processing jobs, big data and analytics workloads, continuous integration and deployment (CI/CD) pipelines, stateless web servers running in autoscaling groups, and compute-heavy tasks. When properly architected to handle interruptions — for example, by using job checkpointing, load balancing, task queues, or via GKE custom compute classes (see more above) — &lt;a href="https://cloud.google.com/solutions/spot-vms?e=48754805&amp;amp;hl=en"&gt;Spot VMs&lt;/a&gt; can play a critical role in minimizing infrastructure costs while maintaining high availability and system resilience. Leveraging Spot VMs in these scenarios lets you scale cost-effectively, especially when compute demand is variable or time-flexible.&lt;/p&gt;&lt;h3 data-block-key="9not2"&gt;&lt;b&gt;8. Use optimization recommendations&lt;/b&gt;&lt;/h3&gt;&lt;p data-block-key="bjf0l"&gt;Google Cloud's &lt;a href="https://cloud.google.com/recommender/docs/recommenders"&gt;Recommenders&lt;/a&gt; are a powerful tool designed to help you optimize your cloud resources efficiently. When browsing the Google Cloud console, you may see lightbulb icons next to specific resources — these indicate potential improvements identified by Google's recommendation engine. By analyzing real-time usage patterns and current resource configurations, the &lt;a href="https://cloud.google.com/recommender/docs/key-concepts#recommenders"&gt;Recommender&lt;/a&gt; delivers actionable insights tailored to each user's unique environment. This intelligent system highlights opportunities not only to reduce costs but also to enhance security, performance, reliability, management efficiency, and environmental sustainability.&lt;/p&gt;&lt;p data-block-key="91nm7"&gt;For example, there are &lt;a href="https://cloud.google.com/compute/docs/instances/idle-vm-recommendations-overview"&gt;idle VM recommendations&lt;/a&gt; to help you identify VM instances that have not been used over the last 1 to 14 days. Common recommendations include switching to more suitable machine types, rightsizing underutilized compute instances, or adopting more cost-effective storage solutions. The tool allows you to apply many of these changes directly, streamlining the optimization process. By continuously evaluating workloads and offering these automated, data-driven suggestions, the Recommendation Hub helps organizations maintain cloud performance while managing costs more effectively.&lt;/p&gt;&lt;h3 data-block-key="35ft8"&gt;&lt;b&gt;9. Take advantage of auto-scaling and scheduling&lt;/b&gt;&lt;/h3&gt;&lt;p data-block-key="5i72v"&gt;Matching your compute resources to actual demand patterns is one of the most effective ways to reduce cloud waste and improve overall cost efficiency. Many organizations over-provision their resources to handle peak workloads, leaving machines underutilized during off-peak periods. By aligning compute capacity more closely with real-time or predictable usage patterns, such as business hours or seasonal trends, you can significantly cut unnecessary spending without sacrificing performance.&lt;/p&gt;&lt;p data-block-key="c7ge7"&gt;&lt;a href="https://cloud.google.com/compute/docs/autoscaler"&gt;Autoscaling&lt;/a&gt; is the key to achieving this efficiency. In fact, customers who leverage Google Compute Engine autoscaling for their virtual machines have seen average infrastructure cost savings of more than 40%.&lt;/p&gt;&lt;p data-block-key="70opn"&gt;You can implement autoscaling strategies to dynamically adjust resources based on CPU utilization, load balancing capacity, or custom application metrics, so that workloads receive the necessary compute power when needed, while scaling down automatically during low-demand periods.&lt;/p&gt;&lt;p data-block-key="cj1lq"&gt;For workloads with predictable patterns, such as those that fluctuate with business hours or planned seasonal events, &lt;a href="https://cloud.google.com/compute/docs/autoscaler/scaling-schedules"&gt;schedule-based scaling&lt;/a&gt; is a particularly powerful tool. This approach allows you to proactively increase resources in anticipation of high demand and scale them down during lulls, for the performance you need without constant over-provisioning.&lt;/p&gt;&lt;p data-block-key="1i6kf"&gt;In addition to autoscaling, several practical implementation techniques can further optimize your resource usage. &lt;a href="https://cloud.google.com/scheduler/docs/start-and-stop-compute-engine-instances-on-a-schedule"&gt;Setting up instance scheduling&lt;/a&gt; lets you automatically start and stop development and test environments according to business hours — a simple yet highly effective approach that can lead to cost savings of up to 70%. You can also leverage maintenance windows to reduce disruptions and resource consumption, by concentrating updates and system changes into low-usage periods. Together, these tactics help maintain high availability and performance while keeping infrastructure costs under control.&lt;/p&gt;&lt;h3 data-block-key="evivu"&gt;&lt;b&gt;10. Understand your spend with detailed billing analysis&lt;/b&gt;&lt;/h3&gt;&lt;p data-block-key="34tqj"&gt;Before implementing any cost-saving strategies in Google Cloud, it’s essential to &lt;a href="https://cloud.google.com/billing/docs/concepts"&gt;understand your current spending in detail&lt;/a&gt;. Google Cloud’s billing panel offers granular visibility into your expenses, including costs broken down by individual SKUs. This level of transparency lets you track where your money is going and identify potential inefficiencies. Begin by regularly reviewing your billing dashboard to monitor usage trends and spot anomalies. Applying labels and tags to your resources can further help categorize and attribute costs accurately, especially in complex environments with multiple projects or departments.&lt;/p&gt;&lt;p data-block-key="een0q"&gt;In addition, &lt;a href="https://cloud.google.com/billing/docs/how-to/budgets"&gt;setting up budget alerts&lt;/a&gt; is a practical way to stay ahead of overspending by notifying you when costs approach or exceed predefined thresholds. It’s also important to identify and eliminate unused or idle resources, such as virtual machines or persistent disks that are no longer in active use — these can often be shut down or deleted to immediately reduce costs. By thoroughly analyzing your cost structure, you can uncover “low-hanging fruit” — resources that provide little or no value — and make data-driven decisions to optimize your cloud usage efficiently.&lt;/p&gt;&lt;h3 data-block-key="9jfbk"&gt;&lt;b&gt;11. Consider serverless alternatives&lt;/b&gt;&lt;/h3&gt;&lt;p data-block-key="e7aev"&gt;Last but not least, Google Cloud's &lt;a href="https://cloud.google.com/discover/what-is-serverless-computing?e=48754805&amp;amp;hl=en"&gt;serverless computing&lt;/a&gt; offerings provide a compelling alternative to traditional virtual machines, can deliver better cost efficiency, simplified operations, and greater scalability. By abstracting away infrastructure management, serverless platforms allow teams to focus on writing and deploying code without worrying about provisioning, scaling, or maintaining servers. This shift can not only reduce operational overhead but also cut costs by aligning compute spending directly with application usage.&lt;/p&gt;&lt;p data-block-key="4c30g"&gt;There are multiple serverless options available, each tailored to different workloads. &lt;a href="https://cloud.google.com/run?e=48754805&amp;amp;hl=en"&gt;Cloud Run&lt;/a&gt; is designed for running containerized applications that need rapid scaling and flexible deployment. &lt;a href="https://cloud.google.com/run/docs/write-event-driven-functions"&gt;Cloud Run Functions&lt;/a&gt; supports lightweight, event-driven code execution for microservices or automation tasks. &lt;a href="https://cloud.google.com/kubernetes-engine/docs/concepts/autopilot-overview"&gt;GKE (Autopilot Mode)&lt;/a&gt; simplifies Kubernetes operations by automatically managing nodes and scaling, allowing you to run Kubernetes workloads without handling the underlying infrastructure. All these options charge based on usage not allocation, significantly reducing costs associated with idle resources and over-provisioning. This makes them especially beneficial for variable or unpredictable workloads. Cloud Run and GKE both support GPU’s and flexibility to move between the two. You can start with &lt;a href="https://www.youtube.com/watch?v=nGFXKTz2jZM&amp;amp;t=2s&amp;amp;pp=ygUabW92ZSBmcm9tIGNsb3VkIHJ1biB0byBHS0U%3D" target="_blank"&gt;Cloud Run then move to GKE&lt;/a&gt; or &lt;a href="https://www.youtube.com/watch?v=x12EOsVt2oU&amp;amp;t=1s&amp;amp;pp=ygUabW92ZSBmcm9tIGNsb3VkIHJ1biB0byBHS0U%3D" target="_blank"&gt;vice-versa&lt;/a&gt;. Some customers also leverage both offerings for workloads. The rule of thumb is to start with GKE if you need access to the Kubernetes API. Otherwise, start with Cloud Run.&lt;/p&gt;&lt;h2 data-block-key="6n8fn"&gt;&lt;b&gt;Start reducing your costs today&lt;/b&gt;&lt;/h2&gt;&lt;p data-block-key="buet9"&gt;Migrate to Google Cloud and optimize your infrastructure costs without compromising on what your workloads need. If you are new to Google Cloud, start with &lt;a href="http://g.co/cloud/assess" target="_blank"&gt;a migration assessment&lt;/a&gt;. Google Cloud’s &lt;a href="https://cloud.google.com/migration-center/docs"&gt;Migration Center&lt;/a&gt; can help you with a clear understanding of your potential savings by migrating to Google Cloud, with detailed recommended paths for your workloads, along with TCO reports. Apply the strategies in this article and unlock substantial cost savings.&lt;/p&gt;&lt;/div&gt;</description><pubDate>Mon, 06 Oct 2025 16:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/compute/cost-saving-strategies-when-migrating-to-google-cloud-compute/</guid><category>Infrastructure Modernization</category><category>Storage &amp; Data Transfer</category><category>Serverless</category><category>Compute</category><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>11 ways to reduce your Google Cloud compute costs today</title><description></description><site_name>Google</site_name><url>https://cloud.google.com/blog/products/compute/cost-saving-strategies-when-migrating-to-google-cloud-compute/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Alex Bestavros</name><title>Group Product Manager</title><department></department><company></company></author><author xmlns:author="http://www.w3.org/2005/Atom"><name>Sai Gopalan</name><title>Product Management, Google Cloud</title><department></department><company></company></author></item><item><title>5 best practices for Managed Lustre on Google Kubernetes Engine</title><link>https://cloud.google.com/blog/products/containers-kubernetes/gke-managed-lustre-csi-driver-for-aiml-and-hpc-workloads/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Google Kubernetes Engine (GKE) is a powerful platform for orchestrating scalable AI and high-performance computing (HPC) workloads. But as clusters grow and jobs become more data-intensive, storage I/O can become a bottleneck. Your powerful GPUs and TPUs can end up idle, while waiting for data, driving up costs and slowing down innovation.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;a href="http://goo.gle/managed-lustre-overview" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Google Cloud Managed Lustre&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; is designed to solve this problem. Many on-premises HPC environments already use parallel file systems, and Managed Lustre makes it easier to bring those workloads to the cloud. With its managed Container Storage Interface (CSI) driver, Managed Lustre and GKE operations are fully integrated.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Optimizing your move to a high-performance parallel file system can help you get the most out of your investment from day one. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Before deploying, it's helpful to know when to use Managed Lustre versus other options like Google Cloud Storage. For most AI and ML workloads, Managed Lustre is the recommended solution. It excels in training and checkpointing scenarios that require very low latency (less than a millisecond) and high throughput for small files, which keeps your expensive accelerators fully utilized. For data archiving or workloads with large files (over 50 MB) that can tolerate higher latency, Cloud Storage FUSE with Anywhere Cache can be another choice.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Based on our work with early customers and the learnings from our teams, here are five best practices to ensure you get the most out of Managed Lustre on GKE.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-aside"&gt;&lt;dl&gt;
    &lt;dt&gt;aside_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;title&amp;#x27;, &amp;#x27;$300 in free credit to try Google Cloud containers and Kubernetes&amp;#x27;), (&amp;#x27;body&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8eef0be0&amp;gt;), (&amp;#x27;btn_text&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;href&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;image&amp;#x27;, None)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;1. Design for data locality &lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;For performance-sensitive applications, you want your compute resources and storage to be as close as possible, ideally within the same zone in a given region. When provisioning volumes dynamically, the &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;volumeBindingMode&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt; parameter in your &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;StorageClass&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt; is your most important tool. We strongly recommend setting it to &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;WaitForFirstConsumer&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt;. GKE provides a built-in StorageClass for Managed Lustre that uses &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;WaitForFirstConsumer&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt; binding mode by default.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Generated yaml:&lt;/strong&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;#x27;apiVersion: storage.k8s.io/v1\r\nkind: StorageClass\r\nmetadata:\r\n  name: lustre-regional-wait\r\nprovisioner: lustre.csi.storage.gke.io\r\nvolumeBindingMode: WaitForFirstConsumer\r\n...&amp;#x27;), (&amp;#x27;language&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8eef00a0&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Why it’s a best practice:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Using &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;WaitForFirstConsumer&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt; instructs GKE to delay &lt;/span&gt;&lt;span style="text-decoration: line-through; vertical-align: baseline;"&gt;the&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; provisioning &lt;/span&gt;&lt;span style="text-decoration: line-through; vertical-align: baseline;"&gt;of&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; the Lustre instance until a pod that needs it is scheduled. The scheduler then uses the pod's topology constraints (i.e., the zone it's scheduled in) to create the Lustre instance in that exact same zone. This guarantees co-location of your storage and compute, minimizing network latency.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;2. Right-size your performance with tiers&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Not all high-performance workloads are the same. Managed Lustre offers multiple &lt;/span&gt;&lt;a href="https://cloud.google.com/managed-lustre/docs/performance"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;performance tiers&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; (read and write throughput in MB/s per TiB of storage) so you can align cost directly with your performance requirements.&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;1000 &amp;amp; 500 MB/s/TiB:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Ideal for throughput-critical workloads like foundation model training or large-scale physics simulations where I/O bandwidth is the primary bottleneck.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;250 MB/s/TiB:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; A balanced, cost-effective tier great for many general HPC workloads and AI inference serving, and data-heavy analytics pipelines.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;125 MB/s/TiB:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Best for large-capacity use cases where having a massive, POSIX-compliant file system is more important than achieving peak throughput. This is also useful for migrating on-premises containerized applications without modification,&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;making it easier to migrate on-premises workloads to the cloud storage.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/image1_JuBQFJn.max-1000x1000.png"
        
          alt="image1"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Why it’s a best practice: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Defaulting to the highest tier isn't always the most cost-effective strategy. By analyzing your workload’s I/O profile, you can significantly optimize your total cost of ownership. &lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;3. Master your networking foundation&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;A parallel file system is a network-attached resource. Getting the networking right up front will save you days of troubleshooting. Before provisioning, ensure your VPC is correctly configured by following the setup steps in our &lt;/span&gt;&lt;a href="https://cloud.google.com/managed-lustre/docs/vpc#create_and_configure_the_vpc"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;documentation&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;This involves three key steps detailed in our documentation:&lt;/span&gt;&lt;/p&gt;
&lt;ol&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Enable Service Networking.&lt;/strong&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Create an IP range&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; for VPC peering.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Create a firewall rule&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; to allow traffic from that range on the Lustre network port (TCP 988 or 6988).&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Why it’s a best practice:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; This is a one-time setup per VPC that establishes the secure peering connection that allows your GKE nodes to communicate with the Managed Lustre service. &lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;4. Use dynamic provisioning for simplicity, static for long-lived shared data&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The Managed Lustre CSI driver supports &lt;/span&gt;&lt;a href="https://cloud.google.com/kubernetes-engine/docs/how-to/persistent-volumes/lustre-csi-driver-new-volume"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;two modes&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; for connecting storage to your GKE workloads.&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Dynamic provisioning:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Use when your storage is tightly coupled to the lifecycle of a specific workload or application. By defining a StorageClass and PersistentVolumeClaim (PVC), GKE will automatically manage the Lustre instance lifecycle for you. This is the simplest, most automated approach.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Static provisioning:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Use when you have a long-lived Lustre instance that needs to be shared across multiple GKE clusters and jobs. You create the Lustre instance once, then create a PersistentVolume (PV) and PVC in your cluster to mount it. This decouples the storage lifecycle from any single workload.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Why it’s a best practice:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Thinking about your data’s lifecycle helps you choose the right pattern. Use dynamic provisioning as your default because of simplicity, and opt for static provisioning when you need to treat your file system as a persistent, shared resource across your organization.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;5. &lt;/strong&gt;&lt;strong style="vertical-align: baseline;"&gt;Architecting for parallelism with Kubernetes Jobs&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Many AI and HPC tasks, like data preprocessing or batch inference, are suited for parallel execution. Instead of running a single, large pod, use the Kubernetes Job resource to divide the work across many smaller pods.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Consider this pattern:&lt;/span&gt;&lt;/p&gt;
&lt;ol&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Create a single PersistentVolumeClaim for your Managed Lustre instance, making it available to your cluster.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Define a Kubernetes job with parallelism set to a high number (e.g., 100).&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Each pod created by the Job mounts the same Lustre PVC.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Design your application so that each pod works on a different subset of the data (e.g., processing a different range of files or data chunks).&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Why it’s a best practice: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;In this pattern, you create a single PVC for your Lustre instance and have each pod created by the Job mount that same PVC. By designing your application so that each pod works on a different subset of the data, you turn your GKE cluster into a powerful, distributed data processing engine. The GKE Job controller acts as the parallel task orchestrator, while Managed Lustre serves as the high-speed data backbone, allowing you to achieve massive aggregate throughput.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Get started today&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;By combining the orchestration power of GKE with the performance of Managed Lustre, you can build a truly scalable and efficient platform for AI and HPC. Following these best practices will help you create a solution that is not only powerful, but also efficient, cost-effective, and easy to manage.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Ready to get started? Explore the &lt;/span&gt;&lt;a href="https://cloud.google.com/managed-lustre/docs/overview"&gt;&lt;span style="vertical-align: baseline;"&gt;Managed Lustre documentation&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, and provision your first instance today.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;</description><pubDate>Fri, 19 Sep 2025 16:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/containers-kubernetes/gke-managed-lustre-csi-driver-for-aiml-and-hpc-workloads/</guid><category>Storage &amp; Data Transfer</category><category>HPC</category><category>Containers &amp; Kubernetes</category><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>5 best practices for Managed Lustre on Google Kubernetes Engine</title><description></description><site_name>Google</site_name><url>https://cloud.google.com/blog/products/containers-kubernetes/gke-managed-lustre-csi-driver-for-aiml-and-hpc-workloads/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Nishtha Jain</name><title>Engineering Manager</title><department></department><company></company></author><author xmlns:author="http://www.w3.org/2005/Atom"><name>Dan Eawaz</name><title>Senior Product Manager</title><department></department><company></company></author></item><item><title>Storage Insights datasets: How to optimize storage spend with deep visibility</title><link>https://cloud.google.com/blog/products/storage-data-transfer/storage-insights-datasets-optimizes-storage-footprint/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Managing vast amounts of data in cloud storage can be a challenge. While Google Cloud Storage offers strong scalability and durability, storage admins sometimes sometimes struggle with questions like: &lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;What’s driving my storage spend? &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Where is all my data in Cloud Storage and how is it distributed?&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;How can I search across my data for specific metadata such as age or size?&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span&gt;&lt;span style="vertical-align: baseline;"&gt;Indeed, to achieve cost optimization, security, and compliance, you need to understand what you have, where it is, and how it's being used. That's where &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/insights/datasets"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Storage Insights datasets&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, a feature of &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/storage-intelligence/overview"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Storage Intelligence&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; for Cloud Storage, comes in. Storage Intelligence is a unified management product that offers multiple powerful capabilities to analyze large storage estates and easily take actions. It helps you explore your data, optimize costs, enforce security, and implement governance policies. Storage insights datasets help you deeply analyze your storage footprint and you can use &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/analyze-data-gemini-cloud-assist"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Gemini Cloud Assist &lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;for quick analysis in natural language. Based on these analyses, you can take action, such as &lt;/span&gt;&lt;a href="https://cloud.google.com/blog/products/storage-data-transfer/storage-insights-datasets-optimizes-storage-footprint"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;relocating buckets&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; and performing large-scale &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/batch-operations/overview"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;batch operations&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;In this blog, we focus on how you can use Insights datasets for &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;cost management and visibility&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;, exploring a variety of common use cases. This is especially useful for cloud administrators and FinOps teams performing cloud cost allocation, monitoring and forecasting.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;What are Storage Insights datasets?&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Storage &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;Insights datasets provide a powerful, automated way to gain deep visibility into your Cloud Storage data. Instead of manual scripts, custom one-off reports for buckets or managing your own collection pipelines, Storage Insights datasets generate comprehensive reports about your Cloud Storage objects and their activities, placing them directly in a BigQuery linked dataset.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Think of it as X-ray vision for your Cloud Storage buckets. It transforms raw storage metadata into structured, queryable data that you can analyze with familiar BigQuery tools to gain crucial insights, with automatic data refreshes delivered every 24hrs (after the initial set up, which could take up to 48hrs for the first load).&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Key features&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Customizable scope: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Set the dataset scope to be at the level of the organization, a folder containing projects, a project / set of projects, or a specific bucket. &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Metadata dataset: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;It provides a queryable dataset that contains bucket and object metadata directly in BigQuery.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Regular updates and retention: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;After the first load, datasets update with metadata every 24 hours and can retain data for up to 90 days.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;&lt;/div&gt;
&lt;div class="block-aside"&gt;&lt;dl&gt;
    &lt;dt&gt;aside_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;title&amp;#x27;, &amp;#x27;Try Google Cloud for free&amp;#x27;), (&amp;#x27;body&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8ef78730&amp;gt;), (&amp;#x27;btn_text&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;href&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;image&amp;#x27;, None)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Use cases&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Calculate routine showback&lt;br/&gt;&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Understanding who/which applications are consuming what storage is often the first step in effective cost management, especially for larger organizations. With Storage Insights datasets, your &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/insights/datasets#data-schema"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;object and bucket metadata&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; is available in BigQuery. You can run SQL queries to aggregate storage consumption by specific teams, projects, or applications. You can then attribute storage consumption by buckets or prefixes for internal chargeback or cost attribution, for example: "Department X used 50TB of storage in &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;gs://my-app-data/department-x/&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt; last month”. This transparency fosters accountability and enables accurate internal showback. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Here’s an example SQL query to determine the total storage per bucket and prefix in the dataset:&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;quot;SELECT\r\n  bucket,\r\n  SPLIT(name, &amp;#x27;/&amp;#x27;)[\r\nOFFSET\r\n  (0)] AS top_level_prefix,\r\n  SUM(size) AS total_size_bytes\r\nFROM\r\n object_attributes_view\r\nGROUP BY\r\n  bucket, top_level_prefix\r\nORDER BY\r\n  total_size_bytes DESC;\r\n//Running queries in Datasets accrue BQ query costs, refer to the pricing details page for further details.&amp;quot;), (&amp;#x27;language&amp;#x27;, &amp;#x27;lang-sql&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8efadac0&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Understand how much data you have across storage classes&lt;br/&gt;&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Storage Insights datasets identifies the storage class for every object in your buckets. By querying &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;storageClass, timeCreated, updated &lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt;in the object metadata view in BigQuery, you can quickly visualize your data distribution across various classes (standard, nearline, coldline, archive) for objects beyond a certain age, as well as when they were last updated. This lets you identify potentially misclassified data. It also provides valuable insights into whether you have entire buckets with coldline or archived data or if your objects unexpectedly moved across storage classes (for example, a file expected to be in archive is now in standard class) using the &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;timeStorageClassUpdated&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt; object metadata.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Here’s an example SQL query to see all objects created two years ago, without any updates since and in standard class:&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;quot;SELECT\r\n bucket,\r\n name,\r\n size,\r\n storageClass,\r\n timeCreated,\r\n updated\r\nFROM object_attributes_latest_snapshot_view\r\nWHERE\r\n EXTRACT(YEAR\r\n FROM\r\n   timeCreated) = EXTRACT(YEAR\r\n FROM\r\n   DATE_SUB(CURRENT_DATE(), INTERVAL 24 MONTH))\r\n AND (updated IS NULL\r\n   OR updated = timeCreated)\r\n AND storageClass = &amp;#x27;STANDARD&amp;#x27;\r\nORDER BY\r\n timeCreated;\r\n//Running queries in Datasets accrue BQ query costs, refer to the pricing details page for further details.&amp;quot;), (&amp;#x27;language&amp;#x27;, &amp;#x27;lang-sql&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8efad4c0&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Set lifecycle and autoclass policies: Automating your savings&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Manual data management is time-consuming and prone to error. Storage Insights datasets helps you identify where the use of &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/lifecycle"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Object Lifecycle Management (OLM)&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; or &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/autoclass"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Autoclass&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; might reduce costs.&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Locate the buckets that don’t have OLM or Autoclass configured:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;  Through Storage Insights datasets, you can query bucket metadata to see which buckets lack defined lifecycle policies by using the field &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;lifecycle, autoclass.enabled&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt;. If a bucket contains data that should naturally transition to colder storage or be deleted after a certain period, but has no policy, you can take the appropriate action by knowing which parts of your estate you need to investigate further. Storage Insights datasets provides the data to flag these "unmanaged" buckets, helping you enforce best practices.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Here’s an example SQL query to see all buckets with lifecycle or autoclass configurations enabled and all those without any active configuration:&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;quot;SELECT\r\n name AS bucket_name,\r\n storageClass AS default_class,\r\n CASE\r\n   WHEN lifecycle = TRUE OR autoclass.enabled = TRUE THEN &amp;#x27;Managed&amp;#x27;\r\n   ELSE &amp;#x27;Unmanaged&amp;#x27;\r\nEND\r\n AS lifecycle_autoclass_status\r\nFROM bucket_attributes_latest_snapshot_view\r\n//Running queries in Datasets accrue BQ query costs, refer to the pricing details page for further details.&amp;quot;), (&amp;#x27;language&amp;#x27;, &amp;#x27;lang-sql&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8eee5fd0&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;ul&gt;
&lt;li role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Evaluate Autoclass impact:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Autoclass automatically transitions objects between storage classes based on a &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/autoclass#transitions"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;fixed access timeline&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. But how do you know if it's working as expected or if further optimization is needed? With Storage Insights datasets, you can find the buckets with autoclass enabled using the &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;autoclass.enabled&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt; field and analyze object metadata by tracking the &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;storageClass, timeStorageClassUpdated &lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt;field over time for&lt;/span&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt; specific objects &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;within Autoclass-enabled buckets. This allows you to evaluate the effectiveness of Autoclass, verify if the objects specified are indeed moving to optimal classes, and understand the real-world impact on your costs. For example, once you configure Autoclass on a bucket, you can visualize the movement of your data between storage classes on Day 31 as compared to Day 1 and understand how autoclass policies take effect on your bucket. &lt;/span&gt;&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Evaluate Autoclass suitability:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Analyze your bucket’s data to determine if it’s appropriate to use Autoclass with it. For example, if you have short-lived data (less than 30 days old) in a bucket (you can assess objects in daily snapshots to determine the average life of an object in your bucket using &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;timeCreated &lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt;and&lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt; timeDeleted&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt;), you may not want to turn on Autoclass.  &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Here’s an example SQL query to find a count of all objects with age more than 30 days and age less than 30 days in bucketA and bucketB:&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;quot;SELECT\r\n SUM(\r\n   CASE\r\n     WHEN TIMESTAMP_DIFF(t1.timeDeleted, t1.timeCreated, DAY) &amp;lt; 30 THEN 1\r\n     ELSE 0\r\n END\r\n   ) AS age_less_than_30_days,\r\n SUM(\r\n   CASE\r\n     WHEN TIMESTAMP_DIFF(t1.timeDeleted, t1.timeCreated, DAY) &amp;gt; 30 THEN 1\r\n     ELSE 0\r\n END\r\n   ) AS age_more_than_30_days\r\nFROM\r\n `object_attributes_view` AS t1\r\nWHERE\r\n t1.bucket IN ( &amp;#x27;bucketA&amp;#x27;, &amp;#x27;bucketB&amp;#x27;)\r\n AND t1.timeCreated IS NOT NULL\r\n AND t1.timeDeleted IS NOT NULL;\r\n//Running queries in Datasets accrue BQ query costs, refer to the pricing details page for further details.&amp;quot;), (&amp;#x27;language&amp;#x27;, &amp;#x27;lang-sql&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8eee5f10&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Proactive cleanup and optimization&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Beyond routine management, Storage Insights datasets can help you proactively find and eliminate wasted storage.&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Quickly find duplicate objects:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Accidental duplicates are a common cause of wasted storage. You can use object metadata like &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;size&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt;, &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;name&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; or even &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;crc32c&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt; checksums in your BigQuery queries to identify potential duplicates. For example, finding multiple objects with the exact same size, checksum and similar names might indicate redundancy, prompting further investigation.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Here’s an example SQL query to list all objects where their size, &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/json_api/v1/objects#resource-representations"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;crc32c checksum field&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; and name are the same values (indicating potential duplicates):&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;#x27;SELECT\r\n name,\r\n bucket,\r\n timeCreated,\r\n crc32c,\r\n size\r\nFROM (\r\n SELECT\r\n   name,\r\n   bucket,\r\n   timeCreated,\r\n   crc32c,\r\n   size,\r\n   COUNT(*) OVER (PARTITION BY name, size, crc32c) AS duplicate_count\r\n FROM\r\n   `object_attributes_latest_snapshot_view` )\r\nWHERE\r\n duplicate_count &amp;gt; 1\r\nORDER BY\r\nsize DESC;\r\n//Running queries in Datasets accrue BQ query costs, refer to the pricing details page for further details.&amp;#x27;), (&amp;#x27;language&amp;#x27;, &amp;#x27;lang-sql&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8eee59d0&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Find temporary objects to be cleaned up:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Many applications generate temporary files that, if not deleted, accumulate over time. Storage Insights datasets allows you to query for objects matching specific naming conventions (e.g., &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;*_temp&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt;, &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;*.tmp&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt;), or located in "temp" prefixes, along with their creation dates. This enables you to systematically identify and clean up orphaned temporary data, freeing up valuable storage space.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Here’s an example SQL query to find all log files created a month ago:&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;#x27;SELECT\r\n    name, bucket, timeCreated, size\r\n  FROM\r\n    \&amp;#x27;object_attributes_latest_snapshot_view\&amp;#x27;\r\n  WHERE\r\n   name LIKE &amp;quot;%.log&amp;quot;\r\nAND DATE(timeCreated) &amp;lt;= DATE_SUB(CURRENT_DATE(), INTERVAL 1 MONTH)\r\nORDER BY\r\nsize DESC;\r\n//Running queries in Datasets accrue BQ query costs, refer to the pricing details page for further details.&amp;#x27;), (&amp;#x27;language&amp;#x27;, &amp;#x27;lang-sql&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8eee5760&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;ul&gt;
&lt;li role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;List all objects older than a certain date for easy actioning:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Need to archive or delete all images older than five years for compliance? Or perhaps you need to clean up logs that are older than 90 days? Storage Insights datasets provides &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;timeCreated &lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt;and&lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt; contentType&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt; for every object. A simple BigQuery query can list all objects older than your specified date, giving you a clear, actionable list of objects for further investigation. You can use Storage Intelligence &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/batch-operations/overview"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;batch operations&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, which allows you to action on billions of objects in a serverless manner.&lt;/span&gt;&lt;/li&gt;
&lt;li role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Check SoftDelete suitability:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Find buckets that have a high storage size of data that has been soft deleted by querying for the presence of &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;softDeleteTime &lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt;and&lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt; size &lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt;in the object metadata tables. In those cases, data seems temporary and you may need to investigate &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/soft-delete#cost-optimization"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;soft delete cost optimization&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; opportunities. &lt;/span&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Taking your analysis further&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The true power of Storage Intelligence Insights datasets lies not just in the raw data it provides, but in the insights you can derive and the subsequent actions you can take. Once your Cloud Storage metadata is in BigQuery, the possibilities for advanced analysis and integration are vast.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;For example, you can use Looker Studio, Google Cloud's no-cost data visualization and dashboarding tool, to directly connect to your BigQuery Insights datasets, transforming complex queries into intuitive, interactive dashboards. Now you can:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Visualize cost trends:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Create dashboards that show storage consumption by project, department, or storage class over time. This allows teams to easily track spending, identify spikes, and forecast future costs.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Track fast-growing buckets:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Analyze the buckets with the most growth in the past week or month, and compare them against known projects for accurate cost attribution. Use Looker's alerting capabilities to notify you when certain thresholds are met, such as a sudden increase in the total size of data in a bucket.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Set up custom charts for common analysis: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;For routine FinOps use cases (such as tracking buckets without OLM policies configured or objects past their retention expiration time), you can generate weekly reports to relevant teams for easy actioning.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;You can also use our template &lt;/span&gt;&lt;a href="https://lookerstudio.google.com/c/u/0/reporting/670eee3f-ad6d-45ea-a169-853ab023dc84/page/p_k94oydxikd" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;here&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; to connect to your dataset for quick analysis or you can create your own custom dashboard. &lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/1_06D3hbo.max-1000x1000.png"
        
          alt="1"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/2_OQSj3ao.max-1000x1000.png"
        
          alt="2"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Get started&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Configure Storage Intelligence and create your dataset to start analyzing your storage estate via a &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/storage-intelligence/30-day-introductory-trial/overview"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;30-day trial&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; today. Please refer to our &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/pricing#storage-intelligence"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;pricing documentation&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; for cost details. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Set up your dataset to a scope of your choosing and start analyzing your data: &lt;/span&gt;&lt;/p&gt;
&lt;ol&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Configure a set of Looker Studio dashboards based on team or departmental usage for monthly analysis by the central FinOps team.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Use BigQuery for ad-hoc analysis and to retrieve specific insights.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;For a complete cost picture, you can integrate your Storage Insights dataset with your Google Cloud billing export to BigQuery. Your billing export provides granular details on all your Google Cloud service costs, including Cloud Storage.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;&lt;/div&gt;</description><pubDate>Wed, 27 Aug 2025 17:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/storage-data-transfer/storage-insights-datasets-optimizes-storage-footprint/</guid><category>BigQuery</category><category>Storage &amp; Data Transfer</category><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>Storage Insights datasets: How to optimize storage spend with deep visibility</title><description></description><site_name>Google</site_name><url>https://cloud.google.com/blog/products/storage-data-transfer/storage-insights-datasets-optimizes-storage-footprint/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Misha Sheth</name><title>Product Manager, Storage</title><department></department><company></company></author><author xmlns:author="http://www.w3.org/2005/Atom"><name>Chris Madden</name><title>EMEA Solution Lead, Storage</title><department></department><company></company></author></item><item><title>Immutable, Air-Gapped, and Integrated: Data Protection for your Cloud SQL instances just got better</title><link>https://cloud.google.com/blog/products/databases/introducing-enhanced-backups-for-cloud-sql/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span&gt;&lt;strong style="font-style: italic; vertical-align: baseline;"&gt;December 17, 2025:&lt;/strong&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt; The Enhanced Backups for Cloud SQL capability is now generally available to protect the data in your production Cloud SQL instances. Additional features include support for Terraform, and billing capabilities. &lt;/span&gt;&lt;/span&gt;&lt;/p&gt;
&lt;hr/&gt;
&lt;p&gt;In a world where data is your most valuable asset, protecting it isn’t just a nice-to-have — it's a necessity. That's why we are thrilled to announce a significant leap forward in protecting the data in your Cloud SQL instances, with &lt;strong&gt;Enhanced Backups for Cloud SQL&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;This powerful new capability integrates &lt;a href="https://cloud.google.com/backup-disaster-recovery?e=48754805&amp;amp;hl=en"&gt;Google Cloud Backup and DR Service&lt;/a&gt; directly into Cloud SQL, providing a robust, centralized, and secure solution to help ensure business continuity for your database workloads. The Backup and DR Service already protects Compute Engine VMs, Persistent Disks, and Hyperdisk, extending its ability to protect all of your workloads.   &lt;/p&gt;
&lt;h3&gt;Modern defense for modern threats&lt;/h3&gt;
&lt;p&gt;Enhanced Backups for Cloud SQL provides advanced protection by storing database backups in logically air-gapped and immutable backup vaults. Managed by Google and completely separate from your source project, these vaults provide a critical defense against threats that could compromise your entire environment.&lt;/p&gt;
&lt;p&gt;For customers like &lt;strong&gt;JFrog&lt;/strong&gt;, Cloud SQL Enhanced Backup with Google Cloud Backup and DR is proving to be a superior and robust alternative:&lt;/p&gt;
&lt;p style="padding-left: 40px;"&gt;&lt;em&gt;"Using this integration will help us significantly bolster our security posture by offering logically air-gapped and immutable backup vaults, creating a vital defense layer against diverse data-loss scenarios.” - &lt;strong&gt;Shiran Melamed, DevOps Group Leader, JFrog&lt;/strong&gt;&lt;/em&gt;&lt;/p&gt;
&lt;h3&gt;Control, compliance, and peace of mind&lt;/h3&gt;
&lt;p&gt;We designed Enhanced Backups to be both powerful and easy to use, giving you fine-grained control over your data protection strategy. These capabilities are now available in Preview for both Cloud SQL Enterprise and Enterprise Plus editions, and offer key features to help ensure your data is always secure and recoverable:&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1"&gt;
&lt;p role="presentation"&gt;&lt;strong&gt;Immutable, air-gapped vaults&lt;/strong&gt;: Protect your data with immutable backups stored in a secure, logically air-gapped vault. Setting minimum enforced retention and retention locks ensure backups cannot be deleted or changed for a predefined period, while a zero-trust access policy provides granular control.&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1"&gt;
&lt;p role="presentation"&gt;&lt;strong&gt;Business continuity&lt;/strong&gt;: Your data is safeguarded against both source-instance and source-project deletion, so you can recover your data even if the source project itself becomes unavailable.&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1"&gt;
&lt;p role="presentation"&gt;&lt;strong&gt;Flexible policies that fit your needs&lt;/strong&gt;: Your business isn't one-size-fits-all, and your backup strategy shouldn't be either. We offer highly customizable backup schedules, including hourly, daily, weekly, monthly, and yearly options. You can store backups for periods ranging from days to decades.&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1"&gt;
&lt;p role="presentation"&gt;&lt;strong&gt;Centralized command and control&lt;/strong&gt;: Manage everything from a single, unified dashboard in the Google Cloud console. Monitor job status, identify unprotected resources, and generate reports, all in one place.&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;But you don't have to take our word for it. See how customers like SQUARE ENIX and Rotoplas are already benefiting from Enhanced Backups for Cloud SQL:&lt;/p&gt;
&lt;p style="padding-left: 40px;"&gt;&lt;em&gt;"At SQUARE ENIX, protecting our users' data is paramount. Google Cloud SQL's Enhanced Backup integrated with the Backup and DR service is essential to our resiliency strategy. Its robust protection against instance- and even project-level deletion, combined with a secure, isolated vault and long-term retention, provides a critical safeguard for our most valuable asset. This capability will give us confidence in our data's integrity and recoverability, allowing our teams to focus on creating the unforgettable experiences our users expect." – &lt;strong&gt;Kazutaka Iga, SRE,SQUARE ENIX &lt;/strong&gt;&lt;/em&gt;&lt;/p&gt;
&lt;p style="padding-left: 40px;"&gt;&lt;em&gt;"Google Cloud SQL's Enhanced Backup feature along with Google Professional Services support is a value add to our backup strategy at Rotoplas. The ability to centralize management, flexibly schedule backups, and store them independent of the source project gives us unprecedented control. This streamlined approach simplifies our operations and enhances security, ensuring our data is always protected and easily recoverable." - &lt;strong&gt;Agustín Chávez Cabrera, Devops manager, Rotoplas&lt;/strong&gt;&lt;/em&gt;&lt;/p&gt;
&lt;h3&gt;Get started with Enhanced Backups&lt;/h3&gt;
&lt;p&gt;Getting started with Enhanced Backups is simple. Here’s how you can enable this enhanced protection for your Cloud SQL instances:&lt;/p&gt;
&lt;p&gt;1. &lt;strong&gt;Create or select a backup vault&lt;/strong&gt;: In the Backup and DR service, either create a new backup vault or use an existing one.&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--medium
      
      
        h-c-grid__col
        
        h-c-grid__col--4 h-c-grid__col--offset-4
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/original_images/1_create_vault.gif"
        
          alt="1"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph"&gt;&lt;p data-block-key="prq2i"&gt;2. &lt;b&gt;Create a backup plan:&lt;/b&gt; Define a backup plan for Cloud SQL within your chosen backup vault, setting your desired backup frequency and retention rules.&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--medium
      
      
        h-c-grid__col
        
        h-c-grid__col--4 h-c-grid__col--offset-4
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/original_images/2_create_plan.gif"
        
          alt="2"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph"&gt;&lt;p data-block-key="prq2i"&gt;3. &lt;b&gt;Apply the backup plan to the Cloud SQL instances:&lt;/b&gt; Apply your new backup plan to existing or new Cloud SQL instances.&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--medium
      
      
        h-c-grid__col
        
        h-c-grid__col--4 h-c-grid__col--offset-4
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/original_images/3_apply_plan.gif"
        
          alt="3"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph"&gt;&lt;p data-block-key="prq2i"&gt;Once you apply a backup plan, your backups will automatically be scheduled and moved to the secure backup vault based on the rules you defined. The entire experience can be managed through the tools you already use — whether it's the Google Cloud console, gcloud command-line tool, or APIs — so there’s no additional infrastructure for you to deploy or manage.&lt;/p&gt;&lt;h3 data-block-key="2nim0"&gt;&lt;b&gt;Protect your data now&lt;/b&gt;&lt;/h3&gt;&lt;p data-block-key="cgrbf"&gt;With Enhanced Backups for Cloud SQL, you can build a superior data protection strategy that enhances security, simplifies operations, and strengthens your overall data resilience for Cloud SQL instances.&lt;/p&gt;&lt;p data-block-key="8mhrm"&gt;Get started and use it yourself. The new features are available now in &lt;a href="https://cloud.google.com/backup-disaster-recovery/docs/concepts/backup-vault#regions"&gt;supported regions&lt;/a&gt;.&lt;/p&gt;&lt;ul&gt;&lt;li data-block-key="6g0tm"&gt;Experience the new management solution in the &lt;a href="https://cloud.google.com/cloud-console"&gt;console&lt;/a&gt;.&lt;/li&gt;&lt;li data-block-key="btnve"&gt;Watch &lt;a href="https://www.youtube.com/watch?v=LFo2fCobOLY" target="_blank"&gt;this demo video&lt;/a&gt; and see the new features in action.&lt;/li&gt;&lt;li data-block-key="dsmv3"&gt;Explore the documentation to learn more about &lt;a href="https://cloud.google.com/sql/docs/mysql/backup-recovery/backups"&gt;Enhanced Backups for Cloud SQL&lt;/a&gt;, &lt;a href="https://cloud.google.com/backup-disaster-recovery/docs/cloud-console/compute/disk-backup"&gt;disk backups&lt;/a&gt;, and &lt;a href="https://cloud.google.com/backup-disaster-recovery/docs/cloud-console/compute/compute-instance-backup"&gt;VM backups&lt;/a&gt;. today.&lt;/li&gt;&lt;/ul&gt;&lt;/div&gt;</description><pubDate>Wed, 06 Aug 2025 16:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/databases/introducing-enhanced-backups-for-cloud-sql/</guid><category>Cloud SQL</category><category>Storage &amp; Data Transfer</category><category>Databases</category><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>Immutable, Air-Gapped, and Integrated: Data Protection for your Cloud SQL instances just got better</title><description></description><site_name>Google</site_name><url>https://cloud.google.com/blog/products/databases/introducing-enhanced-backups-for-cloud-sql/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Vinod Kumar Subramanian</name><title>Senior Product Manager, Cloud SQL</title><department></department><company></company></author><author xmlns:author="http://www.w3.org/2005/Atom"><name>Lisha Sinha</name><title>Product Manager, Backup and DR</title><department></department><company></company></author></item><item><title>Enhancing GKE data protection with cross-project backup and restore</title><link>https://cloud.google.com/blog/products/storage-data-transfer/backup-for-gke-supports-cross-project-backup-and-restore/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;As Google Kubernetes Engine (GKE) deployments grow and scale, adopting a multi-project strategy in Google Cloud becomes a best practice for security and environment organization. Creating clear boundaries by using distinct projects for development, testing, and production environments provides isolation and helps manage access control.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;However, isolation introduces a data protection challenge: How do you effectively manage backups across these project boundaries? Without a native solution, centralizing backups, ensuring a clear separation of duties with IAM, and enabling robust disaster recovery all become  complex tasks, often forcing teams to rely on custom scripts or inefficient manual processes.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Introducing cross-project backup and restore&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;To address this, Backup for GKE, now in preview, supports cross-project backup and restore. This new capability allows you to back up workloads from a GKE cluster in one Google Cloud project, securely store the backups in a second, and restore them to a cluster in a third. This streamlines data protection, enhances your security posture, and offers greater flexibility for your operational workflows.&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;Storing backups in a separate, isolated project and region is essential for modern disaster recovery, safeguarding your recovery capability during a regional outage or a compromise in a primary Google Cloud project — the foundation of a resilient infrastructure. This separation also simplifies regulatory compliance, boosts security by limiting the blast radius of any potential incident, and helps you meet RTO/RPO objectives.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/image1_LRWNdWl.max-1000x1000.png"
        
          alt="image1"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Key benefits of cross-project backup and restore &lt;/strong&gt;&lt;/h3&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Centralized backup management:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Consolidate GKE backups from multiple Google Cloud projects into a single project by pointing the &lt;/span&gt;&lt;a href="https://cloud.google.com/kubernetes-engine/docs/add-on/backup-for-gke/how-to/cross-project-backups"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;backup plan&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; for each cluster to the chosen backup project. This simple configuration provides your team with one control plane to oversee monitoring and manage backup policies.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Enhanced disaster recovery:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Storing GKE backups in a separate project and region provides a vital layer of isolation, boosting your resilience against events like regional outages. If your source region becomes unavailable, you can create a &lt;/span&gt;&lt;a href="https://cloud.google.com/kubernetes-engine/docs/add-on/backup-for-gke/how-to/cross-project-restores"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;restore plan&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; from your backup project to recover your workloads to a cluster in another project.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Streamline operations: seeding, cloning, and collaboration&lt;br/&gt;&lt;br/&gt;&lt;/strong&gt;&lt;span style="font-family: -apple-system, BlinkMacSystemFont, 'Segoe UI', Roboto, Oxygen, Ubuntu, Cantarell, 'Open Sans', 'Helvetica Neue', sans-serif;"&gt;Cross-project capabilities bring agility to your development lifecycle by simplifying how you copy data between environments. You can now leverage production backup data for testing or rapidly clone entire application environments.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;ul&gt;
&lt;li style="list-style-type: none;"&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Seed and clone environments:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; You can populate a staging environment with data from a prior backup or create a sandbox. Create a &lt;/span&gt;&lt;a href="https://cloud.google.com/kubernetes-engine/docs/add-on/backup-for-gke/how-to/cross-project-restores"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;restore plan&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; using an existing backup plan located in the backup project, then select a backup — such as one from production for seeding or a dev environment for cloning — and target a cluster in any other project as your destination. This lets you create test environments and isolated sandboxes.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Simplify cross-team collaboration:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Since all backups are stored in a central backup project, you can grant a developer from another team a role like Delegated Restore Admin, and also provide them with read permission on the specific backup plan and all of its associated backups. They can then use it to restore to their cluster without needing access to the other team's live source project.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Achieve separation of duties for security and compliance&lt;br/&gt;&lt;br/&gt;&lt;/strong&gt;&lt;span style="font-family: -apple-system, BlinkMacSystemFont, 'Segoe UI', Roboto, Oxygen, Ubuntu, Cantarell, 'Open Sans', 'Helvetica Neue', sans-serif;"&gt;Isolating backups in a dedicated project allows you to enforce the principle of least privilege by assigning distinct responsibilities. You can empower your application teams with self-service permissions to back up and restore applications within their own projects, without giving them control over the central backup repository. A central platform or operations team can be granted administrative control over the backup project to govern the entire data lifecycle — from setting retention policies with immutability to conducting audits, all without needing access to live production environments. This separation is key to reducing risk and simplifying audits.&lt;br/&gt;&lt;br/&gt;&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;For detailed guidance on Backup for GKE IAM roles and permissions, see the&lt;/span&gt;&lt;a href="https://cloud.google.com/kubernetes-engine/docs/add-on/backup-for-gke/how-to/roles" style="font-family: -apple-system, BlinkMacSystemFont, 'Segoe UI', Roboto, Oxygen, Ubuntu, Cantarell, 'Open Sans', 'Helvetica Neue', sans-serif;"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt; &lt;/span&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;documentation&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Cross-project backup and restore for GKE helps you protect your containerized workloads across multiple Google Cloud projects. This feature allows you to strengthen your disaster recovery capabilities, improve your security posture, and streamline operational workflows.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Get started today&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;This feature is now generally available. To get started, check out the following guides:&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Learn how to perform &lt;/span&gt;&lt;a href="https://cloud.google.com/kubernetes-engine/docs/add-on/backup-for-gke/how-to/cross-project-backups"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;cross-project backups&lt;/span&gt;&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Learn how to perform &lt;/span&gt;&lt;a href="https://cloud.google.com/kubernetes-engine/docs/add-on/backup-for-gke/how-to/cross-project-restores" style="font-family: -apple-system, BlinkMacSystemFont, 'Segoe UI', Roboto, Oxygen, Ubuntu, Cantarell, 'Open Sans', 'Helvetica Neue', sans-serif;"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;cross-project restores&lt;/span&gt;&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;&lt;/div&gt;
&lt;div class="block-aside"&gt;&lt;dl&gt;
    &lt;dt&gt;aside_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;title&amp;#x27;, &amp;#x27;Try Google Cloud for free&amp;#x27;), (&amp;#x27;body&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8f71f190&amp;gt;), (&amp;#x27;btn_text&amp;#x27;, &amp;#x27;Get started for free&amp;#x27;), (&amp;#x27;href&amp;#x27;, &amp;#x27;https://console.cloud.google.com/freetrial?redirectPath=/welcome&amp;#x27;), (&amp;#x27;image&amp;#x27;, None)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;</description><pubDate>Fri, 11 Jul 2025 16:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/storage-data-transfer/backup-for-gke-supports-cross-project-backup-and-restore/</guid><category>Containers &amp; Kubernetes</category><category>GKE</category><category>Storage &amp; Data Transfer</category><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>Enhancing GKE data protection with cross-project backup and restore</title><description></description><site_name>Google</site_name><url>https://cloud.google.com/blog/products/storage-data-transfer/backup-for-gke-supports-cross-project-backup-and-restore/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Ranjith Kumar Palthi</name><title>Product Manager</title><department></department><company></company></author></item><item><title>Cloud Storage bucket relocation: An industry first for non-disruptive bucket migrations</title><link>https://cloud.google.com/blog/products/storage-data-transfer/introducing-cloud-storage-bucket-relocation/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;As your operational needs change, sometimes you need to move data residing within Google’s Cloud Storage to a new location, to improve resilience, optimize performance, meet compliance needs, or simply to reorganize your infrastructure. Yet moving buckets can be a daunting, complex, risky endeavor that involves manual scripting, painstaking coordination, and the risk of data loss, or worse yet, extended downtime. This can discourage organizations from making the changes they need to their storage environments.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;We recently introduced&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/bucket-relocation/overview"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Cloud Storage bucket relocation&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, a unique feature among leading hyperscalers &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;that makes it easy  to change your bucket’s location. Bucket relocation eliminates the need for complex manual planning and helps prevent extended downtime, for an easy transition with minimal application disruption, and strong &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;data integrity&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;. Your bucket's name, and all the object metadata within it, remain identical throughout the relocation, so there are no path changes, and your applications experience minimal downtime while the underlying storage is moved. Furthermore, your objects retain their original storage class (e.g., Standard, Nearline, Coldline, Archive) and time-in-class in the new location. This is key for many cost efficiency strategies, helping ensure capabilities such as Autoclass continue to operate intelligently to optimize your storage costs post-migration.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Bucket relocation is a key capability within &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/storage-intelligence/overview"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;the Storage Intelligence suite&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, alongside tools like Storage Insights, which  provides deep visibility into your storage landscape and identifies optimization opportunities. Bucket relocation then lets you act on these insights, and move your data between diverse &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/locations"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Cloud Storage locations&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; — regional locations for low latency, dual-regions for high availability and disaster recovery, or multi-regions for global accessibility — to meet your  business, performance, and compliance objectives.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-aside"&gt;&lt;dl&gt;
    &lt;dt&gt;aside_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;title&amp;#x27;, &amp;#x27;Try Google Cloud for free&amp;#x27;), (&amp;#x27;body&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8f664580&amp;gt;), (&amp;#x27;btn_text&amp;#x27;, &amp;#x27;Get started for free&amp;#x27;), (&amp;#x27;href&amp;#x27;, &amp;#x27;https://console.cloud.google.com/freetrial?redirectPath=/welcome&amp;#x27;), (&amp;#x27;image&amp;#x27;, None)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Bucket relocation under the hood&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Bucket relocation relies on two critical techniques. &lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Asynchronous data copy:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Bucket relocation leverages a unique and optimized asynchronous data transfer mechanism that copies data in the background to  minimize impact to ongoing operations. Existing operations like writing, reading, and updating objects continue while the entire dataset is being copied.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Metadata preservation:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Historically, Google Cloud customers moved data with the &lt;/span&gt;&lt;a href="https://cloud.google.com/storage-transfer/docs/overview"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Storage Transfer Service&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, which copied the objects to a new bucket and deleted existing ones. Bucket relocation, on the other hand, automatically and meticulously moves all your bucket’s and objects’ associated metadata, thereby preserving state. This includes information like:&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;ul&gt;
&lt;li aria-level="2" style="list-style-type: circle; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Storage class:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Your objects retain their original storage class (e.g., Standard, Nearline, Coldline, Archive) in the new location.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="2" style="list-style-type: circle; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Bucket and object names:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; The naming structure of your buckets and objects remains identical.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="2" style="list-style-type: circle; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Creation and update timestamps:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; These markers are preserved, so that features like object lifecycle management (OLM) rules continue to operate.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="2" style="list-style-type: circle; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Access Control Lists (ACLs) and IAM policies:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Bucket- and object-level permissions are transferred to help maintain your security posture.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="2" style="list-style-type: circle; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Custom metadata:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Any user-defined metadata associated with your objects is also migrated.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;By handling the complexities of asynchronous data transfer and automatic metadata migration, bucket relocation minimizes the risks and overhead associated with a manual bucket migration. Crucially, because the bucket name is preserved throughout the relocation process, applications accessing the bucket don’t need to be modified.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;Relocate your bucket in a few simple steps&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;With bucket relocation, you can move your Cloud Storage buckets in &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/bucket-relocation/relocate-buckets"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;three simple steps&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. Here's a breakdown:&lt;/span&gt;&lt;/p&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;1. &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;Initiate a dry run:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Before starting the actual relocation, it's highly recommended to perform a dry run. This simulates the process without moving any data, allowing you to identify potential issues early on, such as incompatible configurations.&lt;/span&gt;&lt;/li&gt;
&lt;li role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;The dry run checks for incompatibilities like customer-managed encryption keys (CMEK), locked retention policies, objects with temporary holds, and bucket tags, without you having to manually validate each of them.&lt;/span&gt;&lt;/li&gt;
&lt;li role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Make sure to add the &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;--dry-run&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt; flag!&lt;/span&gt;&lt;/li&gt;
&lt;/ul&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;#x27;gcloud storage buckets relocate gs://BUCKET_NAME --location=LOCATION --dry-run&amp;#x27;), (&amp;#x27;language&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8f748d90&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;Replace &lt;/span&gt;&lt;code style="font-style: italic; vertical-align: baseline;"&gt;BUCKET_NAME&lt;/code&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt; with the name of your bucket and &lt;/span&gt;&lt;code style="font-style: italic; vertical-align: baseline;"&gt;LOCATION&lt;/code&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt; with the desired destination.&lt;/span&gt;&lt;/p&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;2. &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;Start the relocation process:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;This step initiates the actual data transfer from the source bucket to the destination bucket. During this phase, you can still read, modify, and delete objects in the bucket. However, the bucket metadata (i.e., bucket-level  parameters and configurations) is write-locked to prevent changes that could affect the relocation.&lt;br/&gt;&lt;br/&gt;&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;Note:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Removing the &lt;/span&gt;&lt;code style="vertical-align: baseline;"&gt;--dry-run&lt;/code&gt;&lt;span style="vertical-align: baseline;"&gt; flag from the dry-run command initiates the relocation.&lt;/span&gt;&lt;/li&gt;
&lt;/ul&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;#x27;gcloud storage buckets relocate gs://BUCKET_NAME --location=LOCATION&amp;#x27;), (&amp;#x27;language&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8f748490&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;3. &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;Finalize the relocation process:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Once the incremental data copy is complete, you’re ready to trigger the final synchronization step (except when moving between multi-region and configurable dual-region). This involves a brief period where writes to the bucket are disabled to help ensure their data integrity; any last-second changes made to the objects within the bucket while the incremental copy was in progress are copied to the destination. After the data’s integrity is verified, the bucket's location is updated, and all requests are automatically redirected to the new location. During the final synchronization step, attempts to update objects in the bucket will result in an HTTP 412 error.&lt;/span&gt;&lt;/li&gt;
&lt;li role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Do not initiate the final synchronization process until the relocation process progress reaches ~99%. This helps you minimize downtime because most of the data has already been synchronized in the background. &lt;/span&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Note: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;If you’re&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;moving between multi-regions and configurable dual-regions within the same &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/locations#location-dr"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;multi-region code&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, you’re all set — bucket relocation handles the transition in the background, no finalization or downtime required!&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-code"&gt;&lt;dl&gt;
    &lt;dt&gt;code_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;code&amp;#x27;, &amp;#x27;gcloud storage buckets relocate --finalize --operation=projects/_/buckets/BUCKET_NAME/operations/OPERATION_ID&amp;#x27;), (&amp;#x27;language&amp;#x27;, &amp;#x27;&amp;#x27;), (&amp;#x27;caption&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8f748fa0&amp;gt;)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;The &lt;/span&gt;&lt;code style="font-style: italic; vertical-align: baseline;"&gt;OPERATION_ID&lt;/code&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt; is provided as output from Step-2. The &lt;/span&gt;&lt;code style="font-style: italic; vertical-align: baseline;"&gt;OPERATION_ID&lt;/code&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt; is listed with the keyword name. For instance:&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;code style="font-style: italic; vertical-align: baseline;"&gt;name: projects/_/buckets/my-bucket/operations/AbCJYd8jKT1n-Ciw1LCNXIcubwvij_TdqO-ZFjuF2YntK0r74&lt;/code&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;And there you have it — In just three steps, you’ven moved your entire bucket, its data, and metadata, to its new location.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Early users of bucket relocation have had great success with the new feature. &lt;/span&gt;&lt;/p&gt;
&lt;p style="padding-left: 40px;"&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;“With Storage Intelligence and bucket relocation, we effortlessly transitioned to dual-region buckets. The seamless process, powered by the bucket relocation, minimized downtime and ensured data integrity. We migrated the buckets with peace of mind and without the manual headaches.” &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;- Adam Steele, Product Manager, Spotify&lt;/strong&gt;&lt;/p&gt;
&lt;p style="padding-left: 40px;"&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;“We recently utilized the bucket relocation feature of Storage Intelligence to successfully complete a ~300 bucket migration and PBs of data project from multi-region to regional storage, to optimize network data transfer costs. Without bucket relocation, this process would have required extensive automation and scripting, resulting in increased downtime and effort.”&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; - Deepak Mahato, Data Platform Infrastructure Manager, GroupOn&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Experience the ease and efficiency of managing your Cloud Storage buckets with bucket relocation in Storage Intelligence. To learn more, visit the&lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/bucket-relocation/overview"&gt;&lt;span style="vertical-align: baseline;"&gt; &lt;/span&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;bucket relocation documentation&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; and the&lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/storage-intelligence/overview"&gt;&lt;span style="vertical-align: baseline;"&gt; &lt;/span&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Storage Intelligence overview&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;</description><pubDate>Thu, 10 Jul 2025 16:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/storage-data-transfer/introducing-cloud-storage-bucket-relocation/</guid><category>Developers &amp; Practitioners</category><category>Storage &amp; Data Transfer</category><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>Cloud Storage bucket relocation: An industry first for non-disruptive bucket migrations</title><description></description><site_name>Google</site_name><url>https://cloud.google.com/blog/products/storage-data-transfer/introducing-cloud-storage-bucket-relocation/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Vaibhav Khunger</name><title>Senior Product Manager, Google Cloud</title><department></department><company></company></author></item><item><title>Accelerate your AI workloads with the Google Cloud Managed Lustre</title><link>https://cloud.google.com/blog/products/storage-data-transfer/google-cloud-managed-lustre-for-ai-hpc/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Today, we're making it even easier to achieve breakthrough performance for your AI/ML workloads: &lt;/span&gt;&lt;a href="https://cloud.google.com/products/managed-lustre"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Google Cloud Managed Lustre&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; is now GA, and available in four distinct performance tiers that deliver throughput ranging from 125 MB/s, 250 MB/s, 500 MB/s, to 1000 MB/s per TiB of capacity — with the ability to scale up to 8 PB of storage capacity. The Managed Lustre solution is powered by DDN’s EXAScaler, combining DDN's decades of leadership in high-performance storage with Google Cloud's expertise in cloud infrastructure.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Managed Lustre provides a POSIX-compliant, parallel file system that delivers consistently high throughput and low latency, essential for:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;High-throughput inference:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; For applications that require near-real-time inference on large datasets, Lustre provides high parallel throughput and sub-millisecond read latency.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Large-scale model training:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Accelerate the training cycles of deep learning models by providing rapid access to petabytes-sized datasets. Lustre's parallel architecture ensures GPUs and TPUs are fed with data, minimizing idle time.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Checkpointing and restarting large models:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Save and restore the state of large models during training faster, improving goodput and allowing for more efficient experimentation.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Data preprocessing and feature engineering:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Process raw data, extract features, and prepare datasets for training, reducing the time spent on data pipelines.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Scientific simulations and research:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Beyond AI/ML, Lustre excels in traditional HPC scenarios like computational fluid dynamics, genomic sequencing, and climate modeling, where massive datasets and high-concurrency access are critical.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Lustre is designed for the highly parallel and random I/O that characterizes many AI/ML training and inference tasks. This parallel processing capability across multiple clients ensures your compute resources are never starved for data.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Performance tiers and pricing&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Managed Lustre offers flexible pricing and performance tiers designed to meet the diverse needs of your workloads, whether you're focused on capacity or highest throughput density. &lt;/span&gt;&lt;/p&gt;
&lt;div align="left"&gt;
&lt;div style="color: #5f6368; overflow-x: auto; overflow-y: hidden; width: 100%;"&gt;
&lt;div style="color: #5f6368; overflow-x: auto; overflow-y: hidden; width: 100%;"&gt;
&lt;div style="color: #5f6368; overflow-x: auto; overflow-y: hidden; width: 100%;"&gt;
&lt;div style="color: #5f6368; overflow-x: auto; overflow-y: hidden; width: 100%;"&gt;
&lt;div style="color: #5f6368; overflow-x: auto; overflow-y: hidden; width: 100%;"&gt;
&lt;div style="color: #5f6368; overflow-x: auto; overflow-y: hidden; width: 100%;"&gt;
&lt;div style="color: #5f6368; overflow-x: auto; overflow-y: hidden; width: 100%;"&gt;
&lt;div style="color: #5f6368; overflow-x: auto; overflow-y: hidden; width: 100%;"&gt;
&lt;div style="color: #5f6368; overflow-x: auto; overflow-y: hidden; width: 100%;"&gt;
&lt;div style="color: #5f6368; overflow-x: auto; overflow-y: hidden; width: 100%;"&gt;&lt;table style="width: 98.4334%;"&gt;&lt;colgroup&gt;&lt;col style="width: 56.3665%;"/&gt;&lt;col style="width: 43.6335%;"/&gt;&lt;/colgroup&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td style="vertical-align: top; border: 1px solid #000000; padding: 16px;"&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Throughput &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;MB/s&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;per TiB &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;of storage capacity&lt;/span&gt;&lt;/p&gt;
&lt;/td&gt;
&lt;td style="vertical-align: top; border: 1px solid #000000; padding: 16px;"&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Storage pricing per GiB per month&lt;/span&gt;&lt;/p&gt;
&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td style="vertical-align: top; border: 1px solid #000000; padding: 16px;"&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;125&lt;/span&gt;&lt;/p&gt;
&lt;/td&gt;
&lt;td style="vertical-align: top; border: 1px solid #000000; padding: 16px;"&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;$0.145&lt;/span&gt;&lt;/p&gt;
&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td style="vertical-align: top; border: 1px solid #000000; padding: 16px;"&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;250&lt;/span&gt;&lt;/p&gt;
&lt;/td&gt;
&lt;td style="vertical-align: top; border: 1px solid #000000; padding: 16px;"&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;$0.21&lt;/span&gt;&lt;/p&gt;
&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td style="vertical-align: top; border: 1px solid #000000; padding: 16px;"&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;500&lt;/span&gt;&lt;/p&gt;
&lt;/td&gt;
&lt;td style="vertical-align: top; border: 1px solid #000000; padding: 16px;"&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;$0.34&lt;/span&gt;&lt;/p&gt;
&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td style="vertical-align: top; border: 1px solid #000000; padding: 16px;"&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;1000&lt;/span&gt;&lt;/p&gt;
&lt;/td&gt;
&lt;td style="vertical-align: top; border: 1px solid #000000; padding: 16px;"&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;$0.60&lt;/span&gt;&lt;/p&gt;
&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Please see more details at the &lt;/span&gt;&lt;a href="https://cloud.google.com/products/managed-lustre/pricing"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Managed Lustre pricing page&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Irrespective of the aggregate throughput, all tiers come with sub-millisecond read latency, high single-stream throughput, and are perfect for parallel access to many small files.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Driving innovation together: partnering with DDN&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Google Cloud’s Managed Lustre is powered by &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;DDN’s EXAScaler&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;, bringing together two industry leaders in high-performance computing and elastic cloud infrastructure. This partnership represents a joint commitment to simplifying the deployment and management of large-scale AI and HPC workloads in the cloud, thanks to:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Trusted leaders:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; By combining DDN's decades of expertise in high-performance Lustre with Google Cloud's global infrastructure and AI ecosystem, we are delivering a foundational capability that removes storage bottlenecks and helps our customers solve their most complex challenges in AI and HPC.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Fully managed and supported solution:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Enjoy the benefits of a fully managed service from Google, with comprehensive support from both Google and DDN, for seamless operations and peace of mind.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Global availability and ecosystem integration:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Managed Lustre is now globally accessible in &lt;/span&gt;&lt;a href="https://cloud.google.com/managed-lustre/docs/locations"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;multiple Google Cloud regions&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; and integrates with the broader Google Cloud ecosystem, including Google Kubernetes Engine (GKE) and TPUs.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;These benefits caught the attention of one of our largest partners, NVIDIA, who is looking forward to having it as part of its NVIDIA AI platform. &lt;/span&gt;&lt;/p&gt;
&lt;p style="padding-left: 40px;"&gt;&lt;span style="vertical-align: baseline;"&gt;"&lt;/span&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;Enterprises today demand AI infrastructure that combines accelerated computing with high-performance storage solutions to deliver uncompromising speed, seamless scalability and cost efficiency at scale. Google and DDN’s collaboration on Google Cloud Managed Lustre creates a better-together solution uniquely suited to meet these needs. By integrating DDN’s enterprise-grade data platforms and Google’s global cloud capabilities, organizations can readily access vast amounts of data and unlock the full potential of AI with the NVIDIA AI platform (or NVIDIA accelerated computing platform) on Google Cloud — reducing time-to-insight, maximizing GPU utilization, and lowering total cost of ownership.&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;” - Dave Salvator, Director of Accelerated Computing Products, NVIDIA&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Get started today!&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Ready to supercharge your AI/ML and HPC workloads? Getting started with Managed Lustre is simple:&lt;/span&gt;&lt;/p&gt;
&lt;ol&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Navigate to &lt;/span&gt;&lt;a href="https://console.cloud.google.com/managed-lustre/"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Managed Lustre in the Google Cloud console&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Provision your Managed Lustre instance, choosing the performance tier and size that best fits your needs.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Connect your compute instances, GKE clusters to your new high-performance file system.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;For detailed instructions and documentation, please visit the Managed Lustre &lt;/span&gt;&lt;a href="https://cloud.google.com/managed-lustre/docs/overview"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;documentation&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. And if needed, &lt;/span&gt;&lt;a href="https://cloud.google.com/contact"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;reach out to Google Cloud sales specialists&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Watch the Fireside Chat&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Don't miss the opportunity to learn more about the strategic partnership between Google Cloud and DDN, and the unique capabilities of Managed Lustre. Read the official DDN press release &lt;/span&gt;&lt;a href="https://www.ddn.com/press-releases/google-cloud-launches-general-availability-of-managed-lustre-powered-by-ddns-exascaler-technology/" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;here&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Watch the fireside chat with Sameet Agarwal, VP/GM Storage and Sven Oehme, CTO of DDN, &lt;/strong&gt;&lt;a href="https://www.youtube.com/watch?v=i6gEHUzIo1w" rel="noopener" target="_blank"&gt;&lt;strong style="text-decoration: underline; vertical-align: baseline;"&gt;here&lt;/strong&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;</description><pubDate>Tue, 08 Jul 2025 17:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/storage-data-transfer/google-cloud-managed-lustre-for-ai-hpc/</guid><category>AI &amp; Machine Learning</category><category>HPC</category><category>Storage &amp; Data Transfer</category><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>Accelerate your AI workloads with the Google Cloud Managed Lustre</title><description></description><site_name>Google</site_name><url>https://cloud.google.com/blog/products/storage-data-transfer/google-cloud-managed-lustre-for-ai-hpc/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Asad Khan</name><title>Sr. Director of Product Management, Google Cloud</title><department></department><company></company></author><author xmlns:author="http://www.w3.org/2005/Atom"><name>Kirill Tropin</name><title>Group Product Manager</title><department></department><company></company></author></item><item><title>Expanding Z3 family with 9 new VMs and a bare metal instance for storage and I/O intensive workloads</title><link>https://cloud.google.com/blog/products/compute/expanded-z3-vm-portfolio-for-io-intensive-workloads/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Today, we are thrilled to announce the expansion of the Z3 Storage Optimized VM family with the general availability of nine new &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/storage-optimized-machines?_gl=1*2vt8da*_up*MQ..&amp;amp;gclid=CjwKCAiAqfe8BhBwEiwAsne6gduqCwwkpJZbE9aPtQmusSUIJYOzGeKiVzaE-1_M9aml0iqY5L8_IBoCh90QAvD_BwE&amp;amp;gclsrc=aw.ds#z3_machine_types"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Z3 virtual machines&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; that offer local SSD capacity ranging from 3 TiB to 18 TiB per VM, complementing existing Z3 VMs which offer 36TiB of Local SSD per VM. We are also very pleased to launch a &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/instances/bare-metal-instances"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Z3 bare metal instance&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, which includes up to 72 TiB of Local SSDs. Z3 VMs enable customers like Shopify, Tenderly and ScyllaDB to achieve impressive performance improvements for their high performance storage workloads by reducing the IO access latency by up to 35% compared to VM instances using previous-generation local SSDs. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Z3 VMs are designed to run I/O-intensive workloads that require large local storage capacity and high storage performance, including SQL, NoSQL, and vector databases, data analytics, semantic data search and retrieval, and distributed file systems. The Z3 bare metal instance provides direct access to the physical server CPUs and&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; is ideal for workloads that require low-level system access like private and hybrid cloud platforms, custom hypervisors, container platforms, or applications with specialized performance or licensing needs.&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Both Z3 VMs and the bare metal instance are based on &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/local-ssd"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Titanium SSDs&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, which &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;offload local storage processing from CPU resources&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; to deliver real-time data processing, low-latency, high-throughput storage performance and &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;enhanced storage security. Z3 VMs with Titanium SSD offer up to 36 GiB/s of read throughput and up to 9M IOPS, increasing write storage performance by up to 25% compared to previous generation Local SSDs&lt;sup&gt;1&lt;/sup&gt;&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-aside"&gt;&lt;dl&gt;
    &lt;dt&gt;aside_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;title&amp;#x27;, &amp;#x27;$300 in free credit to try Google Cloud infrastructure&amp;#x27;), (&amp;#x27;body&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8f6da8b0&amp;gt;), (&amp;#x27;btn_text&amp;#x27;, &amp;#x27;Start building for free&amp;#x27;), (&amp;#x27;href&amp;#x27;, &amp;#x27;http://console.cloud.google.com/freetrial?redirectPath=/compute&amp;#x27;), (&amp;#x27;image&amp;#x27;, None)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Based on the 4th Gen Intel Xeon scalable processor, Z3 VMs come with up to 176 vCPUs, 1,408 GiB of memory, and 36 TiB of local storage in 11 virtual machine shapes. The Z3 bare metal instance offers 192 vCPUs, 1,536 GiB of memory and 72 TiB of local storage. Z3 VMs and the bare metal instance deliver the connectivity and storage performance that enterprise workloads need, with&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;up to 100 Gbps in standard bandwidth and up to 200 Gbps with Tier1 networking for high-traffic applications.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The expanded Z3 virtual machine portfolio lets you rightsize your infrastructure and scale your clusters to meet workloads requirements by providing larger total local SSD capacity and higher local SSD capacity per vCPU. Z3 offers two different VM types: the &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/storage-optimized-machines?_gl=1*2vt8da*_up*MQ..&amp;amp;gclid=CjwKCAiAqfe8BhBwEiwAsne6gduqCwwkpJZbE9aPtQmusSUIJYOzGeKiVzaE-1_M9aml0iqY5L8_IBoCh90QAvD_BwE&amp;amp;gclsrc=aw.ds#z3_machine_types"&gt;&lt;strong style="text-decoration: underline; vertical-align: baseline;"&gt;standardlssd&lt;/strong&gt;&lt;/a&gt;&lt;strong style="vertical-align: baseline;"&gt; VM types,&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; which include five VM shapes that offer about 200 GiB of local SSD per vCPU. They are optimized for data analytics (OLAP), and SQL databases like MySQL and Postgres workloads. The &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/storage-optimized-machines?_gl=1*2vt8da*_up*MQ..&amp;amp;gclid=CjwKCAiAqfe8BhBwEiwAsne6gduqCwwkpJZbE9aPtQmusSUIJYOzGeKiVzaE-1_M9aml0iqY5L8_IBoCh90QAvD_BwE&amp;amp;gclsrc=aw.ds#z3_machine_types"&gt;&lt;strong style="text-decoration: underline; vertical-align: baseline;"&gt;highlssd&lt;/strong&gt;&lt;/a&gt;&lt;strong style="vertical-align: baseline;"&gt; VM types &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;include&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;six different VM shapes and the Z3 bare metal instance. They offer about 400 GiB of local SSD per vCPU and are optimized for distributed databases, data streaming, large parallel file systems and data search. &lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;What our customers and partners are saying&lt;/strong&gt;&lt;/h3&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/1_n43GSEg.max-1000x1000.png"
        
          alt="1"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;"We are thrilled to announce Nutanix Cloud Clusters coming to Google Cloud at the end of CY25 as part of Nutanix’s commitment to delivering flexible, hybrid cloud solutions. Google Cloud’s Z3 instance types represent a perfect foundation for Nutanix to enable performance and resilience for enterprise applications. We’re excited about our partnership with Google Cloud in empowering our joint customers with greater choice and simplicity in their cloud journey." &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;- Saveen Pakala, Vice President of Product Management, Nutanix&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/2_DVxv1Kn.max-1000x1000.png"
        
          alt="2"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;“OP Labs contributes to the Optimism protocol, which enables orders of magnitude of improved performance and scalability for Ethereum. Z3 reduces p99 block insertion tail latencies by 30-50% for our most I/O-demanding blockchain nodes compared to N2. By migrating our solution to Z3, we will be able to scale our blockchain nodes to handle L2 state growth in a more performant and cost-effective way.”&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; - Zach Howard Senior Staff Engineer, OP Labs&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/3_Us2Q9jh.max-1000x1000.png"
        
          alt="3"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;The launch of Google Cloud's Z3 storage optimized instances with smaller VM shapes  represents a leap forward in performance for high-traffic NoSQL environments. In internal tests and customer projects, ScyllaDB has impressively leveraged the advantages of Z3 including extremely low latencies under high read and write loads, high IOPS capacity enabling the processing of massive amounts of data and excellent cost-performance ratio for large-scale production systems. We are very excited to offer Z3 family servers in ScyllaDB Cloud, including Bring Your Own Account (BYOA)." -&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; Avi Kivity, Co-founder and CTO, ScyllaDB&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/4_Wdb1KfR.max-1000x1000.png"
        
          alt="4"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;"Shopify has found Z3s to be an excellent platform to build our most performance sensitive storage systems on. We experienced a critical need for both large data volumes while remaining sensitive to latency and throughput on the storage side. While Google has a lot of options, local SSD was really the best fit, and Z3s allowed us to achieve the best price/performance along with enhanced stability appropriate for a source of truth Storage workload. Right now we see these storage optimized VMs as our platform of choice for the future."&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; - Mattie Toia, VP Infrastructure, Shopify&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/5_mJ8yPFD.max-1000x1000.png"
        
          alt="5"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;"Tenderly is built to be your go-to for Web3 production and development, bringing all the necessary infrastructure into one place. This allows teams to operate with speed and confidence, making blockchain technology accessible. We've seen impressive results running blockchain workloads on Z3 instances, with a 40% improvement on read latency compared to N2 and N2D instances."&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; - Ilija Petrovic, SRE Lead, Tenderly&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/6_MqCDpn9.max-1000x1000.png"
        
          alt="6"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;“The VAST AI Operating System gives organizations a unified platform to reason over all of their data – structured, unstructured, and streaming through a global namespace that spans cloud and on-prem environments – enabling intelligent agents and applications to operate with full context and real-time speed. ,For customers running on Google Cloud, Z3 VMs complement this vision by providing the ideal storage infrastructure to accelerate these workloads, ensuring AI pipelines run fast and scale effortlessly in the cloud.” &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;- Renen Hallak, Founder &amp;amp; CEO, VAST Data&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Z3 VMs are also the physical foundation of AlloyDB, our flagship PostgreSQL-compatible database service, delivering sophisticated multi-level caching. AlloyDB uses Z3's expansive local SSDs as an ultra-fast cache, holding datasets up to 25x larger than can be stored in memory. Database queries can access these large, cached datasets at latencies that closely approach in-memory performance, particularly when factoring in overall end-to-end application response times. This is a significant advantage for very large databases, including real-time analytical workloads, as AlloyDB’s high-performance columnar engine operates entirely within this massive cache. AlloyDB on Z3 VMs will soon be available in preview, delivering up to &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;3x better performance than N-series VMs for transactional workloads, particularly for large datasets&lt;strong&gt;.&lt;/strong&gt;&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Enhanced maintenance experience&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Z3 instances make it easier for you to plan ahead and schedule maintenance operations at a time of your choosing by providing notice from the system several days in advance of a required maintenance. The new Z3 VMs further enhance the maintenance experience by allowing you to live-migrate an instance during maintenance events for VMs with 18 TiB or less of local SSD storage. For Z3 VMs with 36 TiB of local SSD and for Z3 bare metal instances, you’ll also receive in-place upgrades that preserve your data through the planned maintenance events.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Support for Hyperdisk&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Z3 VMs support &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/hyperdisks"&gt;&lt;strong style="text-decoration: underline; vertical-align: baseline;"&gt;Hyperdisk&lt;/strong&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, Google Cloud’s workload-optimized block storage that lets you optimize the performance for each workload by independently tuning the storage performance and capacity for each instance.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Z3 VMs are compatible &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;with &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/hd-types/hyperdisk-balanced"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Hyperdisk Balanced&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/hd-types/hyperdisk-throughput"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Hyperdisk Throughput&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, and &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/hd-types/hyperdisk-extreme"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Extreme&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/hyperdisks"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Hyperdisk&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; storage for scalable, high-performance network-attached storage&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;, supporting up to 512 TiB of capacity per instance. For general-purpose workloads, Hyperdisk Balanced, with up to &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/hd-types/hyperdisk-balanced#achieve-higher-performance-with-multiple-hyperdisk-balanced-volumes"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;160K IOPS&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; per instance, offers a mix of performance and cost-efficiency. Hyperdisk Extreme delivers ultra-low latency and supports up to &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/hd-types/hyperdisk-extreme#achieve-higher-performance-with-multiple-hyperdisk-extreme-volumes"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;350K IOPS and 5,000 MiB/s throughput&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; per Z3 VM instance and up &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/hd-types/hyperdisk-extreme#achieve-higher-performance-with-multiple-hyperdisk-extreme-volumes"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;500K IOPS and 10,000 MiB/s throughput&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; for the Z3 bare metal instance — making it well-suited for demanding workloads like databases. Using Hyperdisk for persistent storage and Z3 Local SSD for caching creates an optimal storage architecture for high end databases and mission critical workloads&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Get started with Z3 today&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Z3 VMs and bare metal instances are available today in most regions worldwide. To start using Z3 instances, select Z3 under the new Storage-Optimized machine family when creating a new VM or GKE node pool in the Google Cloud console. Learn more at the &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/storage-optimized-machines"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Z3 machine series page&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. Contact your &lt;/span&gt;&lt;a href="https://cloud.google.com/contact?e=48754805&amp;amp;hl=en"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Google Cloud sales&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; representative for more information on regional availability.&lt;/span&gt;&lt;/p&gt;
&lt;hr/&gt;
&lt;p&gt;&lt;sup&gt;&lt;em&gt;1. &lt;span style="vertical-align: baseline;"&gt;Results are based on Google Cloud’s internal benchmarking&lt;/span&gt;&lt;/em&gt;&lt;/sup&gt;&lt;/p&gt;&lt;/div&gt;</description><pubDate>Tue, 08 Jul 2025 16:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/compute/expanded-z3-vm-portfolio-for-io-intensive-workloads/</guid><category>Storage &amp; Data Transfer</category><category>Compute</category><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>Expanding Z3 family with 9 new VMs and a bare metal instance for storage and I/O intensive workloads</title><description></description><site_name>Google</site_name><url>https://cloud.google.com/blog/products/compute/expanded-z3-vm-portfolio-for-io-intensive-workloads/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Garv Sawhney</name><title>Group Product Manager</title><department></department><company></company></author><author xmlns:author="http://www.w3.org/2005/Atom"><name>bob Napaa</name><title>Principal Product Manager</title><department></department><company></company></author></item><item><title>Automate data resilience at scale with Eon and Google Cloud Backup</title><link>https://cloud.google.com/blog/products/storage-data-transfer/data-resilience-eons-approach--google-cloud-best-practices/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Cloud backups were once considered as little more than an insurance policy. Now, your backups should do more! They should be autonomous, cost-efficient, and analytics-ready by default.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;That’s why &lt;/span&gt;&lt;a href="https://www.eon.io/blog/google-cloud-announcement" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Eon built a platform purposefully aligned with Google Cloud&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; to eliminate backup blind spots, simplify recovery, and unlock the value inside backup data without requiring teams to become policy experts or infrastructure wranglers.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Still, no matter what platform you use, it’s critical to understand what resilient cloud backup looks like and how to get there with Google Cloud’s native capabilities.&lt;/span&gt;&lt;/p&gt;
&lt;h2&gt;&lt;span style="vertical-align: baseline;"&gt;What makes cloud backup resilient?&lt;/span&gt;&lt;/h2&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Before diving into tooling, it's worth asking: What does a resilient backup strategy look like in the cloud? In our work with Google Cloud users across industries, we’ve found five common criteria:&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;5 signs your backup posture may be at risk&lt;/span&gt;&lt;/h3&gt;
&lt;ol&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;You can’t easily see what’s backed up (or not)&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Retention policies vary across projects and teams&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Data is duplicated or stored inefficiently, driving up spend&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;a href="https://www.eon.io/blog/cloud-ransomware-guide" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Cloud ransomware&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; protection is reactive rather than policy-driven&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Recovery requires full restores even when you only need one object&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;&lt;/div&gt;
&lt;div class="block-aside"&gt;&lt;dl&gt;
    &lt;dt&gt;aside_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;title&amp;#x27;, &amp;#x27;Try Google Cloud for free&amp;#x27;), (&amp;#x27;body&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f8f733040&amp;gt;), (&amp;#x27;btn_text&amp;#x27;, &amp;#x27;Get started for free&amp;#x27;), (&amp;#x27;href&amp;#x27;, &amp;#x27;https://console.cloud.google.com/freetrial?redirectPath=/welcome&amp;#x27;), (&amp;#x27;image&amp;#x27;, None)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h2&gt;&lt;span style="vertical-align: baseline;"&gt;Best practices for data protection&lt;/span&gt;&lt;/h2&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Google Cloud provides foundational capabilities to protect your data if you configure and use them consistently. Here's how to maximize native protection:&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;1. Versioning and retention: first lines of defense&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Enable Object Versioning in Cloud Storage to retain multiple object versions, making it easier to recover from accidental deletions. Pair this with Retention Policies to enforce minimum storage lifetimes for regulatory or critical datasets.&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong style="vertical-align: baseline;"&gt;Tip:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Use Bucket Lock for write-once-read-many (WORM) protection in the areas where compliance matters most.&lt;/span&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;2. Monitor for gaps in coverage&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Use native services like Cloud SQL backups, GKE snapshots, and Persistent Disk images, but be mindful that backup responsibilities can fall to different teams. Without centralized visibility, coverage becomes inconsistent.&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong style="vertical-align: baseline;"&gt;Tip:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Use Cloud Asset Inventory or scheduled BigQuery queries to audit coverage.&lt;/span&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;h3&gt;&lt;span style="vertical-align: baseline;"&gt;3. Design for granular recovery&lt;/span&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Plan for partial restores since not everything needs a full rollback. Whether it's a single BigQuery table or a specific Cloud Storage object, restoring only what you need saves time and cost.&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong style="vertical-align: baseline;"&gt;Tip&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;: Use Object Lifecycle Management to automatically transition older or less critical Cloud Storage objects to colder storage classes.&lt;/span&gt;&lt;/li&gt;
&lt;/ul&gt;&lt;/div&gt;
&lt;div class="block-video"&gt;



&lt;div class="article-module article-video "&gt;
  &lt;figure&gt;
    &lt;a class="h-c-video h-c-video--marquee"
      href="https://youtube.com/watch?v=i1qrmufbJKU"
      data-glue-modal-trigger="uni-modal-i1qrmufbJKU-"
      data-glue-modal-disabled-on-mobile="true"&gt;

      
        

        &lt;div class="article-video__aspect-image"
          style="background-image: url(https://storage.googleapis.com/gweb-cloudblog-publish/images/Screenshot_2025-06-19_at_11.36.51AM_1.max-1000x1000_QEI2O0T.png);"&gt;
          &lt;span class="h-u-visually-hidden"&gt;Google Cloud New Way to Cloud video interview with Eon CTO Ron Kimchi&lt;/span&gt;
        &lt;/div&gt;
      
      &lt;svg role="img" class="h-c-video__play h-c-icon h-c-icon--color-white"&gt;
        &lt;use xlink:href="#mi-youtube-icon"&gt;&lt;/use&gt;
      &lt;/svg&gt;
    &lt;/a&gt;

    
      &lt;figcaption class="article-video__caption h-c-page"&gt;
        
          &lt;h4 class="h-c-headline h-c-headline--four h-u-font-weight-medium h-u-mt-std"&gt;Watch out New Way to Cloud interview with Eon CTO and co-founder Ron Kimchi.&lt;/h4&gt;
        
        
      &lt;/figcaption&gt;
    
  &lt;/figure&gt;
&lt;/div&gt;

&lt;div class="h-c-modal--video"
     data-glue-modal="uni-modal-i1qrmufbJKU-"
     data-glue-modal-close-label="Close Dialog"&gt;
   &lt;a class="glue-yt-video"
      data-glue-yt-video-autoplay="true"
      data-glue-yt-video-height="99%"
      data-glue-yt-video-vid="i1qrmufbJKU"
      data-glue-yt-video-width="100%"
      href="https://youtube.com/watch?v=i1qrmufbJKU"
      ng-cloak&gt;
   &lt;/a&gt;
&lt;/div&gt;

&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h2&gt;&lt;span style="vertical-align: baseline;"&gt;Automating the complexity away&lt;/span&gt;&lt;/h2&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Managing cloud backup at scale is hard to do manually. From onboarding new workloads to applying consistent policies, human-led approaches don’t scale well.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;That’s why more teams are exploring autonomous &lt;/span&gt;&lt;a href="https://www.eon.io/blog/cloud-backup-posture-management" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Cloud Backup Posture Management&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; (CBPM) solutions, like Eon, that detect new assets in real time, apply smart backup rules automatically, and enforce consistent protection across environments.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;With Eon, you don’t have to tag resources or write custom scripts. Our platform classifies and protects your Google Cloud assets out of the box—whether you're working with GKE, Cloud SQL, BigQuery, or another solution.&lt;/span&gt;&lt;/p&gt;
&lt;h2&gt;&lt;span style="vertical-align: baseline;"&gt;From backups to business insights&lt;/span&gt;&lt;/h2&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Traditionally, backup data was siloed, underused, and only meant to be retrieved in emergencies. But, increasingly, teams are unlocking that data to:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Run analysis directly on backups using BigQuery and Dataproc,&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Feed training and monitoring pipelines via Vertex AI,&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Deliver audit-ready dashboards with Looker, powered by backup snapshots.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;With Eon, this is built-in. We transform backups into zero-ETL data lakes that reduce pipeline costs and provide immediate access to structured data with no reprocessing required.&lt;/span&gt;&lt;/p&gt;
&lt;h2&gt;&lt;span style="vertical-align: baseline;"&gt;What a “mature” backup posture looks like&lt;/span&gt;&lt;/h2&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The end goal for many cloud-native teams is not just to “have backups.” It’s to develop a resilient, intelligent backup strategy that adapts to scale and risk.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Here’s what that looks like:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Automated discovery of new resources&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Policy-driven protection tailored to data type and criticality&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Immutable backups with time-locked retention&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Search-first recovery instead of full snapshot restores&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;Cost-aware tiering and storage deduplication&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Eon helps Google Cloud users reach this level of maturity faster without the burden of custom tooling or constant policy updates.&lt;/span&gt;&lt;/p&gt;
&lt;h2&gt;&lt;span style="vertical-align: baseline;"&gt;Ready to simplify backup?&lt;/span&gt;&lt;/h2&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;If your team spends hours managing scripts, storage tiers, or backup tags across cloud environments, it may be time to rethink your approach.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Eon was built to make cloud backup resilient, autonomous, and actually useful. From ransomware protection to instant, object-level recovery—and now, zero-ETL access to analytics—we’re here to help you unlock the full potential of your backup data.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;a href="https://www.eon.io/get-a-demo" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Book a demo&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; to see how Eon can modernize your Google Cloud data protection strategy.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;To discover how Google Cloud can support your startup, visit our &lt;/span&gt;&lt;a href="http://cloud.google.com/startup/apply?utm_source=google&amp;amp;utm_medium=blog&amp;amp;utm_campaign=FY21-Q1-global-demandgen-website-cs-startup_program_mc&amp;amp;utm_content=blog_Descifra&amp;amp;utm_term="&gt;&lt;span style="font-style: italic; text-decoration: underline; vertical-align: baseline;"&gt;program page&lt;/span&gt;&lt;/a&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;. You can also &lt;/span&gt;&lt;a href="https://docs.google.com/forms/d/e/1FAIpQLSfowlgaSsVDQojZ1JDDhRMfZ5TAFY6do4UPZXqkuToX63K2dQ/viewform" rel="noopener" target="_blank"&gt;&lt;span style="font-style: italic; text-decoration: underline; vertical-align: baseline;"&gt;sign up for our newsletter&lt;/span&gt;&lt;/a&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt; to stay informed about community activities, digital events, special offers, and more.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;</description><pubDate>Wed, 18 Jun 2025 16:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/storage-data-transfer/data-resilience-eons-approach--google-cloud-best-practices/</guid><category>Startups</category><category>Customers</category><category>Storage &amp; Data Transfer</category><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>Automate data resilience at scale with Eon and Google Cloud Backup</title><description></description><site_name>Google</site_name><url>https://cloud.google.com/blog/products/storage-data-transfer/data-resilience-eons-approach--google-cloud-best-practices/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Liore Shai</name><title>Solutions Architect, Eon</title><department></department><company></company></author></item><item><title>Enhancing backup vaults with support for Persistent Disk, Hyperdisk, and multi-regions</title><link>https://cloud.google.com/blog/products/storage-data-transfer/backup-vaults-add-support-for-disk-backup-and-multi-region/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;em&gt;&lt;span style="vertical-align: baseline;"&gt;&lt;strong&gt;August 11, 2025&lt;/strong&gt;: Backup vault support for persistent disk (PD) and hyperdisk backups is now all generally available.&lt;/span&gt;&lt;/em&gt;&lt;/p&gt;
&lt;hr/&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;To help protect against evolving digital threats like ransomware and malicious deletions, last year, we introduced &lt;/span&gt;&lt;a href="https://cloud.google.com/blog/products/storage-data-transfer/backup-and-dr-service-adds-immutable-indelible-backups?e=0"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;backup vault&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; in the Google Cloud Backup and DR service, with support for Compute Engine VM backups. This provided immutable and indelible backup capabilities for mission-critical VMs, for both VM metadata and all their attached disks.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Today, we're&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; announcing two enhancements to backup vaults that can help you protect more types of workloads, better:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Backup vaults now support standalone Persistent Disk (PD) and Hyperdisk backups.&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Now generally available, it enables the direct backup of data on individual disks, providing a granular alternative to backing up the entire virtual machine.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Backup vaults can now be created in multi-region locations. &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Now generally available it supports regional data resilience and helping to meet business continuity requirements.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Immutability and indelibility &lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Traditional backups have a well-known vulnerability. If a malicious actor gains access to your environment, if they attempt to delete or corrupt the backup, preventing recovery and thus causing business loss, there is nothing preventing this from happening. This is where backup vaults fundamentally change the game.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;A backup vault provides a secure, isolated storage environment in Google-managed projects that helps ensure your backups are &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;immutable&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; (secured against data modification)&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt; &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;and &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;indelible &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;(secured against data deletion), providing protection against cyber attacks such as ransomware. When creating a backup vault, you can specify that vaulted backups must be secured against modification and deletion — even by a backup administrator who would traditionally have the ability to expire backups — until the specified minimum &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;enforced retention&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; timeframe has elapsed. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Once a backup is stored in a vault, it's logically air-gapped from your Google Cloud project, and cannot be changed during its user-defined enforced retention period. This means:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;No deletion:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; The backup can’t be accidentally or deliberately deleted before its enforced retention period expires.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;No alteration:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; The backup data cannot be changed, and remains exactly as it was when it was created.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;This gives you the confidence that your crucial recovery points have not been modified, so they are available when you need them.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Backup Vault now supports Persistent Disk and Hyperdisk&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Many applications rely on the durable storage provided by &lt;/span&gt;&lt;a href="https://cloud.google.com/persistent-disk"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Persistent Disk&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; and &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/hyperdisks"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Hyperdisk&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. With support for Persistent Disk and Hyperdisk in addition to Compute Engine VMs, backup vaults now offer a &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;holistic defense strategy for your entire compute environment&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;For your VMs:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; &lt;a href="https://cloud.google.com/backup-disaster-recovery/docs/cloud-console/compute/compute-instance-backup"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Backup vaults can help protect your Compute Engine VMs&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; (including VM metadata and all the attached disks). They can provide rapid and secure recovery of operating systems, configurations, application binaries, and all associated disks.&lt;/span&gt;&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;For critical data disks:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Now you can secure specific Persistent Disks and Hyperdisks that contain application data, databases, and file shares. They can provide granular protection, for scenarios where a full VM backup isn't necessary, or you want to optimize costs.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;This integrated approach ensures that whether you need to restore an entire VM or a specific disk, your recovery points are secured in a backup vault. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Key benefits of unified backup vault protection&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;By centralizing your Compute Engine VM, Persistent Disk, and Hyperdisk backups within backup vaults, you gain a powerful suite of advantages that transform your data protection strategy from reactive to proactively resilient:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Unified interface for easy management:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Easily define and enforce consistent backup policies (including backup frequency and retention period) across your entire organization. Manage backups for your Compute Engine VMs, Persistent Disks, and Hyperdisks from a unified interface, even across multiple Google Cloud projects, simplifying administration.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Comprehensive monitoring and reporting:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Benefit from centralized monitoring, detailed reporting, and timely alerting capabilities that streamline your day-to-day backup management. This enhanced visibility also significantly aids in meeting stringent audit and compliance requirements by providing clear, verifiable records of your backup posture.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Proactive security integration:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Elevate your overall security posture with integration to &lt;/span&gt;&lt;a href="https://cloud.google.com/security/products/security-command-center"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Security Command Center&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, enabling proactive detection of anomalous activities, such as unauthorized backup deletion attempts or suspicious policy changes, so you can respond swiftly and decisively to threats.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Reduced operational complexity:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Consolidate your backup management processes, moving away from disparate, script-based, or manual solutions. Backup and DR service provides a streamlined, fully managed service that simplifies operations, reduces human error, and frees up valuable IT resources, so you can focus on innovation.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Here's how it works&lt;/strong&gt;&lt;/h3&gt;
&lt;ol&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Create a backup vault:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Begin by establishing a secure backup vault. This vault acts as your designated, isolated, and highly protected storage destination for all your managed backups.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Define a backup plan:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Next, create a comprehensive backup plan, specifying parameters such as the desired backup frequency (how often your disks will be backed up), backup retention period, and designating the specific backup vault where the backup data will be stored.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: decimal; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Schedule your backups:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Now you are ready to apply your backup plan to your desired Persistent Disks or Hyperdisks. The Backup and DR service automatically takes incremental crash-consistent backups according to your defined schedule, with no manual intervention on your part.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Once these backups are created and stored in your designated vault, the vault’s enforced retention policy is automatically applied, making the backups immutable and indelible for the specified enforced retention period.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/1_-_persistent-disk-backups-to-backup-vaul.max-1000x1000.png"
        
          alt="1 - persistent-disk-backups-to-backup-vault-for-cyber-resilience"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Secure disaster recovery with multi-region backup vaults&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;In addition, you can now create backup vaults in Google-managed, multi-region locations. When using a multi-region backup vault, data is stored in more than one geographic region, thereby providing the security benefits of backup vault, while also making critical backup data available during unforeseen events. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Using multi-region backup vaults lets you:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Retain data access:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Maintain accessibility and recoverability of critical backup data during a regional service disruption (such as natural disasters, power outages).&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Satisfy business continuity requirements:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Instill confidence in your business operations with your ability to perform on-demand, backup-based recoveries.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Secure your data: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Retain all of the critical security benefits delivered by backup vaults.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Multi-region backup vault storage is generally available and currently supports Compute Engine full VM backups and disk backups to &lt;/span&gt;&lt;a href="https://cloud.google.com/backup-disaster-recovery/docs/concepts/backup-vault#multi-regions"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;supported Locations&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. Complete &lt;/span&gt;&lt;a href="https://docs.google.com/forms/d/e/1FAIpQLSfxmIpvwA57BgYNGwc9A7RdB29a6om1ky2eCdOYQNiJfYwGzw/viewform?usp=header" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;this form&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; to request access to the new feature.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--medium
      
      
        h-c-grid__col
        
        h-c-grid__col--4 h-c-grid__col--offset-4
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/original_images/2_-_Backup_vault_creation_screen_-_multi-region_ZlQQWT1.png"
        
          alt="2 - Backup vault creation screen - multi-region"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Protect all your critical Compute Engine data&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;With the addition of multi-region backup vaults and disk-level backup, Backup and DR service can secure and recover critical Compute Engine data better than ever. Try the new capabilities yourself to optimize your VM data protection strategy.&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;To learn more about disk backup, start &lt;/span&gt;&lt;a href="https://cloud.corp.google.com/backup-disaster-recovery/docs/quickstarts/disk-backup-vault" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;here&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;To learn more about multi-region backup vaults, start &lt;/span&gt;&lt;a href="https://cloud.google.com/backup-disaster-recovery/docs/concepts/backup-vault#multi-regions"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;here&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;To request access to use multi-region backup vaults, please complete &lt;/span&gt;&lt;a href="https://docs.google.com/forms/d/e/1FAIpQLSfxmIpvwA57BgYNGwc9A7RdB29a6om1ky2eCdOYQNiJfYwGzw/viewform?usp=header" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;this form&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;span style="vertical-align: baseline;"&gt;See &lt;/span&gt;&lt;a href="https://cloud.google.com/backup-disaster-recovery/pricing" style="font-family: -apple-system, BlinkMacSystemFont, 'Segoe UI', Roboto, Oxygen, Ubuntu, Cantarell, 'Open Sans', 'Helvetica Neue', sans-serif;"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;here&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; for pricing information relating to the new capabilities.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;&lt;/div&gt;
&lt;div class="block-related_article_tout"&gt;





&lt;div class="uni-related-article-tout h-c-page"&gt;
  &lt;section class="h-c-grid"&gt;
    &lt;a href="https://cloud.google.com/blog/products/storage-data-transfer/backup-and-dr-service-adds-immutable-indelible-backups/"
       data-analytics='{
                       "event": "page interaction",
                       "category": "article lead",
                       "action": "related article - inline",
                       "label": "article: {slug}"
                     }'
       class="uni-related-article-tout__wrapper h-c-grid__col h-c-grid__col--8 h-c-grid__col-m--6 h-c-grid__col-l--6
        h-c-grid__col--offset-2 h-c-grid__col-m--offset-3 h-c-grid__col-l--offset-3 uni-click-tracker"&gt;
      &lt;div class="uni-related-article-tout__inner-wrapper"&gt;
        &lt;p class="uni-related-article-tout__eyebrow h-c-eyebrow"&gt;Related Article&lt;/p&gt;

        &lt;div class="uni-related-article-tout__content-wrapper"&gt;
          &lt;div class="uni-related-article-tout__image-wrapper"&gt;
            &lt;div class="uni-related-article-tout__image" style="background-image: url('')"&gt;&lt;/div&gt;
          &lt;/div&gt;
          &lt;div class="uni-related-article-tout__content"&gt;
            &lt;h4 class="uni-related-article-tout__header h-has-bottom-margin"&gt;Introducing backup vaults for cyber resilience and simplified Compute Engine backups&lt;/h4&gt;
            &lt;p class="uni-related-article-tout__body"&gt;Protect your data with Google Cloud&amp;#x27;s enhanced Backup and DR service, featuring immutable backup vaults and streamlined management for da...&lt;/p&gt;
            &lt;div class="cta module-cta h-c-copy  uni-related-article-tout__cta muted"&gt;
              &lt;span class="nowrap"&gt;Read Article
                &lt;svg class="icon h-c-icon" role="presentation"&gt;
                  &lt;use xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="#mi-arrow-forward"&gt;&lt;/use&gt;
                &lt;/svg&gt;
              &lt;/span&gt;
            &lt;/div&gt;
          &lt;/div&gt;
        &lt;/div&gt;
      &lt;/div&gt;
    &lt;/a&gt;
  &lt;/section&gt;
&lt;/div&gt;

&lt;/div&gt;</description><pubDate>Tue, 17 Jun 2025 17:30:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/storage-data-transfer/backup-vaults-add-support-for-disk-backup-and-multi-region/</guid><category>Security &amp; Identity</category><category>Compute</category><category>Storage &amp; Data Transfer</category><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>Enhancing backup vaults with support for Persistent Disk, Hyperdisk, and multi-regions</title><description></description><site_name>Google</site_name><url>https://cloud.google.com/blog/products/storage-data-transfer/backup-vaults-add-support-for-disk-backup-and-multi-region/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Jaswant Chajed</name><title>Product Manager, Google Cloud</title><department></department><company></company></author><author xmlns:author="http://www.w3.org/2005/Atom"><name>Jerome McFarland</name><title>Product Manager, Google Cloud</title><department></department><company></company></author></item><item><title>Selecting the right Hyperdisk block storage for your workloads</title><link>https://cloud.google.com/blog/products/storage-data-transfer/how-to-choose-the-right-hyperdisk-block-storage-for-your-use-case/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;As you adopt Google Cloud or migrate to the latest Compute Engine VMs or to Google Kubernetes Engine (GKE), selecting the right block storage for your workload is crucial. &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/hyperdisks"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Hyperdisk,&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; Google Cloud's workload-optimized block storage that’s designed for our latest &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/machine-resource"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;VM families&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; (C4, N4, M4, and more), delivers high-performance storage volumes that are cost-efficient, easily managed at scale, and enterprise-ready. In this post, we guide you through the basics and help you choose the optimal Hyperdisk for your environment.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Introduction to Hyperdisk block storage&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;With Hyperdisk, you can independently tune capacity and performance to match your block storage resources to your workload. &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/hyperdisks#when-to-use"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Hyperdisk is available in a few flavors:&lt;/span&gt;&lt;/a&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Hyperdisk Balanced:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Designed to fit most workloads and offers the best combination and balance of price and performance. This is also the boot disk for your compute instances. With Hyperdisk Balanced, you can independently configure the capacity, throughput, and IOPS of each volume. Hyperdisk Balanced is available in High Availability and Multi-writer mode.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Hyperdisk Extreme:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Delivers the highest IOPS of all Hyperdisk offerings and is suited for high-end, performance-critical databases. With Hyperdisk Extreme, you can drive up to 350K IOPS from a single volume. &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Hyperdisk Throughput:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Delivers capacity at the cost of cold object storage with the semantics of a disk. Hyperdisk Throughput offers high throughput for bandwidth and capacity-intensive workloads that do not require low latency. It also can be used to deliver cost-effective disks for cost-sensitive workloads (e.g., cold disks).&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Hyperdisk ML:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Purpose-built for loading static data into your compute clusters. With Hyperdisk ML, you hydrate the disk with a fixed data set (such as model weights or binaries), then connect  &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;up to 2,500 compute instances to the same volume, so a single volume can serve over 150x more compute instances than competitive block storage volumes&lt;/span&gt;&lt;sup&gt;&lt;span style="vertical-align: baseline;"&gt;&lt;span style="vertical-align: super;"&gt;1&lt;/span&gt;&lt;/span&gt;&lt;/sup&gt;&lt;span style="vertical-align: baseline;"&gt; in read-only mode&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;. You get exceptionally high aggregate throughput across all of those nodes, enabling you to accelerate inference startup, train models faster, and ensure your valuable compute resources are highly utilized. &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;You can also leverage Hyperdisk &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/storage-pools"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Storage Pools&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, which lowers TCO and simplifies operations by pre-provisioning an aggregate amount of capacity and performance, which is then dynamically consumed by volumes in that pool. You create a storage pool with the aggregate capacity and performance that your workloads will need, and then create disks in the storage pool. You can then attach the disks to your VMs. When you create the disks, you can create them with a much larger size or provisioned performance limit than is needed. This simplifies planning and provides room for growth later, without needing to change the disk's provisioned size or performance. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;You can also use a set of comprehensive data protection capabilities such as high availability, cross-region replication and recovery, backup, and snapshots to protect your business critical workloads.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;For specifics around capabilities, capacity, machine support, and performance, please &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/hyperdisks#when-to-use"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;visit the documentation&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-aside"&gt;&lt;dl&gt;
    &lt;dt&gt;aside_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;title&amp;#x27;, &amp;#x27;Try Google Cloud for free&amp;#x27;), (&amp;#x27;body&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f903fc3a0&amp;gt;), (&amp;#x27;btn_text&amp;#x27;, &amp;#x27;Get started for free&amp;#x27;), (&amp;#x27;href&amp;#x27;, &amp;#x27;https://console.cloud.google.com/freetrial?redirectPath=/welcome&amp;#x27;), (&amp;#x27;image&amp;#x27;, None)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Recommendations for the most common workloads&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;To make choosing the right Hyperdisk architecture simpler, here are high-level recommendations for some of the most common workloads we see. For an enterprise, the Hyperdisk portfolio lets you optimize an entire three-tier application matching the needs of each component of your application to the different flavors of Hyperdisk.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Enterprise applications including general-purpose databases&lt;/strong&gt;&lt;strong style="vertical-align: baseline;"&gt;:&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/hd-types/hyperdisk-balanced"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Hyperdisk Balanced&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; combined with Storage Pools offers an excellent solution for a wide variety of general-purpose workloads, including common database workloads. Hyperdisk Balanced can meet the IOPS and throughput needs for most databases including Clickhouse, MySQL, and PostgreSQL, at general-purpose pricing. Hyperdisk Balanced offers 160K IOPS per volume — better than AWS EBS gp3 volumes&lt;/span&gt;&lt;sup&gt;&lt;span style="vertical-align: baseline;"&gt;&lt;span style="vertical-align: super;"&gt;2&lt;/span&gt;&lt;/span&gt;&lt;/sup&gt;&lt;span style="vertical-align: baseline;"&gt;. With Storage Pools you can enhance efficiency and radically simplify planning. Storage Pools allows customers to save approximately 20-40% on storage costs for typical database workloads when compared to Hyperdisk Balanced Volumes or AWS EBS gp3 volumes&lt;/span&gt;&lt;sup&gt;&lt;span style="vertical-align: baseline;"&gt;&lt;span style="vertical-align: super;"&gt;3&lt;/span&gt;&lt;/span&gt;&lt;/sup&gt;&lt;span style="vertical-align: baseline;"&gt;. &lt;/span&gt;&lt;/p&gt;
&lt;p style="padding-left: 40px;"&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;“At Sentry.io, a platform used by over 4 million developers and 130,000 teams worldwide to quickly debug and resolve issues, adopting Google Cloud's Hyperdisk has enabled us to create a flexible architecture for the next-generation of our Event Analytics Platform, a product at the core of our business. Hyperdisk Storage Pools with advanced capacity and performance enabled us to reduce our planning cycles from weeks to minutes, while saving 37% in storage costs, compared to persistent disks.” &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;- Dave Rosenthal, CTO, Sentry&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-video"&gt;



&lt;div class="article-module article-video "&gt;
  &lt;figure&gt;
    &lt;a class="h-c-video h-c-video--marquee"
      href="https://youtube.com/watch?v=jjU8F8FjzCM"
      data-glue-modal-trigger="uni-modal-jjU8F8FjzCM-"
      data-glue-modal-disabled-on-mobile="true"&gt;

      
        &lt;img src="//img.youtube.com/vi/jjU8F8FjzCM/maxresdefault.jpg"
             alt="Sentry: Creating the next generation data platform with Hyperdisk and Storage Pools"/&gt;
      
      &lt;svg role="img" class="h-c-video__play h-c-icon h-c-icon--color-white"&gt;
        &lt;use xlink:href="#mi-youtube-icon"&gt;&lt;/use&gt;
      &lt;/svg&gt;
    &lt;/a&gt;

    
  &lt;/figure&gt;
&lt;/div&gt;

&lt;div class="h-c-modal--video"
     data-glue-modal="uni-modal-jjU8F8FjzCM-"
     data-glue-modal-close-label="Close Dialog"&gt;
   &lt;a class="glue-yt-video"
      data-glue-yt-video-autoplay="true"
      data-glue-yt-video-height="99%"
      data-glue-yt-video-vid="jjU8F8FjzCM"
      data-glue-yt-video-width="100%"
      href="https://youtube.com/watch?v=jjU8F8FjzCM"
      ng-cloak&gt;
   &lt;/a&gt;
&lt;/div&gt;

&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p style="padding-left: 40px;"&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;“High Availability is essential for Blackline — we run database failover clustering, at massive scale, for our global and mission-critical deployment of Financial Close Management. We are excited to bring this workload to Google Cloud leveraging Hyperdisk Balanced High Availability to meet the performance, capacity, cost efficiency, and resilience requirements that our customers demand, and helps us address our customer’s financial regulatory needs globally.” &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;- Justin Brodley, SVP of Cloud Engineering and Operations, Blackline&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Tier-0 databases&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;For high-end, performance-critical databases like SAP HANA, SQL Server, and Oracle Database, &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/hd-types/hyperdisk-extreme"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Hyperdisk Extreme&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; delivers uncompromising performance. With Hyperdisk Extreme, you can obtain up to 350K IOPS and 10 GiB/s of throughput from a single volume.     &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;AI, analytics, and scale-out workloads&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Hyperdisk offers excellent solutions for the most demanding next-generation machine learning and high performance computing workloads. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;Dynamically scaling AI and analytics workloads and high-performance file systems&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Workloads with fluctuating demand, and high peak throughput and IOPS, benefit from Hyperdisk Balanced and Storage Pools. These workloads can include customer-managed parallel file systems and scratch disks for accelerator clusters. Storage Pools’ dynamic resource allocation helps ensure that these workloads get the performance they need during peak times without requiring constant manual adjustments or inefficient over-provisioning. Further, once your Storage Pool is set up, planning at the per-disk level is significantly simpler. Note: If you want a fully managed file system, &lt;/span&gt;&lt;a href="https://cloud.google.com/managed-lustre/docs/overview"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Managed Lustre&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; is an excellent option for you to consider.  &lt;/span&gt;&lt;/p&gt;
&lt;p style="padding-left: 40px;"&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;“Combining our use of cutting-edge machine learning in quantitative trading at Hudson River Trading (HRT) with Google Cloud's accelerator-optimized machines, Dynamic Workload Scheduler (DWS) and Hyperdisk has been transformative in enabling us to develop [state-of-the-art] models. Hyperdisk storage pools have delivered substantial cost savings, lowering our storage expenses by approximately 50% compared to standard Hyperdisk while minimizing the amount of planning needed.&lt;/span&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;” &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;- Ragnar Kjørstad, Systems Engineer, Hudson River Trading&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-video"&gt;



&lt;div class="article-module article-video "&gt;
  &lt;figure&gt;
    &lt;a class="h-c-video h-c-video--marquee"
      href="https://youtube.com/watch?v=U1NbkumODpg"
      data-glue-modal-trigger="uni-modal-U1NbkumODpg-"
      data-glue-modal-disabled-on-mobile="true"&gt;

      
        &lt;img src="//img.youtube.com/vi/U1NbkumODpg/maxresdefault.jpg"
             alt="Hudson River Trading: Powering cutting-edge quantitative research models with Google Cloud"/&gt;
      
      &lt;svg role="img" class="h-c-video__play h-c-icon h-c-icon--color-white"&gt;
        &lt;use xlink:href="#mi-youtube-icon"&gt;&lt;/use&gt;
      &lt;/svg&gt;
    &lt;/a&gt;

    
  &lt;/figure&gt;
&lt;/div&gt;

&lt;div class="h-c-modal--video"
     data-glue-modal="uni-modal-U1NbkumODpg-"
     data-glue-modal-close-label="Close Dialog"&gt;
   &lt;a class="glue-yt-video"
      data-glue-yt-video-autoplay="true"
      data-glue-yt-video-height="99%"
      data-glue-yt-video-vid="U1NbkumODpg"
      data-glue-yt-video-width="100%"
      href="https://youtube.com/watch?v=U1NbkumODpg"
      ng-cloak&gt;
   &lt;/a&gt;
&lt;/div&gt;

&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;AI/ML and HPC data-load acceleration&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Hyperdisk ML is specifically optimized for accelerating data load times for inference, training and HPC workloads —  Hyperdisk ML accelerates model load time by 3-5x compared to common alternatives&lt;/span&gt;&lt;sup&gt;&lt;span style="vertical-align: baseline;"&gt;&lt;span style="vertical-align: super;"&gt;4&lt;/span&gt;&lt;/span&gt;&lt;/sup&gt;&lt;span style="vertical-align: baseline;"&gt;. Hyperdisk ML is particularly well-suited for serving tasks compared to other storage services on Google Cloud because it can concurrently provide to many VMs exceptionally high aggregate throughput (up to 1.2 TiB/s of aggregate throughput per volume, offering greater than 100x higher performance than competitive offerings)&lt;/span&gt;&lt;sup&gt;&lt;span style="vertical-align: baseline;"&gt;&lt;span style="vertical-align: super;"&gt;5&lt;/span&gt;&lt;/span&gt;&lt;/sup&gt;&lt;span style="vertical-align: baseline;"&gt;. You write once (up to 64 TiB per disk) and attach multiple VM instances to the same volume in a read-only mode. With Hyperdisk ML you can accelerate data load times for your most expensive compute resources, like GPUs and TPUs. For more, check out &lt;/span&gt;&lt;a href="http://g.co/cloud/storage-design-ai" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;g.co/cloud/storage-design-ai&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. &lt;/span&gt;&lt;/p&gt;
&lt;p style="padding-left: 40px;"&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;“At Resemble AI, we leverage our proprietary deep-learning models to generate high-quality AI audio through text-to-speech and speech-to-speech synthesis. By combining Google Cloud’s A3 VMs with NVIDIA H100 GPUs and Hyperdisk ML, we’ve achieved significant improvements in our training workflows. Hyperdisk ML has drastically improved our data loader performance, enabling 2x faster epoch cycles compared to similar solutions. This acceleration has empowered our engineering team to experiment more freely, train at scale, and accelerate the path from prototype to production." &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;-&lt;/span&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt; &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;Zohaib Ahmed, CEO, Resemble AI&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong style="vertical-align: baseline;"&gt;High-capacity analytics workloads: &lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;For large-scale data analytics workloads like Hadoop and Kafka, which are less sensitive to disk latency fluctuations, Hyperdisk Throughput provides a cost-effective solution with high throughput. Its low cost per GiB and configurable throughput are ideal for processing large volumes of data with low TCO.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;How to size and set up your Hyperdisk&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;To select and size the right Hyperdisk volume types for your workload, answer a few key questions:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Storage management.&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Decide if you want to manage the block storage for your workloads in a pool or individually. If your workload will have more than 10 TiB of capacity in a single project and zone, you should consider using Hyperdisk Storage Pools to lower your TCO and simplify planning. &lt;/span&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;Note that Storage Pools do not affect disk performance; some data protection features such as Replication and High Availability are not supported in Storage Pools. &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Latency.&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; If your workload requires SSD-like latency (i.e., sub-millisecond), it likely should be served by Hyperdisk Balanced or Hyperdisk Extreme. &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;IOPS or throughput.&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; If your application requires less than 160K IOPS or 2.4 GiB/s of throughput from a single volume, Hyperdisk Balanced is a great fit. If it needs more than that, consider Hyperdisk Extreme. &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Sizing performance and capacity.&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Hyperdisk offers independently configurable capacity and performance, allowing you to pay for just the resources you need. You can leverage this capability to lower your TCO by understanding how much capacity your workload needs (i.e., how much data, in GiB or TiB, is stored on the disks which serve this workload) and the peak IOPS and throughput of the disks. If the workload is already running on Google Cloud, you can see many of these metrics in your console under “&lt;/span&gt;&lt;a href="https://cloud.google.com/monitoring/charts/metrics-explorer"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Metrics Explorer&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.” &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Another important consideration is the level of business continuity and data protection required for your workloads. Different workloads have different Recovery Point Objective (RPO) and Recovery Time Objective (RTO) requirements, each with different costs. Think about your workload tiers when making data-protection decisions. The more critical an application or workload, the lower the tolerance for data loss and downtime. Applications critical to business operations likely require zero RPO and RTO in the order of seconds. Hyperdisk business continuity and data protection helps customers meet the performance, capacity, cost efficiency, and resilience requirements they demand, and helps them address their financial regulatory needs globally. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Here are a few questions to consider when selecting which variety of Hyperdisk to use for a workload:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;How do I protect my workloads from attack and malicious insiders? &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;Use&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;a href="https://cloud.google.com/backup-disaster-recovery/docs/concepts/backup-vault"&gt;&lt;strong style="text-decoration: underline; vertical-align: baseline;"&gt;Google Cloud Backup vault&lt;/strong&gt;&lt;/a&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;for&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;cyber resilience, backup immutability, and indelibility for managed backup reporting and compliance. If you want to self-manage your own backups, Hyperdisk standard snapshots are an option for your workloads.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;How do I protect data from user errors and bad upgrades cost efficiently with low RPO / RTO?&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; You can use our &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;point-in-time recovery with &lt;/strong&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/instant-snapshots"&gt;&lt;strong style="text-decoration: underline; vertical-align: baseline;"&gt;Instant Snapshots&lt;/strong&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. This feature minimizes the risk of data loss from user error and bad upgrades with ultra-low RPO and RTO — creating a checkpoint is nearly instantaneous.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;How do I easily deploy my critical workload (e.g., MySQL) with resilience across multiple locations?&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; You can utilize &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;Hyperdisk HA. &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;This is a great fit for scenarios that require high availability and fast failover, such as SQL Server that leverages failover clustering. For such workloads, you can also choose our new capability with &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/hd-types/hyperdisk-balanced-ha"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Hyperdisk Balanced High Availability&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; with &lt;/span&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/sharing-disks-between-vms"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Multi-Writer support&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. This allows you to run clustered compute with workload-optimized storage in two zones with RPO=0 synchronous replication. &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;When a disaster occurs, how do I recover my workload elsewhere quickly and reliably, and run drills to confirm my recovery process?&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; Utilize our &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;disaster recovery capabilities with &lt;/strong&gt;&lt;a href="https://cloud.google.com/compute/docs/disks/hyperdisks#hd-sync-rep"&gt;&lt;strong style="text-decoration: underline; vertical-align: baseline;"&gt;Hyperdisk Async Replication&lt;/strong&gt;&lt;/a&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;which&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;enables cross-region continuous replication and recovery from a regional failure, with fast validation support for disaster recovery drills via cloning. Further, consistency group policies help ensure that workload data that’s distributed across multiple disks is recoverable when a workload needs to fail over between regions.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;In short, Hyperdisk provides a wealth of options to help you optimize your block storage to the needs of your workloads. Further, selecting the right Hyperdisk and leveraging features such as Storage Pools can help you lower your TCO and simplify management. To learn more, please visit our &lt;/span&gt;&lt;a href="https://cloud.google.com/products/block-storage?e=48754805&amp;amp;hl=en"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;website&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. For tailored recommendations, always consult your Google Cloud account team.&lt;/span&gt;&lt;/p&gt;
&lt;hr/&gt;
&lt;p role="presentation"&gt;&lt;sub&gt;&lt;em&gt;&lt;span style="vertical-align: baseline;"&gt;1. As of March 2025 based on published information for &lt;/span&gt;&lt;a href="https://docs.aws.amazon.com/ebs/latest/userguide/ebs-volumes-multi.html#considerations" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Amazon EBS&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, &lt;/span&gt;&lt;a href="https://learn.microsoft.com/en-us/azure/virtual-machines/disks-shared#ultra-disk-ranges" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Azure managed disks&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;br/&gt;2. &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;As of May 2025, compared to &lt;/span&gt;&lt;a href="https://aws.amazon.com/ebs/general-purpose/" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Amazon EBS gp3&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; volumes max iops/volume&lt;br/&gt;3. &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;As of March 2025, at list price, 50 to 150 TiB, peak IOPS of 25K to 75K and 25% compressibility, compared to &lt;/span&gt;&lt;a href="https://aws.amazon.com/ebs/general-purpose/" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Amazon EBS gp3&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; volumes.&lt;br/&gt;&lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;4. As of March 2025, based on internal Google benchmarking, compared to Rapid Storage, GCSFuse with Anywhere Cache, Parallelstore and Lustre for larger node sizes. &lt;br/&gt;5. As of March 2025 based on published performance for &lt;a href="https://learn.microsoft.com/en-us/azure/virtual-machines/disks-types" rel="noopener" target="_blank"&gt;Microsoft Azure Ultra SSD&lt;/a&gt; and &lt;a href="https://aws.amazon.com/ebs/features/" rel="noopener" target="_blank"&gt;Amazon EBS io2 BlockExpress&lt;/a&gt;&lt;/span&gt;&lt;/em&gt;&lt;/sub&gt;&lt;/p&gt;
&lt;p role="presentation"&gt;&lt;em&gt;&lt;sup&gt;The authors would like to thank David Seidman and Ruwen Hess for their contributions on this blog.&lt;/sup&gt;&lt;/em&gt;&lt;/p&gt;&lt;/div&gt;</description><pubDate>Wed, 11 Jun 2025 16:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/storage-data-transfer/how-to-choose-the-right-hyperdisk-block-storage-for-your-use-case/</guid><category>Compute</category><category>Infrastructure Modernization</category><category>Storage &amp; Data Transfer</category><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>Selecting the right Hyperdisk block storage for your workloads</title><description></description><site_name>Google</site_name><url>https://cloud.google.com/blog/products/storage-data-transfer/how-to-choose-the-right-hyperdisk-block-storage-for-your-use-case/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Ben Gitenstein</name><title>Group Product Manager, Google</title><department></department><company></company></author><author xmlns:author="http://www.w3.org/2005/Atom"><name>Sai Gopalan</name><title>Product Management, Google Cloud</title><department></department><company></company></author></item><item><title>Google Cloud’s open lakehouse: Architected for AI, open data, and unrivaled performance</title><link>https://cloud.google.com/blog/products/data-analytics/extending-the-google-data-cloud-lakehouse-architecture/</link><description>&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;The Google Data Cloud is a uniquely integrated platform built on Google’s planet-scale infrastructure, infused with AI, and features an &lt;/span&gt;&lt;a href="https://cloud.google.com/solutions/data-lakehouse"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;open lakehouse architecture&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; for multimodal data. Already, organizations like Snap Inc. credit Google's Data Cloud and open lakehouse architecture with empowering their data engineers and data scientists to do more with their data assets.&lt;/span&gt;&lt;/p&gt;
&lt;p style="padding-left: 40px;"&gt;&lt;span style="font-style: italic; vertical-align: baseline;"&gt;“Partnering with Google Cloud has been instrumental in our journey to build Snap's next-generation, open lakehouse and democratize Spark and Iceberg in our developer community!" &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;- Zhengyi Liu, Senior Manager - Software Engineering, Snap Inc.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Today, we’re excited to announce a series of innovations to our AI-powered lakehouse that sets a new standard for openness, intelligence, and performance. These innovations include:&lt;/span&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;BigLake Iceberg native storage:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; leverages Google’s Cloud Storage (GCS) to provide an enterprise-grade experience for managing and interoperating with Iceberg data. This includes &lt;/span&gt;&lt;a href="https://cloud.google.com/biglake"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;BigLake tables for Apache Iceberg&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; (GA) and &lt;/span&gt;&lt;a href="https://cloud.google.com/bigquery/docs/about-blms"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;BigLake Metastore&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; with a new REST Catalog API (Preview).&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;United operational and analytical engines:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; building on the BigLake foundation, customers can seamlessly interoperate on the same Iceberg open data foundation using BigQuery for analytical workloads (GA) and AlloyDB for PostgreSQL (Preview) to target operational needs.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Performance acceleration for BigQuery SQL: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;delivering a suite of automated SQL engine enhancements for significantly faster and more agile data processing, featuring the BigQuery advanced runtime, a low-latency query API, column metadata indexing, and an order of magnitude speedup for fine-grained updates/deletes.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;High-performance Lightning Engine for Apache Spark: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;our new &lt;/span&gt;&lt;a href="https://cloud.google.com/products/lightning-engine"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Lightning Engine&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; (Preview) is designed to supercharge &lt;/span&gt;&lt;a href="https://cloud.google.com/products/serverless-spark"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Apache Spark&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, leveraging optimized data connectors, efficient columnar shuffle operations, in-built caching, and vectorized execution.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;Dataplex Universal Catalog: &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;extends AI-powered intelligence and unified governance across the Google Cloud data estate by automatically discovering and organizing metadata from data to AI (including BigLake Iceberg, BigQuery, Spanner, Vertex AI models), enabling central policy enforcement via BigLake, and supporting AI-driven curation, data insights and semantic search. &lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li aria-level="1" style="list-style-type: disc; vertical-align: baseline;"&gt;
&lt;p role="presentation"&gt;&lt;strong style="vertical-align: baseline;"&gt;AI-native notebooks and tooling:&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; developer experiences are improved with Gemini-powered notebooks, PySpark code generation, and code extensions for JupyterLab and Visual Studio Code. Additionally, third-party notebook interfaces now offer enhanced and integrated experiences.&lt;/span&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;&lt;/div&gt;
&lt;div class="block-image_full_width"&gt;






  
    &lt;div class="article-module h-c-page"&gt;
      &lt;div class="h-c-grid"&gt;
  

    &lt;figure class="article-image--large
      
      
        h-c-grid__col
        h-c-grid__col--6 h-c-grid__col--offset-3
        
        
      "
      &gt;

      
      
        
        &lt;img
            src="https://storage.googleapis.com/gweb-cloudblog-publish/images/image1_2AhedOI.max-1000x1000.png"
        
          alt="image1"&gt;
        
        &lt;/a&gt;
      
    &lt;/figure&gt;

  
      &lt;/div&gt;
    &lt;/div&gt;
  




&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Let's explore these new innovations.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Expanded BigLake services: Open, unified, and interoperable&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;We are actively reimagining &lt;/span&gt;&lt;a href="https://cloud.google.com/biglake"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;BigLake&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; into a comprehensive storage runtime for Google Data Cloud using &lt;/span&gt;&lt;a href="https://cloud.google.com/storage"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Google's Cloud Storage&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;. This approach lets you build open, managed and high-performance lakehouses that span Google native storage and data stored in open formats. As part of BigLake, we are announcing our new Iceberg native storage, which provides enterprise-grade support for Iceberg on Google’s Cloud Storage through &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;BigLake tables for Apache Iceberg (GA)&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;. BigLake natively supports Google’s Cloud Storage management capabilities and extends these to Iceberg data, enabling you to use storage &lt;/span&gt;&lt;a href="https://cloud.google.com/storage/docs/autoclass"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Autoclass&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; for efficient data tiering to colder storage classes and apply customer-managed encryption keys (CMEK) to your storage buckets. BigLake is also natively supported in our Dataplex Universal Catalog, helping to ensure that centralized governance is consistently enforced across your entire data estate.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Underlying BigLake, the new &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;BigLake Metastore (GA) &lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;with an&lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt; Apache Iceberg REST Catalog API (Preview)&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;, allows you to achieve true openness and interoperability across your data ecosystem while simplifying management and governance. BigLake metastore is built on Google’s planet-scale infrastructure, offering a unified, managed, serverless, and scalable offering, bringing together enterprise metadata that spans BigQuery, Iceberg native storage, and self managed open formats to support analytics, operational querying, streaming, and AI. The BigLake solution enables universal engine interoperability, supporting a range of query engines — including first-party Google Cloud services such as BigQuery, AlloyDB, and &lt;/span&gt;&lt;a href="https://cloud.google.com/products/serverless-spark"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Google Cloud Serverless for Apache Spark&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;, as well as third party and open-source engines— to consistently operate on Iceberg data managed by BigLake. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;In addition, it is now easier than ever to bring data into the Iceberg native storage through our enhanced Migration Services that feature automated &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;Iceberg table and metadata migration from Hadoop/Cloudera (Preview)&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; and a push-button &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;Delta to Iceberg service (Preview)&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;
&lt;div class="block-aside"&gt;&lt;dl&gt;
    &lt;dt&gt;aside_block&lt;/dt&gt;
    &lt;dd&gt;&amp;lt;ListValue: [StructValue([(&amp;#x27;title&amp;#x27;, &amp;#x27;$300 in free credit to try Google Cloud data analytics&amp;#x27;), (&amp;#x27;body&amp;#x27;, &amp;lt;wagtail.rich_text.RichText object at 0x7f3f90b2d2b0&amp;gt;), (&amp;#x27;btn_text&amp;#x27;, &amp;#x27;Start building for free&amp;#x27;), (&amp;#x27;href&amp;#x27;, &amp;#x27;http://console.cloud.google.com/freetrial?redirectPath=/bigquery/&amp;#x27;), (&amp;#x27;image&amp;#x27;, None)])]&amp;gt;&lt;/dd&gt;
&lt;/dl&gt;&lt;/div&gt;
&lt;div class="block-paragraph_advanced"&gt;&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Analytical and operational engines unite on open data&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;When you need to perform deep analytics, BigQuery can now read and write Iceberg data using BigLake tables for Apache Iceberg. BigQuery further enhances Iceberg tables with features traditionally associated with proprietary data warehouses, offering high-throughput streaming for zero-latency queries, enhanced table management with automatic data reclustering, and the ability to build advanced ETL use cases with support for &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;multi-table transactions (Preview)&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;. In addition, you can leverage BigQuery’s built-in AI capabilities (BQML, AI Query Engine, multimodal analysis) directly on your open datasets. Through this integration, you benefit from the openness and data ownership associated with native Iceberg storage, while simultaneously gaining access to BigQuery's expansive capabilities. In fact, customer adoption of BigLake Iceberg usage with BigQuery has grown nearly 3x in 18 months, now managing hundreds of petabytes.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Unified data management extends beyond analytics into the operational heart of your business, with AlloyDB for PostgreSQL, our high-performance operational database, which can now natively query the same BigLake-managed Iceberg data. Now, your operational applications can tap into the richness of BigLake without complex ETL, and you can apply AlloyDB AI capabilities such as semantic search and natural language querying to your Iceberg data.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Customers like &lt;/span&gt;&lt;a href="https://cloud.google.com/blog/products/databases/bayer-uses-alloydb?e=48754805"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Bayer&lt;/span&gt;&lt;/a&gt; &lt;span style="vertical-align: baseline;"&gt;modernized their data cloud to store and analyze vast amounts of observational data using a combination of AlloyDB and BigQuery. They use BigQuery to produce real-time analytics and insights which are operationalized by AlloyDB, delivering 50% better response rates and 5x more throughput than their previous solution. &lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Unleashing high-performance BigQuery SQL and serverless Spark on open data &lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;We’re also excited to deliver new high-performance data processing, so that all data can be activated quickly and intelligently. We continue to innovate on &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;BigQuery's SQL engine&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; with a suite of unique, automated performance enhancements. The &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;BigQuery advanced runtime (Preview)&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;, can automatically accelerate analytical workloads, using enhanced vectorization and short query optimized mode, without requiring any user action or code changes. This is complemented by the &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;BigQuery API &lt;/strong&gt;&lt;a href="https://cloud.google.com/bigquery/docs/running-queries#optional-job-creation"&gt;&lt;strong style="text-decoration: underline; vertical-align: baseline;"&gt;optional job creation mode&lt;/strong&gt;&lt;/a&gt;&lt;strong style="vertical-align: baseline;"&gt; (GA)&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;, which optimizes query paths for short-duration, interactive queries, reducing latency. Further query efficiency is unlocked by the &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;BigQuery column metadata index (CMETA) (GA)&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;, which helps process queries on large tables through more efficient, system-managed data pruning. Other architectural improvements also mean that &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;BigQuery fine-grained updates/deletes (Preview)&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; now operate an order of magnitude faster, increasing agility for large-scale data operations, including on open formats.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Simultaneously, we’re launching an accelerated Apache Spark experience with our new &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;Lightning Engine (Preview)&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; for Apache Spark. The Lightning Engine accelerates Apache Spark performance through highly optimized data connectors for Cloud Storage and BigQuery storage, efficient columnar shuffle operations, and intelligent in-built caching mechanisms. Furthermore, our Lightning Engine leverages vectorized execution built with native C++ libraries (Velox and Gluten), optimized for Apache Spark. This powerful combination delivers &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;3.6x faster Spark performance for TPC-H like benchmarks&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;. In addition, our Spark offering is AI/ML-ready, providing pre-packaged AI libraries, updated ML runtimes, and easy GPU support, establishing Apache Spark–available via our Google Cloud Serverless for Apache Spark offering or via &lt;/span&gt;&lt;a href="https://cloud.google.com/dataproc"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Dataproc&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; cluster deployments–as a first-class, high-performance citizen in a Google Data Cloud lakehouse environment.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Dataplex Universal Catalog: AI-powered intelligence across Google Cloud&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;An effective AI-driven data strategy hinges on having an intelligent and active universal catalog that can operate at any scale. This is what &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;Dataplex Universal Catalog&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; now provides for the Google Data Cloud, transforming your entire distributed data estate into trusted, discoverable, and actionable resources.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;a href="https://cloud.google.com/dataplex"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;Dataplex Universal Catalog&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; automatically discovers, understands, and organizes metadata across your whole analytical and operational landscape. This comprehensive view now includes BigLake-native Iceberg storage, other open formats like Delta and Hudi on Cloud Storage, analytical data in BigQuery, transactional data from databases like Spanner, and metadata from machine learning models in Vertex AI—showcasing pervasive governance across Google's Data Cloud.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;This is also integral to the lakehouse by enabling users to define governance policies centrally and enforce them consistently across multiple data engines through BigLake. This integration supports fine-grained access controls and strengthens governance, across all engines of choice in Google’s Data Cloud. The BigLake solution supports credential vending, which allows users to securely extend centrally defined policies all the way to data in Cloud Storage. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Dataplex Universal Catalog is powered by AI, with a Gemini-enhanced knowledge graph, transforming metadata into dynamic, actionable intelligence. Here, AI automates metadata curation, infers hidden relationships between data elements, proactively recommends insights from data backed by complex queries, and enables semantic search with natural language. It also fuels new AI-powered experiences and autonomous agents. For instance, Gemini-powered assistance using Dataplex Universal Catalog shows 50% greater precision in identifying datasets, significantly accelerating insights. Dataplex Universal Catalog is also the foundation of an open ecosystem with seamless metadata federation to platforms like Collibra, and ensures broad connectivity through Dataplex Universal Catalog &lt;/span&gt;&lt;a href="https://cloud.google.com/dataplex/docs/reference/rest"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;APIs&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;
&lt;h3&gt;&lt;strong style="vertical-align: baseline;"&gt;Empowering practitioners with AI-native notebooks and tooling&lt;/strong&gt;&lt;/h3&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;At Google Cloud, our goal is to revolutionize the data practitioner's experience by embedding sophisticated AI and lakehouse integrations directly into their preferred tools and workflows. This commitment to an open, flexible, and intelligent environment lets data scientists, engineers, and analysts unlock new levels of productivity and innovation.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Making this possible are our next-gen, &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;AI-native BigQuery Notebooks&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;, which offer a unified and interoperable development experience across SQL, Python, and Apache Spark. This experience is enhanced by deeply embedded Gemini assistive capabilities. Gemini acts as an intelligent collaborator, offering advanced &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;PySpark code generation&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;, insightful explanations of complex code, and direct integration with &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;Cloud Assist Investigations for serverless Spark troubleshooting (Preview)&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;, dramatically reducing development friction and accelerating the path from data to insight. &lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Furthermore, new &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;JupyterLab and Visual Studio Code extensions for BigQuery, Dataproc and Google Cloud Serverless for Apache Spark (Preview)&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt; allow developers to connect to Google Cloud's open lakehouse capabilities directly from their preferred IDEs with minimal setup. Users can start developing within minutes with access to all their lakehouse datasets and files in their preferred tool, supporting their end-to-end journey from development to deployment. The consumption of notebooks using serverless Spark more than &lt;/span&gt;&lt;strong style="vertical-align: baseline;"&gt;quadrupled from Q1 2024 to Q1 2025&lt;/strong&gt;&lt;span style="vertical-align: baseline;"&gt;.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;span style="vertical-align: baseline;"&gt;Together, these integrated advancements help deliver an adaptable, intelligent, high-performance Data Cloud anchored on the lakehouse architecture, equipping organizations to connect all of their data to Google's AI, unlock its full potential, and define innovation in the AI era. Join our &lt;/span&gt;&lt;a href="https://cloudonair.withgoogle.com/events/lakehouse-live" rel="noopener" target="_blank"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;virtual customer event on 5/29&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; to learn more about these exciting innovations. &lt;/span&gt;&lt;span style="vertical-align: baseline;"&gt;Click here to &lt;/span&gt;&lt;a href="https://cloud.google.com/biglake?e=48754805&amp;amp;hl=en"&gt;&lt;span style="text-decoration: underline; vertical-align: baseline;"&gt;learn more&lt;/span&gt;&lt;/a&gt;&lt;span style="vertical-align: baseline;"&gt; and sign up for early access to these new capabilities. We're excited to see the solutions you'll build.&lt;/span&gt;&lt;/p&gt;&lt;/div&gt;</description><pubDate>Wed, 28 May 2025 16:00:00 +0000</pubDate><guid>https://cloud.google.com/blog/products/data-analytics/extending-the-google-data-cloud-lakehouse-architecture/</guid><category>Storage &amp; Data Transfer</category><category>Streaming</category><category>Data Analytics</category><og xmlns:og="http://ogp.me/ns#"><type>article</type><title>Google Cloud’s open lakehouse: Architected for AI, open data, and unrivaled performance</title><description></description><site_name>Google</site_name><url>https://cloud.google.com/blog/products/data-analytics/extending-the-google-data-cloud-lakehouse-architecture/</url></og><author xmlns:author="http://www.w3.org/2005/Atom"><name>Andi Gutmans</name><title>VP/GM, Data Cloud, Google Cloud</title><department></department><company></company></author><author xmlns:author="http://www.w3.org/2005/Atom"><name>Yasmeen Ahmad</name><title>Managing Director, Data Cloud, Google Cloud</title><department></department><company></company></author></item></channel></rss>