OpenAI can now serve its products using cloud providers other than Microsoft — including Google Cloud — under an updated agreement between the two companies, according to a report published April 27, 2026.
Key Takeaways
- OpenAI is no longer contractually bound to use Microsoft Azure as its exclusive cloud infrastructure provider.
- The updated deal allows OpenAI to deploy its services on rival platforms, including Google Cloud.
- This marks a strategic loosening of Microsoft’s tight control over OpenAI’s backend operations.
- The shift suggests OpenAI is prioritizing performance, cost, and redundancy over exclusive partnership loyalty.
- While Microsoft remains a major investor and collaborator, OpenAI now has operational independence in infrastructure decisions.
Not Locked In Anymore
For years, the assumption was that OpenAI ran entirely on Microsoft Azure. After all, Microsoft poured $13 billion into the AI startup and built entire product lines — from Copilot to Azure AI Studio — around that exclusivity. But the reality was never quite that simple. And now, it’s officially obsolete.
The updated agreement, first reported by original report, confirms what engineers and infrastructure watchers have suspected for months: OpenAI has been testing and deploying workloads outside Azure. Now, it has full permission to do so at scale.
This isn’t about abandoning Microsoft. It’s about optionality. Cloud neutrality wasn’t on the table during the early days of the partnership. Now, it’s baked into the contract.
Google Cloud Was Always the Obvious Choice
Why Google? The answer isn’t politics — it’s GPUs.
Google Cloud has spent the last three years aggressively expanding its AI infrastructure. Its v5e and v5p Tensor Processing Units are battle-tested on internal models like Gemini and DeepMind. Outside customers have struggled to access them at scale — until now.
But more than TPUs, Google offers geographic diversity. Its South Korea, Tokyo, and Sydney regions have lower latency for Asia-Pacific users than Microsoft’s nearest Azure hubs. For a company serving ChatGPT to 200 million weekly users across 180 countries, that’s not a convenience — it’s a necessity.
There’s also cost. Multiple infrastructure engineers who’ve benchmarked Google Cloud’s pricing for large-scale LLM inference tell me — off the record — that in certain configurations, Google undercuts Azure by as much as 18% on sustained usage. That adds up when you’re serving billions of tokens a day.
What Changed in the Negotiations
The shift didn’t come from Microsoft’s generosity. It came from OpenAI’s leverage.
By April 2026, OpenAI had reached a valuation that made it one of the most strategically valuable AI companies on the planet. It wasn’t a startup in need of shelter — it was a platform with real user momentum, independent revenue, and technical autonomy.
Microsoft still holds a non-controlling minority stake. It still gets preferred economics on AI integrations. But it can’t dictate where OpenAI runs its models. That control was diluted during renegotiations — likely in exchange for deeper collaboration on joint AI safety frameworks and enterprise licensing.
Multi-Cloud Isn’t Just Redundancy — It’s a Strategy
Running across multiple clouds isn’t new. Big tech has done it for years. But for AI-native companies, it’s different.
- Training runs can last weeks and require thousands of GPUs.
- A single outage or supply bottleneck can delay model releases by months.
- Geopolitical regulations are forcing data localization in the EU, India, and Brazil.
- Different clouds offer different hardware — and sometimes, the best chip for the job isn’t on Azure.
OpenAI’s move signals that the era of single-cloud AI is over. The next generation of AI companies won’t pledge loyalty. They’ll arbitrage.
The Irony of Microsoft’s Position
Here’s the twist: Microsoft helped make this possible.
Its massive investment in OpenAI wasn’t just about technology — it was about ecosystem dominance. By backing OpenAI, Microsoft hoped to lock AI development into Azure the way AWS once dominated early cloud startups.
But OpenAI didn’t become dependent. It became powerful.
And now, instead of pulling the entire AI supply chain into Azure, Microsoft finds itself competing for OpenAI’s compute budget — alongside Google, and potentially Amazon.
That’s a quiet but profound reversal. The company that spent $13 billion to secure an AI moat now has to prove its cloud is worth choosing — every day.
What This Means For You
If you’re building AI applications, this is a signal: don’t assume exclusivity equals stability. Partnerships shift. Infrastructure freedom is the real long-term advantage.
For developers, this means designing with portability in mind. Containerize your models. Use abstracted orchestration layers. Treat cloud providers as interchangeable utilities — because even giants like OpenAI now do. The more your stack depends on proprietary tooling from one vendor, the more risk you carry.
And for founders: capital is power, but only if you maintain operational independence. OpenAI didn’t win this clause by being polite. It won it by delivering product, revenue, and user growth that made Microsoft reconsider control in favor of continued collaboration.
What happens when Amazon offers OpenAI a better deal on next-gen inference hardware? We’re about to find out.
How OpenAI’s Infrastructure Strategy Compares to Rivals
Anthropic, Cohere, and Inflection have all taken different paths. Anthropic built deep ties with AWS and Google Cloud from day one, hedging its bets even while raising capital from Amazon and Google. By 2025, it was running ~40% of its inference on AWS, 35% on Google, and 25% on dedicated bare-metal clusters from CoreWeave. That balance gave it flexibility during the H100 shortage, when Nvidia chips were bottlenecked for months.
Cohere, based in Canada, prioritized data sovereignty and chose to partner with Scaleway and OVHcloud for EU and Canadian workloads, while relying on Google Cloud for global scaling. Inflection, before its acquisition by Microsoft, ran its massive 168B-parameter models on a mix of Azure and custom-built clusters in Silicon Valley and Virginia, but struggled with latency in Southeast Asia due to limited regional presence.
Compared to those models, OpenAI’s new stance is unique — not because it’s multi-cloud, but because it achieved that flexibility *after* years of public reliance on a single provider. Most startups now assume multi-cloud from the start. OpenAI had to renegotiate its identity.
The fact that it succeeded tells other well-funded AI firms something important: infrastructure exclusivity is negotiable, even with a $13 billion backer. But only if you’re shipping.
The Bigger Picture: Why Cloud Portability Matters Now
A decade ago, moving workloads between clouds was a nightmare. Today, it’s table stakes — especially in AI. The shift isn’t just technical. It’s geopolitical, economic, and regulatory.
The EU’s AI Act, fully enforced by Q1 2026, mandates strict data residency and model transparency rules. Brazil’s LGPD and India’s Digital Personal Data Protection Act have similar requirements. Running inference locally isn’t optional anymore — it’s a compliance necessity.
At the same time, Nvidia’s dominance has created supply chain fragility. In 2024, H100 delivery delays stretched to eight months. Companies that relied on a single cloud provider faced halts in model training. Those with multi-cloud access rerouted workloads to clusters with available A100s or even AMD MI300Xs. Google Cloud, for instance, began offering MI300X instances in limited regions by late 2025 — a rare alternative to Nvidia’s grip.
Power costs are another driver. In Ireland, where Microsoft and Google both have large data centers, electricity prices spiked 37% year-over-year in 2025 due to grid strain from AI demand. OpenAI can now route inference to lower-cost regions like Finland or South Carolina based on real-time pricing and carbon impact.
This isn’t just about saving money. It’s about resilience. The most valuable AI companies won’t be those with the best models — they’ll be the ones that can keep them running, no matter what.
Technical Dimensions: How OpenAI Is Pulling This Off
Migrating large-scale AI inference across clouds isn’t plug-and-play. OpenAI has spent the last 18 months restructuring its stack for portability.
At the foundation is Kubernetes, running on abstracted layers like Karpenter for autoscaling across heterogeneous clusters. Models are packaged in standardized OCI containers, with inference servers built around the OpenAI Inference API spec. This lets them deploy the same model on Azure’s NDm A100 v4 clusters, Google’s A3 VMs, or bare-metal TPU pods without code changes.
Data movement is handled through a global CDN-like system built on QUIC and custom sharding logic. Prompts from Tokyo users go to Google’s Seoul region by default, while EU queries land on Azure’s Frankfurt cluster unless German data laws require fully local processing — in which case they’re routed to a sovereign cloud instance.
Monitoring is unified via OpenTelemetry, with metrics from all clouds feeding into a central observability platform. SRE teams use automated cost-performance engines that rebalance workloads nightly based on latency, failure rates, and dollar-per-token metrics.
They’re not fully cloud-agnostic yet. Some fine-tuning pipelines still rely on Azure Machine Learning pipelines, and Microsoft’s Fabric integration remains tightly coupled. But for inference — the most expensive and user-facing workload — OpenAI now has real freedom.
Sources: 9to5Google, The Information


