
Zilliz Brings BYOC to Azure, Completing Three-Cloud Coverage
Context and Chronology
Zilliz has enabled its managed vector database to run inside customer cloud accounts on AWS, GCP, and now Azure, closing a strategic gap for organizations that require data residency and tight compliance controls. This release includes an infrastructure-as-code integration via a Terraform provider to automate deployments into corporate subscriptions, letting teams keep billing and governance within their existing contracts. For enterprises standardizing on Microsoft's AI stack, the new option reduces the need to shuttle vectors across cloud boundaries and aligns vector storage with Azure-hosted model runtimes.
Operationally, deploying the vector service inside a customer's tenancy shifts operational responsibility: Zilliz manages the application layer while compute, storage, and network costs flow through the customer's cloud account. That arrangement preserves enterprise reserved capacity and licensing benefits and removes a common procurement blocker during pilots that would otherwise require data export approvals. The product also includes a set of migration paths from several competing vector and search stores to ease switching at scale.
Strategically, completing three-cloud BYOC positions Zilliz as a cross-cloud enabler for retrieval-augmented workloads, giving engineering teams a single platform for vector index management without forcing a vendor-held data plane. Mr. Xie framed the release as removing a key barrier to adoption; the technical reality is that vendors that cannot offer in-account hosting now face a harder sales pitch to regulated customers. Cloud providers gain larger downstream compute consumption as vector operations remain local to customer accounts, while Zilliz gains distribution and lock-in through deep integration with native cloud toolchains.
Not all benefits are absolute: latency and cost gains depend on where model inference runs relative to vector storage, and cross-region replication or multi-account topologies will reintroduce networking trade-offs. Governance teams still need runtime controls for model access and query telemetry even when data never leaves a tenant, and procurement must reconcile managed service convenience with cloud resource governance. Expect most enterprise pilots to migrate to in-account managed vectors when regulatory clarity and predictable cost models are required.
Read Our Expert Analysis
Create an account or login for free to unlock our expert analysis and key takeaways for this development.
By continuing, you agree to receive marketing communications and our weekly newsletter. You can opt-out at any time.
Recommended for you

Perplexity Turns to Microsoft Azure for AI Hosting as Tensions with Amazon Flare
Perplexity has formalized a hosting arrangement with Microsoft Azure to support its AI services while navigating a public fracas with Amazon’s cloud unit. The move underscores a broader industry trend—hyperscalers pairing capital, privileged hosting and commercial ties to shape access to leading models, raising scrutiny over lock‑in and interoperability.
Regolo launches European data path to blunt CLOUD Act exposure
Rising U.S. data-compulsion risk and new state AI rules are forcing firms to rethink cloud jurisdiction and data flows. Regolo offers an EU-hosted, zero-data-retention routing layer to reduce CLOUD Act reach and to complement sovereign-region strategies from incumbents such as Genesys and hyperscalers.

Private cloud regains ground as AI reshapes cloud cost and risk calculus
Enterprises are pushing persistent inference, embedding caches, and retrieval layers into private or localized clouds to tame rising AI inference costs, latency and correlated outage risk, while keeping burst training and large-scale experimentation in public clouds. This hybrid posture is reinforced by shifts in data architecture toward projection-first stores, growing endpoint inference capability, and silicon-market dynamics that favor bespoke, on-prem stacks.

Amazon and Prosus Strike AI Cloud Agreement to Secure Double-Digit Cost Reductions
Amazon Web Services has reached a commercial cloud agreement with Prosus to support its AI workloads, targeting double-digit percentage savings on infrastructure costs. The deal signals continued vendor consolidation for large-scale AI deployments and reinforces AWS’s position as the dominant supplier for enterprise generative-AI projects.

Genesys lines up support for AWS European Sovereign Cloud to address EU data controls
Genesys will make its customer engagement platform available on the AWS European Sovereign Cloud so organisations can keep data and operational control inside EU boundaries. The move targets regulated buyers concerned about cross-border legal access and signals wider momentum for sovereign-region cloud offerings in Europe.

Coinerella Rebuilds Platform on European Cloud Providers
Coinerella migrated its stack to European providers to regain data locality and lower infrastructure spend while accepting more operational responsibility. The shift exposes integration gaps, tooling shortfalls, and a clear path for platform engineering and sovereign-cloud economics to scale.

Coveo launches hosted MCP server to bridge enterprise content and major LLMs
Coveo released a hosted implementation of the Model Context Protocol to let large language models query enterprise content indexes while preserving security and governance. The offering is generally available for major commercial LLMs, is already in use by early customers, and queries count toward existing consumption-based licensing.

Cloud giants' hardware binge tightens markets and nudges users toward rented AI compute
Major cloud providers are concentrating purchases of GPUs, high-density DRAM and related components to support AI workloads, creating retail shortages and higher prices that push smaller buyers toward rented compute. Rapid datacenter buildouts, permitting and power constraints, and changes in supplier allocation and financing compound the risk that scarcity will be monetized into long-term service revenue and reduced market choice.