Listen to this story
Oracle has announced a multi-year agreement with Microsoft to bolster the growth of AI services. Microsoft has adopted Oracle Cloud Infrastructure (OCI) AI infrastructure, alongside Microsoft Azure AI infrastructure, to power daily Bing conversational searches.
This utilisation of AI infrastructure is enabled through Oracle’s Interconnect for Microsoft Azure, and Microsoft leverages managed services like Azure Kubernetes Service (AKS) to orchestrate OCI Compute on a massive scale.
The development aims to support the increasing demand for Bing conversational search, which relies on robust computing infrastructure clusters to evaluate and analyse search results using Bing’s inference model.
Subscribe to our Newsletter
Join our editors every weekday evening as they steer you through the most significant news of the day, introduce you to fresh perspectives, and provide unexpected moments of joy
“Generative AI is a monumental technological leap, and Oracle is facilitating Microsoft and numerous other businesses in creating and running new products with our OCI AI capabilities,” stated Karan Batta, senior vice president of Oracle Cloud Infrastructure. He added, “Our extended collaboration with Microsoft is set to deliver innovative experiences to a global audience.”
Divya Kumar, global head of marketing for Search & AI at Microsoft, emphasised the enhancement this collaboration brings to Microsoft Bing’s search experience: “Microsoft Bing is leveraging the latest advancements in AI to provide a dramatically better search experience for people across the world. Our collaboration with Oracle and the use of Oracle Cloud Infrastructure, in conjunction with Microsoft Azure AI infrastructure, will broaden customer access and enhance the speed of many search results.”
To facilitate inference models, thousands of compute and storage instances and tens of thousands of GPUs operate in parallel as a single supercomputer over a multi-terabit network.
Oracle’s OCI Superclusters offer OCI Compute Bare Metal instances, ultra-low latency RDMA cluster networking, and a choice of HPC storage.
These superclusters can scale up to 4,096 OCI Compute Bare Metal instances with 32,768 A100 GPUs or 16,384 H100 GPUs, coupled with petabytes of high-performance clustered file system storage, ensuring efficient processing of massively parallel applications.