Oracle and NVIDIA Continue to Expand Partnership to Deliver Sovereign AI Worldwide

Oracle and NVIDIA are expanding their collaboration to deliver sovereign AI solutions to customers around the world. Oracle’s distributed cloud, AI infrastructure, and generative AI services, combined with NVIDIA’s accelerated computing and generative AI software, are enabling governments and enterprises to deploy AI factories, according to the companies.

These AI factories can run cloud services locally, and within a country’s or organization’s secure premises with a range of operational controls, supporting sovereign goals of diversifying and boosting economic growth.

“As AI reshapes business, industry, and policy around the world, countries and organizations need to strengthen their digital sovereignty in order to protect their most valuable data,” said Safra Catz, CEO of Oracle. “Our continued collaboration with NVIDIA and our unique ability to deploy cloud regions quickly and locally will ensure societies can take advantage of AI without compromising their security.”

According to the companies, the combination of NVIDIA’s full-stack AI platform with Oracle’s Enterprise AI—deployable across OCI Dedicated Region, Oracle Alloy, Oracle EU Sovereign Cloud, and Oracle Government Cloud—offers customers a state-of-the-art AI solution that provides greater control over operations, location, and security to help support digital sovereignty.

“In an era where innovation will be driven by generative AI, data sovereignty is a cultural and economic imperative,” said Jensen Huang, founder and CEO of NVIDIA. “Oracle’s integrated cloud applications and infrastructure, combined with NVIDIA accelerated computing and generative AI services, create the flexibility and security nations and regions require to control their own destiny.”

The companies’ combined offerings can be deployed via the public cloud or in a customer’s data center in specific locations, with flexible operational controls. Oracle is the only hyperscaler capable of delivering AI and full cloud services locally, anywhere. OCI services and pricing are consistent across deployment types to simplify planning, portability, and management.

Oracle’s cloud services leverage a range of NVIDIA’s stack, including NVIDIA accelerated computing infrastructure and the NVIDIA AI Enterprise software platform, including newly announced NVIDIA NIM™ inference microservices, which are built on the foundation of NVIDIA inference software such as NVIDIA TensorRT, NVIDIA TensorRT-LLM, and NVIDIA Triton Inference Server.

To help customers address the ever-increasing needs of AI models, Oracle plans to take advantage of the latest NVIDIA Grace Blackwell computing platform, announced today at GTC, across OCI Supercluster and OCI Compute. OCI Supercluster will become significantly faster with new OCI Compute bare metal instances, ultra-low-latency RDMA networking, and high-performance storage. OCI Compute will adopt both the NVIDIA GB200 Grace Blackwell Superchip and the NVIDIA Blackwell B200 Tensor Core GPU.

The NVIDIA GB200 Grace Blackwell Superchip will power a new era of computing. GB200 delivers up to 30X faster real-time large language model (LLM) inference, 25X lower TCO, and requires 25X less energy compared to the previous generation of GPUs, supercharging AI training, data processing, and engineering design and simulation. NVIDIA Blackwell B200 Tensor Core GPUs are designed for the most demanding AI, data analytics, and high-performance computing (HPC) workloads.

NVIDIA NIM and CUDA-X microservices, including NVIDIA NeMo Retriever for retrieval augmented generation (RAG) inference deployments, will also help OCI customers bring more insight and accuracy to their generative AI copilots and other productivity tools using their own data.

Oracle and NVIDIA’s sovereign AI solutions are available immediately.

For more information about this news, visit