NVIDIA is asserting its model new AI Foundry Companies initiative to additional speed up AI whereas Microsoft Azure affords the newest Hopper cases.
NVIDIA AI Foundry Companies & Microsoft Azure’s Hopper H100/H200 Situations Propel AI To New Heights
Press Launch: NVIDIA introduced an AI foundry service, a set of NVIDIA AI Basis Fashions, NVIDIA NeMo framework and instruments, and DGX Cloud AI supercomputing and companies that give enterprises an end-to-end answer for creating and optimizing customized
generative AI fashions.
Utilizing the AI foundry service, Amdocs, a number one supplier of software program and companies for communications and media suppliers, will optimize enterprise-grade massive language fashions for the telco and media industries to effectively deploy generative AI use instances throughout their companies, from buyer experiences to community operations and provisioning. The LLMs will run on NVIDIA accelerated computing as a part of the Amdocs amAIz framework.
The collaboration builds on the beforehand introduced Amdocs-Microsoft partnership, enabling service suppliers to undertake these purposes in safe, trusted environments, together with on-premises and within the cloud.
New NVIDIA H100 and H200 Tensor Core GPU Situations Coming to Microsoft Azure to Speed up AI Workloads
As NVIDIA continues to collaborate with Microsoft to construct state-of-the-art AI infrastructure, Microsoft is introducing extra H100-based digital machines to Microsoft Azure to speed up (mid-range) AI workloads.
At its Ignite convention in Seattle as we speak, Microsoft introduced its new NC H100 v5 VM sequence for Azure, the trade’s first cloud cases that includes NVIDIA H100 NVL GPUs. This providing brings collectively a pair of PCIe-based H100 GPUs related by way of NVIDIA NVLink, with almost 4 petaflops of AI compute and 188GB of sooner HBM3 reminiscence. The NVIDIA H100 NVL GPU can ship as much as 12x greater efficiency on GPT-3 175B over the earlier technology and is good for inference and mainstream coaching workloads.
Moreover, Microsoft introduced plans so as to add the NVIDIA H200 Tensor Core GPU to its Azure fleet subsequent 12 months to help bigger mannequin inferencing with no discount in latency. This new providing is purpose-built to speed up the biggest AI workloads, together with LLMs and generative AI fashions. The H200 GPU brings dramatic will increase each in reminiscence capability and bandwidth utilizing the latest-generation HBM3e reminiscence.
In comparison with its predecessor, this new GPU will supply 141GB of HBM3e reminiscence (1.8x extra) and 4.8 TB/s of peak reminiscence bandwidth (a 1.4x improve).
Cloud Computing Will get Confidential
Additional increasing the supply of NVIDIA-accelerated generative AI computing for Azure prospects, Microsoft introduced one other NVIDIA-powered occasion: the NCC H100 v5.
These Azure confidential digital machines (VMs) with NVIDIA H100 Tensor Core GPUs permit Azure prospects to guard the confidentiality and integrity of their knowledge and purposes in use, in reminiscence, whereas accessing the unsurpassed acceleration of H100 GPUs. These GPU-enhanced confidential VMs will likely be coming quickly to non-public preview.