Next-gen Nvidia AI chips coming to cloud platforms - TechCentral

  • 📰 TechCentral
  • ⏱ Reading Time:
  • 87 sec. here
  • 3 min. at publisher
  • 📊 Quality Score:
  • News: 38%
  • Publisher: 71%

Technology Technology Headlines News

Technology Technology Latest News,Technology Technology Headlines

Next-gen Nvidia AI chips coming to cloud platforms

introduced new chips, supercomputing services and a raft of high-profile partnerships on Tuesday intended to showcase how its technology will fuel the next wave of AI breakthroughs.

At the chip maker’s annual developer conference on Tuesday, CEO Jensen Huang positioned Nvidia as the engine behind “the iPhone moment of AI”, as he’s taken to calling this inflection point in computing. Spurred by a boom in consumer and enterprise applications, such as advanced chatbots and eye-popping graphics generators, “generative AI will reinvent nearly every industry”, Huang said.

The idea is to build infrastructure that can make AI apps faster and more accessible to customers. Nvidia’s graphics processing units have become the brains behind ChatGPT and its ilk, helping them digest and process ever-greater sums of training data. Microsoft revealed last week it had to string together tens of thousands of Nvidia’s A100 GPUs in data centres in order to handle the computational workloads in the cloud for OpenAI, ChatGPT’s developer.

These kinds of chip superclusters are part of a push by Nvidia to rent out supercomputing services through a new program called DGX Cloud, hosted by Oracle and soon Microsoft Azure and Google Cloud. Nvidia said the goal is to make accessing an AI supercomputer as easy as opening a webpage, enabling companies to train their models without the need for on-premises infrastructure that’s costly to install and manage.

Nvidia also launched two new chips, one focused on enhancing AI video performance and the other an upgrade to the H100. The latter GPU is designed specifically to improve the deployment of large language models like those used by ChatGPT. Called the H100 NVL, it can perform 12 times faster when handling inferences — that is, how AI responds to real-life queries — compared to the prior generation of A100s at scale in data centres.

 

Thank you for your comment. Your comment will be published after being reviewed.
Please try again later.
We have summarized this news so that you can read it quickly. If you are interested in the news, you can read the full text here. Read more:

 /  🏆 8. in TECHNOLOGY

Technology Technology Latest News, Technology Technology Headlines

Similar News:You can also read news stories similar to this one that we have collected from other news sources.

Apple, Microsoft dominance on tech-stock indexes grows - TechCentralApple and Microsoft, already dominant in technology-stock indexes, are about to become even more so.
Source: TechCentral - 🏆 8. / 71 Read more »