Nvidia CEO
Jensen Huang said on Monday that the company’s next generation of chips is in “full production” saying they can deliver five times the artificial-intelligence computing of the company’s previous chips when serving up chatbots and other AI apps.In a speech at the
Consumer Electronics Show in
Las Vegas, the leader of the world’s most valuable company revealed new details about its chips, which will arrive later this year and which
Nvidia executives are in the company’s labs being tested by AI firms, as
Nvidia faces increasing competition from rivals as well as its own customers.The Vera Rubin platform, made up of six separate
Nvidia chips, is expected to debut later this year, with the flagship server containing 72 of the company’s graphics units and 36 of its new central processors.Huang showed how they can be strung together into “pods” with more than 1,000 Rubin chips and said they could improve the efficiency of generating what are known as “tokens” – the fundamental unit of AI systems – by 10 times.To get the new performance results, however, Huang said the Rubin chips use a proprietary kind of data that the company hopes the wider industry will adopt.“This is how we were able to deliver such a gigantic step up in performance, even though we only have 1.6 times the number of transistors,” Huang said.While
Nvidia still dominates the market for training AI models, it faces far more competition – from traditional rivals such as
Advanced Micro Devices as well as customers like
Alphabet’s
Google – in delivering the fruits of those models to hundreds of millions of users of chatbots and other technologies.Much of Huang’s speech focused on how well the new chips would work for that task, including adding a new layer of storage technology called “context memory storage” aimed at helping chatbots provide snappier responses to long questions and conversations.
Nvidia also touted a new generation of networking switches with a new kind of connection called co-packaged optics. The technology, which is key to linking together thousands of machines into one, competes with offerings from
Broadcom and Cisco Systems.In other announcements, Huang highlighted new software that can help self-driving cars make decisions about which path to take – and leave a paper trail for engineers to use afterward.
Nvidia showed research about software, called Alpamayo, late last year, with Huang saying on Monday it would be released more widely, along with the data used to train it so that automakers can make evaluations.“Not only do we open-source the models, we also open-source the data that we use to train those models, because only in that way can you truly trust how the models came to be,” Huang said from a stage in
Las Vegas. Last month,
Nvidia scooped up talent and chip technology from startup Groq, including executives who were instrumental in helping
Alphabet’s
Google design its own AI chips.While
Google is a major
Nvidia customer, its own chips have emerged as one of
Nvidia’s biggest threats as
Google works closely with Meta Platforms and others to chip away at the company’s lead.During a question-and-answer session with financial analysts after his speech, Huang said the Groq deal “won’t affect our core business” but could result in new products that expand its lineup. At the same time,
Nvidia is eager to show that its latest products can outperform older chips like the H200, which Donald Trump has allowed to flow to China.The chip, which was the predecessor to
Nvidia’s current “Blackwell” chip, is in high demand in China, which has alarmed China hawks across the US political spectrum.