Customize Consent Preferences

We use cookies to help you navigate efficiently and perform certain functions. You will find detailed information about all cookies under each consent category below.

The cookies that are categorized as "Necessary" are stored on your browser as they are essential for enabling the basic functionalities of the site. ... 

Always Active

Necessary cookies are required to enable the basic features of this site, such as providing secure log-in or adjusting your consent preferences. These cookies do not store any personally identifiable data.

No cookies to display.

Functional cookies help perform certain functionalities like sharing the content of the website on social media platforms, collecting feedback, and other third-party features.

No cookies to display.

Analytical cookies are used to understand how visitors interact with the website. These cookies help provide information on metrics such as the number of visitors, bounce rate, traffic source, etc.

No cookies to display.

Performance cookies are used to understand and analyze the key performance indexes of the website which helps in delivering a better user experience for the visitors.

No cookies to display.

Advertisement cookies are used to provide visitors with customized advertisements based on the pages you visited previously and to analyze the effectiveness of the ad campaigns.

No cookies to display.

Nvidia CEO Says ‘Reasoning’ AI Will Depend on Cheaper Computing

Share

Nvidia Chief Executive Officer Jensen Huang said that the future of Artificial Intelligence (AI) will be services that can “reason,” but such a stage requires the cost of computing to come down first.

Next-generation tools will be able to respond to queries by going through hundreds or thousands of steps and reflecting on their own conclusions, he said during a podcast hosted by Arm Holdings Plc CEO Rene Haas. That will give this future software the ability to reason and set it apart from current systems such as OpenAI’s ChatGPT, which Huang said he uses every day. 

Nvidia will set the stage for these advances by boosting its chip performance every year by two to three times, at the same level of cost and energy consumption, Huang said. This will transform the way AI systems handle inference — the ability to spot patterns and draw conclusions.

“We’re able to drive incredible cost reduction for intelligence,” he said. “We all realize the value of this. If we can drive down the cost tremendously, we could do things at inference time like reasoning.”

The Santa Clara, California-based company has more than 90 percent of the market for so-called accelerator chips — processors that speed up AI work. It has also branched out to selling computers, software, AI models, networking and other services — part of a push to get more companies to embrace artificial intelligence. 

Nvidia is facing attempts to loosen its grip on the market. Data center operators such as Amazon.com Inc.’s AWS and Microsoft are developing in-house alternatives. And Advanced Micro Devices, already an Nvidia rival in gaming chips, has emerged as an AI contender. AMD plans to share the latest on its AI products at an event Thursday.

© 2024 Bloomberg LP

(This story has not been edited by NDTV staff and is auto-generated from a syndicated feed.)