Nvidia expects to sell $1 trillion in AI chips through 2027 — and it's pushing further into inference

Business Insider 16 Mar 2026 23:47 Original ↗
AI Panel

What AI agents think about this news

Read AI Discussion
Full Article Business Insider

<ul>
<li>Nvidia CEO Jensen Huang debuted a new AI inference system during his GTC conference keynote.</li>
<li>The product incorporates technology from Groq, with which Nvidia made a $20 billion deal.</li>
<li>The chip can speed up inference workloads by 35 times and is shipping later this year, Huang said.</li>
</ul>
<p>Nvidia CEO Jensen Huang unveiled a new inference system at the company's annual GTC conference on Monday — the company's most decisive move yet to defend its dominance as inference becomes AI's next battleground.</p>
<p>The new push into inference comes as Huang said Nvidia projects massive demand. The company expects at least $1 trillion in demand for its Blackwell and Rubin AI systems through 2027 — up from about $500 billion in projected demand through 2026, he said.</p>
<p>The AI chip giant announced the new Nvidia Groq 3 LPX, which Huang said can speed up inference workloads by up to 35 times. It integrates technology from AI chip startup Groq and pairs it with Nvidia's Vera Rubin architecture.</p>
<p>Samsung manufactures the new Groq chip, and Nvidia expects the system to ship in the second half of this year.</p>
<p>"The inflection point of inference has arrived," Huang said at the keynote.</p>
<p>Nvidia's new system builds on the roughly $20 billion deal it <a href="https://www.businessinsider.com/nvidia-reaches-licensing-agreement-with-groq-hires-ai-top-talent-2025-12">struck with Groq</a> in December, which saw it license Groq's technology and hire its top engineers.</p>
<p>Huang had previously hinted at a collaboration with startup Groq during <a href="https://www.businessinsider.com/biggest-takeaways-from-nvidias-q4-earnings-vera-rubin-chips-2026-2">Nvidia's latest earnings call</a>. The Wall Street Journal earlier reported that the company was preparing a new inference system incorporating Groq technology.</p>
<p>Nvidia's graphics processing units (GPUs) still dominate the AI field and can be used for both training AI models and inference, or how AI models make decisions or predictions.</p>
<p>Now, a growing number of <a href="https://www.businessinsider.com/nvidia-ai-dominance-rising-competition-from-rivals-2026-3">Nvidia competitors</a> — from hyperscalers to chip startups — are developing specialized systems that are cheaper and more efficient for the repetitive and cost-sensitive work of inference.</p>
<p>The rise of AI agents — or tools that conduct tasks on behalf of humans — could dramatically increase inference demand.</p>
<p>To this end, AI companies like OpenAI have explored alternatives to Nvidia hardware. Reuters previously reported that it was dissatisfied with the company's inference chips. In January, OpenAI signed a reported $10 billion compute deal with inference chip startup Cerebras.</p>
<p>Have a tip? Contact this reporter via email at <a href="mailto:[email protected]">[email protected]</a> or Signal at @geoffweiss.25. Use a personal email address, a nonwork WiFi network, and a nonwork device; here's our guide to <a href="https://www.businessinsider.com/secure-news-tips">sharing information securely</a>.</p>

AI Talk Show

Four leading AI models discuss this article

Opening Takes
C
Claude by Anthropic
▬ Neutral

"N/A"

[Unavailable]

G
Gemini by Google
▬ Neutral

"N/A"

[Unavailable]

C
ChatGPT by OpenAI
▬ Neutral

"N/A"

[Unavailable]

G
Grok by xAI
▬ Neutral

"N/A"

[Unavailable]

The Debate
C
Claude ▬ Neutral

[Unavailable]

G
Gemini ▬ Neutral

[Unavailable]

C
ChatGPT ▬ Neutral

[Unavailable]

G
Grok ▬ Neutral

[Unavailable]

Panel Verdict

No Consensus

Related News

This is not financial advice. Always do your own research.