Business · organisation

Jensen Huang on Deepseek

Countering efficiency fears (strong)

TL;DR

Jensen Huang asserts that Deepseek's efficiency advancements do not negate the fundamental, long-term need for vastly more computing power across the AI lifecycle.

Key Points

  • He stated that investors who sold off stock in response to Deepseek’s R1 model advancements had the wrong mental model regarding AI development in early 2025.

  • The CEO characterized the efficiency demonstrated by Deepseek as a 'gift to the world' of AI industry inference, praising its impact.

  • Huang maintained that post-training processes, such as reasoning and inference, remain more critical and compute-heavy than the initial training phase, requiring high-end hardware.

Summary

Jensen Huang directly addressed market concerns and stock sell-offs triggered by the perceived efficiency of the Deepseek AI model. The core of his stance is a rebuttal to the idea that Deepseek’s advancement, built with reportedly weaker chips, signals the end of the demand curve for high-end AI hardware. He argued that investors developed a flawed mental model by overemphasizing training costs while neglecting the critical, compute-intensive post-training phases like reasoning and inference, which still demand substantial computational resources.

He positioned Deepseek’s progress as a positive catalyst, suggesting it accelerates the open-source AI shift and, counterintuitively, drives greater overall compute demand by democratizing access and enabling new use cases. Huang emphasized that efficiency improvements lead to broader adoption, which in turn necessitates scaling larger models and refining them through intensive post-training. This suggests his view is that the entire AI lifecycle, not just initial training, dictates the long-term market for high-performance chips, reinforcing Nvidia’s essential role.

Frequently Asked Questions

Jensen Huang's position is that Deepseek’s achievements, while notable for efficiency, do not undermine the broader, massive demand for Nvidia's computing power. He views the market reaction to Deepseek as a misunderstanding of the full AI computing lifecycle.

No, he strongly disagreed with the implication that Deepseek's cost-efficient model rendered high-end chips unnecessary for the entire industry. He countered this by emphasizing that post-training operations and scaling to serve large user bases still require superior computing power.

He credited the Chinese AI firm with accelerating the open-source AI shift in the industry. Furthermore, he suggested that such efficiency gains ultimately expand the total addressable market for AI compute by enabling broader adoption and innovation.