Jensen Huang: NVIDIA and the AI Revolution
Extreme Co-Design
极端协同设计
Jensen opens with a principle that defines NVIDIA’s approach: “the problem no longer fits inside one computer to be accelerated by one GPU.” Modern AI workloads span thousands of machines, which means optimization must happen across every layer simultaneously — GPUs, CPUs, memory, networking, power, cooling, and software.
This isn’t just a technical observation; it’s an organizational one. NVIDIA’s structure mirrors its product. Jensen has 60+ direct reports, almost all engineers, who participate in group problem-solving sessions rather than isolated one-on-ones. A memory specialist tunes into a discussion about thermal management because everything interconnects. The Vera Rubin rack exemplifies this: 1.3 million components, 200 suppliers, 7 chip types, all optimized together.
His design philosophy: “as complex as necessary, but as simple as possible.” Complexity is not avoided — it is managed through co-design across every boundary that traditional organizations treat as separate.
Four Types of AI Scaling
AI 缩放的四种类型
Jensen lays out a framework of four scaling laws, each compounding on the others:
-
Pre-training scaling: skeptics predicted data would become the bottleneck. Instead, synthetic data generation dissolved this constraint — humans create ground truth, AI augments and regenerates it in a self-perpetuating cycle.
-
Post-training scaling: continued expansion through refinement, fine-tuning, and alignment processes.
-
Test-time scaling: Jensen rejects the notion that inference is “easy” and commoditizable. “Inference is thinking, and thinking is hard.” Reasoning, planning, and problem-solving demand immense compute. This scaling law surprised many but now drives hardware demand.
-
Agentic scaling: agents spawning sub-agents create exponential compute requirements. One agent orchestrating multiple specialized agents scales intelligence faster than linearly.
These form interconnected loops: agents generate experiences that become training data for pre-training, refined through post-training, enhanced at test-time, then deployed agentically — perpetuating endless scaling potential. The implication for hardware: “intelligence is gonna scale by one thing, and that’s compute.”
CUDA and the Install Base Bet
CUDA 与装机量豪赌
One of the most revealing segments covers the decision to put CUDA on GeForce gaming GPUs — a bet that nearly killed the company. It increased costs by 50%, consuming all gross profit on a 35% margin business. NVIDIA’s market cap dropped from ~$8 billion to $1.5 billion.
The reasoning: “install base defines an architecture. Everything else is secondary.” Jensen draws the analogy to x86 — an inelegant instruction set that dominated because developers followed users. Beautiful RISC architectures lost not on technical merit but on adoption. By embedding CUDA in millions of gaming GPUs, NVIDIA ensured researchers would discover it organically — and they did, building the first deep learning clusters from GeForce cards.
This is a pattern worth studying: Jensen repeatedly bets on creating installed ecosystems rather than optimizing isolated products. The cost is existential risk in the short term; the payoff is architectural lock-in that compounds over decades.
Speed of Light Thinking
"光速"思维
Jensen’s decision-making framework centers on what he calls “speed of light” thinking — testing every decision against physics-based limits before considering practical trade-offs. The sequence matters: first establish what’s theoretically possible, then discuss what’s practical.
This prevents incremental thinking. If a manufacturing process takes 74 days, the natural instinct is to optimize and save 2 days. But if you first ask “what is the physical minimum?” and discover it’s 6 days, you realize the entire process needs to be reimagined, not tweaked.
On shaping organizational belief: rather than top-down mandates, Jensen lays “bricks” of information over months or years — at GTC talks, board meetings, management sessions, company-wide communications. When a major pivot is announced (like “go all-in on deep learning” or “acquire Mellanox”), employees feel it was inevitable. The goal is 100% buy-in because everyone has been convinced incrementally, not surprised.
This is manifesting the future through systematic belief construction — arguably more powerful than charismatic leadership because it scales beyond personal presence.
Supply Chain as Strategy
供应链即战略
NVIDIA’s growth isn’t just accelerating — it’s accelerating while increasing market share. Jensen spends substantial time educating upstream partners (TSMC, ASML, memory manufacturers) and downstream infrastructure providers about demands that don’t yet exist.
He convinced DRAM manufacturers that HBM memory — previously niche in supercomputing — would become mainstream. He persuaded them that low-power phone memory (LPDDR5) could scale to data centers. These conversations preceded actual demand by years, requiring billions in capital investments from suppliers.
The shift from DGX-1 to NVLink-72 changed where assembly happens: from data centers to the supply chain itself. Each NVLink-72 rack ships as a complete two-ton supercomputer. This requires supply chain partners to add manufacturing power at a scale measured in gigawatts per week.
Jensen’s approach to these relationships: visit partners personally, reason from first principles, draw pictures, respect their questions, build trust so they act with confidence on multi-billion-dollar bets. The supply chain is not a vendor relationship — it’s a co-engineering partnership.
Elon Musk and Colossus
Elon Musk 与 Colossus
Jensen offers a detailed assessment of how Elon Musk built Colossus — 200,000 GPUs deployed in Memphis in four months. He attributes this to several compounding factors:
Radical minimalism: strip everything to essentials without sacrificing core functionality. “He questions everything: Is it necessary? Does it have to be done this way? Must it take this long?”
Systems thinking: apply minimalism across all disciplines simultaneously, not just one domain.
Ground presence: Elon is physically present at the point of action, detailing cable-routing processes with engineers to eliminate errors. Being there forces problem-solving that remote management cannot.
Urgent leadership: personal urgency cascades through supply chains. Suppliers treat his projects as top priority because he demonstrates it relentlessly.
Jensen sees parallels to NVIDIA’s extreme co-design philosophy but notes Elon’s unique willingness to rebuild entire processes from scratch rather than incrementally improve existing ones. The Colossus timeline was not achieved through better project management — it was achieved through questioning whether the standard process should exist at all.
China's Tech Ecosystem
中国的科技生态
Jensen provides one of the more nuanced assessments of China’s tech competitiveness. “50% of the world’s AI researchers are Chinese.” He identifies several structural advantages:
Competitive federalism: 30+ provinces with mayors competing on economic metrics. This produces dozens of EV companies, AI startups, and tech ventures in ruthless internal competition that forges winners.
Knowledge diffusion: a cultural norm of sharing within extended networks — schoolmates, family connections, former colleagues. “Family first, friends second, company third” means information flows freely across organizational boundaries. Open-sourcing technology feels natural when your schoolmate at the competing company already knows how it works.
Engineering culture: strong math and science education, cultural prestige attached to engineering careers rather than law or finance. Jensen notes: “It’s a builder nation. Their leaders are engineers, ours are lawyers.”
The result is the fastest-innovating country globally in certain domains, driven by rapid knowledge diffusion, intense internal competition, and abundant top talent.
Open Source and Nemotron
开源与 Nemotron
NVIDIA’s open-source strategy with Nemotron 3 reflects a deeper logic than altruism:
-
Co-design intelligence: NVIDIA conducts basic research (SSMs, conditional GANs, diffusion models) to anticipate future computing requirements. Open-sourcing models reveals architectural innovations that inform hardware design.
-
Ecosystem activation: proprietary models serve as products; open models activate every industry, researcher, and country. NVIDIA can afford this because its business model is hardware, not model APIs.
-
Modality diversity: “AI is not just language.” Biology AI, weather prediction, physical AI, chemistry — not everything fits language model architectures. NVIDIA doesn’t build cars but wants every car company accessing great models. Doesn’t discover drugs but wants Eli Lilly having world-class biology AI.
-
Full transparency: NVIDIA open-sourced Nemotron’s weights, data, and creation methodology — not just model weights.
The strategic insight: when your business is selling the pickaxes, you want as many people mining as possible. Open-sourcing models maximizes demand for NVIDIA hardware across every domain.
Agents and Future Computing
智能体与未来计算
Jensen reasons about agents through a thought experiment: what does a capable agent actually need?
- Access to ground truth — file systems, databases, documentation.
- Research capability — because no agent is omniscient.
- Tool use — rather than morphing into different instruments, use existing tools intelligently.
This naturally leads to agents that access file systems, read manuals, and execute code — but with security constraints. NVIDIA developed OpenClaw (agent framework) and NemoClaw (security layer) to demonstrate responsible deployment: agents can access files and execute code, but enterprise policy engines control what combinations of capabilities are available.
Jensen frames agentic computing as “the impact to the future of computing is deeply profound” — not because individual agents are revolutionary, but because agent-spawning-agent dynamics create exponential scaling of intelligence that maps directly to hardware demand. Each layer of agent orchestration multiplies compute requirements, creating a flywheel between software capability and hardware sales.
The Vera Rubin Pod
Vera Rubin Pod
The conversation culminates with the Vera Rubin pod — what Jensen calls “the most complex computer the world has ever made”:
- 1 pod = 1,100+ Rubin GPUs, 60 exaflops, 10 petabytes/second bandwidth
- 1 NVLink-72 rack = 1.3 million components, 1,300 chips
- 7 chip types, 5 rack types, 40 racks per pod
- Nearly 1.2 quadrillion transistors
- Target: 200 pods per week
Vera Rubin adds storage accelerators and a new Rock subsystem for agentic workloads — requirements invisible two years prior but logical through first-principles reasoning about how agents interact with data. Grace Blackwell focused on inference; Vera Rubin anticipates the agent era.
Jensen’s hardware anticipation cycle: model architectures change every 6 months, system architectures every 3 years, hardware every 3+ years. NVIDIA must predict trends 2-3 years ahead through research, industry signals, and principled reasoning — which is why the co-design philosophy is not optional but existential.
Power Grid and Graceful Degradation
电网与优雅降级
Jensen identifies a non-obvious bottleneck: power availability is constrained not by generation capacity but by contractual rigidity. “99% of the time, power grid has excess power” — grids are engineered for worst-case scenarios but typically operate at ~60% capacity.
The problem is a cascade of uptime guarantees: end customers demand six-nines availability, cloud providers replicate those demands, and utilities must then guarantee impossibly high reliability. Jensen’s proposed solution is not engineering but contracting: data centers that voluntarily reduce power during infrastructure emergencies, shift workloads geographically, or accept slightly longer latency.
“We need data centers that gracefully degrade.” This uses idle grid capacity without forcing utilities to overbuild — a systems-thinking approach to what most people frame as a physics problem.