SMIC
Full transcript (Instant)

@Asianometry & Dylan Patel — How the semiconductor industry actually works

ARGUMENT

dwarkesh.com

Gist

1.

ARGUMENT

Original

Continue Reading

Full transcript (Deep)

@Asianometry & Dylan Patel — How the semiconductor industry actually works

ARGUMENT

dwarkesh.com

Gist

1.

SMIC is importing lithography tools into Beijing labeled "for 28nm" and using them to build 5nm chips. That single act of bureaucratic judo captures the real story: the semiconductor supply chain can deliver 100,000× GPT-4's compute by 2029, but only if every layer of the most stratified industry humans have ever built unlocks its bottlenecks in sequence — and right now, China is solving its bottlenecks faster than the US is solving its own.

Logic

2.

China can out-scale any single US lab by 2027 — using American chips

  • Over one million NVIDIA H20 GPUs still ship legally to China per year post-October 2023 sanctions, plus additional Hopper GPUs through semi-legal channels
  • 600,000 Huawei Ascend 910Bs (~400 TFLOPS each) are deployed across Alibaba, ByteDance, and Baidu — with over a million expected next year
  • No Chinese entity operates a 100K-GPU cluster today; DeepSeek's best models come from roughly 10,000–16,000 GPUs
  • Dylan Patel: "China could absolutely have it in 2026–27 if they desire to, just from foreign chips" — the constraint is centralization, not hardware

3.

US export controls accelerated Chinese resolve without stopping Chinese progress

  • China was 48% of ASML's revenue and 45% of Applied Materials' in recent quarters — despite sanctions
  • SMIC's Shanghai fab runs 45–50 high-end immersion lithography tools producing 25,000–35,000 wafers/month of 7nm capacity at 50–80 yielded dies per wafer
  • Huawei shipped 60 million phones on domestic chips performing one year behind Qualcomm — on a process node three to four years old
  • Jon Y: "I don't think you can stop the Chinese semiconductor industry from progressing. That's basically impossible. The sanctions have made China believe in the importance of the semiconductor industry more than anything else."

4.

Nobody understands the full stack — and that's load-bearing

  • An etch PhD told Jon Y "you really know lithography"; lithography people told him "you've done pretty good work in packaging" — nobody knows anything beyond their own well
  • Knowledge transmits master-to-apprentice, not through papers: "In Taiwan, there's a bunch of masters. They teach apprentices, and they just pass this secret, sacred knowledge down"
  • Fab equipment runs on Windows XP; chip design tools run on CentOS 6 — Jon Y: "This thing should not work. It's literally a miracle"
  • Semiconductor papers are paywalled behind IEEE and SPIE; unlike AI, where 18-year-olds break in via ArXiv, no one enters chip manufacturing without at least a master's in chemistry

5.

De-integration made the industry better — and now makes it fragile

  • In the 1960s–70s, engineers spun off, built better tools, sold them back; "the companies that accepted outside products got better stuff and did better"
  • Foundries collapsed from dozens (Silterra, Chartered, Hong Kong and Japanese companies) to effectively TSMC plus Samsung and Intel — each node required aggregating more customer demand to fund the next
  • Jon Y on N2: "You could argue that economically, N2 is a monstrosity that doesn't make sense. It should not exist without the immense concentrated spend of like five players"
  • At 28nm, 80% of world production is in Taiwan — every display driver IC Jon could identify is manufactured there

6.

Losing Taiwan resets civilization, not just leading-edge AI

  • The Magnificent Seven are 60–75% of the S&P 500 and their entire business relies on chips
  • Cars carry 2,000+ chips each — even ICE vehicles; "there are chips in the tires"
  • Six months after Taiwan loss: "the supply of chips for making new cars is gone or sequestered to make military shit"
  • Jon Y: "It's not just leading-edge. No fridges, no automobiles, no weed whackers." His toothbrush has Bluetooth.

7.

The scaling path to 1e30 total FLOPs by 2029 is physically possible — bottleneck by bottleneck

  • 100K GPU clusters exist today (xAI Memphis, OpenAI Arizona); 300–500K next-gen GPUs arrive in 2025, each 2–3× faster — roughly 1 million H100-equivalents in a single linked cluster by year-end
  • Microsoft signed $10+ billion in fiber deals with Lumen and Zayo to connect five regions; permits already filed for digging between data centers
  • TCO breakdown: ~80% GPUs, ~10% data center, ~10% power — power cost is "sub 10% actually," making geographic restrictions less about electricity price and more about generation capacity
  • Dylan's caveat: "Pre-training flops will be irrelevant." The 1e30 figure is total FLOPs across pre-training, post-training, synthetic data generation, and inference-time compute — not a standard training metric

8.

Pascal's Wager is the operating psychology — and the capital structure is not dot-com debt

  • Satya Nadella on his earnings call: "The risk of under-investing is worse than the risk of over-investing." Sundar Pichai and Mark Zuckerberg have said the same.
  • GPT-4 cost ~$500M to train and has generated billions in recurring revenue; GPT-4 API pricing fell from ~$120/million tokens in February 2023 to ~$10 now
  • Private AI capital hit ~$55–60B in 2024 year-to-date; the dot-com bubble ran ~$150B/year in mostly debt-financed telecom CapEx — current AI investment is equity from "the most profitable companies that humanity has ever created"
  • Microsoft CapEx trajectory: $50–80B direct plus $20B via Oracle/CoreWeave plus $10B via data center partners — affordable because they believe OpenAI will deliver

Counter-Argument

9.

The entire trillion-dollar scaling path depends on a model that was late and didn't exist

  • Jon Y, the conversation's persistent skeptic, was never answered: "GPT-5 is not here. GPT-5 is late. We don't know. If GPT-5 sucks, if GPT-5 looks like it doesn't blow people's socks off, this is all void."
  • Dylan's response was belief ("I have tremendous belief in the GPT-5 area") and capital-flow arguments — but never evidence that the next capability jump will arrive; every cluster being built, every fiber deal signed, every sovereign fund committed is a bet on a demo that hasn't happened
  • The 1990s optical/MEMS collapse is the cautionary analogue Jon raised and Dylan laughed off: billions poured into fiber infrastructure, the bubble burst in 2002, and Jon notes "there hasn't been revitalization since — you could freeze AI for another two decades"

Steelman

10.

The bet is asymmetric — and asymmetric bets don't need the next model to be good

  • Both the scaling bulls and the GPT-5 skeptics share one assumption: that the investment thesis depends on whether the next model impresses. It doesn't. The thesis depends on whether not investing is survivable — and for every major tech CEO, it isn't.
  • The dot-com bubble was debt-financed by middling telecoms; this wave is equity-financed by the seven most profitable companies in human history, each of whom watched the others commit. Game theory, not model quality, drives the spend: the cost of sitting out and being wrong dwarfs the cost of building and being early.
  • Even if GPT-5 disappoints and the bubble deflates, the infrastructure — fiber between data centers, gigawatt power connections, advanced packaging capacity, millions of GPUs — becomes the substrate for whatever comes next, exactly as the 1990s fiber overbuild became the backbone of the modern internet. The question isn't whether 1e30 FLOPs arrives on schedule. It's whether any CEO can afford to bet it won't.

Original

Continue Reading

Transcript

@Asianometry & Dylan Patel — How the semiconductor industry actually works

ARGUMENT

dwarkesh.com

Gist

1.

SMIC is importing lithography tools into Beijing labeled "for 28nm" and using them to build 5nm chips. That single act of bureaucratic judo captures the real story: the semiconductor supply chain can deliver 100,000× GPT-4's compute by 2029, but only if every layer of the most stratified industry humans have ever built unlocks its bottlenecks in sequence — and right now, China is solving its bottlenecks faster than the US is solving its own.

Logic

2.

China can out-scale any single US lab by 2027 — using American chips

  • Over one million NVIDIA H20 GPUs still ship legally to China per year post-October 2023 sanctions, plus additional Hopper GPUs through semi-legal channels
  • 600,000 Huawei Ascend 910Bs (~400 TFLOPS each) are deployed across Alibaba, ByteDance, and Baidu — with over a million expected next year
  • No Chinese entity operates a 100K-GPU cluster today; DeepSeek's best models come from roughly 10,000–16,000 GPUs
  • Dylan Patel: "China could absolutely have it in 2026–27 if they desire to, just from foreign chips" — the constraint is centralization, not hardware

3.

US export controls accelerated Chinese resolve without stopping Chinese progress

  • China was 48% of ASML's revenue and 45% of Applied Materials' in recent quarters — despite sanctions
  • SMIC's Shanghai fab runs 45–50 high-end immersion lithography tools producing 25,000–35,000 wafers/month of 7nm capacity at 50–80 yielded dies per wafer
  • Huawei shipped 60 million phones on domestic chips performing one year behind Qualcomm — on a process node three to four years old
  • Jon Y: "I don't think you can stop the Chinese semiconductor industry from progressing. That's basically impossible. The sanctions have made China believe in the importance of the semiconductor industry more than anything else."

4.

Nobody understands the full stack — and that's load-bearing

  • An etch PhD told Jon Y "you really know lithography"; lithography people told him "you've done pretty good work in packaging" — nobody knows anything beyond their own well
  • Knowledge transmits master-to-apprentice, not through papers: "In Taiwan, there's a bunch of masters. They teach apprentices, and they just pass this secret, sacred knowledge down"
  • Fab equipment runs on Windows XP; chip design tools run on CentOS 6 — Jon Y: "This thing should not work. It's literally a miracle"
  • Semiconductor papers are paywalled behind IEEE and SPIE; unlike AI, where 18-year-olds break in via ArXiv, no one enters chip manufacturing without at least a master's in chemistry

5.

De-integration made the industry better — and now makes it fragile

  • In the 1960s–70s, engineers spun off, built better tools, sold them back; "the companies that accepted outside products got better stuff and did better"
  • Foundries collapsed from dozens (Silterra, Chartered, Hong Kong and Japanese companies) to effectively TSMC plus Samsung and Intel — each node required aggregating more customer demand to fund the next
  • Jon Y on N2: "You could argue that economically, N2 is a monstrosity that doesn't make sense. It should not exist without the immense concentrated spend of like five players"
  • At 28nm, 80% of world production is in Taiwan — every display driver IC Jon could identify is manufactured there

6.

Losing Taiwan resets civilization, not just leading-edge AI

  • The Magnificent Seven are 60–75% of the S&P 500 and their entire business relies on chips
  • Cars carry 2,000+ chips each — even ICE vehicles; "there are chips in the tires"
  • Six months after Taiwan loss: "the supply of chips for making new cars is gone or sequestered to make military shit"
  • Jon Y: "It's not just leading-edge. No fridges, no automobiles, no weed whackers." His toothbrush has Bluetooth.

7.

The scaling path to 1e30 total FLOPs by 2029 is physically possible — bottleneck by bottleneck

  • 100K GPU clusters exist today (xAI Memphis, OpenAI Arizona); 300–500K next-gen GPUs arrive in 2025, each 2–3× faster — roughly 1 million H100-equivalents in a single linked cluster by year-end
  • Microsoft signed $10+ billion in fiber deals with Lumen and Zayo to connect five regions; permits already filed for digging between data centers
  • TCO breakdown: ~80% GPUs, ~10% data center, ~10% power — power cost is "sub 10% actually," making geographic restrictions less about electricity price and more about generation capacity
  • Dylan's caveat: "Pre-training flops will be irrelevant." The 1e30 figure is total FLOPs across pre-training, post-training, synthetic data generation, and inference-time compute — not a standard training metric

8.

Pascal's Wager is the operating psychology — and the capital structure is not dot-com debt

  • Satya Nadella on his earnings call: "The risk of under-investing is worse than the risk of over-investing." Sundar Pichai and Mark Zuckerberg have said the same.
  • GPT-4 cost ~$500M to train and has generated billions in recurring revenue; GPT-4 API pricing fell from ~$120/million tokens in February 2023 to ~$10 now
  • Private AI capital hit ~$55–60B in 2024 year-to-date; the dot-com bubble ran ~$150B/year in mostly debt-financed telecom CapEx — current AI investment is equity from "the most profitable companies that humanity has ever created"
  • Microsoft CapEx trajectory: $50–80B direct plus $20B via Oracle/CoreWeave plus $10B via data center partners — affordable because they believe OpenAI will deliver

Counter-Argument

9.

The entire trillion-dollar scaling path depends on a model that was late and didn't exist

  • Jon Y, the conversation's persistent skeptic, was never answered: "GPT-5 is not here. GPT-5 is late. We don't know. If GPT-5 sucks, if GPT-5 looks like it doesn't blow people's socks off, this is all void."
  • Dylan's response was belief ("I have tremendous belief in the GPT-5 area") and capital-flow arguments — but never evidence that the next capability jump will arrive; every cluster being built, every fiber deal signed, every sovereign fund committed is a bet on a demo that hasn't happened
  • The 1990s optical/MEMS collapse is the cautionary analogue Jon raised and Dylan laughed off: billions poured into fiber infrastructure, the bubble burst in 2002, and Jon notes "there hasn't been revitalization since — you could freeze AI for another two decades"

Steelman

10.

The bet is asymmetric — and asymmetric bets don't need the next model to be good

  • Both the scaling bulls and the GPT-5 skeptics share one assumption: that the investment thesis depends on whether the next model impresses. It doesn't. The thesis depends on whether not investing is survivable — and for every major tech CEO, it isn't.
  • The dot-com bubble was debt-financed by middling telecoms; this wave is equity-financed by the seven most profitable companies in human history, each of whom watched the others commit. Game theory, not model quality, drives the spend: the cost of sitting out and being wrong dwarfs the cost of building and being early.
  • Even if GPT-5 disappoints and the bubble deflates, the infrastructure — fiber between data centers, gigawatt power connections, advanced packaging capacity, millions of GPUs — becomes the substrate for whatever comes next, exactly as the 1990s fiber overbuild became the backbone of the modern internet. The question isn't whether 1e30 FLOPs arrives on schedule. It's whether any CEO can afford to bet it won't.

Original

Continue Reading