Discussion about this post

User's avatar
Leon Liao's avatar

It is right to identify why Silicon Valley is becoming increasingly drawn to Chinese AI: China is turning low-cost token generation into a structural advantage. Some Chinese models charge only around $2 to $3 per million output tokens, versus roughly $15 for Anthropic’s Claude Sonnet, a gap of nearly 6x. That cost differential is already pushing some U.S. companies, including Airbnb, to adopt Chinese large models.

That said, this post’s understanding of this advantage is still incomplete. It argues that one reason is lower electricity prices in China, when in fact U.S. industrial electricity prices are generally lower than China’s. Its second point is more convincing: China’s embrace of the MoE path did emerge partly under conditions of compute constraint, and those constraints do force companies to focus more intensely on inference efficiency, model compression, and unit-cost optimization. But the problem is that MoE is not a uniquely Chinese method, nor is it a technological path that naturally implies durable Chinese leadership. It is better understood as a tool that, at least for this stage, amplifies China’s engineering strengths.

I would add two points.

First, this should not be read too optimistically for China AI. What is truly worth paying attention to is not simply that Chinese models are cheaper. It is that China may be developing a broader system-level capability to keep driving inference costs down while diffusing AI rapidly into real industrial and commercial use cases. But that is still very far from proving that the global center of AI power has already shifted. In reality, AI competition needs to be broken down into at least four layers. The first is frontier model capability: who is closest to the technological frontier. The second is inference cost: who can produce tokens more cheaply. The third is systems integration: cloud infrastructure, networking, memory, packaging, power, and data-center orchestration. The fourth is the application ecosystem: whether these capabilities can actually be absorbed by enterprises and consumers at scale. Frontier models, the compute foundation, cloud platforms, enterprise gateways, and global capital markets still remain, to a very large extent, in American hands today. I discussed this in an earlier post as well.(https://leonliao.substack.com/p/low-cost-open-models-are-squeezing?r=731anr&utm_medium=ios)

Second, cheap tokens do not automatically mean value capture will accrue to China. In the AI value chain, the party that produces tokens, the party that consumes them, the party that defines application-layer standards, and the party that controls customer relationships are not necessarily the same. Even if Chinese models provide cheaper inference in certain scenarios, the entities that ultimately capture the profit pool may still be American platform companies, cloud providers, enterprise software firms, or application-layer players that control distribution. This is no different from manufacturing: the lowest-cost producer does not necessarily capture the highest profits. What ultimately determines profit distribution is system control, branding, access points, standards, and customer stickiness. I have discussed that in another post as well.(https://leonliao.substack.com/p/chinas-token-exports-are-explodingwhat?r=731anr&utm_medium=ios)

No posts

Ready for more?