China's AI Compute Power Counterstrike
Article | Sleepy.txt
Eight years ago, ZTE suffered a sudden cardiac arrest.
On April 16, 2018, a ban from the U.S. Department of Commerce's Bureau of Industry and Security brought ZTE, the world's fourth-largest telecommunications equipment provider with 80,000 employees and annual revenue exceeding a hundred billion, to a standstill overnight. The ban was simple: for the next seven years, it prohibited any U.S. company from selling parts, goods, software, or technology to ZTE.
Without Qualcomm's chips, base stations ceased production. Without Google's Android license, phones were left without a usable system. Twenty-three days later, ZTE announced that its main business operations had become impossible.
However, ZTE ultimately survived, but at a cost of 1.4 billion U.S. dollars.
A $1 billion fine, to be paid in a lump sum; a $400 million escrow deposit, placed in a U.S. bank's custody account. In addition, a complete overhaul of all executives, accepting the presence of a U.S. compliance oversight team. In 2018, ZTE incurred a net loss of 7 billion RMB for the full year, with a year-on-year revenue plummet of 21.4%.
In an internal memo, then ZTE Chairman Yin Yimin wrote, "We are in a complex industry that heavily relies on the global supply chain." At that time, this statement was both reflective and resigned.
Eight years later, on February 26, 2026, Chinese AI unicorn DeepSeek announced that its upcoming V4 multimodal large model would prioritize deep collaboration with domestic chip manufacturers, achieving for the first time a full-process non-NVIDIA solution from pre-training to fine-tuning.
To put it simply: we no longer need NVIDIA.
Upon hearing this news, the market's initial reaction was skepticism. With NVIDIA holding over 90% of the global AI training chip market share, is it commercially reasonable to abandon it?
However, behind DeepSeek's choice lies a question much larger than business logic: What kind of computing independence does Chinese AI truly need?
What Was Really Choking Us
Many people think that the chip ban was blocking hardware. But what truly suffocated Chinese AI companies was something called CUDA.
CUDA, short for Compute Unified Device Architecture, is a parallel computing platform and programming model introduced by NVIDIA in 2006. It allows developers to directly harness the computing power of NVIDIA GPUs to accelerate various complex computing tasks.
Before the AI era, this was just a tool for a small number of geeks. But when the wave of deep learning arrived, CUDA became the foundation of the entire AI industry.
The training of AI large models is essentially massive matrix operations, which happens to be GPU's strong suit.
Thanks to NVIDIA's early positioning, CUDA has provided a complete toolchain from low-level hardware to upper-level applications for AI developers worldwide. Today, all mainstream AI frameworks globally, from Google's TensorFlow to Meta's PyTorch, are deeply integrated with CUDA at the core.
An AI Ph.D. student, from day one of enrollment, learns, programs, and experiments in the CUDA environment. Every line of code they write reinforces NVIDIA's moat.

By 2025, the CUDA ecosystem has attracted over 4.5 million developers, covering 3000+ GPU-accelerated applications, with over 40,000 companies worldwide using CUDA. This number means that over 90% of AI developers globally are locked into NVIDIA's ecosystem.
The frightening aspect of CUDA is that it is a flywheel. The more developers use it, the more tools, libraries, and code are generated, making the ecosystem more prosperous. The more prosperous the ecosystem, the more developers it attracts. Once this flywheel starts spinning, it is almost impossible to shake.
The result is that NVIDIA sells you the most expensive shovel and defines the only mining posture. Want to switch shovels? You can. But first, you have to rewrite all the experience, tools, and code accumulated by hundreds of thousands of the world's smartest brains over the past decade in this posture.
Who bears this cost?
So when on October 7, 2022, the first round of BIS regulations landed, restricting the export of NVIDIA A100 and H100 to China, Chinese AI companies collectively experienced a ZTE-style suffocation for the first time. NVIDIA subsequently launched the "China Special Edition" A800 and H800, reducing the interchip bandwidth to barely maintain supply.
However, just one year later, on October 17, 2023, the second round of regulations tightened again, and A800 and H800 were also banned, with 13 Chinese companies being added to the entity list. NVIDIA had to introduce the further castrated H20. By December 2024, in the final round of regulations during the Biden administration's term, even the export of H20 was severely restricted.
Tripartite Control, Layered Escalation.
But this time, the storyline is completely different from ZTE's back then.
An Asymmetric Breakout
Under the ban, everyone thought that China's AI dream of large models would come to an end.
They were all wrong. Faced with the blockade, Chinese companies did not choose a head-on confrontation but instead embarked on a breakout. The first battlefield of this breakout was not in chips but in algorithms.
From late 2024 to 2025, Chinese AI companies collectively shifted to a technical direction: Hybrid Expert Models.
In simple terms, it is splitting a huge model into many small experts, activating only the most relevant ones during tasks, rather than having the entire model in motion.
DeepSeek's V3 is a typical representative of this approach. It has 671 billion parameters, but only activates 37 billion of them during each inference, accounting for just 5.5% of the total. In terms of training costs, it used 2048 NVIDIA H800 GPUs, took 58 days to train, and cost a total of 5.576 million dollars. For comparison, the estimated training cost of GPT-4 is around 78 million dollars. A difference of an order of magnitude.
The extreme optimization on algorithms is directly reflected in the pricing. DeepSeek's API price is only $0.028 to $0.28 per million tokens for input and $0.42 per output, while GPT-4's input price is $5, and output is $15. Claude Opus is even more expensive, with input at $15 and output at $75. By conversion, DeepSeek is 25 to 75 times cheaper than Claude.
This price difference had a significant impact on the global developer market. In February 2026, on OpenRouter, the world's largest AI model API aggregation platform, China's AI model weekly call volume surged by 127% within three weeks, surpassing the United States for the first time. A year ago, China's model share on OpenRouter was less than 2%. A year later, it had grown by 421%, approaching sixty percent.

Behind this data is a structurally overlooked change. Starting in the second half of 2025, the mainstream scenario for AI applications shifted from chat to Agent. In an Agent scenario, the token consumption per task is 10 to 100 times that of simple chat. When token consumption grows exponentially, price becomes a decisive factor. The extreme cost-effectiveness of Chinese models happened to hit this window.
However, the problem is that the reduction in inference costs has not solved the fundamental issue of training. If a large model cannot sustain training and iteration on the latest data, its capability quickly degrades. And training remains the unavoidable power-hungry black hole.
So, where does the "shovel" for training come from?
The Promotion of Reserves
Xinghua, Jiangsu, a small city in central Jiangsu known for stainless steel and health food, had no previous connection to AI. However, in 2025, a 148-meter-long domestic computing power server production line was established and put into operation here, taking only 180 days from contract signing to production.
The core of this production line consists of two fully domestic chips: the Loongson 3C6000 processor and the Taichi Element T100 AI acceleration card. The Loongson 3C6000 was independently developed from instruction set to microarchitecture. The Taichi Element was derived from the National Supercomputing Center in Wuxi and a team from Tsinghua University, utilizing heterogeneous multi-core architecture.
When running at full capacity, this production line can roll out one server every 5 minutes. The total investment in this production line is 1.1 billion RMB, with an estimated annual output of 100,000 units.
More importantly, the ten-thousand-card cluster composed of these domestic chips has begun to undertake real large-scale model training tasks.
In January 2026, Zhipu AI and Huawei jointly released GLM-Image, the first SOTA image generation model to be fully trained relying on domestic chips. In February, China Telecom's billion-scale "Xingchen" large model completed end-to-end training on a domestically produced ten-thousand-card computing power pool in Lingang, Shanghai.

The significance of these cases is that they prove one thing: domestic chips have transitioned from "usable for inference" to "usable for training." This is a qualitative change. Inference only requires running pre-trained models, with relatively low chip requirements; whereas training involves handling massive data, performing complex gradient calculations and parameter updates, with demands on chip computation power, interconnect bandwidth, and software ecosystem an order of magnitude higher.
The core force undertaking these tasks is Huawei's Ascend series chips. By the end of 2025, the number of developers in the Ascend ecosystem had exceeded 4 million, with over 3,000 partners. Forty-three mainstream industry large models had completed pre-training based on Ascend, over 200 open-source models had been adapted. At the MWC conference on March 2, 2026, Huawei also launched a new generation of computing power base SuperPoD for overseas markets.
The FP16 computing power of Ascend 910B has already caught up with NVIDIA A100. Although the gap still exists, it has changed from unavailable to available, and from available to becoming usable. Ecosystem development cannot wait until the chip is perfect before starting. It must be massively deployed in the usable stage, using real business needs to drive the iteration of chips and software. Bytedance, Tencent, and Baidu aim to import domestically made computing power servers, with a widespread doubling in 2026 compared to the previous year. Ministry of Industry and Information Technology data shows that China's smart computing scale has reached 1590 EFLOPS. In 2026, it is becoming the first year of domestic computing power scale deployment.
American Power Shortage and China's Overseas Expansion
In early 2026, Virginia, which carries a large amount of global data center traffic, suspended the approval of new data center construction projects. Georgia followed suit, with approval suspensions extended to 2027. Illinois and Michigan have also introduced restrictive measures.
According to the International Energy Agency, in 2024, U.S. data centers consumed 183 terawatt-hours of electricity, accounting for about 4% of the national total electricity consumption. By 2030, this number is expected to double to 426TWh, with the share possibly exceeding 12%. Arm's CEO even predicts that by 2030, AI data centers will consume 20% to 25% of the U.S.'s electricity.
The U.S. power grid is already overwhelmed. The PJM grid covering 13 states in the eastern U.S. is facing a 6GW capacity shortage. By 2033, the U.S. as a whole will face a 175GW power capacity gap, equivalent to the electricity consumption of 130 million households. Wholesale power costs in data center-concentrated areas have increased by 267% compared to five years ago.
At the end of computing power lies energy. And in this energy dimension, the gap between China and the U.S. is even greater than in chips, just in the opposite direction.
China's annual electricity generation is 10.4 trillion kWh, compared to the U.S.'s 4.2 trillion kWh, making China 2.5 times that of the U.S. More importantly, residential electricity consumption accounts for only 15% of China's total electricity usage, while in the U.S., this proportion is 36%. This means that China has a far larger industrial electricity surplus than the U.S. to invest in computing infrastructure.

In terms of electricity prices, the electricity price in the AI company cluster area in the U.S. ranges from 0.12 to 0.15 USD per kWh, while the industrial electricity price in western China is about 0.03 USD, only a quarter to a fifth of the U.S. price.
China's electricity generation increment has reached 7 times that of the U.S.
Just as the United States was worrying about its electricity supply, China's AI was quietly going global. But this time, what went global was not a product or a factory, but a Token.
Token, the smallest unit for AI model information processing, is becoming a new digital commodity. It is produced in China's computing power factories and transported globally through undersea cables.
DeepSeek's user distribution data tells a compelling story: China accounts for 30.7%, India 13.6%, Indonesia 6.9%, the United States 4.3%, and France 3.2%. It supports 37 languages and is popular in emerging markets like Brazil. Globally, 26,000 companies have opened accounts, and 3,200 institutions have deployed the enterprise version.
By 2025, 58% of new AI startups have integrated DeepSeek into their tech stack. In China, DeepSeek has captured 89% of the market share. In other sanctioned countries, the market share ranges from 40% to 60%.
This scene is strikingly similar to another war over industrial autonomy forty years ago.
In 1986, under intense pressure from the United States, the Japanese government signed the "U.S.-Japan Semiconductor Agreement." The core provisions of the agreement had three main points: Japan was required to open its semiconductor market, U.S. chip market share in Japan had to reach 20% or more; Japan was forbidden from exporting semiconductors below cost; a 100% punitive tariff was imposed on $300 million worth of chips exported from Japan. At the same time, the U.S. vetoed Fujitsu's acquisition of Fairchild Semiconductor.
That year, the Japanese semiconductor industry was at its peak. By 1988, Japan controlled 51% of the global semiconductor market share, while the U.S. had only 36.8%. Out of the world's top ten semiconductor companies, Japan held six positions: NEC at second, Toshiba at third, Hitachi at fifth, Fujitsu at seventh, Mitsubishi at eighth, Panasonic at ninth. In 1985, Intel suffered a $173 million loss in the U.S.-Japan semiconductor battle, teetering on the brink of bankruptcy.
But everything changed after the agreement was signed.
Using measures such as the 301 investigation, the U.S. launched a comprehensive suppression of Japanese semiconductor companies. At the same time, it supported Samsung and Hynix in Korea to impact the Japanese market with lower prices. Japan's DRAM market share plummeted from 80% to 10%. By 2017, Japan's IC market share was only 7%. The once dominant giants were either split up, acquired, or quietly exited amid endless losses.

The tragedy of the Japanese semiconductor industry is that it was content to be the best producer in a global division of labor dominated by a single external force, but never thought to build its own independent ecosystem. When the tide ebbed, it found itself with nothing beyond production itself.
Today's Chinese AI industry is standing at a similar yet entirely different crossroads.
Similar in that we also face tremendous pressure from external forces. The three rounds of chip control, escalating layers of barriers, and the towering walls of the CUDA ecosystem.
What is different is that this time, we have chosen a more difficult path. From algorithm-level optimization to the leap of domestic chips from inference to training, to the accumulation of the Ascend ecosystem's 4 million developers, to the penetration of global markets through Token globalization. Each step on this path is building an independent industrial ecosystem that Japan never had.
Epilogue
On February 27, 2026, performance reports from three local AI chip companies were released on the same day.
Cambricon, revenue surged by 453%, achieving annual profit for the first time. Moore Threads, revenue grew by 243%, but with a net loss of 1 billion. Horizon, revenue increased by 121%, with a net loss of nearly 800 million.
Half flame, half seawater.
The flame is the market's extreme thirst. The 95% gap relinquished by Jensen Huang is being gradually filled inch by inch by the revenue figures of these local companies. Regardless of performance or ecosystem, the market needs a second option beyond NVIDIA. This is a rare structural opportunity torn open by geopolitics.
The seawater is the high cost of ecosystem development. Every penny of loss is real money paid to catch up with the CUDA ecosystem. It is investment in research and development, software subsidies, and the manpower cost of engineers deployed to customer sites solving compilation issues one by one. These losses are not due to mismanagement but are the war tax required to build an independent ecosystem.
These three financial reports paint a more honest picture of the reality of this computational power war than any industry report. It is not a triumphant advance but a fierce battlefield engagement fought with bloodshed.
But the form of war has indeed changed. Eight years ago, we discussed the question of "whether we can survive." Today, we discuss the question of "how high a price we must pay to survive."
The price itself is progress.
You may also like
US & Canada Crypto Tax Season 2026: Official Tax Reporting Support from WEEX × KoinX
Prepare for US & Canada crypto tax season 2026. Learn how to export your WEEX transaction history and access official reporting support through our partnership with KoinX.

Conversation between Tom Lee and "The Big Short" Author: AI has detected bubble signal, crypto correction due to gold liquidity being "siphoned off"

The true reason for Claude's ban, Kraken accessing the Federal Reserve payment system, What is the English community paying attention to?

「Buying the Dip」 of 400,000 BTC: Is $74,000 a Rebound or a Reversal?

OpenClaw, Another Batch of Middle Class Jobless

Morning News | Backpack will launch on-chain IPO subscription service; Predict.fun strategically acquires on-chain prediction platform Probable; SoFi partners with Mastercard for strategic cooperation

Inventorying the Washington power in the crypto space, who is speaking out for U.S. crypto legislation?

650 million dollars, 1.5 billion dollars, 2 billion dollars, the crypto VC landscape has changed!

Why prediction markets are the largest untapped collateral pool in DeFi
500% XAUT Staking, Zero-Fee Gold Futures and $100K Rewards: Why Traders Are Turning to WEEX for Tokenized Gold
Explore WEEX's $100,000+ gold campaign featuring 500% XAUT staking, zero-fee gold contracts, and $30,000 PAXG rewards. Trade tokenized gold today.
AI within artillery range
“The cloud” is a metaphor, but the data center isn’t.

March 4th Market Key Intelligence, How Much Did You Miss?

Taking Stock of Crypto's Washington Power Players: Who is Advocating for US Crypto Regulation?

DDC Enterprise Limited Announces 2025 Unaudited Preliminary Financial Performance: Record Revenue Achieved, Bitcoin Treasury Grows to 2183 Coins
On March 4, 2026, DDC Enterprise Limited (NYSE American: DDC) today announced preliminary, unaudited full-year financial performance for the year ended December 31, 2025. The company expects to achieve record revenue and record positive adjusted EBITDA, primarily driven by continued growth in its core consumer food business and overall margin improvement. The final audited financial report is expected to be released in mid-April 2026.
Revenue: Expected to be between $39 million and $41 million, reaching a new company high.
Organic Growth: Excluding the impact of the company's strategic contraction of its U.S. operations, core revenue is expected to grow 11% to 17% year over year.
Gross Profit Margin: Expected to be between 28% and 30%, reflecting continued operational efficiency improvements.
Adjusted EBITDA: The company expects to achieve a positive full-year result in 2025, a significant improvement from a $3.5 million loss in 2024, mainly due to rigorous cost controls and a higher-margin sales mix.
In 2025, DDC's core consumer food business maintained strong operational performance.
The company also disclosed Core Consumer Food Business Adjusted EBITDA, a metric that further excludes costs related to its Bitcoin reserve strategy and non-cash fair value adjustments related to its Bitcoin holdings from adjusted EBITDA to more accurately reflect the core business performance.
In 2025, Core Consumer Food Business Adjusted EBITDA is expected to be between $5.5 million and $6 million.
In the first half of 2025, DDC initiated a long-term Bitcoin accumulation strategy, holding Bitcoin as its primary reserve asset.
As of December 31, 2025: The company holds 1,183 BTC.
As of February 28, 2026: Holdings increased to 2,118 BTC
Today's additional purchase of 65 BTC brings the company's total holdings to 2,183 BTC
DDC Founder, Chairman, and CEO Norma Chu stated, "We are proud to have closed 2025 with record revenue and positive adjusted EBITDA, demonstrating the steady growth of the company's consumer food business and the ongoing improvement in profitability. We are building a disciplined, growth-oriented food platform and strategically allocating capital to Bitcoin assets with a long-term view, aligning with our core beliefs. We believe that this dual-track model of 'Steady Consumer Business + Strategic Bitcoin Reserve' will help DDC create lasting long-term value for shareholders."
For the full year 2025, the company defines "Adjusted EBITDA" (a non-GAAP financial measure) as: Net income / (loss) excluding the following items:· Interest expense· Taxes· Foreign exchange gains/losses· Long-lived asset impairment· Depreciation and amortization· Non-cash fair value changes related to financial instruments (including Bitcoin holdings)· Stock-based compensation
DDC Enterprise Limited (NYSE: DDC) is actively implementing its corporate Bitcoin Treasury strategy while continuing to strengthen its position as a leading global Asian food platform.
The company has established Bitcoin as a core reserve asset and is executing a prudent, long-oriented accumulation strategy. While expanding its portfolio of food brands, DDC is gradually becoming one of the public company pioneers in integrating Bitcoin into its corporate financial architecture.

Uncovering YZi Labs 229 Investment: Over 18% of the portfolio is already inactive, with an average project transparency score of 78

The business of crypto VC is becoming promising

Global Assets Plunge: Hormuz, Chips, and a South Korean Holiday

Bloomberg has reported twice, Hyperliquid once again in Wall Street's radar
US & Canada Crypto Tax Season 2026: Official Tax Reporting Support from WEEX × KoinX
Prepare for US & Canada crypto tax season 2026. Learn how to export your WEEX transaction history and access official reporting support through our partnership with KoinX.