Vol. 1 · No. 118
Tuesday Edition
Price · Free

The Unvarnished AI Gazette

AI news distilled in its purest form
122Unique stories
← Back to the newspaper

DeepSeek's new models are so efficient they'll run on a toaster ... by which we mean Huawei's NPUs

DeepSeek V4 runs inference on Huawei silicon at a fraction of the cost rivals demand, collapsing the economic moat around closed models.

DeepSeek released V4 in preview on 24 April, an open-weights LLM claiming competitive performance with frontier proprietary models while cutting inference costs dramatically and extending support for Huawei's Ascend accelerators. Architectural redesign handles longer context windows and reduces computational overhead, making the model viable on consumer-grade and domestic Chinese hardware.

Oliver's take: Cost-per-inference collapse is the actual story. V4 runs on Huawei silicon. That's not a benchmark win; that's infrastructure substitution. American closed models are now expensive because they have to be; open Chinese models are cheap because they can be. Margins die first.
← Back to the newspaper
— The Unvarnished AI Gazette · Tuesday, April 28, 2026 · 122 stories from 168 sources · ← Back to the newspaper