DeepSeek's new models are so efficient they'll run on a toaster ... by which we mean Huawei's NPUs

DeepSeek V4 runs inference on Huawei silicon at a fraction of the cost rivals demand, collapsing the economic moat around closed models.
DeepSeek released V4 in preview on 24 April, an open-weights LLM claiming competitive performance with frontier proprietary models while cutting inference costs dramatically and extending support for Huawei's Ascend accelerators. Architectural redesign handles longer context windows and reduces computational overhead, making the model viable on consumer-grade and domestic Chinese hardware.
CN




