×
DeepSeek cuts AI processing costs 50% with new sparse attention tech
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

Chinese AI startup DeepSeek has launched DeepSeek-V3.2-Exp, an experimental model that introduces “sparse attention” technology to cut AI processing costs in half while maintaining performance levels. The release builds on DeepSeek’s reputation for creating efficient AI systems using fewer resources than traditional approaches, though experts question whether the cost-cutting architecture compromises model reliability and safety.

What you should know: DeepSeek’s new experimental model represents a significant shift in AI architecture design, focusing on efficiency over raw computational power.

  • The V3.2-Exp model introduces DeepSeek Sparse Attention (DSA), which selectively processes only the most relevant information rather than analyzing all available data.
  • According to Adina Yakefu, Chinese community lead at Hugging Face (an AI development platform), the technology “cuts the cost of running the AI in half compared to the previous version” while improving handling of long documents and conversations.
  • DeepSeek has made the model’s programming code and tools publicly available, allowing other developers to build upon the technology.

How sparse attention works: The technology functions like an airline route optimizer, filtering out less viable options to reduce processing time and resources.

  • Traditional AI models analyze all available data when making decisions, while sparse attention models exclude information deemed less important for specific tasks.
  • “So basically, you cut out things that you think are not important,” explained Ekaterina Almasque, cofounder and managing partner of BlankPage Capital, a venture capital fund.
  • This approach dramatically reduces computational requirements while maintaining model performance on par with DeepSeek’s V3.1-Terminus version.

The efficiency advantage: Industry experts see significant potential in DeepSeek’s cost-reduction approach for democratizing AI access.

  • “This makes powerful AI more accessible to developers, researchers, and smaller companies, potentially leading to a wave of new and innovative applications,” said Nick Patience, vice president at The Futurum Group, a technology research firm.
  • The models work seamlessly with Chinese-made AI chips like Ascend and Cambricon, enabling domestic hardware deployment without additional setup.
  • As Patience noted, “this is DeepSeek’s value prop all over: efficiency is becoming as important as raw power.”

Safety and reliability concerns: Experts worry that sparse attention’s selective data processing could compromise model accuracy and inclusivity.

  • “The reality is, they [sparse attention models] have lost a lot of nuances,” Almasque warned, questioning whether the exclusion mechanisms properly identify truly unimportant data.
  • The approach raises particular concerns for AI safety and inclusivity, as it may not be “the optimal one or the safest” compared to traditional architectures.
  • DeepSeek acknowledges V3.2-Exp as an “intermediate step toward our next-generation architecture,” suggesting ongoing development is needed.

Competitive implications: DeepSeek’s open-source approach presents both opportunities and challenges for maintaining competitive advantage.

  • The company cannot patent its sparse attention technology due to its open-source nature, potentially limiting defensibility against competitors.
  • Almasque noted that the industry has been “talking about sparse models since 2015,” suggesting the core concept isn’t entirely novel.
  • DeepSeek’s competitive edge must therefore lie in its specific implementation of information selection algorithms rather than the underlying architecture.

What they’re saying: Industry observers emphasize the strategic importance of DeepSeek’s efficiency-first approach in the evolving AI landscape.

  • “DeepSeek is playing the long game to keep the community invested in their progress,” Yakefu observed.
  • “People will always go for what is cheap, reliable, and effective,” she added, highlighting the market appeal of cost-efficient AI solutions.
China's DeepSeek launches next-gen AI model. Here's what makes it different

Recent News

Trump posts AI deepfake of Democratic leaders making racist remarks

Political deepfakes blur the line between fact and fiction in democratic discourse.

OpenAI launches Sora app for AI-generated videos to rival TikTok

Experts worry AI-generated content could undermine how we trust information online.

Friend’s $1M NYC subway ad campaign faces fierce, unfriendly anti-AI vandalism

The 22-year-old CEO deliberately chose New York knowing residents would hate his AI companion device.