DeepSeek V4 - almost on the frontier, a fraction of the price

· ai · Source ↗

TLDR

  • DeepSeek releases V4-Pro (1.6T params, 49B active) and V4-Flash (284B, 13B active) with MIT license, undercutting all frontier model pricing.

Key Takeaways

  • V4-Flash costs $0.14/M input, $0.28/M output – cheaper than GPT-5.4 Nano; V4-Pro at $1.74/$3.48 undercuts Gemini 3.1 Pro and Claude Sonnet 4.6.
  • Both models support 1M token context; V4-Pro achieves only 27% single-token FLOPs and 10% KV cache size vs DeepSeek V3.2 at 1M context.
  • DeepSeek self-reports V4-Pro-Max trailing GPT-5.4 and Gemini-3.1-Pro by roughly 3-6 months on reasoning benchmarks.
  • V4-Pro at 865GB is the largest open-weights model released, exceeding Kimi K2.6 (1.1T) and GLM-5.1 (754B); Flash at 160GB may run quantized on 128GB M5 MacBook Pro.
  • The current API price includes a 75% discount valid through 2026-05-31, meaning post-discount pricing will be roughly 4x higher.

Hacker News Comment Review

  • The 75% API discount is active until May 31, 2026 and was not factored into the article’s pricing table – real post-promo costs change the competitive picture significantly.
  • Practical coding evaluations rate V4-Pro as Sonnet-level for implementation tasks but below Opus 4.7 max thinking for open-ended planning; a max thinking mode does exist in the API.
  • Data privacy risk of routing production codebases through DeepSeek’s API (direct or via OpenRouter) drew little community pushback despite analogous concerns raised about GitHub Copilot training data.

Notable Comments

  • @holysantamaria: Flags that the pricing page lists a 75% discount on V4-Pro extended until 2026-05-31 – not reflected in the article’s comparison table.
  • @zozbot234: Notes V4-Pro has a max thinking mode in the API, relevant for direct capability comparisons with Opus 4.7.

Original | Discuss on HN