DeepSeek releases preview of Open Source V4 AI Model

DeepSeek V4 AI

DeepSeek’s newly released V4 model marks a significant step forward in open‑source AI, combining long‑context capability with major architectural upgrades.

DeepSeek V4 arrives as a preview release, offering two variants — V4‑Pro and V4‑Flash — both designed to push the boundaries of efficiency and reasoning performance.

The headline feature is the one‑million‑token context window, enabling the model to process and retain far larger bodies of information than previous generations.

Positioning

This positions V4 as a strong contender in tasks requiring extended reasoning, research support, and complex agentic workflows.

The V4 series introduces a refined Hybrid Attention Architecture, combining compressed sparse and heavily compressed attention mechanisms to dramatically reduce computational overhead.

DeepSeek claims this approach cuts inference FLOPs and KV‑cache requirements to a fraction of those seen in earlier models, making long‑context operation more practical and cost‑effective.

V4‑Pro, the flagship model, includes a maximum reasoning‑effort mode, which the company says significantly advances open‑source reasoning performance and narrows the gap with leading closed‑source systems.

Meanwhile, V4‑Flash offers a more economical, faster alternative while retaining strong capability across everyday tasks.

Accelerating AI ambition

The release underscores China’s accelerating AI ambitions. DeepSeek’s earlier R1 model shook global markets with its low‑cost, high‑performance profile, and V4 continues that trajectory — now optimised for domestic chips and supported by growing local hardware ecosystems.

With open‑source availability and aggressive efficiency gains, DeepSeek V4 strengthens the company’s position as one of the most closely watched challengers in the global AI race.

And it’s far cheaper than its peers and not so power hungry either.