What V4 Actually Ships
V4 arrives in two configurations. DeepSeek-V4-Pro is a 1.6-trillion-parameter Mixture-of-Experts architecture with 49 billion parameters activated per forward pass. DeepSeek-V4-Flash runs 284 billion total parameters with 13 billion active. Both support the one-million-token context window, placing the family in the same architectural league as Anthropic's Claude and Google's Gemini frontier releases. The open weights are published on Hugging Face and DeepSeek has documented inference recipes for self-hosted deployment.
The pricing is the headline number that will drive boardroom conversations across Frankfurt, Paris, and London. Consider the comparison across the major providers currently on European enterprise evaluation shortlists:
- DeepSeek V4-Flash output: $0.28 per million tokens, open weights, Huawei Ascend 950 native
- DeepSeek V4-Pro output: $3.48 per million tokens, open weights, Huawei Ascend 950 native
- Mistral Large (via Mistral API): approximately $8.00 per million tokens, partial open weights, Nvidia-anchored
- OpenAI GPT-class: $15 to $60 per million tokens, closed weights, Nvidia only
- Anthropic Claude: $15 to $75 per million tokens, closed weights, Nvidia only
That is the table every European chief information officer in financial services will circulate next week. The unit economics are simply not comparable. DeepSeek also claims V4-Pro can run autonomously on multi-file code writing and debugging, signalling explicitly that the company is now optimising for agentic workflows rather than conversational chat. That matters for banks building document-processing pipelines and insurers automating claims triage.
The Huawei Silicon Angle Is the Real Story for European Procurement
The chip story is as significant as the model benchmarks. DeepSeek has confirmed that V4 was trained and is being served on Huawei Ascend 950 clusters connected by Huawei's Supernode interconnect, with Cambricon providing supporting accelerators. This is the first time a globally competitive frontier-class model has been trained and served end-to-end on Chinese-designed silicon.
For European cloud buyers, that creates three distinct procurement implications:
- It removes the Nvidia hardware bottleneck for at least one high-quality open-weights option, which matters as H100 and H200 availability remains constrained across European hyperscaler regions.
- It establishes a proof point that frontier-grade performance does not require US-origin infrastructure, a fact that changes the theoretical supply-chain diversification arguments European firms have been making in risk registers.
- It sharpens a real architectural divergence between the Nvidia-anchored stacks used by most European cloud operators and a Huawei-anchored alternative that is now demonstrably usable for production-grade workloads.
The open weights are simultaneously usable on Nvidia H100, H200, and Blackwell systems, so European firms already invested in Nvidia infrastructure are not locked out. Most enterprise deployments will run a hybrid setup in practice.
Why European Regulators Will Have Opinions Quickly
Simon Willison, the AI researcher and creator of the Datasette tool, has described V4-Pro as "almost on the frontier, at a fraction of the price" in early public evaluations, a characterisation that independent testers have broadly confirmed. But in the European financial services context, benchmark performance is only the first question. Regulatory conformance is the second, and it is far harder.
The EU AI Act classifies a broad range of financial services AI applications, including credit scoring, fraud detection, and customer-facing advice tools, as high-risk systems. High-risk classification triggers mandatory third-party conformity assessments, logging obligations, and human-oversight requirements before deployment. DeepSeek's model card documents training data and architecture, but it does not provide the kind of independent safety evaluation that European notified bodies will require for high-risk use cases.
Dragomir Vatev, a senior technology policy analyst at the European Banking Authority, has previously noted in public statements that open-weights models create a distinct compliance challenge because the deploying institution, rather than the model developer, bears full responsibility for ensuring the system meets the Act's requirements. That means a European bank deploying V4-Pro in a credit-decision pipeline cannot point to DeepSeek's documentation as a substitute for its own conformity assessment.
Equally relevant is the position of the UK Financial Conduct Authority, which under its existing Model Risk Management guidelines and forthcoming AI-specific supervisory statements requires firms to maintain explainability and audit trails for material model decisions. A 1.6-trillion-parameter MoE is not inherently less explainable than any other large model, but the absence of a third-party audit trail from the developer complicates the compliance picture for UK-regulated firms.
Who Wins and Who Has to Move in the European AI Supplier Landscape
The clearest winner is any European enterprise that needs frontier-grade language capability for non-regulated, internally-hosted workloads. Document summarisation, internal knowledge retrieval, software development assistance, and research tooling are all legitimate production targets where the regulatory burden is lower and the cost savings are immediate.
The pressure falls hardest on mid-tier proprietary API providers operating in Europe who have been charging premium rates for moderate-capability models. The price floor has dropped, and it has dropped sharply. Mistral, which occupies a unique position as Europe's only domestically headquartered frontier lab, faces a genuine pricing challenge on its commercial API tiers, though its advantage in EU data-residency guarantees, GDPR-native infrastructure, and regulatory familiarity with French and EU authorities remains structurally valuable in ways that a Chinese open-weights model cannot replicate.
Yann LeCun, Chief AI Scientist at Meta and one of Europe's most prominent voices on open-source AI strategy, has argued consistently that open-weights models are foundational to technological sovereignty. V4's release gives that argument a commercially concrete form: a European bank can now download frontier-grade weights, run them inside its own data centre in Frankfurt or Dublin, and pay no per-token fee to any external provider. That is a genuine sovereignty option, if the regulatory and security questions can be resolved.
The governance catch is real. V4 weights ship under DeepSeek's open licence, but security teams at European financial institutions are already raising questions about training data provenance, potential embedding of behaviour that could be exploited, and the absence of evaluations conducted by bodies recognised under the EU AI Act. The working assumption among compliance officers currently is that V4 will be deployed in air-gapped or self-hosted configurations by enterprises that want the cost savings but cannot route sensitive customer data through DeepSeek's hosted API endpoints.
The Practical Deployment Path for European Financial Firms
The realistic near-term deployment pattern for regulated European firms will follow a clear sequence:
- Non-regulated internal tooling deployed on self-hosted V4-Flash within weeks, leveraging open weights and existing Nvidia or hybrid infrastructure.
- Pilot programmes for medium-risk applications, such as internal research assistance and unstructured document processing, running through Q3 2026 with internal conformity assessments being conducted in parallel.
- High-risk financial services applications, including credit, fraud, and customer-facing advice, held pending EU AI Act notified-body guidance on open-weights model evaluation, which is not expected to be fully clarified before 2027.
The firms that move fastest will be those in less-regulated corners of financial services: trading technology teams evaluating code-generation tooling, data engineering groups building internal data pipelines, and fintech startups outside the scope of high-risk classification. For them, V4-Flash at $0.28 per million tokens is simply the most cost-effective frontier option available today, and the open weights mean the unit economics do not deteriorate as usage scales.
DeepSeek V4 does not solve European AI sovereignty on its own. It introduces a new and compelling variable into a procurement conversation that European financial institutions have been conducting mostly in hypotheticals. The question now is whether European regulators, led by the European Banking Authority and national competent authorities, can provide conformity guidance fast enough for firms to act on an option that is, on pure capability and cost grounds, genuinely compelling.
Comments
Sign in to join the conversation. Be civil, be specific, link your sources.