DeepSeek V4 Preview brings 1M context and open weights

DeepSeek has released DeepSeek-V4 Preview, with the company saying the model is now live, open-sourced, and built around a 1 million token context window. For users in Samoa and across the Pacific, that combination matters because it points to AI systems that can handle far longer documents, larger codebases, and more extended workplace tasks in a single session.

The announcement covers two models. DeepSeek-V4-Pro is described as the larger option, with 1.6 trillion total parameters and 49 billion active parameters. DeepSeek-V4-Flash is the smaller and faster model, with 284 billion total parameters and 13 billion active parameters. DeepSeek says both are available through its API and chat interface.

Why the 1M context window matters

According to DeepSeek’s API docs, 1M context is now the default across its official services. That changes how people can use the system day to day. Instead of splitting a report, policy paper, contract, lesson plan, or software project into many smaller prompts, users can keep more of the material in one place.

For a government team reviewing long consultation documents, that means fewer breaks in the workflow. For educators working with course packs or research notes, it means the model can keep more material in view at once. For businesses, it may help with customer records, internal manuals, and project documentation, especially where consistency matters across long exchanges.

There’s still a practical limit, of course. A larger context window doesn’t remove the need for careful prompting or human review. It does, however, reduce the amount of manual chopping and re-pasting that often slows people down.

DeepSeek-V4-Pro and DeepSeek-V4-Flash at a glance

DeepSeek positions V4-Pro as the stronger model for reasoning, coding, and agentic tasks. The company says it leads current open models in world knowledge and performs strongly in math, STEM, and coding, while rivaling top closed-source models in some areas.

V4-Flash is aimed at speed and cost control. DeepSeek says it comes close to V4-Pro on reasoning, performs similarly on simpler agent tasks, and offers faster responses with lower API pricing. That makes it the more practical choice for users who need frequent, lower-cost interactions rather than the highest possible output on every task.

The split is straightforward:

- V4-Pro for heavier reasoning, coding, and complex document work

- V4-Flash for faster responses, simpler agent tasks, and tighter budgets

That choice will matter for Pacific organisations watching usage costs closely. If a team only needs drafting help, quick analysis, or routine automation, the smaller model may be enough. If the task involves long technical documents or more demanding reasoning, the larger model may be worth the extra cost.

What DeepSeek says about performance and structure

DeepSeek says V4-Pro shows strong results in agentic coding benchmarks and ranks highly among open models in reasoning tasks. The company also says the model uses a new attention design with token-wise compression and DeepSeek Sparse Attention, which it claims improves long-context efficiency while reducing compute and memory costs.

That technical detail matters because long-context models can become expensive to run if they are inefficient. Lower compute use can make a big difference for API pricing, server load, and response times. For smaller organisations, that may be the difference between testing a tool and actually rolling it out.

DeepSeek also says V4 is integrated with leading AI agents such as Claude Code, OpenClaw and OpenCode, and that it is already used in the company’s own coding workflows. For developers, that suggests the model is being positioned as a working tool rather than a demo release.

API access, model names and migration timing

The company says the API is available now. Users keep the same base URL and only need to update the model name to deepseek-v4-pro or deepseek-v4-flash. DeepSeek also says the models support OpenAI ChatCompletions and Anthropic APIs, which should make integration easier for teams already using those formats.

There’s also a migration deadline in the notice. DeepSeek says deepseek-chat and deepseek-reasoner will be retired and inaccessible after 24 July 2026 at 15:59 UTC, although they are currently being routed to DeepSeek-V4-Flash in thinking and non-thinking modes. Teams using those older names will need to plan ahead rather than leave the change until the last moment.

For organisations in Samoa and the Pacific, that means a simple audit is sensible:

  1. Check which DeepSeek model names are in use.
  2. Test V4-Pro and V4-Flash on real work examples.
  3. Review cost, response time, and output quality.
  4. Update internal notes and integrations before the retirement date.

What this could mean for Samoa and the Pacific

The main appeal of this release is practical. A model with a 1M context window can handle long documents more naturally, and the open-weight release gives technical teams more room to experiment. For ministries, schools, law firms, publishers, and local businesses, that could mean better support for document-heavy work.

In Samoa, where many teams work with limited time and small staff numbers, tools that reduce repetitive handling of information can have a real operational effect. A teacher preparing lesson materials, a business owner drafting client responses, or a public sector team comparing policy drafts may all benefit from a model that keeps more context in memory.

At the same time, the announcement is still a vendor release, not an independent benchmark report. DeepSeek’s claims about performance come from DeepSeek itself, so local users should test the models against their own tasks before making decisions about adoption.

A sensible way to test DeepSeek V4

For ARLO+ users, the best next step is to treat V4 as a trial candidate rather than a default replacement. Start with a small set of real tasks that reflect your actual work, then compare the two models side by side.

A practical test set might include:

- A long PDF or policy document that needs summarising

- A drafting task, such as a letter, memo, or lesson outline

- A coding or data task with multiple steps

- A simple customer service prompt

- A longer conversation that needs continuity across many turns

Measure more than just the final answer. Look at response time, cost, consistency, and how often the model loses track of earlier instructions. Those are the details that decide whether a tool is useful in a real office, classroom, or home setting.

DeepSeek’s V4 Preview shows how fast the AI market is moving towards longer context and lower-friction access. For Pacific users, the useful question isn’t whether the release sounds impressive, but whether it handles local work better than the tools already in use.

Sources