DeepSeek;1M-context V4 model;open-sources system

DeepSeek unveils 1M-context V4 model, open-sources system

Flash pricing starts at 0.2 yuan per million tokens for cached input.

by SONG Jianan

DeepSeek on April 24 released and open-sourced a preview of its DeepSeek-V4 model, featuring a 1 million-token context window and stronger reasoning and agent capabilities.

The model comes in two versions — V4-Pro and V4-Flash — both supporting 1M context, well above the 128K–256K range typical of most domestic models.

V4-Pro has 49 billion activated parameters and 33 trillion training tokens, while V4-Flash uses 13 billion parameters with 32 trillion tokens, targeting faster, lower-cost deployment.

The API supports OpenAI- and Anthropic-style interfaces. Existing DeepSeek-chat and DeepSeek-reasoner endpoints will be phased out within three months.

Flash pricing starts at 0.2 yuan per million tokens for cached input. Pro is more expensive and currently limited by high-end compute capacity, though costs may fall as Huawei's Ascend systems scale.

DeepSeek said V4-Pro ranks among the top open-source models in coding and agent benchmarks, approaching leading proprietary systems in some tasks.

The model uses a sparse attention architecture to extend context while reducing compute costs.

The launch lifted China's semiconductor index. Media reports also said Tencent and Alibaba are in talks to invest in DeepSeek at a valuation above $20 billion. The company did not comment.

来源:界面新闻

广告等商务合作,请点击这里

未经正式授权严禁转载本文,侵权必究。

打开界面新闻APP,查看原文
界面新闻
打开界面新闻,查看更多专业报道

热门评论

打开APP,查看全部评论,抢神评席位

热门推荐

    下载界面APP 订阅更多品牌栏目
      界面新闻
      界面新闻
      只服务于独立思考的人群
      打开