DeepSeek has unveiled the preview version of its V4 series open-source models, licensed under MIT, with weights now available on Hugging Face and ModelScope. According to Odaily, the series includes two MoE models: V4-Pro, with approximately 1.6 trillion total parameters and 49 billion parameters activated per token, and V4-Flash, with 284 billion total parameters and 13 billion parameters activated. Both models support a context of 1 million tokens. The official statement highlights that compared to version V3.2, the new models significantly reduce memory usage and computational costs during long-text inference.

