DeepSeek V3 is the newest version of the large language model developed at DeepSeek, representing an enormous improvement in AI capability. This model has as many as 685 billion parameters, one of the largest publicly known models in the world. DeepSeek V3 uses a Mixture of Experts architecture: 256 experts, each activating 8 per token to bring about better performance and efficiency. DeepSeek V3 has performed fantastically on many benchmarks: DeepSeek V3 scored 48.4% in the Aider Polyglot leaderboard, which is second only to O1 and above other famous models such as Claude 3 Sonnet and Gemini, showing that DeepSeek V3 will do great on many programming languages and tasks.
Access to this model is presently being granted via a gradual rollout through chat.deepseek.com and the AIAPILAB. As this development continues to roll out, DeepSeek V3 will become another entrant into the ranks of large language models with increased functionality for a number of use cases.