DeepSeek recently announced its new model, DeepSeek v3, which is a significant upgrade over its predecessor. The new model is reported to be three times faster than v2 and boasts enhanced capabilities and intelligence. Like all previous DeepSeek models, v3 is open source. According to benchmarks, it outperforms existing models, including Claude 3.5 Sonnet, and ChatGPT-4o, especially in math and coding tasks such as HumanEval.
With 671 billion parameters, DeepSeek v3 is the largest open-source language model to date, surpassing the previous record held by LLaMA at 405 billion parameters. The model is now available on Hugging Face and is gradually rolling out to the DeepSeek Chat UI, making it accessible to a wider audience.
In addition to the new model, some hidden features have been discovered in the DeepSeek ecosystem. One notable feature in development is called Deep Roles, which will allow users to explore “roles” created by others in both Chinese and English or design their own. While still in its early stages, this feature appears to function similarly to Custom GPTs, enabling users to add personalized prompts to the DeepSeek LLM and share them publicly. However, the full scope of Deep Roles remains unclear, and further updates are expected as the feature evolves.