DeepSeek recently announced its new model DeepSeek v3, which is a significant improvement over its predecessor. The new model is said to be three times faster than v2 and has improved capabilities and intelligence. Like all previous DeepSeek models, v3 is open source. According to benchmarks, it outperforms existing models, including Claude 3.5 Sonnet and ChatGPT-4o, particularly on math and coding tasks such as HumanEval.
Introducing DeepSeek-V3!
Biggest leap forward so far:
60 tokens/second (3x faster than V2!)
Advanced features
API compatibility intact
Completely open source models and papers
1/n pic.twitter.com/p1dV9gJ2Sd
— DeepSeek (@deepseek_ai) December 26, 2024
With 671 billion parameters, DeepSeek v3 is the largest open source language model to date, surpassing LLaMA’s previous record of 405 billion parameters. The model is now available on Hugging Face and will be gradually rolled out to the DeepSeek chat interface to make it available to a wider audience.
DeepSeek is working on Deep Roles
Users can create their own roles and explore roles created by other users (custom GPTs?).
This feature is currently in early development
pic.twitter.com/0T4pGt93bi
— TestingCatalog news
(@testingcatalog) December 25, 2024
In addition to the new model, some hidden features have been discovered in the DeepSeek ecosystem. A notable feature in development is called Deep Roles, which allows users to explore “roles” created by others in both Chinese and English or design their own. Although this feature is still in its early stages, it appears to work similarly to custom GPTs and allows users to add personalized prompts to the DeepSeek LLM and share them publicly. However, the full scope of Deep Roles remains unclear and further updates are expected as the feature continues to develop.