DeepSeek Upgrades V3 AI Model

Edited by: Veronika Nazarova

Chinese company DeepSeek has released an updated language model, DeepSeek-V3-0324, now available on the Hugging Face platform. The model is built on a Mixture of Experts (MoE) architecture and includes 236 billion active parameters (from a total of 685 billion). This structure allows only a subset of the model to be activated per query, significantly reducing computational costs while maintaining high-quality output.

DeepSeek-V3-0324 was trained on carefully selected English and Chinese datasets, making it highly effective in bilingual contexts. It demonstrates strong contextual understanding, logical reasoning, and programming capabilities, along with improved accuracy when handling complex tasks.

The model supports chatbot functionalities, text generation, and processing, and can be applied across a variety of fields—from education and research to intelligent assistants and software automation. With open access via Hugging Face, developers can integrate it into their applications or use it to explore new AI solutions.

Did you find an error or inaccuracy?

We will consider your comments as soon as possible.