DeepSeek-V3’s development highlights the narrowing gap between open and closed-source AI, offering enterprises diverse options for AI integration. The model’s code is available on GitHub under an MIT license, with the model itself provided under DeepSeek’s license. Enterprises can test it via DeepSeek Chat and access the API for commercial use, with pricing set to change after February 8. This advancement underscores the potential for open-source models to compete with closed-source counterparts, promoting a more competitive and diverse AI landscape.
Key takeaways:
```html
- DeepSeek has released a new ultra-large AI model, DeepSeek-V3, with 671B parameters, using a mixture-of-experts architecture for efficient task handling.
- DeepSeek-V3 outperforms leading open-source models and closely matches the performance of closed models, marking a significant development in closing the gap between open and closed-source AI.
- The model introduces innovations like an auxiliary loss-free load-balancing strategy and multi-token prediction (MTP), enhancing training efficiency and speed.
- DeepSeek-V3 is available under the company's model license on Hugging Face, with the code accessible on GitHub under an MIT license, and enterprises can test it via DeepSeek Chat.