DeepSeek-V3.1 is a large language model that is post-trained based on DeepSeek-V3.1-Base. It has a 128K context length and supports mixed thinking modes, intelligent tool invocation, and code proxy functions. This model performs excellently in multiple benchmark tests, especially showing significant improvements in mathematical reasoning, code generation, and search agent tasks.
Natural Language Processing
TransformersMultiple Languages