On January 29, 2026, SenseTime officially announced the open-source release of its multimodal autonomous reasoning model SenseNova-MARS, along with two versions: 8B and 32B. The launch of this model marks a critical step forward in the field of multimodal large models for autonomous reasoning.

Technical Breakthrough: First Agentic VLM Model

SenseNova-MARS has achieved significant innovation in its technical architecture, making it the first Agentic VLM (Agentive Vision-Language Model) in the industry that integrates dynamic visual reasoning with text-image search.

Autonomous Reasoning: The model not only understands image content but also possesses autonomous planning and reasoning capabilities similar to those of an agent.

Deep Integration: By integrating real-time search capabilities into the visual understanding process, the model can handle complex visual tasks that require external knowledge support.

Industry Impact and Significance

SenseTime's decision to open-source both versions aims to provide global developers with more flexible research tools:

8B Version: Balances performance and efficiency, suitable for deployment on edge devices or in environments with limited computing power.

32B Version: Offers stronger logical reasoning capabilities, meeting the needs of complex industry applications.