Xiaomi has recruited Luo Fuli, a core member of DeepSeek, with a salary reaching millions of yuan. Lei Jun personally promoted the recruitment to accelerate AGI layout. Luo Fuli will lead the team to build general artificial intelligence, reflecting Xiaomi's urgent demand for AI talents. Although Xiaomi open-sourced its reasoning large model MiMo in April, it still strengthens its strategy by introducing top talents to face industry competition.
On November 12, AI Prodigy "Luo Fuli" announced her joining of Xiaomi on WeChat Moments, expressing her excitement for future work. She emphasized her commitment to advancing AI from language to the physical world, and working with her team to build Artificial General Intelligence (AGI). Previously, the rumors about her joining had attracted widespread attention.
Xiaomi HyperAI won the top prize in the smart terminal category at the 'Direct to Wuzhen' global internet competition, beating 1,082 projects from 29 countries with its AI innovations.....
Xiaomi and Peking University co-published a paper on arXiv. Corresponding author Luo Fuli, noted for Lei Jun's high-salary recruitment, is affiliated with PKU's Computational Linguistics Institute, not Xiaomi's model team.....
Xiaomi's first large inference model, MiMo, is open - sourced and designed specifically for inference tasks with excellent performance.
Integrates Xiaomi's AIoT speaker with ChatGPT to create a personalized smart home voice assistant.
A large-scale pre-trained language model developed by Xiaomi with a parameter scale of 64 billion.
Enterprise Intelligent Service Solution
XiaomiMiMo
MiMo Audio is an audio language model developed by Xiaomi, which demonstrates strong few-shot learning ability through large-scale pre-training. This model breaks through the limitations of traditional models that rely on fine-tuning for specific tasks and performs excellently in tasks such as speech intelligence and audio understanding, reaching an advanced level among open-source models.
MiMo Audio is an audio language model based on large-scale pre-training, achieving SOTA performance among open-source models in speech intelligence and audio understanding benchmark tests. This model demonstrates strong few-shot learning ability and can generalize to tasks not included in the training data, supporting various audio tasks such as speech conversion, style transfer, and speech editing.
bartowski
This is a quantized version of XiaomiMiMo's MiMo-VL-7B-SFT-2508 model, optimized using llama.cpp to improve the model's running performance on specific hardware. This model is a vision-language model with 7 billion parameters, supporting image-to-text generation tasks.
This is the GGUF quantized version of the Xiaomi MiMo-VL-7B-RL-2508 model, quantized using the imatrix option of llama.cpp. It supports multiple quantization levels and is suitable for different hardware configurations and performance requirements.
mispeech
MiDashengLM-7B-0804 is a multimodal speech language model with 7B parameters released by Xiaomi, which supports audio understanding and text generation tasks and is suitable for inference and fine-tuning in general scenarios.
allura-forge
MiMo is a series of large language models trained from scratch by Xiaomi specifically for inference tasks. Through optimizing pre-training and post-training strategies, it demonstrates excellent performance in mathematical and code reasoning tasks. The project has open-sourced multiple versions with a 7B parameter scale, including the base model, SFT model, and RL model.
benxh
This is the GGUF quantized version of the XiaomiMiMo/MiMo-VL-7B-RL-2508 model, using the Q6_K quantization level. This model is a multimodal visual language model with a scale of 7B parameters, supporting joint understanding and generation tasks of images and text.
MiMo-VL is a compact and powerful vision-language model that combines a native resolution ViT encoder, an MLP projector, and the MiMo-7B language model. It performs excellently in tasks such as multimodal reasoning. This model shows outstanding performance in multiple benchmark tests, has a chain-of-thought control function, and significantly improves the user experience.
MiMo-VL is a compact and powerful vision-language model that combines a native resolution ViT encoder, an MLP projector, and the MiMo-7B language model optimized for complex reasoning tasks. Through multi-stage pre-training and post-training, it has achieved excellent results in multiple vision-language tasks.
MiMo is a series of 7B parameter models trained from scratch for inference tasks. Through optimized pre-training and post-training strategies, it performs excellently in mathematical and code reasoning tasks.
MiMo-7B is a language model series launched by Xiaomi, specifically designed for reasoning tasks, including base models, SFT models, and RL models, excelling in mathematical and code reasoning tasks.
MiMo-7B-RL is a reinforcement learning model trained based on the MiMo-7B-SFT model, demonstrating outstanding performance in mathematical and code reasoning tasks, comparable to OpenAI o1-mini.
A 7B-parameter specialized inference language model series launched by Xiaomi, significantly enhancing mathematical and code reasoning capabilities through optimized pre-training and post-training strategies
MiMo-7B-RL is a reinforcement learning model trained based on the MiMo-7B-SFT model, achieving performance comparable to OpenAI o1-mini in mathematical and code reasoning tasks.
Tonic
The GemmaX2-28-2B GGUF quantized model is a collection of quantized versions of the GemmaX2-28-2B-v0.1 translation large language model developed by Xiaomi, supporting machine translation tasks in 28 languages.
Implementation of the Xiaomi Cloud Notes MCP Server, supporting complete management functions such as note reading, searching, creation, editing, and deletion, and providing the ability to convert between Markdown and XML formats and upload images