This is a high-performance 4-bit quantized version of the Kimi K2 Instruct model, optimized for running on Apple Silicon (M1/M2/M3/M4) Macs using the MLX framework. The model has 671 billion parameters, supports a 128K context window, and achieves an excellent balance between quality and efficiency, making it an ideal choice for most practical deployments.
Natural Language Processing
MlxMultiple Languages