This is an MXFP4 quantized version based on Huihui-gpt-oss-20b-BF16-abliterated-v2, specifically designed for text generation tasks. The model uses MXFP4 quantization technology, which significantly reduces the model size and inference cost while maintaining performance. It supports multiple deployment methods, including QAT, ollama, and GGUF formats.
Natural Language Processing
Transformers