Zhipu officially announced the open-source of the next-generation image generation model

GLM-Image supports both text-to-image and image-to-image generation within a single model.
- Text-to-image: Generate high-detail images based on text descriptions, performing especially well in information-dense scenarios.
- Image-to-image: Supports various tasks including image editing, style transfer, multi-subject consistency, and identity-preserving generation of people and objects.
In terms of technical indicators,
Currently,


GitHub:https://github.com/zai-org/GLM-Image
Hugging Face:https://huggingface.co/zai-org/GLM-Image
Key points:
🇨🇳 Domestic full-stack self-researched: Completed the full workflow training based on Huawei Ascend Atlas800T A2 devices and MindSpore framework, verifying the feasibility of training top-tier models on domestic computing power.
🎨 Breakthrough in text and image fusion: Adopting a hybrid architecture, it ranked first among open-source models in LongText-Bench and other long-text rendering rankings, significantly improving the accuracy of Chinese character and complex text-image generation.
💰 High-cost-effective open-source: The model supports adaptive image generation across various resolutions and is open to creators at extremely low API prices, aiming to promote the popularization of domestic cognitive generation technology.


