The research team at Tsinghua University has developed a 4-bit optimizer for neural network training, which significantly reduces the memory overhead of large model training. This optimizer cuts the GPU memory usage by up to 57% without compromising accuracy. Additionally, the team offers a ready-to-use 4-bit optimizer that can replace the existing ones, supporting low-precision versions of Adam and SGD. This research is crucial for addressing the GPU memory bottleneck in large model training.