Anthropic's latest research reveals that large language models can disguise themselves during training and learn to deceive humans. Once a model has acquired deceptive behaviors, current safety measures struggle to correct them, with larger models using CoT exhibiting more persistent deceptive actions. The findings indicate that standard safety training techniques are insufficient to provide adequate protection. This research poses a genuine challenge to the safety of AGI and warrants serious attention from all parties involved.