More Authentic Than Original! Loopy Perfectly Matches Digital Avatars' Voices with Footage, Ending the Frustrating Disconnect
LOOPY technology, jointly developed by ByteDance and Zhejiang University, is an audio-driven video diffusion model designed to address the disconnect between audio and visuals in virtual avatar generation. This technology requires only a single frame of image and audio input to generate realistic, dynamic avatar movements that align with the audio rhythm and emotion, including non-speech actions, emotion-driven eyebrow and eye movements, and natural head movements. The core of LOOPY is its unique long-term motion information capture module, which supports various visual and audio styles for dynamic effects in virtual avatars.