A showcase of generated motions driven by the unsupervised style of bird gliding. Our method synthesizes motions by combining textual descriptions of desired motion content with unsupervised style reference motions.
We compare our model to three baseline methods: StableMoFusion+MCM_LDM, StableMoFusion+DecouplingContact, and SMooDi. When the motion involves multiple actions, our method seamlessly integrates distinct styles into corresponding actions.
Style Reference
StableMoFusion+MCM_LDM
StableMoFusion+DecouplingContact
SMooDi
Ours
Style Reference
StableMoFusion+MCM_LDM
StableMoFusion+DecouplingContact
SMooDi
Ours
Even if the content of style motions diverges from texts, our method still produces compelling results.
Style Reference
StableMoFusion+MCM_LDM
StableMoFusion+DecouplingContact
SMooDi
Ours
Style Reference
StableMoFusion+MCM_LDM
StableMoFusion+DecouplingContact
SMooDi
Ours
Here we demonstrate a combination of content texts and style references that are out of distribution.
Style Reference
StableMoFusion+MCM_LDM
StableMoFusion+DecouplingContact
SMooDi
Ours
Our approach enables motion style transfer while ensuring superior preservation of content.
Content Reference
Style Reference
MCM_LDM
DecouplingContact
SMooDi
Ours
Our approach also allows for stylized motion in-between. Orange frames represent keyframes, and purple frames represent generated results.
MDM
Ours (+ style from keyframes)
Without explicit style control, previous diffusion methods use the statistically probable motions to reach target keyframes, disrupting the “old man” style. Our method, on the other hand, can use the style from keyframe sequences to create motion that retains the “old man” style.
Style Reference
Ours (+ style from reference)
Given a reference motion with a relaxed style, our approach enables a style transition from an “old man” pace to a relaxed pace.
If the style characteristics conflict with the content texts, our method prioritizes the content.
Content Reference
Style Reference