Scalable Motion Style Transfer with Constrained Diffusion Generation (2024)

Wenjie Yin1,Yi Yu2,Hang Yin3,Danica Kragic1,Mårten Björkman111footnotemark: 1Corresponding authors

Abstract

Current training of motion style transfer systems relies on consistency losses across style domains to preserve contents, hindering its scalable application to a large number of domains and private data. Recent image transfer works show the potential of independent training on each domain by leveraging implicit bridging between diffusion models, with the content preservation, however, limited to simple data patterns. We address this by imposing biased sampling in backward diffusion while maintaining the domain independence in the training stage. We construct the bias from the source domain keyframes and apply them as the gradient of content constraints, yielding a framework with keyframe manifold constraint gradients (KMCGs). Our validation demonstrates the success of training separate models to transfer between as many as ten dance motion styles. Comprehensive experiments find a significant improvement in preserving motion contents in comparison to baseline and ablative diffusion-based style transfer models. In addition, we perform a human study for a subjective assessment of the quality of generated dance motions. The results validate the competitiveness of KMCGs.

Introduction

Human motion style is a complex and important aspect of human behavior, which can be perceived as motion characteristics that convey personality and temper or articulate socio-cultural factors. The ability to accurately capture and replicate motion style is crucial in numerous applications, such as video games, and choreography.Style transfer systems can streamline creation of various media, including images(Gatys, Ecker, and Bethge 2016), music(Brunner etal. 2018), and indeed human movements(Dong etal. 2020). For instance, in the realm of video games, unique modes of action correspond to various player operations and character states. Similarly, in the field of dance choreography, each dance genre has its own movement patterns, and the application of style transfer can assist choreographers in creating variations of particular movements.

Scalable Motion Style Transfer with Constrained Diffusion Generation (1)

Early methods for motion style transfer(Hsu, Pulli, and Popović 2005; Taylor and Hinton 2009; Smith etal. 2019) primarily relied on supervised learning, requiring paired and annotated data. However, obtaining matched human motion sequences is challenging and often involves a laborious preprocessing phase. Consequently, the predominant style transfer methods today are unpaired translations that do not demand direct mappings between motion samples. Common methods among these are the motion transfer systems based on CycleGAN(Dong etal. 2020; Yin etal. 2023b) for paired style domains and StarGAN(Chan, Irimia, and Ho 2020; Yin etal. 2023c) for a set of domains.

While these systems are capable of generating high-quality human motion, they exhibit considerable limitations when it comes to scalability towards new domains. Specifically, CycleGAN-based methods are trained on distinct pairs of motion styles, where each model is tailored for transfer between a specific pair of styles(Dong etal. 2020; Yin etal. 2023b), or establish a shared domain bridging multiple styles via StarGAN-based architecture(Chan, Irimia, and Ho 2020; Yin etal. 2023c).

The former implies training a quadratically increasing number of models, making it an impractical solution when transfer between a large number of style domains is considered. The latter faces finding such a shared domain, which is an enormously challenging task. Both methods also show limited incremental scalability and adaptability; a new style requires retraining of the models on the entire dataset. The requirement of simultanous access to the entire dataset highlights an additional drawback of these methods when parts of the data are sensitive due to privacy concerns, such as in rehabilitation therapy. Developing methods that address these prominent challenges is thus critical for building practical motion style transfer systems.

In this paper, we propose a method based on Dual Diffusion Implicit Bridges (DDIBs)(Su etal. 2022) to mitigate both issues of system scalability and data privacy. DDIBs adopt the Schrödinger bridge perspective of diffusion models, showing certain information can be passed between diffusion latent spaces and used for image translation. Our system leverages this for a motion style transfer system with a two-step process. Given a source and a target model, the system initially utilizes the source model to obtain a latent encoding of motion at the final diffusion time step. This latent encoding is subsequently fed as the starting condition to the target model to generate the target motion (See Figure 1). The source and target domain models are completely decoupled, which allows for training the models separately. As a result, our system alleviates the need for a paired dataset that is typical for CycleGAN- and StarGAN-based methods, while upholding data privacy.

Our core technical contribution is improving content preservation of DDIBs in the motion style transfer domain. We find DDIBs struggle to retain the motion content faithfully when the analogy between source and target domains is low, consistent with the observations in image translation(Su etal. 2022). This is probably due to uncontrolled information encoding and decoding in independent diffusion processes. To this end, we propose a Keyframe Manifold Constraint Gradients (KMCGs) framework to improve content coherence of the target domain during inference.KMCGs uses keyframes from the source domain as context constraints and employ Manifold Constrained Gradient(Chung etal. 2022) to enforce these constraints during the second phase of DDIBs. Our experiments and quantative analysis on the 100STYLE(Mason, Starke, and Komura 2022) locomotion database, and the AIST++(Tsuchida etal. 2019) dance database find KMCGs achieves successful style transfer and better content preservation, reflected on the cycle consistency property on these two motion datasets and two probabilistic divergence-based metrics. Additionally, our human study as a subjective evaluation find samples generated from KMCGs are preferred in general. Overall, these evaluations demonstrate that our system significantly outperforms baseline methods.A video of the summary and examples can be accessed via https://youtu.be/98wEWavjnxI.

In summary, our contributions are:

  • A motion style transfer system that generates motions ranging from fundamental human locomotion to sophisticated dance movements. Our system pioneers in terms of system scalability and data privacy, demonstrating efficient and independent training over ten styles.

  • A technical method KMCGs that mitigates the content coherence issue of dual diffusion implicit bridges, which is found prominent in transferring complex motions.

  • A comprehensive evaluation of the proposed style transfer system including both objective metrics and subjective human study, showing significant performances boost compared to baseline and ablative models.

Related Work

We now review relevant prior works, including diffusion-based motion synthesis, as well as motion style transfer.

Diffusion-based Motion Synthesis

In light of recent groundbreaking advances made possible by diffusion models(Ho, Jain, and Abbeel 2020; Song and Ermon 2019), there has been a surge of interest in extending these techniques to the 3D motion domain. The recent MotionDiffuse(Zhang etal. 2022a) is regarded as the pioneering diffusion-based framework for text-driven motion generation. Similar to MotionDiffuse, the concurrent MDM(Tevet etal. 2022) and FLAME(Kim, Kim, and Choi 2022) integrate diffusion models and pre-trained language models such as CLIP(Radford etal. 2021) and RoBERTa(Liu etal. 2019) for generating motion from natural language descriptions. In a parallel development, speech audio is considered a model input(Zhang etal. 2023; Alexanderson etal. 2022) for gesture synthesis. Compared with gesture synthesis, dance generation is perhaps a more complex and challenging, but relatively under-explored field. Alexanderson etal. (2022) pioneers diffusion models with Conformer(Zhang etal. 2022b) for music to dance generation.EDGE(Tseng, Castellon, and Liu 2022) and Magic(Li etal. 2022) are Transformer-based diffusion models.EDGE incorporates auxiliary losses to encourage physical realism. Our work explores diffusion-based motion transfer on both human locomotion and dance databases.

One line of study focuses on manipulating partial body movements and styles. The modeling is hence not only about what contents are expressed by gestures but also how they are executed. HumanMAC(Chen etal. 2023) achieves controllable prediction of any part of the body. FLAME(Kim, Kim, and Choi 2022) and MDM(Tevet etal. 2022) enable body part editing both frame-wise and joint-wise by adapting diffusion inpainting to motion data. Tevet etal. (2022) design a multi-task architecture to unify human motion synthesis and stylization. Alexanderson etal. (2022) control the style and strength of motion expression by guided diffusion(Dhariwal and Nichol 2021). Yin etal. (2023a) integrate multimodal transformer and autoregressive diffusion models for controllable motion generation and reconstruction. These works often require joint training of models for body parts and styles while this is not desired in our work for the application of dual diffusion implicit bridges(Su etal. 2022).

Motion Style Transfer

The field of computer animation has long grappled with the challenge of motion style transfer, a process that entails transferring animation from a source style to a target style while retaining key content aspects, such as structure, timing, spatial relationships, etc.Early research in motion style transfer has depended on handcrafted features(Amaya, Bruderlin, and Calvert 1996; Unuma, Anjyo, and Takeuchi 1995; Witkin and Popovic 1995).As style is an elusive concept to define precisely, contemporary studies tend to endorse data-driven methodologies for feature extraction.Typical models utilized for style transfer include CycleGAN(Dong etal. 2020), AdaIN(Aberman etal. 2020), autoregressive flows(Wen etal. 2021), and etc.Additionally, some research concentrates on addressing real-time style transfer(Xia etal. 2015; Smith etal. 2019; Mason, Starke, and Komura 2022). Nonetheless, it is essential to highlight that these studies focus on relatively simplistic human motions, such as exercise and locomotion, where stylistic variation is restricted. CycleDance(Yin etal. 2023b) and StarDance(Yin etal. 2023c) address the transfer of dance movements that exhibit a substantial degree of complexity in terms of postures, transitions, rhythms, and artistic styles. However, CycleDance and StarDance suffer from severe drawbacks in rapid adaptation to alternative domains. As diffusion models advance, Alexanderson etal. (2022) demonstrate the ability to control the style and intensity of dance motion expression using a classifier-free guided diffusion model. Raab etal. (2023) propose the diffusion-based SinMDM to transfer the style of a given reference motion to learned motion motifs. Our proposed framework explores diffusion models with manifold constraints to facilitate the transfer of dance style. The process of style transfer hinges on the diffusion models independently trained within each respective domain. This method overcomes the adaptability limitations of CycleDance and StarDance. We further boost the performance of style transfer by imposing keyframe manifold constraints.

Method

In this section, we formulate the problem and provide preliminaries of diffusion models and Dual Diffusion Implicit Bridges (DDIBs).After this, the proposed system with Keyframe Manifold Constraint Gradients (KMCGs) is presented.

Scalable Motion Style Transfer with Constrained Diffusion Generation (2)
Scalable Motion Style Transfer with Constrained Diffusion Generation (3)

Problem Formulation

Our study aims to develop a translation method across multiple motion domains, denoted as source domain D(i)superscript𝐷𝑖D^{(i)}italic_D start_POSTSUPERSCRIPT ( italic_i ) end_POSTSUPERSCRIPT and target domain D(j)superscript𝐷𝑗D^{(j)}italic_D start_POSTSUPERSCRIPT ( italic_j ) end_POSTSUPERSCRIPT, without relying on jointly training over multiple data domains. In our scenario, we focus on conditional motion sequences. Given a sample sequence 𝒙(i)superscript𝒙𝑖\bm{x}^{(i)}bold_italic_x start_POSTSUPERSCRIPT ( italic_i ) end_POSTSUPERSCRIPT in the source domain and a conditional signal 𝒄𝒄\bm{c}bold_italic_c, our purpose is to transfer this sequence to the target domain, resulting in a sample sequence 𝒙(j)superscript𝒙𝑗\bm{x}^{(j)}bold_italic_x start_POSTSUPERSCRIPT ( italic_j ) end_POSTSUPERSCRIPT with the style of the target domain while striving to retain the content of the source domain.

Dual Diffusion Motion Transfer

To tackle the problem described above, we employ a strategy inspired by DDIB which use two separately trained Denoising Diffusion Implicit Models (DDIMs)(Song, Meng, and Ermon 2020) for image-to-image translation. With DDIMs, DDIBs achieve exact cycle consistency. We further boost the transfer performance by imposing keyframe context from the source motion as a manifold constraint.

Scalable Motion Style Transfer with Constrained Diffusion Generation (4)

Deep Diffusion Models.

Continuous-time diffusion processes {𝒙(t),t[0,1]}𝒙𝑡𝑡01\{\bm{x}(t),t\in[0,1]\}{ bold_italic_x ( italic_t ) , italic_t ∈ [ 0 , 1 ] } of diffusion models are defined as forward Stochastic Differential Equations (SDEs)(Song etal. 2020),

d𝒙=𝒇(𝒙,t)dt+g(t)d𝒘,𝑑𝒙𝒇𝒙𝑡𝑑𝑡𝑔𝑡𝑑𝒘d\bm{x}=\bm{f}(\bm{x},{t})d{t}+g({t})d\bm{w},italic_d bold_italic_x = bold_italic_f ( bold_italic_x , italic_t ) italic_d italic_t + italic_g ( italic_t ) italic_d bold_italic_w ,(1)

where 𝒘𝒘\bm{w}bold_italic_w is the standard Wiener process running backward in time, f(𝒙,t)𝑓𝒙𝑡f(\bm{x},t)italic_f ( bold_italic_x , italic_t ) is the drift term, g(t)𝑔𝑡g(t)italic_g ( italic_t ) is the scalar diffusion coefficient.The backward-time SDEs of Eqn. (1) are

d𝒙=[𝒇(𝒙,t)g2𝒙logpt(𝒙)]dt+g(t)d𝒘,𝑑𝒙delimited-[]𝒇𝒙𝑡superscript𝑔2subscript𝒙subscript𝑝𝑡𝒙𝑑𝑡𝑔𝑡𝑑𝒘d\bm{x}=[\bm{f}(\bm{x},{t})-g^{2}\nabla_{\bm{x}}\log p_{\it{t}}(\bm{x})]\,d{t}%+g({t})\,d\bm{w},italic_d bold_italic_x = [ bold_italic_f ( bold_italic_x , italic_t ) - italic_g start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT ∇ start_POSTSUBSCRIPT bold_italic_x end_POSTSUBSCRIPT roman_log italic_p start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ( bold_italic_x ) ] italic_d italic_t + italic_g ( italic_t ) italic_d bold_italic_w ,(2)

where 𝒙logpt(𝒙)subscript𝒙subscript𝑝𝑡𝒙\nabla_{\bm{x}}\log p_{t}(\bm{x})∇ start_POSTSUBSCRIPT bold_italic_x end_POSTSUBSCRIPT roman_log italic_p start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ( bold_italic_x ) is the score function of the noise perturbed data distribution at time t𝑡titalic_t. The actual implementations of diffusion models are sample discrete times. The time horizon t[0,1]𝑡01t\in[0,1]italic_t ∈ [ 0 , 1 ] is split up into T𝑇Titalic_T discretization segments as {𝒙k}k=0Tsuperscriptsubscriptsubscript𝒙𝑘𝑘0𝑇\{\bm{x}_{{k}}\}_{k=0}^{T}{ bold_italic_x start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT } start_POSTSUBSCRIPT italic_k = 0 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_T end_POSTSUPERSCRIPT, where k𝑘kitalic_k is an integer ranging from 00 to T𝑇Titalic_T. The forward and backward diffusion process can be defined as:

𝒙k=ak𝒙0+bkz,subscript𝒙𝑘subscript𝑎𝑘subscript𝒙0subscript𝑏𝑘𝑧\bm{x}_{{k}}={a}_{{k}}\bm{x}_{0}+{b}_{k}z,bold_italic_x start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT = italic_a start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT bold_italic_x start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT + italic_b start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT italic_z ,(3)
𝒙k1=f(𝒙k,𝒔θ)+g(𝒙k)z.subscript𝒙𝑘1𝑓subscript𝒙𝑘subscript𝒔𝜃𝑔subscript𝒙𝑘𝑧\bm{x}_{{k}-{1}}={f}(\bm{x}_{{k}},\bm{s}_{\theta})+{g}(\bm{x}_{k})z.bold_italic_x start_POSTSUBSCRIPT italic_k - 1 end_POSTSUBSCRIPT = italic_f ( bold_italic_x start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT , bold_italic_s start_POSTSUBSCRIPT italic_θ end_POSTSUBSCRIPT ) + italic_g ( bold_italic_x start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ) italic_z .(4)

We base our diffusion model on the EDGE architecture(Tseng, Castellon, and Liu 2022) to generate human motion contingent upon conditional signals. This architecture utilizes a transformer-based diffusion model that accepts conditional feature vectors as input. It then generates corresponding motion sequences as depicted in Figure2. The model incorporates a cross-attention mechanism, following(Saharia etal. 2022). We optimize the θ𝜃\thetaitalic_θ-parameterized score networks 𝒔θsubscript𝒔𝜃\bm{s}_{\theta}bold_italic_s start_POSTSUBSCRIPT italic_θ end_POSTSUBSCRIPT with paired conditional signal 𝒄𝒄\bm{c}bold_italic_c. The objective function is simplified as:

𝔼𝒙,k𝒙𝒔θ(𝒙k,k,𝒄)22.subscript𝔼𝒙𝑘subscriptsuperscriptnorm𝒙subscript𝒔𝜃subscript𝒙𝑘𝑘𝒄22\mathbb{E}_{\bm{x},{k}}\left\|\bm{x}-\bm{s}_{\theta}(\bm{x}_{{k}},{k},\bm{c})%\right\|^{2}_{2}.blackboard_E start_POSTSUBSCRIPT bold_italic_x , italic_k end_POSTSUBSCRIPT ∥ bold_italic_x - bold_italic_s start_POSTSUBSCRIPT italic_θ end_POSTSUBSCRIPT ( bold_italic_x start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT , italic_k , bold_italic_c ) ∥ start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT .(5)

Dual Diffusion Implicit Bridges.

Unlike GAN-based approaches for style transfer(Yin etal. 2023b, c), which do not inherently ensure cycle consistency, the DDIBs(Su etal. 2022) leverage the connections between score-based generative models(Song etal. 2020) and the Schrödinger Bridges Problem(Chen, Georgiou, and Pavon 2016). To optimize cycle consistency across two domains, GAN-based methods require additional components during training, necessitating simultaneous access to multiple domains. This imposes constraints on both the system scalability and data privacy. In contrast, DDIBs establish deterministic bridges between distributions, operating as a form of entropy-regularized optimal transport. Consequently, cycle consistency is ensured up to the discretization errors of the ODE solvers, avoiding simultaneous access to multiple domains for consistency.

The DDIBs-based strategy is depicted in Figure 3 and Algorithm 1. The motion transfer process involves encoding source motion 𝒙(i)superscript𝒙𝑖\bm{x}^{(i)}bold_italic_x start_POSTSUPERSCRIPT ( italic_i ) end_POSTSUPERSCRIPT and conditional signal 𝒄𝒄\bm{c}bold_italic_c with the source diffusion model 𝒔θ(i)superscriptsubscript𝒔𝜃𝑖\bm{s}_{\theta}^{(i)}bold_italic_s start_POSTSUBSCRIPT italic_θ end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ( italic_i ) end_POSTSUPERSCRIPT to latent encoding 𝒙(l)superscript𝒙𝑙\bm{x}^{(l)}bold_italic_x start_POSTSUPERSCRIPT ( italic_l ) end_POSTSUPERSCRIPT at the end time t=1𝑡1t=1italic_t = 1, and then decoding the latent encoding using the target model 𝒔θ(j)superscriptsubscript𝒔𝜃𝑗\bm{s}_{\theta}^{(j)}bold_italic_s start_POSTSUBSCRIPT italic_θ end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ( italic_j ) end_POSTSUPERSCRIPT to construct target motion 𝒙(j)superscript𝒙𝑗\bm{x}^{(j)}bold_italic_x start_POSTSUPERSCRIPT ( italic_j ) end_POSTSUPERSCRIPT at time t=0𝑡0t=0italic_t = 0. Both steps are defined via ODEs. We then employ ODE solvers to solve the ODE and construct 𝒙(j)superscript𝒙𝑗\bm{x}^{(j)}bold_italic_x start_POSTSUPERSCRIPT ( italic_j ) end_POSTSUPERSCRIPT at different times. In our experiment, we adopt DDIMs as the ODE solver. DDIMs generalize DDPMs via a class of non-Markovian diffusion processes that lead to the same training objective.

Input: source motion sample 𝒙(i)pD(i)similar-tosuperscript𝒙𝑖superscriptsubscript𝑝𝐷𝑖\bm{x}^{(i)}{\sim}p_{D}^{(i)}bold_italic_x start_POSTSUPERSCRIPT ( italic_i ) end_POSTSUPERSCRIPT ∼ italic_p start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ( italic_i ) end_POSTSUPERSCRIPT, conditional signal 𝒄𝒄\bm{c}bold_italic_c, source model 𝒔θ(i)superscriptsubscript𝒔𝜃𝑖\bm{s}_{\theta}^{(i)}bold_italic_s start_POSTSUBSCRIPT italic_θ end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ( italic_i ) end_POSTSUPERSCRIPT, and target model 𝒔θ(j)superscriptsubscript𝒔𝜃𝑗\bm{s}_{\theta}^{(j)}bold_italic_s start_POSTSUBSCRIPT italic_θ end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ( italic_j ) end_POSTSUPERSCRIPT

Transfer:

source to latent: 𝒙(l)=superscript𝒙𝑙absent\bm{x}^{(l)}=bold_italic_x start_POSTSUPERSCRIPT ( italic_l ) end_POSTSUPERSCRIPT = ODESolver(𝒙(i);𝒄,𝒔θ(i),0,1)superscript𝒙𝑖𝒄superscriptsubscript𝒔𝜃𝑖01(\bm{x}^{(i)};\bm{c},\bm{s}_{\theta}^{(i)},0,1)( bold_italic_x start_POSTSUPERSCRIPT ( italic_i ) end_POSTSUPERSCRIPT ; bold_italic_c , bold_italic_s start_POSTSUBSCRIPT italic_θ end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ( italic_i ) end_POSTSUPERSCRIPT , 0 , 1 )

latent to target: 𝒙(j)=superscript𝒙𝑗absent\bm{x}^{(j)}=bold_italic_x start_POSTSUPERSCRIPT ( italic_j ) end_POSTSUPERSCRIPT = ODESolver(𝒙(l);𝒄,𝒔θ(j),1,0)superscript𝒙𝑙𝒄superscriptsubscript𝒔𝜃𝑗10(\bm{x}^{(l)};\bm{c},\bm{s}_{\theta}^{(j)},1,0)( bold_italic_x start_POSTSUPERSCRIPT ( italic_l ) end_POSTSUPERSCRIPT ; bold_italic_c , bold_italic_s start_POSTSUBSCRIPT italic_θ end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ( italic_j ) end_POSTSUPERSCRIPT , 1 , 0 )

Output: 𝒙(j)superscript𝒙𝑗\bm{x}^{(j)}bold_italic_x start_POSTSUPERSCRIPT ( italic_j ) end_POSTSUPERSCRIPT, transferred motion in target domain.

The trained diffusion models in DDIBs can be regarded as a summary of the domains of datasets. When translating, these models strive to generate motions in the target domain that is closest to the source motion in terms of optimal transport distances. This process is both a strength and a limitation of DDIBs. In image translation, when source and target domains share similarities, DDIBs typically succeed in identifying correct content. A similar phenomenon is observed in human locomotion domains, as patterns of locomotion across different domains are similar. DDIBs generally transfer correctly on human gaits. However, when datasets show less similarity (e.g., birds and dogs), DDIBs may struggle to produce translation results that accurately retain the postures. In human motion, dance movements present a greater complexity and lesser similarity. Consequently, when employing DDIBs for dance style transfer, the system sometimes fails to preserve the content of the movements while transferring the style.

Content Constrained Style Transfer.

We propose incorporating additional corrective content to address the limitations of DDIBs discussed above:

𝒚=𝑯𝒙i+ϵ,𝒚𝑯superscript𝒙𝑖italic-ϵ\bm{y}=\bm{Hx}^{i}+\epsilon,bold_italic_y = bold_italic_H bold_italic_x start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT + italic_ϵ ,(6)

where 𝒚𝒚\bm{y}bold_italic_y is the measured content for motion correction, andϵitalic-ϵ\epsilonitalic_ϵ is the noise in the measurement. In our implementation, we extract keyframes from the source motion as corrective content to retain the postures from the source motion in the target motion. The keyframes are selected based on the acceleration of body joints.

The most straightforward method to incorporate additional information is explicitly adding these frames as part of model contexts. However, this approach has obvious drawbacks. It necessitates retraining the model, which negatively impacts the scalability of systems. Moreover, the inclusion of original motion frames as high-dimensional context can exacerbate the training stability. Nonetheless, we suggest this as an alternative way of enforcing domain information. A comparison to our proposed strategy can be found in the experimental evaluation section.

To avoid extra training, we use an additional corrective term by manifold constraints(Chung etal. 2022). We take the keyframes from the source motion sample as a constraint, which leads to the defined keyframe manifold constrained gradients (KMCGs).This term can synergistically operate with previous solvers to steer the reverse diffusion process closer to the context manifold.The reverse diffusion process in equation 4 can be replaced by:

𝒙k1=𝒇(𝒙k,sθ)α𝒙k(𝒚𝑯𝒙^0(𝒙k))22+g(𝒙k)𝒛,subscriptsuperscript𝒙𝑘1𝒇subscript𝒙𝑘subscript𝑠𝜃𝛼subscript𝒙𝑘superscriptsubscriptnorm𝒚𝑯subscript^𝒙0subscript𝒙𝑘22𝑔subscript𝒙𝑘𝒛\bm{x}^{{}^{\prime}}_{k-1}=\bm{f}(\bm{x}_{k},s_{\theta})-\alpha\frac{\partial}%{\partial\bm{x}_{k}}\|(\bm{y}-\bm{H}\hat{\bm{x}}_{0}(\bm{x}_{k}))\|_{2}^{2}+{g%}(\bm{x}_{k})\bm{z},bold_italic_x start_POSTSUPERSCRIPT start_FLOATSUPERSCRIPT ′ end_FLOATSUPERSCRIPT end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_k - 1 end_POSTSUBSCRIPT = bold_italic_f ( bold_italic_x start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT , italic_s start_POSTSUBSCRIPT italic_θ end_POSTSUBSCRIPT ) - italic_α divide start_ARG ∂ end_ARG start_ARG ∂ bold_italic_x start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT end_ARG ∥ ( bold_italic_y - bold_italic_H over^ start_ARG bold_italic_x end_ARG start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT ( bold_italic_x start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ) ) ∥ start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT + italic_g ( bold_italic_x start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ) bold_italic_z ,(7)
𝒙k1=𝑨𝒙k1+𝒃k,subscript𝒙𝑘1𝑨subscriptsuperscript𝒙𝑘1subscript𝒃𝑘\bm{x}_{k-1}=\bm{A}\bm{x}^{{}^{\prime}}_{{k}-1}+\bm{b}_{k},bold_italic_x start_POSTSUBSCRIPT italic_k - 1 end_POSTSUBSCRIPT = bold_italic_A bold_italic_x start_POSTSUPERSCRIPT start_FLOATSUPERSCRIPT ′ end_FLOATSUPERSCRIPT end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_k - 1 end_POSTSUBSCRIPT + bold_italic_b start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ,(8)

where 𝒙^0subscriptbold-^𝒙0\bm{\hat{x}}_{0}overbold_^ start_ARG bold_italic_x end_ARG start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT is estimated 𝒙0subscript𝒙0\bm{x}_{0}bold_italic_x start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT. α𝛼\alphaitalic_α depends on the noise covariance. Other parameters are defined by 𝑨=𝑰𝑯T𝑯𝑨𝑰superscript𝑯𝑇𝑯\bm{A}=\bm{I}-\bm{H}^{T}\bm{H}bold_italic_A = bold_italic_I - bold_italic_H start_POSTSUPERSCRIPT italic_T end_POSTSUPERSCRIPT bold_italic_H, 𝒃k=𝑯T𝒚subscript𝒃𝑘superscript𝑯𝑇𝒚\bm{b}_{k}=\bm{H}^{T}\bm{y}bold_italic_b start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT = bold_italic_H start_POSTSUPERSCRIPT italic_T end_POSTSUPERSCRIPT bold_italic_y.Here we omit the target domain index j𝑗jitalic_j in the reverse process for the notation brevity.

Although keyframes in the target domain are not available in style translation problem formulation, we found that the keyframe context in the source motion as a manifold constraint can also significantly improve the performance of dance style transfer.We illustrate our scheme in Figure4. The gradient term 𝒙k𝑾(𝒚𝑯𝒙^0(𝒙k))22subscript𝒙𝑘superscriptsubscriptnorm𝑾𝒚𝑯subscript^𝒙0subscript𝒙𝑘22\frac{\partial}{\partial\bm{x}_{k}}\|\bm{W}(\bm{y}-\bm{H}\hat{\bm{x}}_{0}(\bm{%x}_{k}))\|_{2}^{2}divide start_ARG ∂ end_ARG start_ARG ∂ bold_italic_x start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT end_ARG ∥ bold_italic_W ( bold_italic_y - bold_italic_H over^ start_ARG bold_italic_x end_ARG start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT ( bold_italic_x start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ) ) ∥ start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT incorporates the information of 𝒚𝒚\bm{y}bold_italic_y so the gradient of corrective term stays on the context manifold.

Experiments

To validate the capabilities of our motion style transfer system, we are eager to answer the questions: 1) can the system achieve scalable training? 2) can KMCGs enhance content preservation? 3) will the introduction of KMCGs compromise the strength of style transfer? In what follows, we present both objective and subjective evaluations of the system on domain standard human motion datasets with results favouring the proposed system and KMCGs.

Experimental Setting and Datasets

Experimental Setting.

We compare our proposed system with KMCGs (DDIBs-gradient) to the baseline DDIBs-vanilla. For a fair comparison, we also add another ablation setting, DDIBs-explicit, which also has a direct access to source domain keyframes but incoporated through cross-attention. To remain indpendent training, DDIBs-explicit trains each style model with its own keyframes.To evaluate the system performance, we quantitatively assess the transfer strength and content preservation based on two probabilistic divergence-based metrics. The concrete metrics are presented below. Qualitative user studies are also important in the evaluation of generative models as well. As a complementary, we performed a subjective evaluation in the form of online survey to evaluate human perceived quality in terms of naturalness and content preservation.

Datasets.

We evaluate our system on the 100STYLE(Mason, Starke, and Komura 2022) locomotion database and the AIST++(Tsuchida etal. 2019) dance database. We downsample both motion datasets to 30 fps and use 150-frame clips for experiments.

For the 100STYLE dataset, we employ forward walking movements that encompass 100 diverse styles, such as angry, neutral, stiff, etc. We transform the motion context into a form that includes 24 body joints with 3D rotations (72-dimensional vector), supplemented with frame-wise delta translation and delta rotation of the root joint (3-dimensional feature) as control signals.

The AIST++ dataset includes ten dance genres: break, pop, locking, waacking, middle hip-hop, LA hip-hop, house, krump, street-jazz, and ballet-jazz dance.We used the same pose representation as in(Tseng, Castellon, and Liu 2022) that represent dance as sequences of poses in 24-joint SMPL format(Loper etal. 2015) using the 6-DOF rotation representation and 4-dimensional binary foot contact label (151-dimensional vector). The music features are extracted with a frozen Jukebox model(Dhariwal etal. 2020), resulting in a 4800-dimensional feature.

Evaluations

Cycle Consistency.

A desirable feature of the motion style transfer system is the cycle consistency property, which means transforming a motion sequence from the source domain to the target domain, and then back to the source, will recover the original data point in the source domain.

Scalable Motion Style Transfer with Constrained Diffusion Generation (5)

Figure 5 illustrates the cycle consistency property guaranteed by DDIBs-based systems. It concerns the locomotion dataset. Starting from the source domain, DDIBs first obtain the latent encodings and construct the motion in the target domain. Next, DDIBs do translations in the reverse direction, transforming the motion back to the latent and the source domain. After this round trip, motions are approximately mapped to their original patterns. This cycle consistency can also be observed in the dance dataset as shown in Figure 6. DDIBs-based transfer systems restore almost the exact same motion patterns as the original ones, which implies that noised latent encoding indeed carried information that may recover content.Table 1 reports quantitative results on cycle-consistent translation among both locomotion and dance styles, by DDIB-vanilla and DDIB-gradient. The reported values are negligibly small and endorse the cycle-consistent property in the human motion domain.

Scalable Motion Style Transfer with Constrained Diffusion Generation (6)

Method/dataset100STYLEAIST++
DDIBs-vanilla0.0198±0.0083plus-or-minus0.01980.00830.0198\pm 0.00830.0198 ± 0.00830.0264±0.0035plus-or-minus0.02640.00350.0264\pm 0.00350.0264 ± 0.0035
DDIBs-gradient0.0192±0.0079plus-or-minus0.01920.00790.0192\pm 0.00790.0192 ± 0.00790.0232±0.0032plus-or-minus0.02320.00320.0232\pm 0.00320.0232 ± 0.0032

Transfer Performance.

The primary objective of this system is to transfer the motion style from a specified source domain to a particular target domain. We conduct assessments from both objective and subjective perspectives to ensure a comprehensive evaluation of the complex motion patterns common in dance. For the objective evaluation, we analyze the transfer strength and content preservation by 90 dance sequences for each style. Style transfer was performed on each ablated system and among all possible style pairs. Two metrics based on Fréchet distance are adopted(Yin etal. 2023c), which is computed by Equation 9:

FID=μrμg22+Tr(Σr+Σg2ΣrΣg),FIDsuperscriptsubscriptnormsubscript𝜇𝑟subscript𝜇𝑔22TrsubscriptΣ𝑟subscriptΣ𝑔2subscriptΣ𝑟subscriptΣ𝑔\displaystyle\text{FID}=\|\mu_{r}-\mu_{g}\|_{2}^{2}+\text{Tr}(\Sigma_{r}+%\Sigma_{g}-2\sqrt{\Sigma_{r}\Sigma_{g}}),FID = ∥ italic_μ start_POSTSUBSCRIPT italic_r end_POSTSUBSCRIPT - italic_μ start_POSTSUBSCRIPT italic_g end_POSTSUBSCRIPT ∥ start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT + Tr ( roman_Σ start_POSTSUBSCRIPT italic_r end_POSTSUBSCRIPT + roman_Σ start_POSTSUBSCRIPT italic_g end_POSTSUBSCRIPT - 2 square-root start_ARG roman_Σ start_POSTSUBSCRIPT italic_r end_POSTSUBSCRIPT roman_Σ start_POSTSUBSCRIPT italic_g end_POSTSUBSCRIPT end_ARG ) ,(9)

where (μr,Σr)subscript𝜇𝑟subscriptΣ𝑟(\mu_{r},\Sigma_{r})( italic_μ start_POSTSUBSCRIPT italic_r end_POSTSUBSCRIPT , roman_Σ start_POSTSUBSCRIPT italic_r end_POSTSUBSCRIPT ) and (μg,Σg)subscript𝜇𝑔subscriptΣ𝑔(\mu_{g},\Sigma_{g})( italic_μ start_POSTSUBSCRIPT italic_g end_POSTSUBSCRIPT , roman_Σ start_POSTSUBSCRIPT italic_g end_POSTSUBSCRIPT ) are, respectively, the mean and the covariance matrix of the real and generated dance movement distribution.

MethodFPD\downarrowFMD\downarrow
StarDance0.41380.1816
DDIBs-vanilla0.29040.1295
DDIBs-explicit0.17480.1313
DDIBs-gradient0.12080.1214

Scalable Motion Style Transfer with Constrained Diffusion Generation (7)

The Fréchet motion distance (FMD) quantifies transfer strength, which measures the extent that the motion is transferred from the source domain to the target domain. FMD calculates the distance between the true and generated dance motion distributions. The body joint acceleration and velocity are adopted as style-correlated features.The Fréchet pose distance (FPD) evaluates content preservation, which measures how well the salient poses of the source motion sequence are preserved after the transfer. The salient poses are detected by local maxima in joint acceleration and are normalized with respect to the hip-centric origin.

The quantitative results for the proposed DDIBs-gradient and baselines across various dance style pairings are presented in Table 2.For StarDance, establishing a shared domain that bridges multiple styles presents a significant challenge. Other methods exhibit comparable performance on transfer strength, shows the introduced constraints do not compromise style transfer. However, in terms of content preservation, the baseline system DDIBs-vanilla encounters difficulties, as indicated by the higher FPD value relative to DDIBs-explicit and DDIBs-gradient.The improvement in performance can be attributed to the explicit imposition of context information or through manifold constraint gradients.An example of synthesized motion sequences showcasing the transfer of dance style from locking to krump is provided in Figure 7. The middle sequence is generated by DDIBs-vanilla, whereas the bottom sequence is created by DDIBs-gradient, which integrates keyframe context through manifold constraint gradients. By comparing the poses in each column, the DDIBs-gradient achieves a higher similarity in body postures to the source gesture in terms of body orientations and limb/body shapes, thus preserving more content.Interestingly, DDIBs-gradient outperforms DDIBs-explicit. One possible explaination is that DDIBs-gradient gets to impose source domain related information across multiple intermediate diffusion steps. DDIBs-gradient is even more favoured for the possibility of directly using pre-trained models without requiring additional training and hyperparameter searching.

DatasetAdult2ChildAIST+100STYLEPowerWash
Number of Styles21010011080
Number of Models
CycleDance2909900122901720
DDIBs-based21010011080

Scalability and Data Privacy

We further compare our proposed method with CycleDance concerning the scalability.Table3 demonstrates that as the diversity of styles increases, our method scales the number of models linearly. In contrast, CycleDance surges the number of models quadratically, which would strain resources prohibitively.Consequently, CycleDance is appropriate for scenarios with a limited range of styles. As the scope of styles expands, our approach, characterized by its improved scalability, becomes increasingly advantageous.In instances where data sensitivity is a priority due to privacy issues, such as with the Emopain@Home(Olugbade etal. 2023) dataset in rehabilitation therapy or the PowerWash(Vuorre etal. 2023) dataset in player behavior analysis. Our method secures data privacy by decoupling the training procedure.

User Study.

To obtain a more comprehensive evaluation of our system and the baseline system, we conducted a user study in addition to the objective evaluation, asking participants to rate three aspects: motion naturalness, transfer strength, and content preservation. An online survey was performed to evaluate the transfer tasks. In this study, 21 participants between were recruited. Participants were between 24 and 35 years of age, 71.4% male and 28.6% female.We blend videos for each source and target dance sequence. During the survey, participants were presented with a source dance video clip followed by a generated target dance clip. To avoid potential order effects, the order of the target dance clips was randomly shuffled. Each target dance clip was generated either from the DDIBs-vanilla or DDIBs-gradient system. The participants were allowed to view the clips multiple times.

For motion naturalness and transfer strength, the participants were asked to rate how much they agree with the statement that the motion is natural or transferred to the target style, using a 1-5 Likert scale, where “1” means strongly disagree, and “5” means strongly agree. Figure 8 indicates that the performance on these two aspects is similar.We use the two one-sided tests to determine if the means of the two systems evaluations are equivalent.The equivalence margin was set at δ=0.5𝛿0.5\delta=0.5italic_δ = 0.5.The differences are 0.2 and 0.35, which fall within the specified equivalence margin.The results showed there is no statistically significant difference between the means for the motion naturalness and transfer strength of DDIBs-vanilla and the DDIBs-gradient, which indicates that the constraints from source motion content will not lead to a decline in the motion naturalness and style transfer.

Scalable Motion Style Transfer with Constrained Diffusion Generation (8)
Scalable Motion Style Transfer with Constrained Diffusion Generation (9)

For content preservation, the participants were asked to identify the features they believe are preserved between the source and the target motion, including orientations, limbs shape, body shape, and rhythmic patterns(Newlove and Dalby 2004). Figure 9 presents the overall statistics for the four aspects.To assess the statistical significance of the differences between the DDIBs-vanilla and DDIBs-gradient systems, we conducted a McNemar test. The test showed no significant differences for ‘limb shape’ (p=0.25𝑝0.25p=0.25italic_p = 0.25), and ‘rhythmic’ (p=0.47𝑝0.47p=0.47italic_p = 0.47). However, there was a statistically significant difference for ‘orientations’ (p=0.07𝑝0.07p=0.07italic_p = 0.07) and ‘body shape’ (p=0.0000961𝑝0.0000961p=0.0000961italic_p = 0.0000961). Both systems scored high on ‘rhythmic’, attributable to the DDIBs-based transfer system, which naturally preserved the rhythmic patterns due to its grounding in pre-trained music-conditioned diffusion models. In terms of preserving body trunk shape, the DDIBs-gradient outperformed DDIBs-vanilla because the posture information was provided as a constraint. As limb movements are more complex, participants reported that this content was not well preserved in either system.

Conclusions

This study tackles the challenging task of motion style transfer. We propose a dual diffusion-based method with keyframe manifold constrain gradients. Our solution first converts the source motion into a latent encoding by pretrained source model, and then produces motion in the target domain by the pretrained target model.This framework addresses the scalability and data privacy issues associated with GAN-based motion style transfer systems. We improve the transfer performance on content preservation by guiding the transfer process with keyframe manifold constraint gradients. Extensive evaluations demonstrate the efficacy and superior performance of the proposed method.

Acknowledgments

We extend our gratitude to Ruibo Tu and Xuejiao Zhao for all the inspiring discussions and their valuable feedback.

This research received partial support from the National Institute of Informatics (NII) in Tokyo. This work has been supported by the European Research Council (BIRD-884807) and H2020 EnTimeMent (no. 824160).This work benefited from access to the HPC resources provided by the Swedish National Infrastructure for Computing (SNIC), partially funded by the Swedish Research Council through grant agreement no. 2018-05973.

References

  • Aberman etal. (2020)Aberman, K.; Weng, Y.; Lischinski, D.; Cohen-Or, D.; and Chen, B. 2020.Unpaired motion style transfer from video to animation.ACM Transactions on Graphics (TOG), 39(4): 64–1.
  • Alexanderson etal. (2022)Alexanderson, S.; Nagy, R.; Beskow, J.; and Henter, G.E. 2022.Listen, denoise, action! Audio-driven motion synthesis with diffusion models.arXiv preprint arXiv:2211.09707.
  • Amaya, Bruderlin, and Calvert (1996)Amaya, K.; Bruderlin, A.; and Calvert, T. 1996.Emotion from motion.In Graphics interface, volume96, 222–229. Toronto, Canada.
  • Brunner etal. (2018)Brunner, G.; Wang, Y.; Wattenhofer, R.; and Zhao, S. 2018.Symbolic music genre transfer with cyclegan.In 2018 ieee 30th international conference on tools with artificial intelligence (ictai), 786–793. IEEE.
  • Chan, Irimia, and Ho (2020)Chan, J.C.; Irimia, A.-S.; and Ho, E. 2020.Emotion Transfer for 3D Hand Motion using StarGAN.
  • Chen etal. (2023)Chen, L.-H.; Zhang, J.; Li, Y.; Pang, Y.; Xia, X.; and Liu, T. 2023.HumanMAC: Masked Motion Completion for Human Motion Prediction.arXiv preprint arXiv:2302.03665.
  • Chen, Georgiou, and Pavon (2016)Chen, Y.; Georgiou, T.T.; and Pavon, M. 2016.On the relation between optimal transport and Schrödinger bridges: A stochastic control viewpoint.Journal of Optimization Theory and Applications, 169: 671–691.
  • Chung etal. (2022)Chung, H.; Sim, B.; Ryu, D.; and Ye, J.C. 2022.Improving diffusion models for inverse problems using manifold constraints.Advances in Neural Information Processing Systems, 35: 25683–25696.
  • Dhariwal etal. (2020)Dhariwal, P.; Jun, H.; Payne, C.; Kim, J.W.; Radford, A.; and Sutskever, I. 2020.Jukebox: A generative model for music.arXiv preprint arXiv:2005.00341.
  • Dhariwal and Nichol (2021)Dhariwal, P.; and Nichol, A. 2021.Diffusion models beat gans on image synthesis.Advances in Neural Information Processing Systems, 34: 8780–8794.
  • Dong etal. (2020)Dong, Y.; Aristidou, A.; Shamir, A.; Mahler, M.; and Jain, E. 2020.Adult2child: Motion style transfer using cyclegans.In Motion, Interaction and Games, 1–11.
  • Gatys, Ecker, and Bethge (2016)Gatys, L.A.; Ecker, A.S.; and Bethge, M. 2016.Image style transfer using convolutional neural networks.In Proceedings of the IEEE conference on computer vision and pattern recognition, 2414–2423.
  • Ho, Jain, and Abbeel (2020)Ho, J.; Jain, A.; and Abbeel, P. 2020.Denoising diffusion probabilistic models.Advances in Neural Information Processing Systems, 33: 6840–6851.
  • Hsu, Pulli, and Popović (2005)Hsu, E.; Pulli, K.; and Popović, J. 2005.Style translation for human motion.In ACM SIGGRAPH 2005 Papers, 1082–1089.
  • Kim, Kim, and Choi (2022)Kim, J.; Kim, J.; and Choi, S. 2022.Flame: Free-form language-based motion synthesis & editing.arXiv preprint arXiv:2209.00349.
  • Li etal. (2022)Li, R.; Zhao, J.; Zhang, Y.; Su, M.; Ren, Z.; Zhang, H.; and Li, X. 2022.Magic: Multi Art Genre Intelligent Choreography Dataset and Network for 3D Dance Generation.arXiv preprint arXiv:2212.03741.
  • Liu etal. (2019)Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; and Stoyanov, V. 2019.Roberta: A robustly optimized bert pretraining approach.arXiv preprint arXiv:1907.11692.
  • Loper etal. (2015)Loper, M.; Mahmood, N.; Romero, J.; Pons-Moll, G.; and Black, M.J. 2015.SMPL: A skinned multi-person linear model.ACM transactions on graphics (TOG), 34(6): 1–16.
  • Mason, Starke, and Komura (2022)Mason, I.; Starke, S.; and Komura, T. 2022.Real-Time Style Modelling of Human Locomotion via Feature-Wise Transformations and Local Motion Phases.arXiv preprint arXiv:2201.04439.
  • Newlove and Dalby (2004)Newlove, J.; and Dalby, J. 2004.Laban for all.Taylor & Francis US.
  • Olugbade etal. (2023)Olugbade, T.; Buono, R.; Potapov, K.; Bujorianu, A.; Williams, A.; deOssornoGarcia, S.; Gold, N.; Holloway, C.; and Berthouze, N. 2023.The EmoPain@ Home Dataset: Capturing Pain Level and Activity Recognition for People with Chronic Pain in Their Homes.
  • Raab etal. (2023)Raab, S.; Leibovitch, I.; Tevet, G.; Arar, M.; Bermano, A.H.; and Cohen-Or, D. 2023.Single Motion Diffusion.arXiv preprint arXiv:2302.05905.
  • Radford etal. (2021)Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; etal. 2021.Learning transferable visual models from natural language supervision.In ICML, 8748–8763. PMLR.
  • Saharia etal. (2022)Saharia, C.; Chan, W.; Saxena, S.; Li, L.; Whang, J.; Denton, E.L.; Ghasemipour, K.; GontijoLopes, R.; KaragolAyan, B.; Salimans, T.; etal. 2022.Photorealistic text-to-image diffusion models with deep language understanding.Advances in Neural Information Processing Systems, 35: 36479–36494.
  • Smith etal. (2019)Smith, H.J.; Cao, C.; Neff, M.; and Wang, Y. 2019.Efficient neural networks for real-time motion style transfer.Proceedings of the ACM on Computer Graphics and Interactive Techniques, 2(2): 1–17.
  • Song, Meng, and Ermon (2020)Song, J.; Meng, C.; and Ermon, S. 2020.Denoising diffusion implicit models.arXiv preprint arXiv:2010.02502.
  • Song and Ermon (2019)Song, Y.; and Ermon, S. 2019.Generative modeling by estimating gradients of the data distribution.Advances in neural information processing systems, 32.
  • Song etal. (2020)Song, Y.; Sohl-Dickstein, J.; Kingma, D.P.; Kumar, A.; Ermon, S.; and Poole, B. 2020.Score-based generative modeling through stochastic differential equations.arXiv preprint arXiv:2011.13456.
  • Su etal. (2022)Su, X.; Song, J.; Meng, C.; and Ermon, S. 2022.Dual diffusion implicit bridges for image-to-image translation.arXiv preprint arXiv:2203.08382.
  • Taylor and Hinton (2009)Taylor, G.W.; and Hinton, G.E. 2009.Factored conditional restricted Boltzmann machines for modeling motion style.In Proceedings of the 26th annual international conference on machine learning, 1025–1032.
  • Tevet etal. (2022)Tevet, G.; Raab, S.; Gordon, B.; Shafir, Y.; Cohen-Or, D.; and Bermano, A.H. 2022.Human Motion Diffusion Model.arXiv.
  • Tseng, Castellon, and Liu (2022)Tseng, J.; Castellon, R.; and Liu, C.K. 2022.EDGE: Editable Dance Generation From Music.arXiv preprint arXiv:2211.10658.
  • Tsuchida etal. (2019)Tsuchida, S.; f*ckayama, S.; Hamasaki, M.; and Goto, M. 2019.AIST Dance Video Database: Multi-Genre, Multi-Dancer, and Multi-Camera Database for Dance Information Processing.In ISMIR, volume1, 6.
  • Unuma, Anjyo, and Takeuchi (1995)Unuma, M.; Anjyo, K.; and Takeuchi, R. 1995.Fourier principles for emotion-based human figure animation.In Proceedings of the 22nd annual conference on Computer graphics and interactive techniques, 91–96.
  • Vuorre etal. (2023)Vuorre, M.; Magnusson, K.; Johannes, N.; Butlin, J.; and Przybylski, A.K. 2023.An intensive longitudinal dataset of in-game player behaviour and well-being in PowerWash Simulator.Scientific Data, 10(1): 622.
  • Wen etal. (2021)Wen, Y.-H.; Yang, Z.; Fu, H.; Gao, L.; Sun, Y.; and Liu, Y.-J. 2021.Autoregressive Stylized Motion Synthesis with Generative Flow.In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 13612–13621.
  • Witkin and Popovic (1995)Witkin, A.; and Popovic, Z. 1995.Motion warping.In Proceedings of the 22nd annual conference on Computer graphics and interactive techniques, 105–108.
  • Xia etal. (2015)Xia, S.; Wang, C.; Chai, J.; and Hodgins, J. 2015.Realtime style transfer for unlabeled heterogeneous human motion.ACM Transactions on Graphics (TOG), 34(4): 1–10.
  • Yin etal. (2023a)Yin, W.; Tu, R.; Yin, H.; Kragic, D.; Kjellström, H.; and Björkman, M. 2023a.Controllable Motion Synthesis and Reconstruction with Autoregressive Diffusion Models.arXiv preprint arXiv:2304.04681.
  • Yin etal. (2023b)Yin, W.; Yin, H.; Baraka, K.; Kragic, D.; and Björkman, M. 2023b.Dance style transfer with cross-modal transformer.In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 5058–5067.
  • Yin etal. (2023c)Yin, W.; Yin, H.; Baraka, K.; Kragic, D.; and Björkman, M. 2023c.Multimodal dance style transfer.Machine Vision and Applications, 34(4): 1–14.
  • Zhang etal. (2023)Zhang, F.; Ji, N.; Gao, F.; and Li, Y. 2023.DiffMotion: Speech-Driven Gesture Synthesis Using Denoising Diffusion Model.arXiv preprint arXiv:2301.10047.
  • Zhang etal. (2022a)Zhang, M.; Cai, Z.; Pan, L.; Hong, F.; Guo, X.; Yang, L.; and Liu, Z. 2022a.Motiondiffuse: Text-driven human motion generation with diffusion model.arXiv.
  • Zhang etal. (2022b)Zhang, M.; Liu, C.; Chen, Y.; Lei, Z.; and Wang, M. 2022b.Music-to-Dance Generation with Multiple Conformer.In Proceedings of the 2022 International Conference on Multimedia Retrieval, 34–38.
Scalable Motion Style Transfer with Constrained Diffusion Generation (2024)
Top Articles
Latest Posts
Article information

Author: Stevie Stamm

Last Updated:

Views: 6765

Rating: 5 / 5 (80 voted)

Reviews: 95% of readers found this page helpful

Author information

Name: Stevie Stamm

Birthday: 1996-06-22

Address: Apt. 419 4200 Sipes Estate, East Delmerview, WY 05617

Phone: +342332224300

Job: Future Advertising Analyst

Hobby: Leather crafting, Puzzles, Leather crafting, scrapbook, Urban exploration, Cabaret, Skateboarding

Introduction: My name is Stevie Stamm, I am a colorful, sparkling, splendid, vast, open, hilarious, tender person who loves writing and wants to share my knowledge and understanding with you.