[PDF] Controllable Motion Synthesis and Reconstruction with Autoregressive Diffusion Models | Semantic Scholar (2024)

Skip to search formSkip to main contentSkip to account menu

Semantic ScholarSemantic Scholar's Logo
@article{Yin2023ControllableMS, title={Controllable Motion Synthesis and Reconstruction with Autoregressive Diffusion Models}, author={Wenjie Yin and Ruibo Tu and Hang Yin and Danica Kragic and Hedvig Kjellstr{\"o}m and M{\aa}rten Bj{\"o}rkman}, journal={2023 32nd IEEE International Conference on Robot and Human Interactive Communication (RO-MAN)}, year={2023}, pages={1102-1108}, url={https://api.semanticscholar.org/CorpusID:258048451}}
  • Wenjie Yin, Ruibo Tu, Mårten Björkman
  • Published in IEEE International Symposium… 3 April 2023
  • Computer Science, Engineering

MoDiff, an autoregressive probabilistic diffusion model over motion sequences conditioned on control contexts of other modalities, is introduced and the benefits of diffusion data dropout for robust synthesis and reconstruction of high-fidelity motion close to recorded data are shown.

1 Citation

Background Citations

1

Figures and Tables from this paper

  • figure 1
  • figure 2
  • figure 3
  • figure 4
  • figure 5
  • figure 6
  • table I
  • table II

Topics

Motion Sequences (opens in a new tab)Human Motion Synthesis (opens in a new tab)Autoregressive Diffusion Models (opens in a new tab)

Ask This Paper

BETA

AI-Powered

Our system tries to constrain to information found in this paper. Results quality may vary. Learn more about how we generate these answers.

Feedback?

One Citation

Scalable Motion Style Transfer with Constrained Diffusion Generation
    Wenjie YinYi YuHang YinDanica KragicMårten Björkman

    Computer Science

    AAAI

  • 2024

This work constructs the bias from the source domain keyframes and applies them as the gradient of content constraints, yielding a framework with keyframe manifold constraint gradients (KMCGs) and performs a human study for a subjective assessment of the quality of generated dance motions.

28 References

Graph-based Normalizing Flow for Human Motion Generation and Reconstruction
    Wenjie YinHang YinD. KragicMårten Björkman

    Computer Science

    2021 30th IEEE International Conference on Robot…

  • 2021

A probabilistic generative model is proposed to synthesize and reconstruct long horizon motion sequences conditioned on past information and control signals, such as the path along which an individual is moving, to reconstructing missing markers and achieving comparable results on generating realistic future poses.

A Recurrent Variational Autoencoder for Human Motion Synthesis
    I. HabibieDaniel HoldenJonathan SchwarzJ. YearsleyT. Komura

    Computer Science

    BMVC

  • 2017

A novel generative model of human motion that can be trained using a large motion capture dataset, and allows users to produce animations from high-level control signals that can predict the movements of the human body over long horizons more accurately than state-of-the-art methods.

  • 138
  • PDF
Combining Recurrent Neural Networks and Adversarial Training for Human Motion Synthesis and Control
    Zhiyong WangJinxiang ChaiShi-hong Xia

    Computer Science, Engineering

    IEEE Transactions on Visualization and Computer…

  • 2021

A new generative deep learning network for human motion synthesis and control that is compact, contact-aware, and can generate an infinite number of naturally looking motions with infinite lengths is introduced.

Human Motion Diffusion Model
    Guy TevetSigal RaabBrian GordonYonatan ShafirDaniel Cohen-OrAmit H. Bermano

    Computer Science

    ICLR

  • 2023

Motion Diffusion Model (MDM), a carefully adapted classifier-free diffusion-based generative model for the human motion domain, is introduced, a transformer-based approach, enabling different modes of conditioning, and different generation tasks.

Listen, Denoise, Action! Audio-Driven Motion Synthesis with Diffusion Models
    Simon AlexandersonRajmund NagyJ. BeskowG. Henter

    Computer Science

    ACM Trans. Graph.

  • 2023

This work adapts the DiffWave architecture to model 3D pose sequences, putting Conformers in place of dilated convolutions for improved modelling power and demonstrates control over motion style, using classifier-free guidance to adjust the strength of the stylistic expression.

On Human Motion Prediction Using Recurrent Neural Networks
    Julieta MartinezMichael J. BlackJ. Romero

    Computer Science

    2017 IEEE Conference on Computer Vision and…

  • 2017

It is shown that, surprisingly, state of the art performance can be achieved by a simple baseline that does not attempt to model motion at all, and a simple and scalable RNN architecture is proposed that obtains state-of-the-art performance on human motion prediction.

BeLFusion: Latent Diffusion for Behavior-Driven Human Motion Prediction
    Germán BarqueroSergio EscaleraCristina Palmero

    Computer Science

    2023 IEEE/CVF International Conference on…

  • 2023

BeLFusion is a model that, for the first time, leverages latent diffusion models in HMP to sample from a behavioral latent space where behavior is disentangled from pose and motion, and its predictions display a variety of behaviors that are significantly more realistic, and coherent with past motion than the state of the art.

Auto-Conditioned Recurrent Networks for Extended Complex Human Motion Synthesis
    Yi ZhouZimo LiShuangjiu XiaoC. HeZeng HuangHao Li

    Computer Science

    ICLR

  • 2018

This work is the first to the knowledge that demonstrates the ability to generate over 18,000 continuous frames (300 seconds) of new complex human motion w.r.t. different styles.

  • 202
  • PDF
MotionDiffuse: Text-Driven Human Motion Generation with Diffusion Model
    Mingyuan ZhangZhongang Cai Ziwei Liu

    Computer Science

    IEEE transactions on pattern analysis and machine…

  • 2024

The proposed MotionDiffuse is one of the first diffusion model-based text-driven motion generation frameworks, which demonstrates several desired properties over existing methods, and outperforms existing SoTA methods by convincing margins on text-driven motion generation and action-conditioned motion generation.

Deep Representation Learning for Human Motion Prediction and Classification
    Judith BütepageMichael J. BlackD. KragicH. Kjellström

    Computer Science

    2017 IEEE Conference on Computer Vision and…

  • 2017

The results show that deep feedforward networks, trained from a generic mocap database, can successfully be used for feature extraction from human motion data and that this representation can be used as a foundation for classification and prediction.

...

...

Related Papers

Showing 1 through 3 of 0 Related Papers

    [PDF] Controllable Motion Synthesis and Reconstruction with Autoregressive Diffusion Models | Semantic Scholar (2024)
    Top Articles
    Latest Posts
    Article information

    Author: Catherine Tremblay

    Last Updated:

    Views: 6779

    Rating: 4.7 / 5 (47 voted)

    Reviews: 86% of readers found this page helpful

    Author information

    Name: Catherine Tremblay

    Birthday: 1999-09-23

    Address: Suite 461 73643 Sherril Loaf, Dickinsonland, AZ 47941-2379

    Phone: +2678139151039

    Job: International Administration Supervisor

    Hobby: Dowsing, Snowboarding, Rowing, Beekeeping, Calligraphy, Shooting, Air sports

    Introduction: My name is Catherine Tremblay, I am a precious, perfect, tasty, enthusiastic, inexpensive, vast, kind person who loves writing and wants to share my knowledge and understanding with you.