DROP: Dynamics Responses from Human Motion Prior and Projective Dynamics

September 24, 2023 Β· Entered Twilight Β· πŸ› ACM SIGGRAPH Conference and Exhibition on Computer Graphics and Interactive Techniques in Asia

πŸ’€ TWILIGHT: Eternal Rest
Repo abandoned since publication

"No code URL or promise found in abstract"
"Derived repo from GitHub Pages (backfill)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, README.md, index.html, static

Authors Yifeng Jiang, Jungdam Won, Yuting Ye, C. Karen Liu arXiv ID 2309.13742 Category cs.GR: Graphics Cross-listed cs.CV Citations 14 Venue ACM SIGGRAPH Conference and Exhibition on Computer Graphics and Interactive Techniques in Asia Repository https://github.com/stanford-tml/drop Last Checked 1 month ago
Abstract
Synthesizing realistic human movements, dynamically responsive to the environment, is a long-standing objective in character animation, with applications in computer vision, sports, and healthcare, for motion prediction and data augmentation. Recent kinematics-based generative motion models offer impressive scalability in modeling extensive motion data, albeit without an interface to reason about and interact with physics. While simulator-in-the-loop learning approaches enable highly physically realistic behaviors, the challenges in training often affect scalability and adoption. We introduce DROP, a novel framework for modeling Dynamics Responses of humans using generative mOtion prior and Projective dynamics. DROP can be viewed as a highly stable, minimalist physics-based human simulator that interfaces with a kinematics-based generative motion prior. Utilizing projective dynamics, DROP allows flexible and simple integration of the learned motion prior as one of the projective energies, seamlessly incorporating control provided by the motion prior with Newtonian dynamics. Serving as a model-agnostic plug-in, DROP enables us to fully leverage recent advances in generative motion models for physics-based motion synthesis. We conduct extensive evaluations of our model across different motion tasks and various physical perturbations, demonstrating the scalability and diversity of responses.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

πŸ“œ Similar Papers

In the same crypt β€” Graphics

R.I.P. πŸ‘» Ghosted

Everybody Dance Now

Caroline Chan, Shiry Ginosar, ... (+2 more)

cs.GR πŸ› ICCV πŸ“š 820 cites 7 years ago
R.I.P. πŸ‘» Ghosted

Animating Human Athletics

Jessica K. Hodgins, Wayne L. Wooten, ... (+2 more)

cs.GR πŸ› SIGGRAPH πŸ“š 765 cites 3 years ago