VisemeNet: Audio-Driven Animator-Centric Speech Animation

May 24, 2018 Β· Declared Dead Β· πŸ› ACM Transactions on Graphics

πŸ‘» CAUSE OF DEATH: Ghosted
No code link whatsoever

"No code URL or promise found in abstract"

Evidence collected by the PWNC Scanner

Authors Yang Zhou, Zhan Xu, Chris Landreth, Evangelos Kalogerakis, Subhransu Maji, Karan Singh arXiv ID 1805.09488 Category cs.GR: Graphics Citations 159 Venue ACM Transactions on Graphics Last Checked 1 month ago
Abstract
We present a novel deep-learning based approach to producing animator-centric speech motion curves that drive a JALI or standard FACS-based production face-rig, directly from input audio. Our three-stage Long Short-Term Memory (LSTM) network architecture is motivated by psycho-linguistic insights: segmenting speech audio into a stream of phonetic-groups is sufficient for viseme construction; speech styles like mumbling or shouting are strongly co-related to the motion of facial landmarks; and animator style is encoded in viseme motion curve profiles. Our contribution is an automatic real-time lip-synchronization from audio solution that integrates seamlessly into existing animation pipelines. We evaluate our results by: cross-validation to ground-truth data; animator critique and edits; visual comparison to recent deep-learning lip-synchronization solutions; and showing our approach to be resilient to diversity in speaker and language.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

πŸ“œ Similar Papers

In the same crypt β€” Graphics

R.I.P. πŸ‘» Ghosted

Everybody Dance Now

Caroline Chan, Shiry Ginosar, ... (+2 more)

cs.GR πŸ› ICCV πŸ“š 820 cites 7 years ago
R.I.P. πŸ‘» Ghosted

Animating Human Athletics

Jessica K. Hodgins, Wayne L. Wooten, ... (+2 more)

cs.GR πŸ› SIGGRAPH πŸ“š 765 cites 3 years ago

Died the same way β€” πŸ‘» Ghosted