Papers
arxiv:2412.01106

One Shot, One Talk: Whole-body Talking Avatar from a Single Image

Published on Dec 2
· Submitted by xiangjun-xj on Dec 5
Authors:
,
,
,

Abstract

Building realistic and animatable avatars still requires minutes of multi-view or monocular self-rotating videos, and most methods lack precise control over gestures and expressions. To push this boundary, we address the challenge of constructing a whole-body talking avatar from a single image. We propose a novel pipeline that tackles two critical issues: 1) complex dynamic modeling and 2) generalization to novel gestures and expressions. To achieve seamless generalization, we leverage recent pose-guided image-to-video diffusion models to generate imperfect video frames as pseudo-labels. To overcome the dynamic modeling challenge posed by inconsistent and noisy pseudo-videos, we introduce a tightly coupled 3DGS-mesh hybrid avatar representation and apply several key regularizations to mitigate inconsistencies caused by imperfect labels. Extensive experiments on diverse subjects demonstrate that our method enables the creation of a photorealistic, precisely animatable, and expressive whole-body talking avatar from just a single image.

Community

Paper author Paper submitter

One Shot, One Talk: Whole-body Talking Avatar from a Single Image

It enables the expressive 3D talking avatar reconstruction from a one-shot image (e.g., your favorite photo), which fully inherits the identity and supports realistic animation, including vivid body gestures and natural expression changes.

Project page: https://ustc3dv.github.io/OneShotOneTalk/
Arxiv: https://arxiv.org/abs/2412.01106

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2412.01106 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2412.01106 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2412.01106 in a Space README.md to link it from this page.

Collections including this paper 2