Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions regarding blendshapes #18

Open
HoiM opened this issue Mar 15, 2022 · 1 comment
Open

Questions regarding blendshapes #18

HoiM opened this issue Mar 15, 2022 · 1 comment

Comments

@HoiM
Copy link

HoiM commented Mar 15, 2022

Thank you for your great work. After reading the paper, I still have some questions. So I hope you or anyone else can answer me if possible.

In the paper, the Residual Deformation Branch learns to predict blendshapes for each individual character. I'm wondering how these blendshapes are defined.

I'm not familiar with body blendshapes but as far as I know, blendshapes for facial expressions like jawOpen, eyeBlinkLeft, smileRight, etc., are semantically defined. In the paper[1], personalized facial blendshapes are learned via blendshape gradient loss function, which forces each of the generated blendshapes to have specific semantic meaning.

Another way to use blendshapes for facial expression is like what was done in MetaHuman[2] (Unreal Engine), where expressions are produced by bones. Blendshapes (called morph targets in Unreal Engine) are used to refine the face, which add more details. I think this is more similar to your work.

So I would like to know some details on your blendshapes: how they are defined, how they are learned, etc.

I really appreciate it if you could answer my questions.

Ref:
[1] Personalized Face Modeling for Improved Face Reconstruction and Motion Retargeting
[2] MetaHuman

@PeizhuoLi
Copy link
Owner

Thanks for your great question! In general cases, each blend shape corresponds to some semantic meaning (smile, sad, etc.). But here we extend it such that the residual deformation is expressed as a linear combination of blend shapes, where each blend shape does not correspond to any semantic interpretation. I would say the "semantic" definition and how to learn them are somehow coupled: the blend shapes are an optimization target, such that the linear combination of them approximates the ground-truth best. For more details you might need to refer to the paper.

Hope it helps!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants