EMOCA: Emotion Driven Monocular Face Capture and Animation (CVPR 2022)

  Рет қаралды 5,374

Michael Black

Michael Black

2 жыл бұрын

As 3D facial avatars become more widely used for communication, it is critical that they faithfully convey emotion. Unfortunately, the best recent methods that regress parametric 3D face models from monocular images are unable to capture the full spectrum of facial expression, such as subtle or extreme emotions. We find the standard reconstruction metrics used for training (landmark reprojection error, photometric error, and face recognition loss) are insufficient to capture high-fidelity expressions. The result is facial geometries that do not match the emotional content of the input image. We address this with EMOCA (EMOtion Capture and Animation), by introducing a novel deep perceptual emotion consistency loss during training, which helps ensure that the reconstructed 3D expression matches the expression depicted in the input image. While EMOCA achieves 3D reconstruction errors that are on par with the current best methods, it significantly outperforms them in terms of the quality of the reconstructed expression and the perceived emotional content. We also directly regress levels of valence and arousal and classify basic expressions from the estimated 3D face parameters. On the task of in-the-wild emotion recognition, our purely geometric approach is on par with the best image-based methods, highlighting the value of 3D geometry in analyzing human behavior.
Project: emoca.is.tue.mpg.de/
Code: github.com/radekd91/emoca
Paper: ps.is.mpg.de/uploads_file/att...
arXiv: arxiv.org/abs/2204.11312
Reference:
@inproceedings{EMOCA:CVPR:2021,
title = {{EMOCA}: {E}motion Driven Monocular Face Capture and Animation},
author = {Danecek, Radek and Black, Michael J. and Bolkart, Timo},
booktitle = {Conference on Computer Vision and Pattern Recognition (CVPR)},
year = {2022}
}

Пікірлер: 17
@lelluc
@lelluc 2 жыл бұрын
This is wild, well done!
@cedricvillani8502
@cedricvillani8502 2 жыл бұрын
He failed the test by the way, she was going back to her fiancé not her husband
@cedricvillani8502
@cedricvillani8502 2 жыл бұрын
Just curious why after 6:49 you decided not to show a continuous sentiment label or emotional label I guess, or the overall analytics at the end to show how many times a label appeared? It would have been nice to see what the overall clip was estimating, was this a fight or argument, was there deception going on, did this person seem to be in danger or something else ❤. Ya know. Anyway absolutely amazing work, congratulations 🎉🎉 I love all the work you and your team does, I follow and read everything completely out of your faculty, wish I was there 😮😢, 😂🎉😂
@MichaelBlackMPI
@MichaelBlackMPI 2 жыл бұрын
Super question. We were more focused on using emotion to get accurate 3D than on the emotion recognition per se. But you are right that we could just show this, even though the processing is all single frame and doesn't take the temporal nature into account. Emotions really evolve over time and so I think it is important to model that. My very first work on facial expressions with Yaser Yacoob, used a very simple parametric model of face motion. From the parameters of the model over time, we recognized expressions surprisingly well for 1995! Here's the old video: kzbin.info/www/bejne/kJ-mmo2Ng9OBeZY
@mancumbus
@mancumbus 7 ай бұрын
Hello guys, but may be do you have video how to install code if you a not programmer? Because i 3d character animator in maya, and very interesting to try! OR maybe detailed instructions? Thank you!
@starstar-cr2hr
@starstar-cr2hr Жыл бұрын
Thank you for the amazing work! I’m wondering if there’s a way to apply this code to create a lively animated face, similar to Apple's Memoji, to replace man head in a video?.
@starstar-cr2hr
@starstar-cr2hr Жыл бұрын
Like, I will create a 3D animated character by analyzing the features of a person's face in a video. Using your code, I'll then map the appropriate facial expressions to this 3D character and replace the person's face with this animated figure. Does this sound feasible to you? Thanks in advance!
@boulimermoz9111
@boulimermoz9111 2 жыл бұрын
hello, thank you very much for your amazing work. Just asking : is there a way to apply this code and try this mocap system on my 3d characters ? thank you very much
2 жыл бұрын
in principle it's possible yes. but our code does not have this functionality. you would have to attach the FLAME face model (which is what we use) on your characters in place of the characters head. this is not trivial as there would probably be discontinuities around the neck which would then also have to be taken care of. btw, of you're interested in full body capture, be sure to check out projects such as PIXIE or Simplify-X.
@boulimermoz9111
@boulimermoz9111 2 жыл бұрын
@ Thank you very much, really appreciate
@ericmlevy
@ericmlevy Жыл бұрын
1. Can the model be exported without cropping to the roi box? 2. What can be done to improve the temporal stability / shakiness? Thank you!
@MichaelBlackMPI
@MichaelBlackMPI Жыл бұрын
1. The result is a full 3D FLAME head model. This cropping is only for display here. 2. EMOCA v2 is more stable (github.com/radekd91/emoca) and you can always run a 1-Euro filter if you still want more but it's pretty stable. 3. but also check out MICA, which is very stable. justusthies.github.io/posts/mica/
@liam9519
@liam9519 Жыл бұрын
Is this just DECA + an extra emotion detection model-based loss term?
@MichaelBlackMPI
@MichaelBlackMPI Жыл бұрын
Basically, yes. We take the DECA loss and add a term that says that the emotional content of the rendered image should match that of the original image. This is enough to improve the 3D realism of the mesh, without any explicit 3D training. This is what I find exciting. Emotion is a form of semantic "side information" (ie weak supervision) that is easy to get and can improve 3D shape estimation.
@liam9519
@liam9519 Жыл бұрын
​@@MichaelBlackMPI thanks for the response! I was having a read through the supplementary material and it seems this was not nearly as simple as my initial comment perhaps made it out to be :D Appreciate you open-sourcing the code too!
@MichaelBlackMPI
@MichaelBlackMPI Жыл бұрын
@@liam9519 no worries. Happy to help.
@phizc
@phizc 11 ай бұрын
To me it looks like it totally loses the identity compared to DECA. Also, the expressions also look exaggerated and not like in the original image. It would have been interesting to see the rendered deformed mesh with the extracted textures.
Useful gadget for styling hair 🤩💖 #gadgets #hairstyle
00:20
FLIP FLOP Hacks
Рет қаралды 9 МЛН
Finger Heart - Fancy Refill (Inside Out Animation)
00:30
FASH
Рет қаралды 25 МЛН
Alex hid in the closet #shorts
00:14
Mihdens
Рет қаралды 15 МЛН
[SIGGRAPH ASIA 2022] Video-driven Neural Physically-based Facial Asset for Production
5:14
Neural Head Avatars from Monocular RGB Videos (CVPR 2022)
3:36
Matthias Niessner
Рет қаралды 6 М.
Real-Time High-Fidelity Facial Performance Capture
4:33
DisneyResearchHub
Рет қаралды 455 М.
Semantic Deep Face Models
6:25
DisneyResearchHub
Рет қаралды 13 М.
АЙФОН 20 С ФУНКЦИЕЙ ВИДЕНИЯ ОГНЯ
0:59
КиноХост
Рет қаралды 1,2 МЛН
Какой ноутбук взять для учёбы? #msi #rtx4090 #laptop #юмор #игровой #apple #shorts
0:18