The official repository of the paper EDTalk: Efficient Disentanglement for Emotional Talking Head Synthesis
The official repository of the paper EDTalk: Efficient Disentanglement for Emotional Talking Head Synthesis
Paper | Project Page | Code
Given an identity source, EDTalk synthesizes talking face videos characterized by mouth shapes, head poses, and expressions consistent with mouth GT, pose source and expression source. These facial dynamics can also be inferred directly from driven audio. Importantly, EDTalk demonstrates superior efficiency in disentanglement training compared to other methods.
@article{tan2024edtalk,
title={EDTalk: Efficient Disentanglement for Emotional Talking Head Synthesis},
author={Tan, Shuai and Ji, Bin and Bi, Mengxiao and Pan, Ye},
journal={arXiv preprint arXiv:2404.01647},
year={2024}
}
Some figures in the paper is inspired by:
The README.md template is borrowed from SyncTalk
Thanks for these great projects.