DeX-Portrait: Disentangled and Expressive Portrait Animation via Explicit and Latent Motion Representations

CVPR 2026 (ACCEPT)


Yuxiang Shi1,*,†, Zhe Li2,*, Yanwen Wang3,†, Hao Zhu3, Xun Cao3, Ligang Liu1,✉️
1University of Science and Technology of China, 2Central Media Technology Institute, Huawei, 3Nanjing University,

Abstract

We introduce DeX-Portrait, a novel approach capable of generating expressive portrait animation driven by disentangled pose and expression signals. Specifically, we represent the pose as an explicit global transformation and the expression as an implicit latent code. In our framework, we firstly design a powerful motion trainer to learn both pose and expression encoders for extracting precise and decomposed driving signals. Then we propose to inject the pose transformation into the diffusion model through a dual-branch conditioning mechanism, and the expression latent through cross attention. Finally, we design a progressive hybrid classifier-free guidance for more faithful identity consistency. Experiments show that our method outperforms state-of-the-art baselines.



Cross-Reenactment





Disentangled-Reenactment




Expression-Only Editing




Pose-Only Editing




Citation


@article{shi2025dex,
  title={DeX-Portrait: Disentangled and Expressive Portrait Animation via Explicit and Latent Motion Representations},
  author={Shi, Yuxiang and Li, Zhe and Wang, Yanwen and Zhu, Hao and Cao, Xun and Liu, Ligang},
  journal={arXiv preprint arXiv:2512.15524},
  year={2025}
}