Cosh-DiT: Co-Speech Gesture Video Synthesis via
Hybrid Audio-Visual Diffusion Transformers

Yasheng Sun1,3*   Zhiliang Xu2*   Hang Zhou2   Jiazhi Guan4   Quanwei Yang5  
Kaisiyuan Wang2   Borong Liang2   Yingying Li2   Haocheng Feng2   Jingdong Wang2   Ziwei Liu6   Koike Hideki1
1. Center of Excellence for Generative AI, King Abdullah University of Science and Technology
2. Department of Computer Vision Technology (VIS), Baidu Inc.,
3. Tokyo Institute of Technology,
4. Department of Computer Science and Technology, Tsinghua University,
5. Department of Electronic Engineering and Information Science, University of Science and Technology of China,
6. S-Lab, Nanyang Technological University.

Abstract


Co-speech gesture video synthesis is a challenging task that requires both probabilistic modeling of human gestures and the synthesis of realistic images that align with the rhythmic nuances of speech. To address these challenges, we propose Cosh-DiT, a Co-speech gesture video system with hybrid Diffusion Transformers that perform audio-to-motion and motion-to-video synthesis using discrete and continuous diffusion modeling, respectively. First, we introduce an audio Diffusion Transformer (Cosh-DiT-A) to synthesize expressive gesture dynamics synchronized with speech rhythms. To capture upper body, facial, and hand movement priors, we employ vector-quantized variational autoencoders (VQ-VAEs) to jointly learn their dependencies within a discrete latent space. Then, for realistic video synthesis conditioned on the generated speech-driven motion, we design a visual Diffusion Transformer (Cosh-DiT-V) that effectively integrates spatial and temporal contexts. Extensive experiments demonstrate that our framework consistently generates lifelike videos with expressive facial expressions and natural, smooth gestures that align seamlessly with speech.

Demo Video



Materials



Citation


@misc{https://doi.org/10.48550/arxiv.2302.06857,
	
	
	author = {Yasheng Sun, Zhiliang Xu, Hang Zhou, Jiazhi Guan, Quanwei Yang, Kaisiyuan Wang, Borong Liang, Yingying Li, Haocheng Feng, Jingdong Wang, Ziwei Liu, Koike Hideki},
	title = {Cosh-DiT: Co-Speech Gesture Video Synthesis via Hybrid Audio-Visual Diffusion Transformers},
	publisher = {arXiv},
	year = {2025},
	copyright = {Creative Commons Attribution 4.0 International}
  }