Embodied AI · Motion Generation · World Models

Xiangyue ZHANG (Ian)

M.Sc. student in Computer Applications, Wuhan University

Hi there 👋. I am Xiangyue ZHANG (章湘粤), an incoming Ph.D. student in Mechano-Informatics at The University of Tokyo, supervised by Prof. Tatsuya Harada.

My research focuses on Embodied AI, 3D/2D motion generation, and world models, with a long-term interest in how intelligent agents perceive, move, and interact in physical environments.

Outside research, music is where I learn timing without equations. I sing for Xia Qian Yu (夏千嶼), an indie-rock band, and have performed at music festivals, bars, campus shows, and solo stages. Music keeps my research from becoming too mechanical: it asks me to listen before moving, to feel rhythm before explaining it, and to remember that expressive motion is something lived before it is modeled.

I am open to remote or on-site internship and visiting opportunities around embodied intelligence, motion generation, and human-centered AI systems.

News

Mar 31, 2026
🎉 MACE-Dance was accepted by SIGGRAPH 2026.
Nov 8, 2025
🎉 GlobalDiff was accepted by AAAI 2026.
Jul 5, 2025
🎉 EchoMask was accepted by ACM MM 2025.
Jun 26, 2025
🎉 SemTalk was accepted by ICCV 2025.
Apr 16, 2025
🎉 One paper has been accepted by IEEE T-CSVT.
Dec 22, 2024
🥂 Our band performed successfully at the Hua Young Music Festival! Cheers! Check More to see pictures.


Featured Open-Source System

  1. auto-deep-researcher-banner.png
    Technical Report
    Deep Researcher Agent: An Autonomous Framework for 24/7 Deep Learning Experimentation with Zero-Cost Monitoring
    Xiangyue Zhang
    arXiv preprint arXiv:2604.05854, 2026


Selected Publications

  1. PersonaGesture.png
    arXiv 2026
    PersonaGesture: Single-Reference Co-Speech Gesture Personalization for Unseen Speakers
    Xiangyue Zhang, Yiyi Cai, Kunhang Li, Kaixing Yang, You Zhou, Zhengqing Li, Xuangeng Chu, Jiaxu Zhang, and Haiyang Liu
    arXiv, 2026
  2. GlobalDiff.png
    AAAI 2026
    Mitigating Error Accumulation in Co-Speech Motion Generation via Global Rotation Diffusion and Multi-Level Constraints
    Xiangyue Zhang*, Jianfang Li*†, Jianqiang Ren, and Jiaxu Zhang
    Annual AAAI Conference on Artificial Intelligence (AAAI), 2026
  3. SemTalk.png
    ICCV 2025
    SemTalk: Holistic Co-speech Motion Generation with Frame-level Semantic Emphasis
    Xiangyue Zhang*, Jianfang Li*, Jiaxu Zhang, Ziqiang Dang, Jianqiang Ren, Liefeng Bo, and Zhigang Tu†
    International Conference on Computer Vision (ICCV), 2025
  4. EchoMask.png
    ACM MM 2025
    EchoMask: Speech-Queried Attention-based Mask Modeling for Holistic Co-Speech Motion Generation
    Xiangyue Zhang*, Jianfang Li*, Jiaxu Zhang, Jianqiang Ren, Liefeng Bo, and Zhigang Tu†
    ACM International Conference on Multimedia (ACM MM), 2025
  5. 2D3_SkelAct.png
    T-CSVT 2025
    Robust 2D Skeleton Action Recognition via Decoupling and Distilling 3D Latent Features
    Xiangyue Zhang*, Yifan Jia*, Jiaxu Zhang, Yijie Yang, and Zhigang Tu†
    IEEE Transactions on Circuits and Systems for Video Technology (T-CSVT), 2025


Experience & Education

2025.12 - 2026.03Shenzhen

ByteDance

Research Intern, Intelligent Creation Team

Advisor: Youjiang Xu. Working on large streaming motion generation models.



Awards & Honors

JAN, 2026
Wang Zhizhuo Scholarship for Innovative Talents (top 0.3%)
OCT, 2025
National Scholarship (top 3%)
JUN, 2023
Outstanding Graduate Award


Service

Reviewer
Reviewer Service: NeurIPS, AAAI, ACM MM, T-CSVT