Xin Chen

Xin Chen

陈欣 | Senior Research Scientist

ByteDance | Tech Lead

University of Chinese Academy of Sciences

I am a Senior Research Scientist and Tech Leader at ByteDance in Bay Area, USA, working on large-scale video generation, including Seedance. Previously, I was a Research Scientist T10 at Tencent working with Dr. Gang Yu. I received my Ph.D. from the Chinese Academy of Sciences under the supervision of Prof. Jingyi Yu, and was earlier supervised by Prof. Youyi Zheng at ShanghaiTech University.

My research interests focus on large-scale video generation, audio-visual foundation models, and unified multi-modal models. I have a great passion on new things and new ideas, my goal is to create Generative AI which is about humans, used for humans, and benefits humans.

Interests
  • Large-scale Video Generation
  • Multi-modal Generation
  • Human Motion Generation
  • Embodied Agents
  • Computer Vision for Graphics
Services
  • [Conference Reviewer]
    SIGGRAPH, CVPR, ICCV, ECCV, AAAI
  • [Journal Reviewer]
    TPAMI, TIP, IJCV, TMM
Collaborations
  • Welcome discussions from academia and industry, especially regarding technology implementation and real-world impact. Feel free to reach out via email.
  • I currently have internship positions available with the goal of conducting cutting-edge research in artificial intelligence. If you are interested, please send me an email.
News

Publications

.js-id-selected
Seedance 2.0
FlowAct-R1: Towards Interactive Humanoid Video Generation
Bridging Your Imagination with Audio-Video Generation via a Unified Director
Seedance 1.5 pro: A Native Audio-Visual Joint Generation Foundation Model
InterAgent: Physics-based Multi-agent Command Execution via Diffusion on Interaction Graphs
InterAgent: Physics-based Multi-agent Command Execution via Diffusion on Interaction Graphs
ViBES: A Conversational Agent with Behaviorally-Intelligent 3D Virtual Body
ViBES: A Conversational Agent with Behaviorally-Intelligent 3D Virtual Body
Video-As-Prompt: Unified Semantic Control for Video Generation
Video-As-Prompt: Unified Semantic Control for Video Generation
Motion2motion: Cross-topology Motion Transfer with Sparse Correspondence
MotionGPT3: Human Motion as a Second Modality
MotionGPT3: Human Motion as a Second Modality
MeshXL: Neural Coordinate Field for Generative 3D Foundation Models
Paint3D: Paint Anything 3D with Lighting-less Texture Diffusion Models
TightCap: 3D Human Shape Capture with Clothing Tightness Field
Neural Free-Viewpoint Performance Rendering under Complex Human-object Interactions
Sparse Photometric 3D Face Reconstruction Guided by Morphable Models

Experiences

 
 
 
 
 
ByteDance | 字节跳动
Senior Research Scientist
July 2024 – Present San Jose, USA
Senior Research Scientist focusing on large-scale video generation and multi-modal humanoid agents. Recognized as a 2025 Seed Spot Bonus Winner for outstanding contributions to Seedance 2.0.
 
 
 
 
 
Tencent | 腾讯
Research Scientist
February 2022 – June 2024 Shanghai
Research Scientist at QQ Image Lab, received two Outstanding performances and Tencent STAR Award 2023.
 
 
 
 
 
Tencent | 腾讯
Research Scientist Intern
November 2020 – May 2021 Shanghai
Research Scientist Intern at Tencent YouTuLab in 2020, focusing on the 3D human body reconstruction.
 
 
 
 
 
DGene Digital Technology Inc. | 叠境数字
Research Engineer Intern
July 2018 – December 2019 Shanghai
Research Engineer Intern at DGene Digital Technology Inc., my supervisor’s start-up, for world digitalization. I won the Best Outstanding Intern Award in 2018 for leading the mobile virtual fitting project.

Contact

20,454 Total Pageviews