|
Haotian Yang | 杨皓天 I'm a Research Scientist at ByteDance working on video generation. Previously, I was a Research Engineer at Kuaishou Technology, where I worked on both video generation (as a core contributor to Kling video generation model) and 3D avatars (leading the development of a Light Stage system). I received the M.S. degree (under the supervision of Xun Cao and Hao Zhu) and B.S. degree from Nanjing University, in 2021 and 2018 respectively. My research interests lie in computer vision and computer graphics, including video generation and editing, human digitization, and 3D vision. |
|
|
VIVA: VLM-Guided Instruction-Based Video Editing with Reward Optimization |
|
TGT: Text-Grounded Trajectories for Locally Controlled Video Generation |
|
VMoBA: Mixture-of-Block Attention for Video Diffusion Models |
|
Physically Based Facial Texture Generation in the Wild |
|
Imbalance in Balance: Online Concept Balancing in Generation Models |
|
Koala-36M: A Large-scale Video Dataset Improving Consistency between Fine-grained Conditions and Video Content |
|
VideoTetris: Towards Compositional
Text-To-Video Generation |
|
VRMM: A Volumetric Relightable
Morphable Head Model |
|
Towards Practical Capture of
High-Fidelity Relightable Avatars |
|
Facescape: 3d facial dataset and
benchmark for single-view 3d face reconstruction |
|
Detailed facial geometry recovery from
multi-view images by learning an implicit function |
|
Detailed avatar recovery from single
image |
|
Facescape: a large-scale high quality
3d face dataset and detailed riggable 3d face prediction |
This template is a modification of Jon Barron's website