Runyi (Ingrid) YuPhD Student
Hong Kong University of Science and Technology |
|
I am a first-year phd student at HKUST, supervised by Qifeng Chen. I obtained my Master's Degree from Peking University in 2024 and Bachelor's Degree from Beijing Normal University in 2021. Currently, I am interning at Unitree, and used to be an intern at MSRA.
My research interests include Computer Vision, Machine Learning, Robotics and their intersections. From 2021 to 2022, I focused on the Transformer-based Understanding researches. From 2023 to 2024, I worked on image/video generation and talking head generation. Now, I am exploring the intergration of Computer Vision and Embodied AI.
Cooperation and discussion are welcomed.
Microsoft Research Asia Apr. 2023 - Mar. 2024, Beijing, China Topic: Talking Head Generation |
|
Unitree Apr. 2024 - Present, Unitree, R&D departments, Hangzhou, China closely worked with Yinhuai Wang Topic: Real World Humanoid-Object Interaction |
|
The Hong Kong University of Science and Technology, Hong Kong PhD Student in Visual Intelligence Lab, HKUST Advisor: Prof. Qifeng Chen Sep. 2024 - Future |
|
Peking University, China Master of Science in Computer Science Advisor: Prof. Jie Chen Sep. 2021 - Jun. 2024 |
|
Beijing Normal Univesity, China Bachelor of Management in Information Systems Sep. 2017 - Jun. 2021
|
/*Preprints*/ | |
SkillMimic: Learning Reusable Basketball Skills from Demonstrations Yinhuai Wang*, Qihan Zhao*, Runyi Yu*, Ailing Zeng, Jing Lin, Zhengyi Luo, Hok Wai Tsui, Jiwen Yu, Xiu Li, Qifeng Chen, Jian Zhang, Lei Zhang, Ping Tan [paper] [project page] [code] |
|
Make Your Actor Talk: Generalizable and High-Fidelity Lip Sync with Motion and Appearance Disentanglement Runyi Yu, Tianyu He, Ailing Zhang, Yuchi Wang, Junliang Guo, Xu Tan, Chang Liu, Jie Chen, Jiang Bian [paper] [project page] |
|
InstructAvatar: Text-Guided Emotion and Motion Control for Avatar Generation Yuchi Wang, Junliang Guo, Jianhong Bai, Runyi Yu, Tianyu He, Xu Tan, Xu Sun, Jiang Bian [paper] [project page] |
|
/*Conference*/ | |
Local Action-Guided Motion Diffusion Model for Text-to-Motion Generation Peng Jin, Hao Li, Zesen Cheng, Kehan Li, Runyi Yu, Chang Liu, Xiangyang Ji, Li Yuan, Jie Chen ECCV 2024[paper] [project page] |
|
GAIA: Data-driven Zero-shot Talking Avatar Generation Tianyu He*, Junliang Guo*, Runyi Yu*, Yuchi Wang*, Jialiang Zhu, Kaikai An, Leyi Li, Xu Tan, Chunyu Wang, Han Hu, HsiangTao Wu, Sheng Zhao, Jiang Bian ICLR 2024[paper] [project page] |
|
LaPE: Layer-adaptive Position Embedding for Vision Transformers with Independent Layer Normalization Runyi Yu*, Zhennan Wang*, Yinhuai Wang*, Kehan Li, Chang Liu, Haoyi Duan, Xiangyang Ji, Jie Chen ICCV 2023 |
|
ACSeg: Adaptive Conceptualization for Unsupervised Semantic Segmentation Kehan Li, Zhennan Wang, Zesen Cheng, Runyi Yu, Yian Zhao, Guoli Song, Chang Liu, Li Yuan, Jie Chen CVPR 2023 Hightlight[paper] [project page] |
|
Unlimited-Size Diffusion Restoration Yinhuai Wang, Jiwen Yu, Runyi Yu, Jian Zhang CVPR Workshop 2023 Oral |
|
Locality guidance for improving vision transformers on tiny datasets Kehan Li*, Runyi Yu*, Zhennan Wang, Li Yuan, Guoli Song, Jie Chen ECCV 2022 |
|