Runyi (Ingrid) Yu

PhD Student

Hong Kong University of Science and Technology
Clear Water Bay, Kowloon, HongKong.

Email: ingrid.yu [at] connect [dot] ust [dot] hk


Biography

I am a first-year phd student at HKUST, supervised by Qifeng Chen. I obtained my Master's Degree from Peking University in 2024 and Bachelor's Degree from Beijing Normal University in 2021. Currently, I am interning at Unitree, and used to be an intern at MSRA.

My research interests include Computer Vision, Machine Learning, Robotics and their intersections. From 2021 to 2022, I focused on the Transformer-based Understanding researches. From 2023 to 2024, I worked on image/video generation and talking head generation. Now, I am exploring the intergration of Computer Vision and Embodied AI.

Cooperation and discussion are welcomed.

News

Internship

               

Microsoft Research Asia

Apr. 2023 - Mar. 2024, Beijing, China

closely worked with Tianyu He and Xu Tan

Topic: Talking Head Generation

Unitree

Apr. 2024 - Present, Unitree, R&D departments, Hangzhou, China

closely worked with Yinhuai Wang

Topic: Real World Humanoid-Object Interaction

Education

The Hong Kong University of Science and Technology, Hong Kong

PhD Student in Visual Intelligence Lab, HKUST

Advisor: Prof. Qifeng Chen

Sep. 2024 - Future

Peking University, China

Master of Science in Computer Science

Advisor: Prof. Jie Chen

Sep. 2021 - Jun. 2024

Beijing Normal Univesity, China

Bachelor of Management in Information Systems

Sep. 2017 - Jun. 2021

Selected Publications

                                             
/*Preprints*/

SkillMimic: Learning Reusable Basketball Skills from Demonstrations

Yinhuai Wang*, Qihan Zhao*, Runyi Yu*, Ailing Zeng, Jing Lin, Zhengyi Luo, Hok Wai Tsui, Jiwen Yu, Xiu Li, Qifeng Chen, Jian Zhang, Lei Zhang, Ping Tan

[paper] [project page] [code]

Make Your Actor Talk: Generalizable and High-Fidelity Lip Sync with Motion and Appearance Disentanglement

Runyi Yu, Tianyu He, Ailing Zhang, Yuchi Wang, Junliang Guo, Xu Tan, Chang Liu, Jie Chen, Jiang Bian

[paper] [project page]

InstructAvatar: Text-Guided Emotion and Motion Control for Avatar Generation

Yuchi Wang, Junliang Guo, Jianhong Bai, Runyi Yu, Tianyu He, Xu Tan, Xu Sun, Jiang Bian

[paper] [project page]

/*Conference*/

Local Action-Guided Motion Diffusion Model for Text-to-Motion Generation

Peng Jin, Hao Li, Zesen Cheng, Kehan Li, Runyi Yu, Chang Liu, Xiangyang Ji, Li Yuan, Jie Chen

ECCV 2024

[paper] [project page]

GAIA: Data-driven Zero-shot Talking Avatar Generation

Tianyu He*, Junliang Guo*, Runyi Yu*, Yuchi Wang*, Jialiang Zhu, Kaikai An, Leyi Li, Xu Tan, Chunyu Wang, Han Hu, HsiangTao Wu, Sheng Zhao, Jiang Bian

ICLR 2024

[paper] [project page]

LaPE: Layer-adaptive Position Embedding for Vision Transformers with Independent Layer Normalization

Runyi Yu*, Zhennan Wang*, Yinhuai Wang*, Kehan Li, Chang Liu, Haoyi Duan, Xiangyang Ji, Jie Chen

ICCV 2023

[paper] [code]

ACSeg: Adaptive Conceptualization for Unsupervised Semantic Segmentation

Kehan Li, Zhennan Wang, Zesen Cheng, Runyi Yu, Yian Zhao, Guoli Song, Chang Liu, Li Yuan, Jie Chen

CVPR 2023 Hightlight

[paper] [project page]

Unlimited-Size Diffusion Restoration

Yinhuai Wang, Jiwen Yu, Runyi Yu, Jian Zhang

CVPR Workshop 2023 Oral

[paper] [code]

Locality guidance for improving vision transformers on tiny datasets

Kehan Li*, Runyi Yu*, Zhennan Wang, Li Yuan, Guoli Song, Jie Chen

ECCV 2022

[paper] [code]

Community Services



© Ingrid Yu