Runyi (Ingrid) Yu

PhD Student

Hong Kong University of Science and Technology
Clear Water Bay, Kowloon, HongKong.

Email: ingrid.yu [at] connect [dot] ust [dot] hk


Biography

I am a first-year phd student at HKUST, supervised by Qifeng Chen. I obtained my Master's Degree from Peking University in 2024 and Bachelor's Degree from Beijing Normal University in 2021. I used to be an intern at Unitree, and MSRA.

My research interests include Computer Vision, Machine Learning, Robotics and their intersections. From 2021 to 2022, I focused on the Transformer-based Understanding researches. From 2023 to 2024, I worked on image/video generation and talking head generation. Now, I am exploring the intergration of Computer Vision and Embodied AI.

Cooperation and discussion are welcomed.

I'm looking for an Embodied AI related internship, please feel free to contact me if you're interested.

News

Internship

               

Unitree

Apr. 2024 - Aug. 2024, Unitree, R&D departments, Hangzhou, China

closely worked with Yinhuai Wang

Topic: Real World Humanoid-Object Interaction

Microsoft Research Asia

Apr. 2023 - Mar. 2024, Beijing, China

closely worked with Tianyu He and Xu Tan

Topic: Talking Head Generation

Education

The Hong Kong University of Science and Technology, Hong Kong

PhD Student in Visual Intelligence Lab, HKUST

Advisor: Prof. Qifeng Chen

Sep. 2024 - Future

Peking University, China

Master of Science in Computer Science

Advisor: Prof. Jie Chen

Sep. 2021 - Jun. 2024

Beijing Normal Univesity, China

Bachelor of Management in Information Systems

Sep. 2017 - Jun. 2021

Selected Publications

                                             
/*Preprints*/

SkillMimic: Learning Reusable Basketball Skills from Demonstrations

Yinhuai Wang*, Qihan Zhao*, Runyi Yu*, Ailing Zeng, Jing Lin, Zhengyi Luo, Hok Wai Tsui, Jiwen Yu, Xiu Li, Qifeng Chen, Jian Zhang, Lei Zhang, Ping Tan

[paper] [project page] [code]

Make Your Actor Talk: Generalizable and High-Fidelity Lip Sync with Motion and Appearance Disentanglement

Runyi Yu, Tianyu He, Ailing Zhang, Yuchi Wang, Junliang Guo, Xu Tan, Chang Liu, Jie Chen, Jiang Bian

[paper] [project page]

InstructAvatar: Text-Guided Emotion and Motion Control for Avatar Generation

Yuchi Wang, Junliang Guo, Jianhong Bai, Runyi Yu, Tianyu He, Xu Tan, Xu Sun, Jiang Bian

[paper] [project page]

/*Conference*/

Local Action-Guided Motion Diffusion Model for Text-to-Motion Generation

Peng Jin, Hao Li, Zesen Cheng, Kehan Li, Runyi Yu, Chang Liu, Xiangyang Ji, Li Yuan, Jie Chen

ECCV 2024

[paper] [project page]

GAIA: Data-driven Zero-shot Talking Avatar Generation

Tianyu He*, Junliang Guo*, Runyi Yu*, Yuchi Wang*, Jialiang Zhu, Kaikai An, Leyi Li, Xu Tan, Chunyu Wang, Han Hu, HsiangTao Wu, Sheng Zhao, Jiang Bian

ICLR 2024

[paper] [project page]

LaPE: Layer-adaptive Position Embedding for Vision Transformers with Independent Layer Normalization

Runyi Yu*, Zhennan Wang*, Yinhuai Wang*, Kehan Li, Chang Liu, Haoyi Duan, Xiangyang Ji, Jie Chen

ICCV 2023

[paper] [code]

ACSeg: Adaptive Conceptualization for Unsupervised Semantic Segmentation

Kehan Li, Zhennan Wang, Zesen Cheng, Runyi Yu, Yian Zhao, Guoli Song, Chang Liu, Li Yuan, Jie Chen

CVPR 2023 Hightlight

[paper] [project page]

Unlimited-Size Diffusion Restoration

Yinhuai Wang, Jiwen Yu, Runyi Yu, Jian Zhang

CVPR Workshop 2023 Oral

[paper] [code]

Locality guidance for improving vision transformers on tiny datasets

Kehan Li*, Runyi Yu*, Zhennan Wang, Li Yuan, Guoli Song, Jie Chen

ECCV 2022

[paper] [code]

Community Services



© Ingrid Yu