Weihan Xu

Researcher at MIT EECS and MIT Media Lab.

Weihan Xu profile photo

I am a researcher at MIT EECS and MIT Media Lab. I am broadly interested in multimodal learning in creative application. My research vision is to build multimodal systems with user-friendly interfaces that can reason, retrieve, generate, and edit content, ultimately making advanced AI capabilities accessible and useful in everyday life.

I have been working closely with researchers at Sony Research, MIT EECS, and Berkeley AI Research on multimodal content stylization.

I have been working with Dr. Julian McAuley, Dr. Taylor Berg-Kirkpatrick, Dr. Hao-Wen Dong, Dr. Paul Liang, and Dr. Shlomo Dubnov on multimodal retrieval and content generation.

I was fortunate to work with Dr. Cynthia Rudin on interpretable music analysis, and Dr. Pardis Emami-Naeini on human-AI interaction.

Previously, I did my undergraduate at the University of Michigan with a double major in computer science (honors) and data science, where I worked with Dr. Sardar Ansari and Dr. Kayvan Najarian on time-series analysis in clinical settings. I also worked with Dr. Gongjun Xu on statistical analysis of educational assessment data.

I began studying music at six and have training in piano, violin, and the French horn.

Explore

Publications

Research papers, preprints, demos, and links to code.

View publications

Projects

Selected projects and research systems.

View projects

Teaching

Courses and teaching experience.

View teaching

Professional Service

Journal Reviewer: International Journal of Computer Vision

Volunteer Experience

Organization: Shanghai Adream Charitable Foundation
Year: 2020

Miscellaneous

Language: English, Chinese

Interests: Music, Travel, Roller Coasters, Karting

Music Instruments: Piano, French Horn, Violin

Adventures: Bungee Jumping in 2019, Skydiving in 2022, Visit to the Arctic Circle in 2022