Multimodal Egocentric Computer Vision Understanding of Human Interaction
Next-generation Computer Vision goes beyond recognizing actions to understand interactions. In this project, two participants with Project Aria glasses will be recorded during natural activities. Building on our recent work on egoPPG (ICCV'25) and egoEMOTION (NeurIPS'25), you will explore how multimodal egocentric signals, video, gaze, motion, and context, can be fused to model human interaction, coordination, and behavioral dynamics. All hardware (Project Aria glasses + extensions) will be provided.



