This project involves iteratively refining a participative Unity-based installation where users experience immersive sound and visual feedback with a motion capture suit and VR headset. The immersive application directs gallery-visitors’ attention in various ways with sound and particle effects (for example: https://www.youtube.com/watch?v=wb_jM2kbqTA). The practical goal would be to refine the development of a complete gallery installation experience. The research goal would be to understand how the application affects sensory perception, embodiment, and subjective experiences. Over the year, we would customise the application based on stakeholder feedback.
This project is connected to three others that focus on: 1) an embodied AI agent, 2) mixed reality responsive arduino controls, and 3) a Māori creative story (pūrākau) scene.
Undergraduate
developing an immersive installation with responsive audio and visual effects to physical movement
user study interviewing participants and measuring aspects of their behaviour in VR
supporting dancers in creating a performance with the application
SOFTENG350, SOFTENG702
HCI Lab (303.521, Lab)