by Steve DiPaola, Meehae Song, Nilay Yalcin, Suk Choi, Ulysses Bernardet, Maryam Saberi

BioSensing 2D / 3D / VR Systems

About :: The Research :: Setup and Results :: Downloads and Links :: Contact

Our lab has extensive experience in using different sensing technology including eye tracking and facial emotion recognition (DiPaola et al 2013), as well as gesture tracking and bio sensing heart rate and EDA (Song & DiPaola, 2015) which both affect the generative system and can be used to understand the reception to the generated graphics (still, video, VR).

The Research
Emotional facial tracking using camera and AI software. Motion, gesture and body tracking using overhead cameras and MS Kinect. Hand tracking via our own data gloves and Leap Controller. Eye tracking via our Pupil eye tracker. Bio sensing ( heart rate and EDA) via our Empatica E4 watch.

Setup and Results
Some examples of our tracking systems. All our 2d, 3d and VR systems have an abstraction layer with software modules to support several advanced input technologies such as emotion tracking, motion tracking, and bio-sensors. The first shows heart rate and breath sensing on VR particle systems.

Downloads and Links

Papers / Posters
  PDF Poster: Stanford 2017 Stanford Poster from Stanford's " VR and Behavoiral Change Conference".
  PDF: (Alt) CHI 15 Eye Tracking: Does Observation Reflect Haptic Metaphors in Art Drawing?
  PDF: EVA 15 - VR Nav Exploring Different Ways of Navigating Emotionally-responsive Artwork in Immersive Virtual Environments

Additional Media
  EVA video EVA 16 video


Steve DiPaola :: sdipaola @
cell phone (Vancouver, BC) 604.719.6579