Separating the direct and global components of radiance can aid shape recovery algorithms and can provide useful information about materials in a scene. Practical methods for finding the direct and global components use multiple images captured under varying illumination patterns and require the scene, light source and camera to remain stationary during the image acquisition process. In this talk, we describe a motion compensation method that relaxes this condition and allows direct- global separation to be performed on video sequences of dynamic scenes captured by moving projector-camera systems. Key to our method is being able to register frames in a video sequence to each other in the presence of time varying, high frequency active illumination patterns. Our method compares favorably to alternative approaches like single shot separation and frame interleaving. We present results on challenging video sequences that include various types of motions and deformations in scenes that contain complex materials like fabric, skin, leaves and wax.
Supreeth Achar is a PhD student in the Robotics Institute at Carnegie Mellon University. He is advised by Dr Srinivasa Narasimhan. His research interests include physics based methods for computer vision and projector-camera systems.
Host: Kris Kitani