Selected Projects (complete list of publications)

Samarjit Das

Office : EDSH 215

Phone: 515-451-0528

Email : samarjit@cs.cmu.edu

I am a postdoctoral research fellow at the Robotics Institute, Carnegie Mellon University. I work with Prof. Jessica Hodgins and Prof. Fernando De la Torre. I am a member of CMU Computer Vision Group as well as CMU Graphics Lab. Prior to this, I finished my PhD in Electrical and Computer Engineering at Iowa State University working with Prof. Namrata Vaswani. Here’s a copy of my CV and a complete list of publications.

Research Interests

Deformable Shape Models for Motion Activities


Proposed novel stochastic shape deformation models for human body activity tracking in videos with application such as automatic landmark extraction and abnormality detection.


IEEE  PAMI


I am interested in statistical signal processing and machine learning applications for perception tasks involving visual data as well as other sensor modalities such as inertial measurement units, motion capture and physiological sensors. In general, my work addresses the challenges involved in sensing and intelligent decision making for real-world cyber physical systems (CPS). My past and present research topics include:


Computer Vision: Visual contexts for motion interpretation, motion activity modeling and tracking from visual inputs, abnormality detection, robust visual tracking under dynamic scene parameters


Statistical Signal Processing: sequential Monte Carlo techniques and Bayesian filtering on large dimensional state spaces, stochastic deformable shape models, and sparse tracking with sequential compressive sensing


Machine Learning: weakly supervised learning algorithms for discriminative pattern segmentation and classification in time series data, model switching and change detection in complex high dimensional systems 




Particle Filter with Mode Tracking for Tracking Across Illumination Changes


Proposed efficient Particle Filtering algorithms in order to deal with large dimensional state-spaces and multimodality of the observation likelihood with applications to robust visual tracking under dynamic illumination conditions. (with Siemens Research)


IEEE  TIP






Wearable Sensing for Computational Emergency Medicine


Developing cyber physical systems for improve emergency care under challenging scenarios (e.g. road-side, air-ambulance). Used motion capture, wearable IMUs and video camera enabled medical instruments for visuo-motor analysis of paramedic performances.


In collaboration with UPMC and STAT Air-ambulance


SIH Medical  Journal, IEEE ICASSP






Multiple Instance Learning for Surprising Movement

Pattern Detection in Uncontrolled Environments


The wearable IMU based system relied on a weakly supervised approach owing to the incomplete ground truth information which is a major bottleneck in adapting supervised learning approaches for wearable monitoring during daily living.



IEEE  EMBC






Visual Context for Motion Interpretation: Visuo-motor Analysis in Wearable CPS for Activity Monitoring


Working with wearable cameras and vision based functional object categorization for providing visual contexts towards motion understanding. Our method aims to utilize Segmentation-Classification support vector machine for weakly supervised segmentation and clustering of temporal events.   


On going work






Sparse Tracking with MCMC and Sequential Compressive Sensing: Robust Visual Tracking


This project is a marriage between sequential sampling based Monte Carlo techniques and the idea of compressive sensing under sparsity constraints. We demonstrate robust tracking of complex illumination patterns in real-life videos.  


IEEE TIP (under review),  Asilomar






Vision-based Video Content Analysis for Interactive Laryngoscopy


Video laryngoscopy can be useful for paramedics performing Endotracheal Intubation (ETI), a critical emergency procedure. We are developing a system that integrates computer vision capabilities to video laryngoscopes for better real-time guidance.


NAEMSP 






CPR Assessments Without Body Contact: Smart Phone For Bystander Assistance in Medical Emergencies


We have developed a system that leverage low level vision based motion features from raw smart-phone videos and uses that to classify chest compression rate of CPR in real-time to assist the provider. The phone stays on the ground and no straps are needed. 


NAEMSP                                   Patent  under filing process




  



Information Hiding Inside Structured Shapes


We developed a low level image processing and vision based detector to extract hidden information embedded within structured shapes as subtle imperceptible deformations. Computer graphics tools like Adaptively sampled Distance Fields (ADF) was used for rendering the data embedded shapes. (at MERL)


IEEE ICASSP                                    Patent  filed




  

Motion Capture For Quantitative Measurement of Movement Anomalies


This project used a full-body optical motion capture system with 16 infrared cameras for quantitative assessment and predicting movement abnormalities in people with Deep Brain Stimulator implanted in their brains.


IEEE EMBC                                  




  

Particle Filters on large Dimensional State Spaces and Applications in Computer Vision


Advisor: Namrata Vaswani, Iowa State University


                                 




  

Dissertation