Finding the camera pose is an important step in many egocentric video applications. It has been widely reported that, state of the art SLAM algorithms fail on egocentric videos. In this paper, we propose a robust method for camera pose estimation, designed specifically for egocentric videos. In an egocentric video, the camera views the same scene point multiple times as the wearer’s head sweeps back and forth. We use this specific motion profile to perform short loop closures aligned with wearer’s footsteps. For egocentric videos, depth estimation is usually noisy. In an important departure, we use 2D computations for rotation averaging which do not rely upon depth estimates. The two modification results in much more stable algorithm as is evident from our experiments on various egocentric video datasets for different egocentric applications. The proposed algorithm resolves a long standing problem in egocentric vision and unlocks new usage scenarios for future applications.

WACV 2017 Paper - Computing Egomotion with Local Loop Closures for Egocentric Videos




  title     = {Computing Egomotion with Local Loop Closures for Egocentric Videos},
  author    = {Suvam Patra and Himanshu Aggarwal and Himani Arora and Chetan Arora and Subhashis Banerjee},
  booktitle = {Proceedings of the IEEE Winter Conference on Applications of Computer Vision (WACV)},
  pages = {454--463},
  year      = {2017}




Github link

IITD EgoMotion Dataset

Our dataset with 4 videos comprised of footage shot by us with GPS tags.

* Equal contribution