Real-time tracking of people’s location, orientation and activities is increasingly important for designing novel ubiquitous computing applications. Top-view camera-based tracking avoids occlusion when tracking people while collaborating, but often requires complex tracking systems and advanced computer vision algorithms. To facilitate the prototyping of ubiquitous computing applications for interactive spaces, we developed EagleSense, a real-time human posture and activity recognition system with a single top-view depth sensing camera.

EagleSense real time tracking of people’s location, postures and activities in interactive spaces.

We contribute our novel algorithm and processing pipeline, including details for calculating silhouette-extremities features and applying gradient tree boosting classifiers for activity recognition optimised for top-view depth sensing. EagleSense provides easy access to the real-time tracking data and includes tools for facilitating the integration into custom applications. We report the results of a technical evaluation with 12 participants and demonstrate the capabilities of EagleSense with application case studies.

Videos and Pictures



  • EagleSense: Tracking People and Devices in Interactive Spaces using Real-Time Top-View Depth-Sensing
  • Chi-Jui Wu, S. Houben and N. Marquardt
  • In Proceedings of the 2017 ACM SIGCHI Conference on Human Factors in Computing Systems (CHI’17).
  • [~25% acceptance rate; 2424 submissions]