The perception of articulated motion: recognizing moving light displays
The perception of articulated motion: recognizing moving light displays
A Model of Saliency-Based Visual Attention for Rapid Scene Analysis
IEEE Transactions on Pattern Analysis and Machine Intelligence
Lucas/Kanade meets Horn/Schunck: combining local and global optic flow methods
International Journal of Computer Vision
Histograms of Oriented Gradients for Human Detection
CVPR '05 Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Volume 1 - Volume 01
Object Recognition with Features Inspired by Visual Cortex
CVPR '05 Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Volume 2 - Volume 02
Matching Shape Sequences in Video with Applications in Human Movement Analysis
IEEE Transactions on Pattern Analysis and Machine Intelligence
An Integrated Model of Top-Down and Bottom-Up Attention for Optimizing Detection Speed
CVPR '06 Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Volume 2
Statistical Analysis Quick Reference Guidebook: With SPSS Examples
Statistical Analysis Quick Reference Guidebook: With SPSS Examples
Energy-Based Models in Document Recognition and Computer Vision
ICDAR '07 Proceedings of the Ninth International Conference on Document Analysis and Recognition - Volume 01
Learning features of intermediate complexity for the recognition of biological motion
ICANN'05 Proceedings of the 15th international conference on Artificial Neural Networks: biological Inspirations - Volume Part I
Machine Recognition of Human Activities: A Survey
IEEE Transactions on Circuits and Systems for Video Technology
Modeling and recognition of complex multi-person interactions in video
Proceedings of the 1st ACM international workshop on Multimodal pervasive video analysis
The human action image and its application to motion recognition
Proceedings of the Seventh Indian Conference on Computer Vision, Graphics and Image Processing
Physics-based activity modelling in phase space
Proceedings of the Seventh Indian Conference on Computer Vision, Graphics and Image Processing
Modeling multi-object activities in phase space
ACCV'10 Proceedings of the 2010 international conference on Computer vision - Volume Part I
Large-scale multimedia content analysis using scientific workflows
Proceedings of the 21st ACM international conference on Multimedia
Structured analysis of the ISI Atomic Pair Actions dataset using workflows
Pattern Recognition Letters
Hi-index | 0.00 |
In this paper, we propose a computational framework for integrating the physics of motion with the neurobiological basis of perception in order to model and recognize human actions and object activities. The essence, or gist, of an action is intrinsically related to the motion of the scene's objects. We define the Hamiltonian Energy Signature (HES) and derive the S-Metric to yield a global representation of the motion of the scene's objects in order to capture the gist of the activity. The HES is a scalar time-series that represents the motion of an object over the course of an activity and the S-Metric is a distance metric which characterizes the global motion of the object, or the entire scene, with a single, scalar value. The neurobiological aspect of activity recognition is handled by casting our analysis within a framework inspired by Neuromorphic Computing (NMC), in which we integrate a Motion Energy model with a Form/Shape model. We employ different Form/Shape representations depending on the video resolution but use our HES and S-Metric for the Motion Energy approach in either case. As the core of our Integration mechanism, we utilize variants of the latest neurobiological models of feature integration and biased competition, which we implement within a Multiple Hypothesis Testing (MHT) framework. Experimental validation of the theory is provided on standard datasets capturing a variety of problem settings: single agent actions (KTH), multi-agent actions, and aerial sequences (VIVID).