A Model of Saliency-Based Visual Attention for Rapid Scene Analysis
IEEE Transactions on Pattern Analysis and Machine Intelligence
The computer for the 21st century
ACM SIGMOBILE Mobile Computing and Communications Review - Special issue dedicated to Mark Weiser
Sentiment analysis: capturing favorability using natural language processing
Proceedings of the 2nd international conference on Knowledge capture
Beyond Bags of Features: Spatial Pyramid Matching for Recognizing Natural Scene Categories
CVPR '06 Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Volume 2
ACM Computing Surveys (CSUR)
International Journal of Computer Vision
From game design elements to gamefulness: defining "gamification"
Proceedings of the 15th International Academic MindTrek Conference: Envisioning Future Media Environments
Intelligent Interface Architectures for Folksonomy Driven Structure Network
CISIS '12 Proceedings of the 2012 Sixth International Conference on Complex, Intelligent, and Software Intensive Systems (CISIS)
Object-Centric spatial pooling for image classification
ECCV'12 Proceedings of the 12th European conference on Computer Vision - Volume Part II
Journal of Theoretical and Applied Electronic Commerce Research
Hi-index | 0.00 |
This paper describes a partially developed method on going project for enabling an Intelligent Interactive Multimedia System based on ontological interaction on video clip shown on ubiquitous systems as a computer monitor, mobile or tablet. The paper aims to sketch a theoretical framework on a method for extracting and tracking objects in videos, based on various semantic attributes. It tackles a novel space (cyber-physical) of interaction between human and machine, in a creative way. We use a layered representation based on semantics-driven information to obtain spatiotemporal attributes of objects. The interface is created by extracting object information from the video with a Human Based Computation to obtain a richer semantics of attribute to bridge the semantic gap between words describing an image and its visual features. Users can navigate and manipulate objects displayed on video by associating semantic attributes and comments evaluated by the data and sentiment extraction. Folksonomy tags are extracted from users' comments to be used in a dynamical driven system (Folksodriven). We show some example applications of the proposed method like: advertisement inside the objects displayed on a video, an interface based on objects of interest video navigation, mask layer on an object of interest and a visual interaction for Smart City.