On clustering and retrieval of video shots
MULTIMEDIA '01 Proceedings of the ninth ACM international conference on Multimedia
Semantic Annotation of Sports Videos
IEEE MultiMedia
MDC: A Software Tool for Developing MPEG Applications
ICMCS '99 Proceedings of the IEEE International Conference on Multimedia Computing and Systems - Volume 2
A mid-level representation framework for semantic sports video analysis
MULTIMEDIA '03 Proceedings of the eleventh ACM international conference on Multimedia
Semantic and structural analysis of TV diving programs
Journal of Computer Science and Technology
A fusion scheme of visual and auditory modalities for event detection in sports video
ICME '03 Proceedings of the 2003 International Conference on Multimedia and Expo - Volume 2
A semantic framework for video genre classification and event analysis
Image Communication
Proceedings of the ACM International Conference on Image and Video Retrieval
Time warp sports for internet television
ACM Transactions on Computer-Human Interaction (TOCHI)
Temporal relation analysis in audiovisual documents for complementary descriptive information
AMR'05 Proceedings of the Third international conference on Adaptive Multimedia Retrieval: user, context, and feedback
Automated stroke classification in Tennis
ICIAR'07 Proceedings of the 4th international conference on Image Analysis and Recognition
Hi-index | 0.00 |
In this demonstration, we present a unified framework for semantic shot classification in sports videos. Unlike previous approaches, which focus on clustering by aggregating shots with similar low-level features, the proposed scheme makes use of domain knowledge of specific sport to perform a top-down video shot classification, including identification of video shots classes for each sport, and supervised learning and classification of given sports video with low-level and middle-level features extracted from the sports video. It's observed that for each sport we can predefine a small number of semantic shot classes, 5--10, which cover 90 to 95 % of sports broadcasting video. With supervised learning method, we can map the low-level features to middle-level semantic video shot attributes such as dominant object motion (a player), camera motion patterns, and court shape, etc. On the basis of the appropriate fusion of those middle-level shot attributes, we classify video shots into the predefined video shot classes, each of which has a clear semantic meaning. The proposed method has been tested over 3 types of sports videos: tennis, basketball, and soccer. Good classification results ranging from 80~95% have been achieved. The proposed framework provides a generic solution for sports video semantic shot classification, which can be adapted to a new sport type easily. With correctly classified sports video shots further structural and temporal analysis will be greatly facilitated.