By Topic

Indexing of player events using multimodal cues in golf videos

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$33 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

2 Author(s)
Hyoung-Gook Kim ; Intelligent Multimedia Signal Processing Lab, Kwangwoon University, Korea ; Jin-Ho Lee

In this paper, we propose a system for indexing and retrieving player event scenes using multimodal cues in golf videos. Play scenes and audio classes are detected independently from video and audio tracks at the time of indexing. The audio track is semantically segmented into basic audio categories (studio speech, field speech, music, applause, swing sound and others) by means of audio classification and semantic occupation ratios. The visual play-start scene and the excited reaction of the audience are combined to extract event scenes. The player name related to each event is indexed by the spoken descriptors. At retrieval time, the user selects a text query with the player name on the screen. The lists of each query term are retrieved through a description matcher to identify full and partial phrase hits related to event scenes. Experimental results show that the implemented system achieves an average 82.5% accuracy rate.

Published in:

2011 IEEE International Conference on Multimedia and Expo

Date of Conference:

11-15 July 2011