Skip to Main Content
This paper presents an intelligent guiding bulletin board system (iGBBS), which is based on vision-interactive and multiple key word-spotting technology. The system is aimed to provide different kinds of multimedia human-computer interaction (MMHCI) for users under different requirements. At first, a real-time front-view face detection using Harr-like features is used to decide when iGBBS should wake up and become interactive with the user. After system initialization, some feature points within the detected face area are going to be found. Then the orientation of user's head will be estimated via pyramidal Lucas-Kanade optical flow tracking. In addition, spotting the keyword from user's utterance with some related augmented reality responses would be provided as well. The performance of vision-interaction in iGBBS could be reached to 20 fps under Pentium IV 1G Hz PC. The error rate of multiple key word-spotting interaction in iGBBS is about 36.2% and people can get the right response in 2.76 times search averagely. With the comparison to the traditional guiding system, bulletin board, or other non-vision-based input devices system, such like gloves or markers, our system offers a simple, useful and economical solution for the realtime interaction between the user and computer
Note: PDF Not Yet Available In IEEE Xplore. The document that should appear here is not currently available. IEEE Xplore is working to obtain a replacement PDF. That PDF will be posted as soon as it is available. We regret any inconvenience in the meantime.