Multimodal Information Processing For Interface And Multimedia Technologies
Speaker: Jong-Seok LEE from Ecole Polytechnique F´ed´erale de Lausanne (EPFL)
Abstract: Humans are inherently multimodal information processors, i.e. they tend to use multiple sensory channels simultaneously for perception whenever possible. In fact, information perceived by humans often takes multiple channels, e.g. audio, visual, haptic, etc. This talk deals with multimodal information processing techniques applied to implementation of intelligent interfaces and multimedia analysis techniques. First, the concept and motivation of multimodal information processing is briefly introduced. Then, its specific applications are described, including audio-visual speech recognition, perceptual video compression based on audio-visual focus of attention, biosignal-based emotion recognition, and content-tag-user analysis in social networks.