IGI Global, 2012. — 387 p.
Human Computer interaction (HCI) deals with the relationships among people and computers. As the digital world is getting multi-modal, the information space is getting more and more complex. In order to navigate this information space and to capture and apply the implicit and explicit knowledge to appropriate use an effective interaction is required. Such an effective interaction is only possible if computers can understand and respond to important modalities of human perception and cognition, i.e., speech, image and language, including other modalities, e.g., haptic, olfactory and brain signals. HCI researchers have to respond to these challenges by developing innovative concepts, models and techniques. There have been efforts in the areas of language, speech, vision and signal processing. However, the general techniques may not be applicable to or may degrade in HCI environment. This book attempts to bring all relevant technologies (for Language, Speech, Image and other signal processing) at one place in an interaction framework.
The signal processing community and HCI researchers can refer the book to improve their understanding of the state-of-the-art in HCI and broaden their research spheres. It will help postgraduate and doctoral students in identifying new and challenging research problems in the HCI areas.
We invited chapter proposals for
Speech, Image and Language Processing for Human Computer Interaction: Multimodal Advancements and received around forty proposals. After three rounds of reviews, revisions, suggestions and editing we are herewith fifteen chapters in this book. The book is neither a handbook consisting of numerous research papers nor a textbook describing each aspect of HCI coherently. It is not possible to cover all the aspects of HCI in a single book, especially when HCI is still not a matured area. The 15 chapters in the book provide contribution on models, techniques and applications of HCI and are organized in three sections which are as follows: 1) Modeling Interaction 2) Interaction based on Speech, Image and Language, and 3) Multimodal Developments. We had to include the first section, which describes cognitive user models and underlines the need of a general framework for effective interaction. Chapters contained in the second section present methods of audio, visual and lingual processing for multimodal interaction. The third section introduces issues related to multimodal interfaces and conversational systems. It describes current efforts to develop intelligent, adaptive, proactive, portable and affective multimodal interfaces and discusses mobile vision and health-care applications.
Section 1 Modeling InteractionA Brief Survey on User Modeling in Human Computer Interaction
Working Together with Computers: Towards a General Framework for Collaborative Human Computer Interaction
Interactive and Cognitive Models in the Social Network Environment for Designing
Fuzzy Linguistic Modeling in Multi Modal Human Computer Interaction: Adaptation to Cognitive Styles using Multi Level Fuzzy Granulation Method
Section 2 Speech, Image and Language-Based InteractionImproving Audio Spatialization Using Customizable Pinna Based Anthropometric Model of Head-Related Transfer Functions
Recent Trends in Speech Recognition Systems
Issues in Spoken Dialogue Systems for Human- Computer Interaction
Enhancing Robustness in Speech Recognition using Visual Information
On Applying the Farey Sequence for Shape Representation in Z
2Multi Finger Gesture Recognition and Classification in Dynamic Environment under Varying Illumination upon Arbitrary Background
Human Computer Interaction for Effective Metasearching
Science of Emoticons: Research Framework and State of the Art in Analysis of Kaomoji-Type Emoticons
Yoshio Momouchi, Hokkai-Gakuen University, Japan
Section 3 Multimodal DevelopmentsOn the Development of Adaptive and User-Centred Interactive Multimodal Interfaces
A Survey of Mobile Vision Recognition Applications
Speech Disorders Recognition using Speech Analysis