
- English
- PDF
- Available on iOS & Android
Interactive Multimodal Information Management
About this book
In the past twenty years, computers and networks have gained a prominent role in supporting human communications. This book presents recent research in multimodal information processing, which demonstrates that computers can achieve more than what telephone calls or videoconferencing can do. The book offers a snapshot of current capabilities for the analysis of human communications in several modalities - audio, speech, language, images, video, and documents - and for accessing this information interactively. The book has a clear application goal, which is the capture, automatic analysis, storage, and retrieval of multimodal signals from human interaction in meetings. This goal provides a controlled experimental framework and helps generating shared data, which is required for methods based on machine learning. This goal has shaped the vision of the contributors to the book and of many other researchers cited in it. It has also received significant long-term support through a series of projects, including the Swiss National Center of Competence in Research (NCCR) in Interactive Multimodal Information Management (IM2), to which the contributors to the book have been connected.
Frequently asked questions
- Essential is ideal for learners and professionals who enjoy exploring a wide range of subjects. Access the Essential Library with 800,000+ trusted titles and best-sellers across business, personal growth, and the humanities. Includes unlimited reading time and Standard Read Aloud voice.
- Complete: Perfect for advanced learners and researchers needing full, unrestricted access. Unlock 1.4M+ books across hundreds of subjects, including academic and specialized titles. The Complete Plan also includes advanced features like Premium Read Aloud and Research Assistant.
Please note we cannot support devices running on iOS 13 and Android 7 or earlier. Learn more about using the app.
Information
Table of contents
- Contents
- List of contributors
- 1 Interactive Multimodal Information Management: Shaping the Vision
- PART I: HUMAN-COMPUTER INTERACTION AND HUMAN FACTORS
- 2 Human Factors in Multimodal Information Management
- 3 User Attention During Mobile Video Consumption
- 4 Wizard of Oz Evaluations of the Archivus Meeting Browser
- 5 Document-Centric and Multimodal Meeting Assistants
- 6 Semantic Meeting Browsers and Assistants
- 7 Multimedia InformationRetrieval
- PART II: VISUAL AND MULTIMODAL ANALYSIS OF HUMAN APPEARANCE AND BEHAVIOR
- 8 Face Recognition for Biometrics
- 9 Facial Expression Analysis for Emotion Recognition and Perception Modeling
- 10 Software for Automatic Gazeand Face/Object Tracking and its Use for Early Diagnosis of Autism Spectrum Disorders
- 11 Learning to Learn New Models of Human Activities in Indoor Settings
- 12 Nonverbal Behavior Analysis
- 13 Quality and Reliability in Multimodal and Multi-classifier Biometric Person Recognition
- 14 Medical Image Annotation
- PART III: SPEECH, LANGUAGE, AND DOCUMENT PROCESSING
- 15 Speech Processing
- 16 Research Trends in Speaker Diarization
- 17 Speaker Diarization of Large Corpora
- 18 Language Processing in Dialogues
- 19 From Isolated Words to Unconstrained Documents: Bringing Offine Handwriting Recognition to the Meeting Room
- 20 Online Handwriting Analysis and Recognition
- 21 Analysis of Printed Documents
- PART IV: ASSESSMENTS
- 22 It was Worth it! Assessment of the Impact of IM2
- 23 Technology Transfer: Turning Science into Products
- 24 Conclusion and Perspectives