default search action
4th ICMI 2002: Pittsburgh, Pennsylvania, USA
- 4th IEEE International Conference on Multimodal Interfaces (ICMI 2002), 14-16 October 2002, Pittsburgh, PA, USA. IEEE Computer Society 2002, ISBN 0-7695-1834-6
Perceptual Computing Systems and Methods
- Nuria Oliver, Eric Horvitz, Ashutosh Garg:
Layered Representations for Human Activity Recognition. 3-8 - Christian Bauckhage, Jannik Fritsch, Katharina J. Rohlfing, Sven Wachsmuth, Gerhard Sagerer:
Evaluating Integrated Speech- and Image Understanding. 9-14 - Dan Maynes-Aminzade, Randy Pausch, Steven M. Seitz:
Techniques for Interactive Audience Participation. 15-20 - Paulo Barthelmess, Clarence A. Ellis:
Perceptual Collaboration in Neem. 21-26 - Ediz Polat, Mohammed Yeasin, Rajeev Sharma:
A Tracking Framework for Collaborative Human Computer Interaction. 27-32 - Federico Fontana, Davide Rocchesso, Laura Ottaviani:
A Structural Approach to Distance Rendering in Personal Auditory Displays. 33-38 - Andrea Fusiello, Antonello Panuccio, Vittorio Murino, Federico Fontana, Davide Rocchesso:
A Multimodal Electronic Travel Aid Device. 39-46
Speech Generation and Recognition
- Ivica Rogina, Thomas Schaaf:
Lecture and Presentation Tracking in an Intelligent Meeting Room. 47-52 - Ryuta Taguma, Tatsuhiro Moriyama, Koji Iwano, Sadaoki Furui:
Parallel Computing-Based Architecture for Mixed-Initiative Spoken Dialogue. 53-58 - Satoshi Nakamura, Panikos Heracleous:
3-D N-Best Search for Simultaneous Recognition of Distant-Talking Speech of Multiple Talkers. 59-63 - Pui-Fung Wong, Man-Hung Siu:
Integration of Tone Related Feature for Chinese Speech Recognition. 64-68 - Marc Mersiol, Noël Chateau, Valérie Maffiolo:
Talking Heads: Which Matching between Faces and Synthetic Voices?. 69-74 - Ye Tian, Ji Wu, Zuoying Wang, Dajin Lu:
Robust Noisy Speech Recognition with Adaptive Frequency Bank Selection. 75-80 - Ziqiang Wang, Yang Liu, Peng Ding, Bo Xu:
Covariance-Tied Clustering Method In Speaker Identification. 81-86
Dialogue Processing and Language Understanding
- Joyce Y. Chai, Shimei Pan, Michelle X. Zhou, Keith Houck:
Context-Based Multimodal Input Understanding in Conversational Systems. 87-92 - Helen Wright Hastie, Michael Johnston, Patrick Ehlen:
Context-Sensitive Help for Multimodal Dialogue. 93-98 - Frédéric Landragin, Nadia Bellalem, Laurent Romary:
Referring to Objects with Spoken and Haptic Modalities. 99-104 - Deb Roy:
Towards Visually-Grounded Spoken Language Acquisition. 105-110 - Christian Elting, Gregor Möhler:
Modeling Output in the EMBASSI Multimodal Dialog System. 111-116 - Aseel Ibrahim, Pontus Johansson:
Multimodal Dialogue Systems for Interactive TVApplications. 117-122 - Candace L. Sidner, Myroslava O. Dzikovska:
Human - Robot Interaction: Engagement between Humans and Robots for Hosting Activities. 123-128 - Jack Mostow, Joseph E. Beck, Raghu Chalasani, Andrew Cuneo, Peng Jia:
Viewing and Analyzing Multimodal Human-computer Tutorial Dialogue: A Database Approach. 129-134 - Sorin Dusan, James L. Flanagan:
Adaptive Dialog Based upon Multimodal Language Acquisition. 135-140 - Hartwig Holzapfel, Christian Fügen, Matthias Denecke, Alex Waibel:
Integrating Emotional Cues into a Framework for Dialogue Management. 141-148
Signing, Gesturing, and Writing
- Haifeng Li, Thierry Artières, Patrick Gallinari:
Data Driven Design of an ANN/HMM System for On-line Unconstrained Handwritten Character Recognition. 149-154 - Lei Chen, Mary P. Harper, Francis K. H. Quek:
Gesture Patterns during Speech Repairs. 155-160 - Sanshzar Kettebekov, Mohammed Yeasin, Rajeev Sharma:
Prosody Based Co-analysis for Continuous Recognition of Coverbal Gestures. 161-166 - Aleix M. Martínez, Ronnie B. Wilbur, Robin Shay, Avinash C. Kak:
Purdue RVL-SLLL ASL Database for Automatic Recognition of American Sign Language. 167-172 - Frédéric Landragin:
The Role of Gesture in Multimodal Referring Actions. 173-178 - Yingen Xiong, Francis K. H. Quek, David McNeill:
Hand Gesture Symmetric Behavior Detection and Analysis in Natural Conversation. 179-184 - Jose L. Hernandez-Rebollar, Robert W. Lindeman, Nicholas Kyriakopoulos:
A Multi-Class Pattern Recognition System for Practical Finger Spelling Translation. 185-190 - Andrea Corradini, Richard M. Wesson, Philip R. Cohen:
A Map-Based System Using Speech and 3D Gestures for Pervasive Computing. 191-196 - Edward Lin, Andy Cassidy, Dan Hook, Avinash Baliga, Tsuhan Chen:
Hand Tracking Using Spatial Gesture Modeling and Visual Feedback for a Virtual DJ System. 197-202 - Haifeng Li, Thierry Artières, Patrick Gallinari:
State Sharing in a Hybrid Neuro-Markovian On-Line Handwriting Recognition System through a Simple Hierarchical Clustering Algorithm. 203-210
Translation and Multilingual Interfaces
- Ryosuke Isotani, Kiyoshi Yamabana, Shinichi Ando, Ken Hanazawa, Shin-ya Ishikawa, Tadashi Emori, Ken-ichi Iso, Hiroaki Hattori, Akitoshi Okumura, Takao Watanabe:
An Automatic Speech Translation System on PDAs for Travel Conversation. 211-216 - Jing Zhang, Xilin Chen, Jie Yang, Alex Waibel:
A PDA-Based Sign Translator. 217-222 - Loredana Taddei, Erica Costantini, Alon Lavie:
The NESPOLE! Multimodal Interface for Cross-lingual Communication - Experience and Lessons Learned . 223-228 - Dequan Zheng, Yi Hu, Tiejun Zhao, Hao Yu, Sheng Li:
Research of Machine Learning Method for Specific Information Recognition on the Internet. 229-234 - Erica Costantini, Fabio Pianesi, Susanne Burger:
The Added Value of Multimodality in the NESPOLE! Speech-to-Speech Translation System: an Experimental Study. 235-240 - Shigeo Morishima, Satoshi Nakamura:
Multi-Modal Translation System and Its Evaluation. 241-246 - Zhirong Wang, Umut Topkara, Tanja Schultz, Alex Waibel:
Towards Universal Speech Recognition. 247-252 - Fei Huang, Stephan Vogel:
Improved Named Entity Translation and Bilingual Named Entity Extraction. 253-260
Gaze Tracking and Lipreading
- Rowel Atienza, Alexander Zelinsky:
Active Gaze Tracking for Human-Robot Interaction. 261-266 - David Demirdjian, Trevor Darrell:
3-D Articulated Pose Tracking for Untethered Diectic Reference. 267-272 - Rainer Stiefelhagen:
Tracking Focus of Attention in Meetings. 273-280 - Qiang Wang, Haizhou Ai, Guangyou Xu:
A Probabilistic Dynamic Contour Model for Accurate and Robust Lip Tracking. 281-286 - Chen Yu, Dana H. Ballard, Shenghuo Zhu:
Attentional Object Spotting by Integrating Multimodal Input. 287-292 - Zhilin Wu, Petar S. Aleksic, Aggelos K. Katsaggelos:
Lip Tracking for MPEG-4 Facial Animation. 293-298 - Taeyoon Kim, Yongsung Kang, Hanseok Ko:
Achieving Real-Time Lip Synch via SVM-Based Phoneme Classification and Lip Shape Refinement. 299-304 - Satoshi Nakamura, Ken'ichi Kumatani, Satoshi Tamura:
Multi-Modal Temporal Asynchronicity Modeling by Product HMMs for Robust. 305-312
Applications of Multimodal Interfaces
- Elena V. Zudilova, Peter M. A. Sloot, Robert G. Belleman:
A Multi-Modal Interface for an Interactive Simulated Vascular Reconstruction System. 313-318 - Helen Petrie, Wendy Fisher, Ine Langer, Gerhard Weber, Keith Gladstone, Cathy Rundle, Liesbeth Pyfers:
Universal Interfaces to Multimedia Documents. 319-324 - Ali Zandifar, Ramani Duraiswami, Antoine Chahine, Larry S. Davis:
A Video Based Interface to Textual Information for the Visually Impaired. 325-330 - Rajarathinam Arangarasan, George N. Phillips Jr.:
Modular Approach of Multimodal Integration in a Virtual Environment. 331-336 - Georg Niklfeld, Michael Pucher, Robert Finan, Wolfgang Eckhart:
Mobile Multi-Modal Data Services for GPRS Phones and Beyond. 337-342 - Brad A. Myers, Robert G. Malkin, Michael Bett, Alex Waibel, Ben Bostwick, Robert C. Miller, Jie Yang, Matthias Denecke, Edgar Seemann, Jie Zhu, Choon Hong Peck, Dave Kong, Jeffrey Nichols, William L. Scherlis:
Flexi-Modal and Multi-Machine User Interfaces. 343-348 - Nils Krahnstoever, Sanshzar Kettebekov, Mohammed Yeasin, Rajeev Sharma:
A Real-Time Framework for Natural Multimodal Interaction with Large Screen Displays. 349-354 - Anoop K. Sinha, James A. Landay:
Embarking on Multimodal Interface Design. 355-360 - Lars Bo Larsen, Morten Damm Jensen, Wisdom Kobby Vodzi:
Multi Modal User Interaction in an Automatic Pool Trainer. 361-366 - Daniel P. Siewiorek, Asim Smailagic, Matthew Hornyak:
Multimodal Contextual Car-Driver Interface. 367-376
Sensors, Tools, and Platforms for Multimodal Interfaces
- Jeffrey Nichols, Brad A. Myers, Thomas K. Harris, Roni Rosenfeld, Stefanie Shriver, Michael Higgins, Joseph Hughes:
Requirements for Automatically Generating Multi-Modal Interfaces for Complex Appliances. 377-382 - Huazhong Ning, Liang Wang, Weiming Hu, Tieniu Tan:
Articulated Model Based People Tracking Using Motion Models. 383-388 - Kevin W. Wilson, Vibhav Rangarajan, Neal Checka, Trevor Darrell:
Audiovisual Arrays for Untethered Spoken Interfaces. 389-394 - Sen Wang, Wei Wei Zhang, Yangsheng Wang:
Fingerprint Classification by Directional Fields. 395-399 - Dirk Focken, Rainer Stiefelhagen:
Towards Vision-Based 3-D People Tracking in a Smart Room. 400-405 - Helena M. Mentis, Geri Gay:
Using TouchPad Pressure to Detect Negative Affect. 406-410 - Marc Erich Latoschik:
Designing Transition Networks for Multimodal VR-Interactions Using a Markup Language. 411-416 - Tomoko Yonezawa, Kenji Mase:
Musically Expressive Doll in Face-to-Face Communication. 417-422 - Xilin Chen, Jie Yang:
Towards Monitoring Human Activities Using an Omnidirectional Camera. 423-428 - Weikai Xie, Yuanchun Shi, Guangyou Xu, Yanhua Mao:
Smart Platform - A Software Infrastructure for Smart Space (SISS). 429-436
User Studies and System Evaluation
- Rob Gray, Hong Z. Tan, J. Jay Young:
Do Multimodal Signals Need to Come from the Same Place? Crossmodal Attentional Links Between Proximal and Distal Surfaces. 437-441 - Jan Kleindienst, Ladislav Serédi, Pekka Kapanen, Janne Bergman:
CATCH-2004 Multi-Modal Browser: Overview Description with Usability Analysis. 442-447 - Philip R. Cohen, Rachel Coulston, Kelly Krout:
Multimodal Interaction During Multiparty Dialogues: Initial Results. 448-453 - Yasmine Arafa, Abe Mamdani:
Multi-Modal Embodied Agents Scripting. 454-459 - Jason Williams, Georg Michelitsch, Gregor Möhler, Stefan Rapp:
A Methodology for Evaluating Multimodality in a Home Entertainment System. 460-465 - Changseok Cho, Huichul Yang, Gerard Jounghyun Kim, Sung Ho Han:
Body-Based Interfaces. 466-472 - Jérôme Grosjean, Jean-Marie Burkhardt, Sabine Coquillart, Paul Richard:
Evaluation of the Command and Control Cube. 473-478 - Ernesto Arroyo, Ted Selker, Alexandre Stouffs:
Interruptions as Multimodal Outputs: Which are the Less Disruptive?. 479-482 - Gregory Aist, Barry Kort, Rob Reilly, Jack Mostow, Rosalind W. Picard:
Experimentally Augmenting an Intelligent Tutoring System with Human-Supplied Capabilities: Adding Human-Provided Emotional Scaffolding to an Automated Reading Tutor that Listens. 483-490 - Jeffrey F. Cohn, Karen L. Schmidt, Ralph Gross, Paul Ekman:
Individual Differences in Facial Expression: Stability over Time, Relation to Self-Reported Emotion, and Ability to Inform Person Identification. 491-498
Face Detection, Recognition, and Synthesis
- Michael M. Cohen, Dominic W. Massaro, Rashid Clark:
Training a Talking Head. 499-504 - Piero Cosi, Emanuela Magno Caldognetto, Giulio Perin, Claudio Zmarich:
Labial Coarticulation Modeling for Realistic Facial Animation. 505-510 - Ziyou Xiong, Yunqiang Chen, Roy Wang, Thomas S. Huang:
Improved Information Maximization based Face and Facial Feature Detection from Real-time Video and Application in a Multi-Modal Person Identification System. 511-516 - Dalong Jiang, Wen Gao, Zhiguo Li, Zhaoqi Wang:
Animating Arbitrary Topology 3D Facial Model Using the MPEG-4 FaceDefTables. 517-522 - Wei Wang, Shiguang Shan, Wen Gao, Bo Cao, Baocai Yin:
An Improved Active Shape Model for Face Alignment. 523-528 - Beat Fasel:
Head-Pose Invariant Facial Expression Recognition Using Convolutional Neural Networks. 529-534 - Wenjun Lao, Dehui Kong, Baocai Yin:
An Improved Algorithm for Hairstyle Dynamics. 535-540
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.