Abstract
Natural communication with machines is a crucial factor in bringing the benefits of networked computers to mass markets. In particular, the sensory dimensions of sight, sound and touch are comfortable and convenient modalities for the human user. New technologies are now emerging in these domains that can support human/machine communication with features that emulate face-to-face interaction. A current challenge is how to integrate the, as yet, imperfect technologies to achieve synergies that transcend the benefit of a single modality. Because speech is a preferred means for human information exchange, conversational interaction with machines will play a central role in collaborative knowledge work mediated by networked computers. Utilizing speech in combination with simultaneous visual gesture and haptic signaling requires software agents able to fuse the error-susceptible sensory information into reliable interpretations that are responsive to (and anticipatory of) human user intentions. This report draws a perspective on research in human/machine communication technologies aimed to support computer conferencing and collaborative problem solving.
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
J. Flanagan and I. Marsic, “Issues in measuring the benefits of multimodal interfaces”, Proc. IEEE Int’l. Conf. Acoustics, Speech and Signal Processing (ICASSP’97), Munich, Germany, April 1997.
A. Medl, I. Marsic, V. Popescu, A. Shaikh, M. Andre, C. Kulikowski and J. Flanagan, “Multimodal User Interface for Mission Planning”, submitted to the ACM SIGCHI Conference on Human Factors in Computing Systems (CHI’98), Los Angeles, CA, April 18–23, 1998.
G. Burdea, Force and Touch Feedback for Virtual Reality, John Wiley & Sons, New York, 1996.
J. Flanagan, “Technologies for multimedia information systems”, Proc. IEEE, v. 84, no. 4, pp. 590–603, April 1994.
J. Flanagan and E. Jan, “Sound capture with three-dimensional selectivity, Acustica, v. 83, no. 4, pp. 644–652, July/August 1997.
D. Rabinkin, D. Macomber, R. Renomeron and J.L. Flanagan, “Optimal Truncation Time for Matched Filter Array Processing”, In Proc. Of ICASSP’98, v. VI, pp. 3629–32, Seattle, WA, May 1998.
R. Renomeron, D. Rabinkin, J. French, and J.L. Flanagan, “Small-Scale Matched Filter Array Processing for Spatially Selective Sound Capture”, J. Acoust. Soc. Am., v. 102, No. 5Pt. 2, p. 3208, Nov. 1997.
A. Waibel, M. Vo, P. Duchnowski, and S. Manke, “Multimodal interfaces”, Art. Intell. Rev., v. 10, nos. 3–4, 1995.
J. Wilder, P.J. Phillips, C. Jiand, and S. Wiener, “Comparison of Visible and Infra-Red Imagery for Face Recognition”, In Proc. Of the 2nd Int. Conf. On Automatic Face and Gesture Recognition, pp. 182–187, Killington, VT, 1996.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1999 Springer-Verlag London Limited
About this paper
Cite this paper
Flanagan, J., Marsic, I., Medl, A., Burdea, G., Wilder, J. (1999). Multimodal Human/Machine Communication. In: De Natale, F., Pupolin, S. (eds) Multimedia Communications. Springer, London. https://doi.org/10.1007/978-1-4471-0859-7_41
Download citation
DOI: https://doi.org/10.1007/978-1-4471-0859-7_41
Publisher Name: Springer, London
Print ISBN: 978-1-85233-135-1
Online ISBN: 978-1-4471-0859-7
eBook Packages: Springer Book Archive