InfoScope: Link from Real World to Digital Information Space
We describe an information augmentation system (infoScope) and applications integrating handheld device with a color camera to provide enhanced information perception services to users. InfoScope uses a color camera as an input device to capture scene images from the real world and utilize computer vision techniques to extract information from real world, convert them into digital world as text information and augment them back to the original scene location. The user can see both the real world and information together on display of the handheld device. We have implemented two applications: First one is an automatic sign/text translation for foreign travelers where a user may use infoScope whenever they want to see texts or signs in their own language where they are originally written in foreign language in the scene and extracted from scene images automatically by using computer vision techniques. The second application is “Information Augmentation in the City” where a user can see information associated with building, or a place, overlaid onto real scene images on their PDA's display.
KeywordsGlobal Position System Discrete Wavelet Transform Augmented Reality Text Line Scene Image
Unable to display preview. Download preview PDF.
- 4.A. Jain and B. Yu Automatic Text Location in Images and Video Frames in Proc. IEEE Pattern Recognition, Vol 31, 1998Google Scholar
- 5.R. Lienhart Automatic text recognition for video indexing In Proc. ACM Multimedia 96, pp: 11–20. Nov, 1996Google Scholar
- 6.J. Ohya, A. Shio, and S. Akamatsu, Recognizing Characters in Scene Images IEEE Trans. On Pattern Analysis and Machine Intelligence, Vol 16 Feb, 1994Google Scholar
- 7.J. Rekimoto, NaviCam: A Magnifying Glass Approach to Augmented Reality, MIT Presence (August 1997)Google Scholar
- 8.J. Rekimoto, Y. Ayatsuka, and K. Hayashi, Augment-able Reality: Situated Communication Through Physical and Digital Spaces, Wearable Computers: The Second International Symposium on Wearable Computers, Pittsburgh, PA (October 19–20, 1998), pp. 68–75.Google Scholar
- 9.T. Ridler and S. Calvard Picture thresholding using an interactive selection method IEEE Transactions on Systems, Man and Cybernatics 8(8), 1978Google Scholar
- 10.J. C. Shim, C. Dorai, R. Bolle, Automatic Text Extraction from Video for Content-Based Annotation and Retrieval In. Proc. Of International Conference on Pattern Recognition, pp. 1998.Google Scholar
- 11.J. Spohrer, Information In Places, IBM System Journal, Vol 38, No. 4-Pervasive computingGoogle Scholar
- 12.T. Starner, B. Schiele, and A. Pentland, Visual Contextual Awareness in Wearable Computing, Wearable Computers: The Second International Symposium on Wearable Computers, Pittsburgh, PA (October 19–20, 1998), pp. 50–57.Google Scholar
- 13.M. Swain and D. Ballard, Color Indexing Interotional Journal of Computer Vision, 7(1), 1991Google Scholar
- 14.T. Westman, D. Harwood, T. Laitinen, and M. Pietikainen. Color Segmentation By Hierarchical Connected Components Analysis with Image Enhancement by Symmetric Neighborhood Filters. In Proceedings of the 10th International Conference on Pattern Recognition, pages 796, June 1990.Google Scholar