Abstract
Entirely watching separate video segments of interest or their summary might not be smooth enough nor comprehensible for viewers since contextual information between those segments may be lost. A unified framework for context-preserving video retrieval and summarization is proposed in order to solve this problem. Given a video database and ontologies specifying relationships among concepts used in MPEG-7 annotations, the objective is to identify according to a user query relevant segments together with summaries of contextual segments. Two types of contextual segments are defined: intra-contextual segments intended to form semantically coherent segments, and inter-contextual segments intended to semantically link together two separate segments. Relationships among verbs [3] are exploited to identify contextual segments as the relationships can provide the knowledge about events, causes and effects of actions over time. A query model and context-preserving video summarization are also presented.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Bailer, W., et al.: Content-based Video Retrieval and Summarization using MPEG-7. In: Proceedings of SPIE, vol. 5304, pp. 1–12 (2004)
Browne, P., Smeaton, A.F.: Video Information Retrieval Using Objects and Ostensive Relevance Feedback. In: ACM Symposium on Applied Computing (2004)
Fellbaum, C.: A Semantic Network of English Verbs. In: WordNet: An Electronic Lexical Database, pp. 69–104. MIT Press, Cambridge (1998)
Gaughan., G., et al.: Design, Implementation and Testing of an Interactive Video Retrieval System. In: MIR 2003 (2003)
Graves, A., Lalmas, M.: Video Retrieval using an MPEG-7 Based Inference Network. In: SIGIR 2002 (2002)
Manjunath, B.S., Salembier, P., Sikora, T.: Introduction to MPEG-7: Multimedia Content Description Interface (2002)
Pradhan, S., Tajima, K., Tanaka, K.: A Query Model to Synthesize Answer Intervals from Indexed Video Units. IEEE TKDE 13(5) (2001)
Sistla, A.P., Yu, C., Venkatasubrahmanian, R.: Similarity Based Retrieval of Videos. In: Proc. of the 13th International Conference on Data Engineering (1997)
Tseng, B.L., Lin, C., Smith, J.R.: Using MPEG-7 and MPEG-21 for Personalizing Video. IEEE Multimedia 11(1), 42–52 (2004)
Tsinaraki, C., Polydoros, P., Christodoulakis, S.: Interoperability Support for Ontology-Based Video Retrieval Applications. In: Enser, P.G.B., Kompatsiaris, Y., O’Connor, N.E., Smeaton, A., Smeulders, A.W.M. (eds.) CIVR 2004. LNCS, vol. 3115, pp. 582–591. Springer, Heidelberg (2004)
TV-Anytime Forum (2003), http://www.tv-anytime.org
Wuwongse, V., Akama, K., Anutariya, C., Nantajeewarawat, E.: A Data Model for XML Databases. JIIS 20(1), 63–80 (2003)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2005 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Pattanasri, N., Chatvichienchai, S., Tanaka, K. (2005). Towards a Unified Framework for Context-Preserving Video Retrieval and Summarization. In: Fox, E.A., Neuhold, E.J., Premsmit, P., Wuwongse, V. (eds) Digital Libraries: Implementing Strategies and Sharing Experiences. ICADL 2005. Lecture Notes in Computer Science, vol 3815. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11599517_14
Download citation
DOI: https://doi.org/10.1007/11599517_14
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-30850-8
Online ISBN: 978-3-540-32291-7
eBook Packages: Computer ScienceComputer Science (R0)