Skip to main content

MOSRO: Enabling Mobile Sensing for Real-Scene Objects with Grid Based Structured Output Learning

  • Conference paper
MultiMedia Modeling (MMM 2014)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 8325))

Included in the following conference series:

Abstract

Visual objects in mobile photos are usually captured in uncontrolled conditions, such as various viewpoints, positions, scales, and background clutter. In this paper, therefore, we developed a MObile Sensing framework for robust Real-scene Object recognition and localization (MOSRO). By extending the conventional structured output learning with the proposed grid based representation as the output structure, MOSRO is not only able to locate the visual objects precisely but also achieve real-time performances. The experimental results showed that the proposed framework outperforms the state-of-the-art methods on public real-scene image datasets. Further, to demonstrate its effectiveness for practical applications, the proposed MOSRO framework was implemented on Android mobile platforms as a prototype system for sensing various business signs on the street and instantly retrieving relevant information of the recognized businesses.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Girod, B., Chandrasekhar, V., Grzeszczuk, R., Reznik, Y.A.: Mobile visual search: Architectures, technologies, and the emerging mpeg standard. In: IEEE MultiMedia (2011)

    Google Scholar 

  2. You, C.-W., Cheng, W.-H., Wen Tsui, A., Tsai, T.-H., Campbell, A.: Mobilequeue: an image-based queue card management system through augmented reality phones. In: UbiComp (2012)

    Google Scholar 

  3. Girod, B., Chandrasekhar, V., Chen, D.M., Cheung, N.-M., Grzeszczuk, R., Reznik, Y., Takacs, G., Tsai, S.S., Vedantham, R.: Mobile visual search. IEEE Signal Processing Magazine (2011)

    Google Scholar 

  4. Blaschko, M.B., Lampert, C.H.: Learning to localize objects with structured output regression. In: Forsyth, D., Torr, P., Zisserman, A. (eds.) ECCV 2008, Part I. LNCS, vol. 5302, pp. 2–15. Springer, Heidelberg (2008)

    Chapter  Google Scholar 

  5. Everingham, M., Zisserman, A., Williams, C.K.I., Van Gool, L.: The PASCAL Visual Object Classes Challenge 2006 (VOC 2006) Results (2006), http://www.pascal-network.org/challenges/VOC/voc2006/results.pdf

  6. Ji, R., Duan, L.-Y., Chen, J., Yang, S., Huang, T., Yao, H., Gao, W.: Pkubench: A context rich mobile visual search benchmark. In: ICIP (2011)

    Google Scholar 

  7. OPS data set, http://mclab.citi.sinica.edu.tw/dataset/ops62/ops62.html

  8. Yu, F.X., Ji, R., Chang, S.-F.: Active query sensing for mobile location search. In: ACM Multimedia (2011)

    Google Scholar 

  9. Kuo, Y.-H., Lee, W.-Y., Hsu, W.H., Cheng, W.-H.: Augmenting mobile city-view image retrieval with context-rich user-contributed photos. In: ACM Multimedia (2011)

    Google Scholar 

  10. Lampert, C.H., Blaschko, M.B., Hofmann, T.: Beyond sliding windows: Object localization by efficient subwindow search. In: CVPR (2008)

    Google Scholar 

  11. Yeh, T., Lee, J.J., Darrell, T.: Fast concurrent object localization and recognition. In: CVPR (2009)

    Google Scholar 

  12. Zhang, Z., Cao, Y., Salvi, D., Oliver, K., Waggoner, J., Wang, S.: Free-shape subwindow search for object localization. In: CVPR (2010)

    Google Scholar 

  13. Vijayanarasimhan, S., Grauman, K.: Efficient region search for object detection. In: CVPR (2011)

    Google Scholar 

  14. Tsochantaridis, I., Joachims, T., Hofmann, T., Altun, Y.: Large margin methods for structured and interdependent output variables. J. Mach. Learn. Res. (2005)

    Google Scholar 

  15. Fergus, R., Li, F.-F., Perona, P., Zisserman, A.: Learning object categories from internet image searches. Proceedings of the IEEE (2010)

    Google Scholar 

  16. Nowozin, S., Lampert, C.H.: Global connectivity potentials for random field models. In: CVPR (2009)

    Google Scholar 

  17. Everingham, M., Van Gool, L., Williams, C.K.I., Winn, J., Zisserman, A.: The PASCAL Visual Object Classes Challenge 2008 (VOC 2008) Results (2008), http://www.pascal-network.org/challenges/VOC/voc2008/workshop/index.html

  18. Nister, D., Stewenius, H.: Scalable recognition with a vocabulary tree. In: CVPR (2006)

    Google Scholar 

  19. van de Sande, K., Gevers, T., Snoek, C.: Evaluating color descriptors for object and scene recognition. IEEE Trans. Pattern Anal. Mach. Intell. (2010)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2014 Springer International Publishing Switzerland

About this paper

Cite this paper

Chi, HY., Cheng, WH., Chen, MS., Tsui, A.W. (2014). MOSRO: Enabling Mobile Sensing for Real-Scene Objects with Grid Based Structured Output Learning. In: Gurrin, C., Hopfgartner, F., Hurst, W., Johansen, H., Lee, H., O’Connor, N. (eds) MultiMedia Modeling. MMM 2014. Lecture Notes in Computer Science, vol 8325. Springer, Cham. https://doi.org/10.1007/978-3-319-04114-8_18

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-04114-8_18

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-04113-1

  • Online ISBN: 978-3-319-04114-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics