Abstract
Detailed technical presentation of our contributions that are related to Active Scene Recognition. This includes our approaches to Object Pose Prediction and Next-Best-View estimation.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
During indirect search, we only look for objects that miss in the available scene models. Objects from scene categories for which no partial scene models have been estimated yet are being looked for during direct search.
- 2.
We used the library SMACH [2] for implementing ASR.
- 3.
We look at Fig. 4.3 in landscape format.
- 4.
Since in relation to learning, the contributions in this thesis are restricted to that of scene category models, such a simplification makes sense.
- 5.
We designate a scene model from a scene category as incomplete when the model does not contain all objects that belong to the category.
- 6.
The pose of the robot is composed of the pose of its mobile base and the orientation of its PTU.
- 7.
Please keep in mind that “NBV estimation” selects both the objects that shall be searched from a viewpoint and the viewpoint itself.
- 8.
Without loss of generality, all variations are supposed to be visible from the same set of robot head poses.
- 9.
We present all operations as performed on the view of the left camera.
- 10.
The confidence in detecting objects from a given viewpoint depends on the absolute number of predictions within the viewing frustum the given viewpoint induces.
- 11.
Please bear in mind that scene model instance and scene category instance are synonyms.
- 12.
This kind of visualization is further detailed out in Sect. 4.6.
- 13.
The subset \([\mathbf {I}_{\{\mathbf {S}\}}^{r}]\) is a multiset [10, p. 29].
- 14.
The color of a cloud changes from red to green with increasing confidence of the represented instance.
- 15.
Instances with low, average and high ratings are displayed as red, yellow and green clouds.
- 16.
This picture displays a dark green frustum of a view at position \(\mathbf {p}\) that contains two predicted poses for a spherical grey object. The same lines of sight, visualized as dark red arrows, are assigned to each of both predictions.
- 17.
For the sake of simplicity, we only consider the left camera on the sensor head of the abstract robot from Sect. 4.4 in this and the following two sections.
- 18.
Throughout the experiments we conducted for this thesis, we set frustum parameters as follows: \(\text {fov}_x = {30}^{\circ }\), \(\text {fov}_y = {20}^{\circ }\), ncp \(= 0.4\) m, fcp \(= 1.5\) m.
- 19.
The dimensionality of the parameter space of our objective function is \(|\{o\}_{P}| + 6\).
- 20.
As earlier in this thesis, we limit the description of our contributions to design decisions that reduce problem complexity, leaving out canonical algorithm optimization strategies such as parallelization.
- 21.
Here, we ignore the specificities of hex grids.
References
Aumann-Cleres, F.: Markerbasiertes Kalibrieren der kinematischen Kette und Aufstellen der Rückwärtstransformation zwischen der Basis und dem Sensorkopf eines mobilen Roboters. Bachelor’s thesis, Advisor: P. Meißner, Reviewer: R. Dillmann, Karlsruhe Institute of Technology (2016)
Bohren, J., Cousins, S.: The smach high-level executive. IEEE Robot. Autom. Mag. (2013)
Bourke, P.: Frustum culling. http://paulbourke.net/miscellaneous/frustum (2000). Accessed 01 Dec 2017
Bronshtein, I., Semendyayev, K., Musiol, G., Muehlig, H.: Handbook of Mathematics, 5th edn. Springer, Berlin (2007)
Devert, A.: Spreading points on a disc and on a sphere—Marmakoide’s Blog. http://blog.marmakoide.org/?p=1 (2012). Accessed 14 Nov 2017
Dillmann, R., Huck, M.: Informationsverarbeitung in der Robotik. Springer, Berlin (1991)
Eidenberger, R., Grundmann, T., Schneider, M., Feiten, W., Fiegert, M., Wichert, G.V., Lawitzky, G.: Scene analysis for service robots. In: Towards Service Robots for Everyday Environments, pp. 181–213. Springer, Berlin (2012)
Gamma, E., Johnson, R., Vlissides, J., Helm, R.: Design Patterns: Elements of Reusable Object-Oriented Software. Addison-Wesley, Boston (1995)
Garvey, T.D.: Perceptual strategies for purposive vision. Tech-Technical Note 117, SRI International (1976)
Hein, J.L.: Discrete Mathematics, 2nd edn. Jones and Bartlett Publishers, Inc, Burlington (2002)
Isard, M., Blake, A.: Condensation–conditional density propagation for visual tracking. Int. J. Comput. Vis. 29(1), 5–28 (1998)
Karrenbauer, O.: Realisierung und komparative Analyse von alternativen Methoden zum uninformierten Generieren optimaler Folgen von Ansichten für die 3D-Objektsuche. Bachelor’s thesis, Advisor: P. Meißner, Reviewer: R. Dillmann, Karlsruhe Institute of Technology (2017)
Kunze, L., Doreswamy, K.K., Hawes, N.: Using qualitative spatial relations for indirect object search. In: Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), pp. 163–168. IEEE (2014)
Lehmann, A., Leibe, B., Van Gool, L.: Fast prism: branch and bound hough transform for object class detection. Int. J. Comput. Vis. 94(2), 175–197 (2011)
Lorbach, M., Hofer, S., Brock, O.: Prior-assisted propagation of spatial information for object search. In: Proceedings of the 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2014), pp. 2904–2909. IEEE (2014)
Meißner, P., Reckling, R., Wittenbeck, V., Schmidt-Rohr, S., Dillmann, R.: Active scene recognition for programming by demonstration using next-best-view estimates from hierarchical implicit shape models. In: Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), pp. 5585–5591. IEEE (2014)
Meißner, P., Schleicher, R., Hutmacher, R., Schmidt-Rohr, S., Dillmann, R.: Scene recognition for mobile robots by relational object search using next-best-view estimates from hierarchical implicit shape models. In: Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 137–144. IEEE (2016)
Patel, A.: Hexagonal grids. https://www.redblobgames.com/grids/hexagons (2013 & 2015). Accessed 11 Nov 2017
Potthast, C., Sukhatme, G.S.: A probabilistic framework for next best view estimation in a cluttered environment. J. Vis. Commun. Image Represent. 25(1), 148–164 (2014)
Quigley, M., Conley, K., Gerkey, B., Faust, J., Foote, T., Leibs, J., Wheeler, R., Ng, A.Y.: Ros: an open-source robot operating system. In: ICRA Workshop on Open Source Software, Kobe, p. 5 (2009)
Russell, S.J., Norvig, P.: Artificial Intelligence: A Modern Approach, 3rd international edn. Prentice Hall Press, Upper Saddle River (2010)
Siciliano, B., Khatib, O.: Springer Handbook of Robotics. Springer Science + Business Media, Berlin (2008)
Thrun, S., Burgard, W., Fox, D.: Probabilistic Robotics (Intelligent Robotics and Autonomous Agents). The MIT Press, Cambridge (2005)
Vasquez-Gomez, J.I., Sucar, L.E., Murrieta-Cid, R.: View planning for 3d object reconstruction with a mobile manipulator robot. In: Proceedings of the 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2014), pp. 4227–4233. IEEE (2014)
Wixson, L.E., Ballard, D.H.: Using intermediate objects to improve the efficiency of visual search. Int. J. Comput. Vis. 12(2–3), 209–230 (1994)
Ye, Y., Tsotsos, J.K.: Sensor planning for 3d object search. Comput. Vis. Image Underst. 73(2), 145–168 (1999)
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Meißner, P. (2020). Active Scene Recognition. In: Indoor Scene Recognition by 3-D Object Search. Springer Tracts in Advanced Robotics, vol 135. Springer, Cham. https://doi.org/10.1007/978-3-030-31852-9_4
Download citation
DOI: https://doi.org/10.1007/978-3-030-31852-9_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-31851-2
Online ISBN: 978-3-030-31852-9
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)