Abstract
Research in machine learning concentrates on the study of learning single concepts from examples. In this framework the learner attempts to learn a single hidden function from a collection of examples, assumed to be drawn independently from some unknown probability distribution. However,in many cases—as in most natural language and visual processing situations—decisions depend on the outcomes of several different but mutually dependent classifiers. The classifiers’ outcomes need to respect some constraints that could arise from the sequential nature of the data or other domain specific conditions, thus requiring a level of inference on top the predictions.
We will describe research and present challenges related to Inference with Classifiers—a paradigm in which we address the problem of using the outcomes of several different classifiers in making coherent inferences—those that respect constraints on the outcome of the classifiers. Examples will be given from the natural language domain.
Paper written to accompany an invited talk at ECML’02. This research is supported by NSF grants IIS-99-84168,ITR-IIS-00-85836 and an ONR MURI award.
Download to read the full chapter text
Chapter PDF
Similar content being viewed by others
References
D. Heckerman, D. M. Chickering, C. Meek, R. Rounthwaite, and C. M. Kadie. Dependency networks for inference,collaborative filtering,and data visualization. Journal of Machine Learning Research,1:49–75, 2000.
R. Khardon and D. Roth. Learning to reason. Journal of the ACM,44(5):697–725, Sept. 1997.
J. Lafferty, A. McCallum,and F. Pereira. Conditional random fields: Probabilistic models for segmenting and labeling sequence data. In Proceedings of the 18th International Conference on Machine Learning,Williamstown, MA, 2001.
J. McCarthy. Programs with common sense. In R. Brachman and H. Levesque, editors, Readings in Knowledge Representation, 1985. Morgan-Kaufmann,1958.
M. Munoz, V. Punyakanok, D. Roth,and D. Zimak. A learning approach to shallow parsing. In EMNLP-VLC’99, the Joint SIGDAT Conference on Empirical Methods in Natural Language Processing and Very Large Corpora,pages 168–178, June 1999.
K. P. Murphy, Y. Weiss, and M. I. Jordan. Loopy belief propagation for approximate inference: An empirical study. In Proceedings of Uncertainty in AI,pages 467–475,1999.
N. J. Nilsson. Logic and artificial intelligence. Artificial Intelligence,47:31–56, 1991.
V. Punyakanok and D. Roth. The use of classifiers in sequential inference. In NIPS-13; The 2000 Conference on Advances in Neural Information Processing Systems, pages 995–1001. MIT Press,2001.
D. Roth. Learning to reason: The non-monotonic case. In Proc. of the International Joint Conference on Artificial Intelligence,pages 1178–1184, 1995.
D. Roth. Learning to resolve natural language ambiguities: A unified approach. In Proc. of the American Association of Artificial Intelligence,pages 806–813,1998.
D. Roth and W.-T. Yih. Probabilistic reasoning for entity and relation recognition. In COLING 2002, The 19th International Conference on Computational Linguistics,2002.
L. G. Valiant. Robust logic. In Proceedings of the Annual ACM Symp. on the Theory of Computing,1999.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2002 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Roth, D. (2002). Reasoning with Classifiers. In: Elomaa, T., Mannila, H., Toivonen, H. (eds) Machine Learning: ECML 2002. ECML 2002. Lecture Notes in Computer Science(), vol 2430. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-36755-1_43
Download citation
DOI: https://doi.org/10.1007/3-540-36755-1_43
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-44036-9
Online ISBN: 978-3-540-36755-0
eBook Packages: Springer Book Archive