Skip to main content

Extracting rules from artificial neural networks with kernel-based representations

  • Artificial Neural Nets Simulation and Implementation
  • Conference paper
  • First Online:
Engineering Applications of Bio-Inspired Artificial Neural Networks (IWANN 1999)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 1607))

Included in the following conference series:

Abstract

In Neural Networks models the knowledge synthesized from the training process is represented in a subsymbolic fashion (weights, kernels, combination of numerical descriptions) that makes difficult its interpretation. The interpretation of the internal representation of a successful Neural Network can be useful to understand the nature of the problem and its solution, to use the Neural “model” as a tool that gives insights about the problem solved and not just as a solving mechanism treated as a black box. The internal representation used by the family of kernel-based Neural Networks (including Radial Basis Functions, Support Vector machines, Coulomb potential methods, and some probabilistic Neural Networks) can be seen as a set of positive instances of classification and, thereafter, used to derive fuzzy rules suitable for explanation or inference processes. The probabilistic nature of the kernel-based Neural Networks is captured by the membership functions associated to the components of the rules extracted. In this work we propose a method to extract fuzzy rules from trained Neural Networks of the family mentioned; comparing the quality of the knowledge extracted by different methods using known machine learning benchmarks.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. J.A. Alexander and M.C. Mozer. Template-based algorithms for connectionist rule extraction. Advances in Neural Information Processing Systems, 7:609–616, 1995.

    Google Scholar 

  2. C. Bachman, Cooper L., Dembo A., and Zeitouni O. A relaxation model for memory with high storage density. Proceedings of the National Academy of Science, 21:609–616, 1995.

    Google Scholar 

  3. J.M. Benitez, J.L. Castro, and I. Requena. Are artificial neural networks black boxes? IEEE Transaction on Neural Networks, 8:1156–1164, 1997.

    Article  Google Scholar 

  4. B. Boser, I. Guyon, and V. Vapnik. A training algorithm for optimal margin classification. In Proceedings 5th annual Workshop on Computational Learning Theory, pages 144–152, ACM, 1992.

    Google Scholar 

  5. E. Keogh C. Blake and C.J. Merz. UCI repository of machine learning databases, 1998. http://www.ics.uci.edu/~mlearn/MLRepository.html.

    Google Scholar 

  6. C. Cortes and V. Vapnik. Support vector networks. Machine Learning, 20:273–297, 1995.

    MATH  Google Scholar 

  7. M.W. Craven and J.W. Shavlik. Extracting tree-structured representations of trained networks. Advances in Neural Information Processing Systems, 8:24–30, 1996.

    Google Scholar 

  8. S. I. Gallant. Connectionist expert systems. Communications of the ACM, 31:152–169, 1988.

    Article  Google Scholar 

  9. M.J. Healy and T.P. Caudell. Acquiring rule sets as a product of learning in a logical neural architecture. IEEE Transaction on Neural Networks, 8:461–474, 1997.

    Article  Google Scholar 

  10. J. J. Hopfield. Neural networks and physical systems with emergent collective computational abilities. In Proceedings of the National Aca. of Science USA, volume 81, pages 3088–3092, 1984.

    Google Scholar 

  11. M. Ishikawa. Neural network approach to rule extraction. In Proceedings of the 2nd New Sealand International Conference on Artificial Neural Networks and Fuzzy Systems, pages 6–9. IEEE Computer Society, 1995.

    Google Scholar 

  12. J.S. Roger Jang and C.T. Sun. Functional equivalence between radial basis function networks and fuzzy inference systems. IEEE Transaction on Neural Networks, 4:156–159, 1993.

    Article  Google Scholar 

  13. B. Kosko. Neural Networks and Fuzzy Systems: A dynamical approach to machine intelligence. Prentice-Hall, 1992.

    Google Scholar 

  14. S. Lee and R. Kil. A gaussian potential function network with hierarchically self-organizing learning. Neural Networks, 4(2):207–224, 1991.

    Article  Google Scholar 

  15. C. McMillan, M.C. Mozer, and P. Smolensky. Rule induction through integrated symbolic and subsymbolic processing. Advances in Neural Information Processing Systems, 4:969–1497, 1992.

    Google Scholar 

  16. J.E. Moody and C. Darken. Fast learning networks of locally-tuned processing units. Neural Computation, 1(2):281–294, 1989.

    Google Scholar 

  17. C.W. Omlin and C.L. Giles. Extraction of rules from discrete-time recurrent neural networks. Technical Report 92-23, Department of Computer Science, Rensselaer Polytechnic Institute, 1992.

    Google Scholar 

  18. T. Poggio and F. Girosi. Networks for approximation and learning. Proceedings IEEE, 78:1481–1497, 1990.

    Article  Google Scholar 

  19. J. Ramírez. MHC-an evolutive connectionist model for hybrid training. In Lecture Notes in Computer Science 686: New trends in neural computation, pages 223–229. Springer-Verlag, 1993.

    Google Scholar 

  20. D. Reilly, L. Cooper, and C. Elbaum. A neural model for category learning. Biological Cybernetics, 45:35–41, 1982.

    Article  Google Scholar 

  21. A. Saha and J. Keeler. Algorithms for better representation and faster learning in radial basis function networks. Advances in Neural Information Processing Systems, 2:482–489, 1990.

    Google Scholar 

  22. C. Scofield. Learning internal representation in the coulomb energy network. In Proceedings of the IEEE International Conference on Neural Networks, 1988.

    Google Scholar 

  23. D. Specht. Generation of polynomial discrimination functions for pattern recognition. PhD thesis, Stanford University, 1966.

    Google Scholar 

  24. D. Sreerupa and M. Mozer. A unified gradient-descent/clustering architecture for finite state machine induction. Advances in Neural Information Processing Systems, 6:19–26, 1994.

    Google Scholar 

  25. S. Thrun. Extracting rules from artificial neural networks with distributed representations. Advances in Neural Information Processing Systems, 7:505–512, 1995.

    Google Scholar 

  26. G. Towell and J. Shavlik. Interpretation of artificial neural networks: mapping knowledge-based neural networks into rules. Advances in Neural Information Processing Systems, 4:977–984, 1992.

    Google Scholar 

  27. V. Vapnik. The nature of Statistical Learning Theory. Springer-Verlag, 1995.

    Google Scholar 

  28. L.X. Wang. Adaptive Fuzzy Systems and Control—Desing and stability analysis. Englewood Cliffs, 1994.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to José M. Ramírez .

Editor information

José Mira Juan V. Sánchez-Andrés

Rights and permissions

Reprints and permissions

Copyright information

© 1999 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Ramírez, J.M. (1999). Extracting rules from artificial neural networks with kernel-based representations. In: Mira, J., Sánchez-Andrés, J.V. (eds) Engineering Applications of Bio-Inspired Artificial Neural Networks. IWANN 1999. Lecture Notes in Computer Science, vol 1607. Springer, Berlin, Heidelberg . https://doi.org/10.1007/BFb0100473

Download citation

  • DOI: https://doi.org/10.1007/BFb0100473

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-66068-2

  • Online ISBN: 978-3-540-48772-2

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics