Abstract
Problems requiring the application of neural network technology are typically not those for which standard statistics can be applied, though within the world of statistics it is not uncommon to combine the benefits of neural networks with that of certain statistical methods such as multivariate analysis, to yield solutions that are more accurate than traditional methods. In this chapter, it is shown that one particular type of neural network is used to analyze problems in possession of sets of constraints that impose a series of conditions on solutions. Constraints that are both linear and nonlinear can be satisfied at one time using the constraint satisfaction (CS) artificial neural network (ANN). The ANN is a single layer model for which the weights matrix connections are symmetric and there is no association with local geography. Each node may have its own bias, and the analysis is performed using the auto-associative backpropagation network. It is shown that problems of resource optimization and data profiling are particularly applicable to this advancement.
Keywords
- Constraint Satisfaction (CS)
- Constraint Satisfaction Neural Network
- Connection Weight Matrix
- Hidden Units
- Visible Units
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Buscema, M. (1995). Self-reflexive networks, theory, topology, applications. Quality & Quantity, 29(4), 339 403.
Buscema, M. (1998a). Constraint satisfaction neural networks. Substance Use & Misuse, 33(2), 389–408.
Buscema, M. (1998b). Back propagation neural networks. Substance Use & Misuse, 33(2), 233–270.
Hebb, D. O. (1949). The organization of behavior. New York: Wiley.
Hopfield, J. J. (1982). Neural networks and physical systems with emergent collective computational abilities. Proceedings of the National Academy of Sciences, USA, 79, 2554–2558.
Hopfield, J. J. (1984). Neurons with graded response have collective computational properties like those of two-state neurons. Proceedings of the National Academy of Sciences, USA, 81, 3088–3092.
Rumelhart, D. E., Hinton, G. E., & Williams, R. J. (1986a). Learning internal representations by error propagation. In D. E. Rumelhart & J. L. McClelland (Eds.), Parallel distributed processing (Foundations, explorations in the microstructure of cognition, Vol. 1). Cambridge, MA/London: MIT Press.
Rumelhart, D. E., Smolensky, P., McClelland, J. L., & Hinton, G. E. (1986b). Schemata and sequential thought processes in PDP models. In J. L. McClelland & D. E. Rumelhart (Eds.), PDP, exploration in the microstructure of cognition (Vol. II, pp. 7–57). Cambridge, MA: MIT Press.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer Science+Business Media Dordrecht
About this chapter
Cite this chapter
Buscema, M. (2013). Theory of Constraint Satisfaction Neural Networks. In: Buscema, M., Tastle, W. (eds) Intelligent Data Mining in Law Enforcement Analytics. Springer, Dordrecht. https://doi.org/10.1007/978-94-007-4914-6_13
Download citation
DOI: https://doi.org/10.1007/978-94-007-4914-6_13
Published:
Publisher Name: Springer, Dordrecht
Print ISBN: 978-94-007-4913-9
Online ISBN: 978-94-007-4914-6
eBook Packages: Humanities, Social Sciences and LawSocial Sciences (R0)