# High-Dimensional Mapping

• Bipin Kumar Tripathi
Chapter
Part of the Studies in Computational Intelligence book series (SCI, volume 571)

## Abstract

The complex plane is the geometric representation of complex numbers established by the real axis and the orthogonal imaginary axis. A point on the complex plane can be viewed as a complex number with X and Y coordinates regarded as real and imaginary parts of the number. It can be thought of as a modified Cartesian plane, where real part is represented by a displacement along the X-axis and imaginary part by a displacement along the Y-axis. The set of complex numbers is a Field equipped with basic algebraic properties of addition and multiplication operations [1], and hence gives a perfect platform of operation. The properties of the complex plane are different from those of the real line. A complex number have a nonnegative modulus and an argument (Arg) associated with it that locates the complex number uniquely on the plane. It is natural to represent a nonzero complex number with a directed line segment or vector on the complex plane. The extension of traditional real-valued neuron on complex plane has varied its structure from single dimension to two dimensions. Real-valued neuron administers motion on real line, while learning with a complex-valued neuron applies a linear transformation, called 2D motion, [2, 3] to each input signal (complex number on plane). Thus, learning in a complex-valued neural network (CVNN) is characterized with the complex-valued signals flowing through the network, and has ability to capture two dimension patterns naturally. Therefore, the concept of complex plane allows a geometric interpretation of complex numbers in CVNN. The present chapter investigates and explores the mapping properties of the CVNN through some problems of mapping to bring forth the differences between CVNN and ANN, where the stress was on problems that CVNN solves and ANN does not.

## Keywords

Complex-valued Neural Networks (CVNN) Basic Algebraic Properties Bilinear Transformation Proposed Neuron Conventional Neurons
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

## References

1. 1.
Brown, J.W., Churchill, R.V.: Complex Variables and Applications, 7th edn. McGraw-Hill, New York (2003)Google Scholar
2. 2.
Nitta, T.: An analysis of the fundamental structure of complex-valued neurons. Neural Process. Lett. 12, 239–246 (2000)
3. 3.
Tripathi, B.K., Kalra, P.K.: The novel aggregation function based neuron models in complex domain. Soft Comput. 14(10), 1069–1081 (2010)Google Scholar
4. 4.
Saff, E.B., Snider, A.D.: Fundamentals of Complex Analysis with Applications to Engineering and Science. Prentice-Hall. Englewood Cliffs (2003)Google Scholar
5. 5.
Nitta, T.: An extension of the back-propagation algorithm to complex numbers. Neural Netw. 10(8), 1391–1415 (1997)

## Authors and Affiliations

1. 1.Computer Science and EngineeringHarcourt Butler Technological InstituteKanpurIndia