Abstract
In elementary probability, if Ω is finite, say Ω = {1, 2,…, n}, F is the collection of all subsets of Ω, and if P gives weight a i ≥ 0 to {i} with Σn i=1 a i = 1, then the (mathematical) expectation E[X] of a random variable X on Ω (i.e., a function X : Ω → ℝ) is defined to be Σn i=1 X(i) a i , = Σn i=1 X(i)P({i{). When a i = 1/n for each i, this expectation is the usual average value of X. Heuristically, this number is what we expect as the average of a large number of “observations” of X — see the weak law of large numbers in Chapter IV. Also, if X is non-negative, then E[X] can be conceived as the “area” under the graph of X: over each i one may imagine a rectangle of width P({a i }) and height X(i); then Σn i=1 X(i)P({i}) is the sum of the “areas” of the rectangles that make up the set under the graph.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 1997 Springer Science+Business Media New York
About this chapter
Cite this chapter
Taylor, J.C. (1997). Integration. In: An Introduction to Measure and Probability. Springer, New York, NY. https://doi.org/10.1007/978-1-4612-0659-0_2
Download citation
DOI: https://doi.org/10.1007/978-1-4612-0659-0_2
Publisher Name: Springer, New York, NY
Print ISBN: 978-0-387-94830-0
Online ISBN: 978-1-4612-0659-0
eBook Packages: Springer Book Archive