Jump to content
Main menu
Main menu
move to sidebar
hide
Navigation
Main page
Recent changes
Random page
Help about MediaWiki
Special pages
Niidae Wiki
Search
Search
Appearance
Create account
Log in
Personal tools
Create account
Log in
Pages for logged out editors
learn more
Contributions
Talk
Editing
Expected value
(section)
Page
Discussion
English
Read
Edit
View history
Tools
Tools
move to sidebar
hide
Actions
Read
Edit
View history
General
What links here
Related changes
Page information
Appearance
move to sidebar
hide
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
==Uses and applications== The expectation of a random variable plays an important role in a variety of contexts. In [[statistics]], where one seeks [[Estimator|estimates]] for unknown [[Statistical parameter|parameters]] based on available data gained from [[Sampling (statistics)|samples]], the [[sample mean]] serves as an estimate for the expectation, and is itself a random variable. In such settings, the sample mean is considered to meet the desirable criterion for a "good" estimator in being [[Bias of an estimator|unbiased]]; that is, the expected value of the estimate is equal to the [[true value]] of the underlying parameter. {{See also|Estimation theory}} For a different example, in [[decision theory]], an agent making an optimal choice in the context of incomplete information is often assumed to maximize the expected value of their [[von Neumann–Morgenstern utility function|utility function]]. It is possible to construct an expected value equal to the probability of an event by taking the expectation of an [[indicator function]] that is one if the event has occurred and zero otherwise. This relationship can be used to translate properties of expected values into properties of probabilities, e.g. using the [[law of large numbers]] to justify estimating probabilities by [[Frequency (statistics)|frequencies]]. The expected values of the powers of ''X'' are called the [[moment (mathematics)|moments]] of ''X''; the [[moment about the mean|moments about the mean]] of ''X'' are expected values of powers of {{math|''X'' − E[''X'']}}. The moments of some random variables can be used to specify their distributions, via their [[moment generating function]]s. To empirically estimate the expected value of a random variable, one repeatedly measures observations of the variable and computes the [[arithmetic mean]] of the results. If the expected value exists, this procedure estimates the true expected value in an unbiased manner and has the property of minimizing the sum of the squares of the [[errors and residuals in statistics|residuals]] (the sum of the squared differences between the observations and the estimate). The law of large numbers demonstrates (under fairly mild conditions) that, as the [[Sample size|size]] of the sample gets larger, the [[variance]] of this estimate gets smaller. This property is often exploited in a wide variety of applications, including general problems of [[Estimation theory|statistical estimation]] and [[machine learning]], to estimate (probabilistic) quantities of interest via [[Monte Carlo methods]], since most quantities of interest can be written in terms of expectation, e.g. <math>\operatorname{P}({X \in \mathcal{A}}) = \operatorname{E}[{\mathbf 1}_{\mathcal{A}}],</math> where <math>{\mathbf 1}_{\mathcal{A}}</math> is the indicator function of the set <math>\mathcal{A}.</math> [[File:Beta first moment.svg|thumb|The mass of probability distribution is balanced at the expected value, here a Beta(α,β) distribution with expected value α/(α+β).]] In [[classical mechanics]], the [[center of mass]] is an analogous concept to expectation. For example, suppose ''X'' is a discrete random variable with values ''x<sub>i</sub>'' and corresponding probabilities ''p<sub>i</sub>.'' Now consider a weightless rod on which are placed weights, at locations ''x<sub>i</sub>'' along the rod and having masses ''p<sub>i</sub>'' (whose sum is one). The point at which the rod balances is E[''X'']. Expected values can also be used to compute the variance, by means of the computational formula for the variance <math display="block">\operatorname{Var}(X)= \operatorname{E}[X^2] - (\operatorname{E}[X])^2.</math> A very important application of the expectation value is in the field of [[quantum mechanics]]. The [[Expectation value (quantum mechanics)|expectation value of a quantum mechanical operator]] <math>\hat{A}</math> operating on a [[quantum state]] vector <math>|\psi\rangle</math> is written as <math>\langle\hat{A}\rangle = \langle\psi|\hat{A}|\psi\rangle.</math> The [[uncertainty principle|uncertainty]] in <math>\hat{A}</math> can be calculated by the formula <math>(\Delta A)^2 = \langle\hat{A}^2\rangle - \langle \hat{A} \rangle^2</math>.
Summary:
Please note that all contributions to Niidae Wiki may be edited, altered, or removed by other contributors. If you do not want your writing to be edited mercilessly, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource (see
Encyclopedia:Copyrights
for details).
Do not submit copyrighted work without permission!
Cancel
Editing help
(opens in new window)
Search
Search
Editing
Expected value
(section)
Add topic