<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://en.formulasearchengine.com/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=38.109.87.242</id>
	<title>formulasearchengine - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://en.formulasearchengine.com/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=38.109.87.242"/>
	<link rel="alternate" type="text/html" href="https://en.formulasearchengine.com/wiki/Special:Contributions/38.109.87.242"/>
	<updated>2026-05-02T01:10:09Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.43.0-wmf.28</generator>
	<entry>
		<id>https://en.formulasearchengine.com/index.php?title=Observed_information&amp;diff=16563</id>
		<title>Observed information</title>
		<link rel="alternate" type="text/html" href="https://en.formulasearchengine.com/index.php?title=Observed_information&amp;diff=16563"/>
		<updated>2013-06-18T18:34:52Z</updated>

		<summary type="html">&lt;p&gt;38.109.87.242: /* Fisher information */ Elaboration that Fisher information corresponds to a single observation distributed according to the hypothetical model.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Refimprove|date=February 2008}}&lt;br /&gt;
&lt;br /&gt;
The &#039;&#039;&#039;sample mean&#039;&#039;&#039; or &#039;&#039;&#039;empirical mean&#039;&#039;&#039; and the &#039;&#039;&#039;sample covariance&#039;&#039;&#039; are [[statistic]]s computed from a collection of data on one or more [[random variables]]. The sample mean is a [[vector (mathematics)|vector]] each of whose elements is the sample mean of one of the random variables{{spaced ndash}}that is, each of whose elements is the [[arithmetic average]] of the observed values of one of the variables. The sample covariance matrix is a square [[Matrix (mathematics)|matrix]] whose &#039;&#039;i, j&#039;&#039; element is the sample covariance (an estimate of the population covariance) between the sets of observed values of two of the variables and whose &#039;&#039;i, i&#039;&#039; element is the sample variance of the observed values of one of the variables. If only one variable has had values observed, then the sample mean is a single number (the arithmetic average of the observed values of that variable) and the sample covariance matrix is also simply a single value (the sample variance of the observed values of that variable).&lt;br /&gt;
&lt;br /&gt;
==Sample mean==&lt;br /&gt;
{{main|Arithmetic mean|date=February 2013}}&lt;br /&gt;
&lt;br /&gt;
Let &amp;lt;math&amp;gt;x_{ij}&amp;lt;/math&amp;gt; be the &#039;&#039;i&#039;&#039;&amp;lt;sup&amp;gt;th&amp;lt;/sup&amp;gt; independently drawn observation (&#039;&#039;i=1,...,N&#039;&#039;) on the &#039;&#039;j&#039;&#039;&amp;lt;sup&amp;gt;th&amp;lt;/sup&amp;gt; random variable (&#039;&#039;j=1,...,K&#039;&#039;). These observations can be arranged into &#039;&#039;N&#039;&#039;&lt;br /&gt;
column vectors, each with &#039;&#039;K&#039;&#039; entries, with the &#039;&#039;K&#039;&#039; ×1 column vector giving the &#039;&#039;i&#039;&#039;&amp;lt;sup&amp;gt;th&amp;lt;/sup&amp;gt; observations of all variables being denoted &amp;lt;math&amp;gt;\mathbf{x}_i&amp;lt;/math&amp;gt; (&#039;&#039;i=1,...,N&#039;&#039;). &lt;br /&gt;
&lt;br /&gt;
The &#039;&#039;&#039;sample mean vector&#039;&#039;&#039; &amp;lt;math&amp;gt;\mathbf{\bar{x}}&amp;lt;/math&amp;gt; is a column vector whose &#039;&#039;j&#039;&#039;&amp;lt;sup&amp;gt;th&amp;lt;/sup&amp;gt; element &amp;lt;math&amp;gt;\bar{x}_{j}&amp;lt;/math&amp;gt; is the average value of the &#039;&#039;N&#039;&#039; observations of the &#039;&#039;j&#039;&#039;&amp;lt;sup&amp;gt;th&amp;lt;/sup&amp;gt; variable:&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt; \bar{x}_{j}=\frac{1}{N}\sum_{i=1}^{N}x_{ij},\quad j=1,\ldots,K. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Thus, the sample mean vector contains the average of the observations for each variable, and is written&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt; \mathbf{\bar{x}}=\frac{1}{N}\sum_{i=1}^{N}\mathbf{x}_i. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Sample covariance==&lt;br /&gt;
{{move portions|Estimation of covariance matrices|section=y|small=left|date=February 2013}}&lt;br /&gt;
The &#039;&#039;&#039;sample covariance matrix&#039;&#039;&#039; is a &#039;&#039;K&#039;&#039;-by-&#039;&#039;K&#039;&#039; [[Matrix (mathematics)|matrix]] &amp;lt;math&amp;gt;\textstyle \mathbf{Q}=\left[  q_{jk}\right]  &amp;lt;/math&amp;gt; with entries &lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt; q_{jk}=\frac{1}{N-1}\sum_{i=1}^{N}\left(  x_{ij}-\bar{x}_j \right)  \left( x_{ik}-\bar{x}_k \right), &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;q_{jk}&amp;lt;/math&amp;gt; is an estimate of the [[covariance]] between the {{math|j}}&amp;lt;sup&amp;gt;th&amp;lt;/sup&amp;gt;&lt;br /&gt;
variable and the {{math|k}}&amp;lt;sup&amp;gt;th&amp;lt;/sup&amp;gt; variable of the population underlying the data.&lt;br /&gt;
In terms of the observation vectors, the sample covariance is&lt;br /&gt;
:&amp;lt;math&amp;gt;\mathbf{Q} = {1 \over {N-1}}\sum_{i=1}^N (\mathbf{x}_i-\mathbf{\bar{x}}) (\mathbf{x}_i-\mathbf{\bar{x}})^\mathrm{T},&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Alternatively, arranging the observation vectors as the columns of a matrix, so that&lt;br /&gt;
:&amp;lt;math&amp;gt;\mathbf{F} = \begin{bmatrix}\mathbf{x}_1 &amp;amp; \mathbf{x}_2 &amp;amp; \dots &amp;amp; \mathbf{x}_N \end{bmatrix}&amp;lt;/math&amp;gt;,&lt;br /&gt;
which is a matrix of &#039;&#039;K&#039;&#039; rows and &#039;&#039;N&#039;&#039; columns.&lt;br /&gt;
Here, the sample covariance matrix can be computed as&lt;br /&gt;
:&amp;lt;math&amp;gt;\mathbf{Q} = \frac{1}{N-1}( \mathbf{F} - \mathbf{\bar{x}} \,\mathbf{1}_N^\mathrm{T} ) ( \mathbf{F} - \mathbf{\bar{x}} \,\mathbf{1}_N^\mathrm{T} )^\mathrm{T}&amp;lt;/math&amp;gt;,&lt;br /&gt;
where &amp;lt;math&amp;gt;\mathbf{1}_N&amp;lt;/math&amp;gt; is an &#039;&#039;N&#039;&#039; by {{math|1}}  vector of ones. &lt;br /&gt;
If the observations are arranged as rows instead of columns, so &amp;lt;math&amp;gt;\mathbf{\bar{x}}&amp;lt;/math&amp;gt; is now a 1×&#039;&#039;K&#039;&#039; row vector and &amp;lt;math&amp;gt;\mathbf{M}=\mathbf{F}^\mathrm{T}&amp;lt;/math&amp;gt; is an &#039;&#039;N&#039;&#039;×&#039;&#039;K&#039;&#039; matrix whose column &#039;&#039;j&#039;&#039; is the vector of &#039;&#039;N&#039;&#039; observations on variable &#039;&#039;j&#039;&#039;, then applying transposes &lt;br /&gt;
in the appropriate places yields&lt;br /&gt;
:&amp;lt;math&amp;gt;\mathbf{Q} = \frac{1}{N-1}( \mathbf{M} -  \mathbf{1}_N \mathbf{\bar{x}} )^\mathrm{T} ( \mathbf{M} - \mathbf{1}_N \mathbf{\bar{x}} ).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Discussion==&lt;br /&gt;
The sample mean and the sample covariance matrix are [[Bias of an estimator|unbiased estimates]] of the [[mean]] and the [[covariance matrix]] of the [[random vector]] &amp;lt;math&amp;gt;\textstyle \mathbf{X}&amp;lt;/math&amp;gt;, a row vector whose &#039;&#039;j&#039;&#039;&amp;lt;sup&amp;gt;th&amp;lt;/sup&amp;gt; element (&#039;&#039;j = 1, ..., K&#039;&#039;) is one of the random variables.&amp;lt;ref name=&amp;quot;JohnsonWichern2007&amp;quot;&amp;gt;{{cite book|author1=Richard Arnold Johnson|author2=Dean W. Wichern|title=Applied Multivariate Statistical Analysis|url=http://books.google.com/books?id=gFWcQgAACAAJ|accessdate=10 August 2012|year=2007|publisher=Pearson Prentice Hall|isbn=978-0-13-187715-3}}&amp;lt;/ref&amp;gt; The sample covariance matrix has &amp;lt;math&amp;gt;\textstyle N-1&amp;lt;/math&amp;gt; in the denominator rather than &amp;lt;math&amp;gt;\textstyle N&amp;lt;/math&amp;gt; due to a variant of [[Bessel&#039;s correction]]:  In short, the sample covariance relies on the difference between each observation and the sample mean, but the sample mean is slightly correlated with each observation since it&#039;s defined in terms of all observations. If the population mean &amp;lt;math&amp;gt;\operatorname{E}(\mathbf{X})&amp;lt;/math&amp;gt; is known, the analogous unbiased estimate &lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt; q_{jk}=\frac{1}{N}\sum_{i=1}^N \left(  x_{ij}-\operatorname{E}(X_j)\right)  \left( x_{ik}-\operatorname{E}(X_k)\right), &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
using the population mean, has &amp;lt;math&amp;gt;\textstyle N&amp;lt;/math&amp;gt; in the denominator. This is an example of why in probability and statistics it is essential to distinguish between [[random variable]]s (upper case letters) and [[Realization (probability)|realizations]] of the random variables (lower case letters).&lt;br /&gt;
&lt;br /&gt;
The [[maximum likelihood]]  [[Estimation of covariance matrices|estimate of the covariance]]&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt; q_{jk}=\frac{1}{N}\sum_{i=1}^N \left(  x_{ij}-\bar{x}_j \right)  \left( x_{ik}-\bar{x}_k \right) &amp;lt;/math&amp;gt; &lt;br /&gt;
&lt;br /&gt;
for the [[Gaussian distribution]] case has &#039;&#039;N&#039;&#039; in the denominator as well. The ratio of 1/&#039;&#039;N&#039;&#039; to 1/(&#039;&#039;N&#039;&#039;&amp;amp;nbsp;&amp;amp;minus;&amp;amp;nbsp;1) approaches 1 for large&amp;amp;nbsp;&#039;&#039;N&#039;&#039;, so the maximum likelihood estimate approximately equals the unbiased estimate when the sample is large.&lt;br /&gt;
&lt;br /&gt;
==Variance of the sample mean==&lt;br /&gt;
{{main|Standard error of the mean}}&lt;br /&gt;
For each random variable, the sample mean is a good [[estimator]] of the population mean, where a &amp;quot;good&amp;quot; estimator is defined as being efficient and unbiased. Of course the estimator will likely not be the true value of the [[Statistical population|population]] mean since different samples drawn from the same distribution will give different sample means and hence different estimates of the true mean. Thus the sample mean is a [[random variable]], not a constant, and consequently has its own distribution. For a random sample of &#039;&#039;N&#039;&#039; observations on the &#039;&#039;j&#039;&#039;&amp;lt;sup&amp;gt;th&amp;lt;/sup&amp;gt; random variable, the sample mean&#039;s distribution itself has mean equal to the population mean &amp;lt;math&amp;gt;E(X_j)&amp;lt;/math&amp;gt; and variance equal to  &amp;lt;math&amp;gt; \frac{\sigma^2_j}{N},&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;\sigma^2_j&amp;lt;/math&amp;gt; is the variance of the random variable &#039;&#039;X&#039;&#039;&amp;lt;sub&amp;gt;j&amp;lt;/sub&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
==Weighted samples==&lt;br /&gt;
{{main|Weighted mean|date=February 2013}}&lt;br /&gt;
{{move portions|Weighted mean|section=y|small=left|date=February 2013}}&lt;br /&gt;
&lt;br /&gt;
In a weighted sample, each vector &amp;lt;math&amp;gt;\textstyle \textbf{x}_{i}&amp;lt;/math&amp;gt; (each set of single observations on each of the &#039;&#039;K&#039;&#039; random variables) is assigned a weight &amp;lt;math&amp;gt;\textstyle w_i \geq0&amp;lt;/math&amp;gt;. Without loss of generality, assume that the weights are [[Normalizing constant|normalized]]:&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt; \sum_{i=1}^{N}w_i = 1. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
(If they are not, divide the weights by their sum).&lt;br /&gt;
Then the [[weighted mean]] vector &amp;lt;math&amp;gt;\textstyle \mathbf{\bar{x}}&amp;lt;/math&amp;gt; is given by &lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt; \mathbf{\bar{x}}=\sum_{i=1}^N w_i \mathbf{x}_i.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
and the elements &amp;lt;math&amp;gt;q_{jk}&amp;lt;/math&amp;gt; of the weighted covariance matrix &amp;lt;math&amp;gt;\textstyle \mathbf{Q}&amp;lt;/math&amp;gt; are&lt;br /&gt;
&amp;lt;ref name=&amp;quot;Galassi-2007-GSL&amp;quot;&amp;gt;Mark Galassi, Jim Davies, James Theiler, Brian Gough, Gerard Jungman, Michael Booth, and Fabrice Rossi. [http://www.gnu.org/software/gsl/manual GNU Scientific Library - Reference manual, Version 1.15], 2011. &lt;br /&gt;
[http://www.gnu.org/software/gsl/manual/html_node/Weighted-Samples.html Sec. 21.7 Weighted Samples]&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt; q_{jk}=\frac{\sum_{i=1}^{N}w_i}{\left(\sum_{i=1}^{N}w_i\right)^2-\sum_{i=1}^{N}w_i^2}&lt;br /&gt;
\sum_{i=1}^N w_i \left(  x_{ij}-\bar{x}_j \right)  \left( x_{ik}-\bar{x}_k \right)  . &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If all weights are the same, &amp;lt;math&amp;gt;\textstyle w_{i}=1/N&amp;lt;/math&amp;gt;, the weighted mean and covariance reduce to the sample mean and covariance above.&lt;br /&gt;
&lt;br /&gt;
==Criticism==&lt;br /&gt;
The sample mean and sample covariance are widely used in statistics and applications, and are extremely common measures of [[Location parameter|location]] and [[statistical dispersion|dispersion]], respectively, likely the most common: they are easily calculated and possess desirable characteristics.&lt;br /&gt;
&lt;br /&gt;
However, they suffer from certain drawbacks; notably, they are not [[robust statistics]], meaning that they are sensitive to [[outliers]]. As robustness is often a desired trait, particularly in real-world applications, robust alternatives may prove desirable, notably [[quantile]]-based statistics such the [[sample median]] for location,&amp;lt;ref&amp;gt;[http://www.edge.org/q2008/q08_16.html#kosko The World Question Center 2006: The Sample Mean], [[Bart Kosko]]&amp;lt;/ref&amp;gt; and [[interquartile range]] (IQR) for dispersion. Other alternatives include [[Trimmed estimator|trimming]] and [[Winsorising]], as in the [[trimmed mean]] and the [[Winsorized mean]].&lt;br /&gt;
&lt;br /&gt;
==See also==&lt;br /&gt;
&lt;br /&gt;
*[[Unbiased estimation of standard deviation]]&lt;br /&gt;
*[[Estimation of covariance matrices]]&lt;br /&gt;
*[[Scatter matrix]]&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Category:Covariance and correlation]]&lt;br /&gt;
[[Category:Estimation for specific parameters]]&lt;br /&gt;
[[Category:Summary statistics]]&lt;br /&gt;
[[Category:Matrices]]&lt;br /&gt;
[[Category:U-statistics]]&lt;/div&gt;</summary>
		<author><name>38.109.87.242</name></author>
	</entry>
</feed>