Analysis 1: Difference between revisions

From Advanced Labs Wiki
Jump to navigation Jump to search
No edit summary
No edit summary
Line 13: Line 13:
<math>\sigma^2 = \langle (x-\mu)^2 \rangle. </math>
<math>\sigma^2 = \langle (x-\mu)^2 \rangle. </math>


This is the "second moment" of the distribution of data. Note that the square is essential because, from the definition of the mean, <math> \langle (x-\mu) \rangle = 0 </math>. Analogously to the mean, we can compute the ''sample variance'' from our dataset <math>x_i</math>:
This is the "second moment" of the distribution of data. Note that the square is essential because, from the definition of the mean, <math> \langle (x-\mu) \rangle = 0 </math>. The square root of the variance is the ''standard deviation'' </math>\sigma</math>. Analogously to the mean, we can compute the ''sample variance'' from our dataset <math>x_i</math>:


<math>s^2 = \frac{1}{N-1} \sum^{N}_i (x_i - m)^2 </math>.
<math>s^2 = \frac{1}{N-1} \sum^{N}_i (x_i - m)^2 </math>.


Why do we use $N-1$ instead of $N$ in the denominator for the sample variance? We'll come back to that later.
Why do we use $N-1$ instead of $N$ in the denominator for the sample variance? We'll come back to that later.

Revision as of 14:19, 8 February 2012

Mean and Variance

The measurements you make in the lab are subject to error. We can formalize this idea by thinking of a measurement of<math>x</math>. This may be the number of counts per minute from the ML or the RS labs, or it might be a reading of the Hall voltage, etc etc. Nature determines the expectation value of the measurement. We'll use the mean as the expectation value (instead of, say, the median):

<math>\mu = \langle x \rangle. </math>

If you take a lot of data, such that you have <math>N</math> measurements of <math>x</math> -- <math>x_i</math>, then the sample mean is given by

<math>m = \frac{1}{N} \sum^N_i x_i </math>

which should be familiar. In order to quantify the error associated with our data, we introduce the variance $\sigma^2$

<math>\sigma^2 = \langle (x-\mu)^2 \rangle. </math>

This is the "second moment" of the distribution of data. Note that the square is essential because, from the definition of the mean, <math> \langle (x-\mu) \rangle = 0 </math>. The square root of the variance is the standard deviation </math>\sigma</math>. Analogously to the mean, we can compute the sample variance from our dataset <math>x_i</math>:

<math>s^2 = \frac{1}{N-1} \sum^{N}_i (x_i - m)^2 </math>.

Why do we use $N-1$ instead of $N$ in the denominator for the sample variance? We'll come back to that later.