#### The coefficient of variation, variance, and standard deviation are the most widely used measures of variability. We’ll discuss each of these in turn, finishing off with the coefficient of variation.

Variance measures the dispersion of a set of data points around their mean value. Population variance, denoted by sigma squared, is equal to the sum of squared differences between the observed values and the population mean, divided by the total number of observations. Sample variance, on the other hand, is denoted by s squared and is equal to the sum of squared differences between observed sample values and the sample mean, divided by the number of sample observations minus 1.

While variance is a common measure of data dispersion, in most cases the figure you will obtain is pretty large and hard to compare as the unit of measurement is squared. The easy fix is to calculate its square root and obtain a statistic known as standard deviation. In most analyses you perform, standard deviation will be much more meaningful than variance.

Alright. The other measure we still have to introduce is the coefficient of variation. It is equal to the standard deviation, divided by the mean. Another name for the term is relative standard deviation. This is an easy way to remember its formula – it is simply the standard deviation relative to the mean. As you probably guessed, there is a population and sample formula once again.

## Add comment