Moments In Statistics (2012)

Introduction to Moments in Statistics

The measure of central tendency (location) and the measure of dispersion (variation) are useful for describing a data set. Both the measure of central tendencies and the measures of dispersion fail to tell anything about the shape of the distribution. We need some other certain measure called the moments. Moments in Statistics are used to identify the shape of the distribution known as skewness and kurtosis.

Moments are fundamental statistical tools for understanding the characteristics of any dataset. They provide quantitative measures that describe the data:

  • Central tendency: The “center” of the data. It is the most common measure of central tendency, but other moments can also be used.
  • Spread: Indicates how scattered the data is around the central tendency. Common measures of spread include variance and standard deviation.
  • Shape: Describes the overall form of the data distribution. For instance, is it symmetrical? Does it have a long tail on one side? Higher-order moments like skewness and kurtosis help analyze the shape.

Moments about Mean

The moments about the mean are the mean of deviations from the mean after raising them to integer powers. The $r$th population moment about the mean is denoted by $\mu_r$ is

\[\mu_r=\frac{\sum\limits^{N}_{i=1}(y_i – \bar{y} )^r}{N}\]

where $r=1,2,\cdots$

The corresponding sample moment denoted by $m_r$ is

\[\mu_r=\frac{\sum\limits^{n}_{i=1}(y_i – \bar{y} )^r}{n}\]

Note that if $r=1$ i.e. the first moment is zero as $\mu_1=\frac{\sum\limits^{n}_{i=1}(y_i – \bar{y} )^1}{n}=0$. So the first moment is always zero.

If $r=2$ then the second moment is variance i.e. \[\mu_2=\frac{\sum\limits^{n}_{i=1}(y_i – \bar{y} )^2}{n}\]

Similarly, the 3rd and 4th moments are

\[\mu_3=\frac{\sum\limits^{n}_{i=1}(y_i – \bar{y} )^3}{n}\]

\[\mu_4=\frac{\sum\limits^{n}_{i=1}(y_i – \bar{y} )^4}{n}\]

For grouped data, the $r$th sample moment  about the sample mean $\bar{y}$ is

\[\mu_r=\frac{\sum\limits^{n}_{i=1}f_i(y_i – \bar{y} )^r}{\sum\limits^{n}_{i=1}f_i}\]

where $\sum\limits^{n}_{i=1}f_i=n$

Moments about Arbitrary Value

The $r$th sample sample moment about any arbitrary origin “a” denoted by $m’_r$ is
\[m’_r = \frac{\sum\limits^{n}_{i=1}(y_i – a)^2}{n} = \frac{\sum\limits^{n}_{i=1}D^r_i}{n}\]
where $D_i=(y_i -a)$ and $r=1,2,\cdots$.

therefore
\begin{eqnarray*}
m’_1&=&\frac{\sum\limits^{n}_{i=1}(y_i – a)}{n}=\frac{\sum\limits^{n}_{i=1}D_i}{n}\\
m’_2&=&\frac{\sum\limits^{n}_{i=1}(y_i – a)^2}{n}=\frac{\sum\limits^{n}_{i=1}D_i ^2}{n}\\
m’_3&=&\frac{\sum\limits^{n}_{i=1}(y_i – a)^3}{n}=\frac{\sum\limits^{n}_{i=1}D_i ^3}{n}\\
m’_4&=&\frac{\sum\limits^{n}_{i=1}(y_i – a)^4}{n}=\frac{\sum\limits^{n}_{i=1}D_i ^4}{n}
\end{eqnarray*}

The $r$th sample moment for grouped data about any arbitrary origin “a” is

$$m’_r=\frac{\sum\limits^{n}_{i=1}f_i(y_i – a)^r}{\sum\limits^{n}_{i=1}f} = \frac{\sum f_i D_i ^r}{\sum f}$$

The moments about the mean are usually called central moments and the moments about any arbitrary origin “a” are called non-central moments or raw moments.

One can calculate the moments about mean from the following relations by calculating the moments about arbitrary value

\begin{eqnarray*}
m_1&=& m’_1 – (m’_1) = 0 \\
m_2 &=& m’_2 – (m’_1)^2\\
m_3 &=& m’_3 – 3m’_2m’_1 +2(m’_1)^3\\
m_4 &=& m’_4 -4 m’_3m’_1 +6m’_2(m’_1)^2 -3(m’_1)^4
\end{eqnarray*}

Moments about Zero

If variable $y$ assumes $n$ values $y_1, y_2, \cdots, y_n$ then $r$th moment about zero can be obtained by taking $a=0$ so the moment about arbitrary value will be
\[m’_r = \frac{\sum y^r}{n}\]

where $r=1,2,3,\cdots$.

therefore
\begin{eqnarray*}
m’_1&=&\frac{\sum y^1}{n}\\
m’_2 &=&\frac{\sum y^2}{n}\\
m’_3 &=&\frac{\sum y^3}{n}\\
m’_4 &=&\frac{\sum y^4}{n}\\
\end{eqnarray*}

The third moment is used to define the skewness of a distribution

\[{\rm Skew ness} = \frac{\sum\limits^{i=1}_n (y_i-\overline{y})^3} {ns^3}\]

If the distribution is symmetric then the skewness will be zero. Skewness will be positive if there is a long tail in the positive direction and skewness will be negative if there is a long tail in the negative direction.

The fourth moment is used to define the kurtosis of a distribution

\[{\rm Kurtosis} = \frac{\sum\limits^{i=1}_{n} (y_i -\overline{y})^4}{ns^4}\]

Moments in Statistics

In summary, moments are quantitative measures that describe the distribution of a dataset around its central tendency. Different types of moments, provide specific information about the shape and characteristics of data. By understanding and utilizing moments, one can get a deeper understanding of the data and make more informed decisions in statistical analysis.

FAQS about Moments in Statistics

  1. Define moments in Statistics.
  2. What is the use of moments?
  3. How moments are used to understand the characteristics of the data?
  4. What is meant by moments about mean?
  5. What are moments about arbitrary value?
  6. What is meant by moments about zero?
  7. Define the different types of moments.
Moments In Statistics (2012)

Online MCQs Test Preparation Website

Skewness Formula, Introduction, Interpretation (2012)

Skewness is the degree of asymmetry or departure from the symmetry of the distribution of a real-valued random variable.

Positive Skewed
If the frequency curve of distribution has a longer tail to the right of the central maximum than to the left, the distribution is said to be skewed to the right or to have positively skewed. In a positively skewed distribution, the mean is greater than the median and the median is greater than the mode i.e. $$Mean > Median > Mode$$

Negative Skewed
If the frequency curve has a longer tail to the left of the central maximum than to the right, the distribution is said to be skewed to the left or to be negatively skewed. In a negatively skewed distribution, the mode is greater than the median and the median is greater than the mean i.e. $$Mode > Median > Mean$$

Measure of Skewness Formulation

In a symmetrical distribution, the mean, median, and mode coincide. In a skewed distribution, these values are pulled apart.

Skewness Formula

Pearson’s Coefficient of Skewness Formula

Karl Pearson, (1857-1936) introduced a coefficient to measure the degree of skewness of distribution or curve, which is denoted by $S_k$ and defined by

\begin{eqnarray*}
S_k &=& \frac{Mean – Mode}{Standard Deviation}\\
S_k &=& \frac{3(Mean – Median)}{Standard Deviation}\\
\end{eqnarray*}
Usually, this coefficient varies between –3 (for negative) to +3 (for positive) and the sign indicates the direction of skewness.

Bowley’s Coefficient of Skewness Formula (Quartile Coefficient)

Arthur Lyon Bowley (1869-1957) proposed a measure of skewness based on the median and the two quartiles.

\[S_k=\frac{Q_1+Q_3-2Median}{Q_3 – Q_1}\]
Its values lie between 0 and ±1.

Moment Coefficient of Skewness Formula

This measure of skewness is the third moment expressed in standard units (or the moment ratio) thus given by

\[S_k=\frac{\mu_3}{\sigma^3} \]
Its values lie between -2 and +2.

If $S_k$ is greater than zero, the distribution or curve is said to be positively skewed. If $S_k$ is less than zero the distribution or curve is said to be negatively skewed. If $S_k$ is zero the distribution or curve is said to be symmetrical.

The skewness of the distribution of a real-valued random variable can easily be seen by drawing a histogram or frequency curve.

The skewness may be very extreme and in such a case these are called J-shaped distributions.

Skewness: J-Shaped Distribution

FAQs about Skewness

  1. What is the degree of asymmetry called?
  2. What is a departure from symmetry?
  3. If a distribution is negatively skewed then what is the relation between mean, median, and mode?
  4. If a distribution is positively skewed then what is the relation between mean, median, and mode?
  5. What is the relation between mean, median, and mode for a symmetrical distribution?
  6. What is the range of the moment coefficient of skewness?

Learn R Frequently Asked Questions

Correlation Coefficient: A Comprehensive Guide

The correlation is a measure of the co-variability of variables. It is used to measure the strength between two quantitative variables. It also tells the direction of a relationship between the variables. The positive value of the correlation coefficient indicates that there is a direct (supportive or positive) relationship between the variables while the negative value indicates there is a negative (opposite or indirect) relationship between the variables.

Correlation as Interdependence Between Variables

By definition, Pearson’s correlation is the interdependence between two quantitative variables. The causation (known as) cause and effect, is when an observed event or action appears to have caused a second event or action. Therefore, It does not necessarily imply any functional relationship between the variables concerned. Correlation theory does not establish any causal relationship between the variables as it is interdependence between the variables. Knowledge of the value of Pearson’s correlation coefficient $r$ alone will not enable us to predict the value of $Y$ from $X$.

High Correlation Coefficient does not Indicate Cause and Effect

Sometimes there is a high Relationship between unrelated variables such as the number of births and the number of murders in a country. This is a spurious correlation.

For example, suppose there is a positive correlation between watching violent movies and violent behavior in adolescence. The cause of both these could be a third variable (extraneous variable) say, growing up in a violent environment which causes the adolescents to watch violence-related movies and to have violent behavior.

Correlation Coefficient

Other Examples

  • The number of absences from class lectures decreases the grades.
  • As the weather gets colder, air conditioning costs decrease.
  • As the speed of the train (car, bus, or any other vehicle) is increased the length of time to get to the final point will also decrease.
  • As the age of a chicken increases the number of eggs it produces also decreases.
Statistics Help https://itfeature.com, Correlation Coefficient

R Frequently Asked Questions

Partial Correlation Coefficient (2012)

The Partial Correlation Coefficient measures the relationship between any two variables, where all other variables are kept constant i.e. controlling all other variables or removing the influence of all other variables. Partial correlation aims to find the unique variance between two variables while eliminating the variance from the third variable. The partial correlation technique is commonly used in “causal” modeling of fewer variables. The coefficient is determined in terms of the simple correlation coefficient among the various variables involved in multiple relationships.

Assumptions for computing the Partial Correlation Coefficient

The assumption for partial correlation is the usual assumption of Pearson Correlation:

  1. Linearity of relationships
  2. The same level of relationship throughout the range of the independent variable i.e. homoscedasticity
  3. Interval or near-interval data, and
  4. Data whose range is not truncated.

We typically conduct correlation analysis on all variables so that you can see whether there are significant relationships amongst the variables, including any “third variables” that may have a significant relationship to the variables under investigation.

This type of analysis helps to find the spurious correlations (i.e. correlations that are explained by the effect of some other variables) as well as to reveal hidden correlations – i.e. correlations masked by the effect of other variables. The partial correlation coefficient $r_{xy.z}$ can also be defined as the correlation coefficient between residuals $dx$ and $dy$ in this model.

Suppose we have a sample of $n$ observations $(x1_1,x2_1,x3_1),(x1_2,x2_2,x3_2),\cdots,(x1_n,x2_n,x3_n)$ from an unknown distribution of three random variables. We want to find the coefficient of partial correlation between $X_1$ and $X_2$ keeping $X_3$ constant which can be denoted by $r_{12.3}$ is the correlation between the residuals $x_{1.3}$ and $x_{2.3}$. The coefficient $r_{12.3}$ is a partial correlation of the 1st order.

\[r_{12.3}=\frac{r_{12}-r_{13} r_{23}}{\sqrt{1-r_{13}^2 } \sqrt{1-r_{23}^2 } }\]

Partial Correlation Coefficient

The coefficient of partial correlation between three random variables $X$, $Y$, and $Z$ can be denoted by $r_{x,y,z}$ and also be defined as the coefficient of correlation between $\hat{x}_i$ and $\hat{y}_i$ with
\begin{align*}
\hat{x}_i&=\hat{\beta}_{0x}+\hat{\beta}_{1x}z_i\\
\hat{y}_i&=\hat{\beta}_{0y}+\hat{\beta}_{1y}z_i\\
\end{align*}
where $\hat{\beta}_{0x}$ and $\hat{\beta_{1x}}$ are the least square estimators obtained by regressing $x_i$ on $z_i$ and $\hat{\beta}_{0y}$ and $\hat{\beta}_{1y}$ are the least square estimators obtained by regressing $y_i$ on $z_i$. Therefore by definition, the partial correlation between of $x$ and $y$ by controlling $z$ is
\[r_{xy.z}=\frac{\sum(\hat{x}_i-\overline{x})(\hat{y}_i-\overline{y})}{\sqrt{\sum(\hat{x}_i-\overline{x})^2}\sqrt{\sum(\hat{y}_i-\overline{y})^2}}\]

Partial Correlation Analysis

The coefficient of partial correlation is determined in terms of the simple correlation coefficients among the various variables involved in a multiple relationship. It is a very helpful tool in the field of statistics for understanding the true underlying relationships between variables, especially when you are dealing with potentially confounding factors.

Reference
Yule, G. U. (1926). Why do we sometimes get non-sense correlations between time series? A study in sampling and the nature of time series. J. Roy. Stat. Soc. (2) 89, 1-64.

Learn R Programming Language