Pareto Chart Easy Guide (2012)

A Pareto chart named after Vilfredo Pareto (an Italian Economist) is a bar chart in which all bars are ordered from largest to the smallest along with a line showing the cumulative percentage and count of the bars. The left vertical axis has the frequency of occurrence (number of occurrences), or some other important unit of measure such as cost. The right vertical axis contains the cumulative percentage of the total number of occurrences or the total of the particular unit of measure such as total cost. For the Pareto chart, the cumulative function is concave because the bars (representing the reasons) are in decreasing order. A Pareto chart is also called a Pareto distribution diagram.

The Pareto chart is also known as the 80/20 rule chart. These charts offer several benefits for data analysis and problem-solving.

A Pareto chart can be used when the following questions have their answer is “yes”

  1. Can data be arranged into categories?
  2. Is the rank of each category important?

Pareto charts are often used to analyze defects in a manufacturing process or the most frequent reasons for customer complaints to help determine the types of defects that are most prevalent (important) in a process. So a Company can focus on improving its efforts in particular important areas where it can make the largest gain or the lowest loss by eliminating causes of defects. So it’s easy to prioritize the problem areas using Pareto charts. The categories in the “tail” of the Pareto chart are called the insignificant factors.

Pareto Chart Example

Pareto Chart

The Pareto chart given above shows the reasons for consumer complaints against airlines in 2004. Here each bar represents the number (frequency) of each complaint received. The major complaints received are related to flight problems (such as cancellations, delays, and other deviations from the schedule). The 2nd largest complaint is about customer service (rude or unhelpful employees, inadequate meals or cabin service, treatment of delayed passengers, etc.). Flight problems account for 21% of the complaints, while both flight problems and customer service account for 40% of the complaints. The top three complaint categories account for 55% of the complaints. So, to reduce the number of complaints, airlines should need to work on flight delays, customer service, and baggage problems.

By incorporating Pareto-charts into data analysis, one can get valuable insights, prioritize effectively, and make data-driven decisions.

Charts and Graphs

References:

  • Nancy R. Tague (2004). “Seven Basic Quality Tools”. The Quality Toolbox. Milwaukee, Wisconsin: American Society for Quality. p. 15. Retrieved 2010-02-05.
  • http://en.wikipedia.org/wiki/Pareto_chart

See more about Charts and Graphs

Online MCQs Intermediate Mathematics (Matrices and Determinants)

Graphs in R Language

Goldfeld Quandt Test: Comparison of Variances of Error Terms

The Goldfeld Quandt test is one of two tests proposed in a 1965 paper by Stephen Goldfeld and Richard Quandt. Both parametric and nonparametric tests are described in the paper, but the term “Goldfeld–Quandt test” is usually associated only with the parametric test.
Goldfeld-Quandt test is frequently used as it is easy to apply when one of the regressors (or another r.v.) is considered the proportionality factor of heteroscedasticity. Goldfeld-Quandt test is applicable for large samples. The observations must be at least twice as many as the parameters to be estimated. The test assumes normality and serially independent error terms $u_i$.

The Goldfeld Quandt test compares the variance of error terms across discrete subgroups. So data is divided into h subgroups. Usually, the data set is divided into two parts or groups, and hence the test is sometimes called a two-group test.

Goldfeld Quandt Test: Comparison of Variances of Error Terms

Before starting how to perform the Goldfeld Quand Test, you may read more about the term Heteroscedasticity, the remedial measures of heteroscedasticity, Tests of Heteroscedasticity, and Generalized Least Square Methods.

Goldfeld Quandt Test Procedure:

The procedure for conducting the Goldfeld-Quandt Test is;

  1. Order the observations according to the magnitude of $X$ (the independent variable which is the proportionality factor).
  2. Select arbitrarily a certain number (c) of central observations which we omit from the analysis. (for $n=30$, 8 central observations are omitted i.e. 1/3 of the observations are removed). The remaining $n-c$ observations are divided into two sub-groups of equal size i.e. $\frac{(n-2)}{2}$, one sub-group includes small values of $X$ and the other sub-group includes the large values of $X$, and a data set is arranged according to the magnitude of $X$.
  3. Now Fit the separate regression to each of the sub-groups, and obtain the sum of squared residuals from each of them.
    So $\sum c_1^2$ shows the sum of squares of Residuals from a sub-sample of low values of $X$ with $(n – c)/2 – K$ df, where K is the total number of parameters.$\sum c_2^2$ shows the sum of squares of Residuals from a sub-sample of large values of $X$ with $(n – c)/2 – K$ df, where K is the total number of parameters.
  4. Compute the Relation $F^* = \frac{RSS_2/df}{RSS_2/df}=\frac{\sum c_2^2/ ((n-c)/2-k)}{\sum c_1^2/((n-c)/2-k) }$

If variances differ, F* will have a large value. The higher the observed value of the F*-ratio the stronger the heteroscedasticity of the $u_i$.

Goldfeld Quandt Test of

References

  • Goldfeld, Stephen M.; Quandt, R. E. (June 1965). “Some Tests for Homoscedasticity”. Journal of the American Statistical Association 60 (310): 539–547
  • Kennedy, Peter (2008). A Guide to Econometrics (6th ed.). Blackwell. p. 116

Numerical Example of the Goldfeld-Quandt Test.

R Programming and Data Analysis in R

Online MCQs Test Website

Heteroscedasticity Definition, Reasons, Consequences (2012)

Heteroscedasticity Definition

An important assumption of OLS is that the disturbances $u_i$ appearing in the population regression function are homoscedastic (Error terms have the same variance).

The variance of each disturbance term $u_i$, conditional on the chosen values of explanatory variables is some constant number equal to $\sigma^2$. $E(u_{i}^{2})=\sigma^2$; where $i=1,2,\cdots, n$.
Homo means equal and scedasticity means spread.

Consider the general linear regression model
\[y_i=\beta_1+\beta_2 x_{2i}+ \beta_3 x_{3i} +\cdots + \beta_k x_{ki} + \varepsilon\]

If $E(\varepsilon_{i}^{2})=\sigma^2$ for all $i=1,2,\cdots, n$ then the assumption of constant variance of the error term or homoscedasticity is satisfied.

If $E(\varepsilon_{i}^{2})\ne\sigma^2$ then the assumption of homoscedasticity is violated and heteroscedasticity is said to be present. In the case of heteroscedasticity, the OLS estimators are unbiased but inefficient.

Examples:

  1. The range in family income between the poorest and richest families in town is the classical example of heteroscedasticity.
  2. The range in annual sales between a corner drug store and a general store.
Heteroscedasticity Definition, Reasons, Consequences

Reasons for Heteroscedasticity

There are several reasons why the variances of error term $u_i$ may be variable, some of which are:

  1. Following the error learning models, as people learn their errors of behavior become smaller over time. In this case $\sigma_{i}^{2}$ is expected to decrease. For example the number of typing errors made in a given period on a test to the hours put in typing practice.
  2. As income grows, people have more discretionary income, and hence $\sigma_{i}^{2}$ is likely to increase with income.
  3. As data-collecting techniques improve, $\sigma_{i}^{2}$ is likely to decrease.
  4. Heteroscedasticity can also arise as a result of the presence of outliers. The inclusion or exclusion of such observations, especially when the sample size is small, can substantially alter the results of regression analysis.
  5. Heteroscedasticity arises from violating the assumption of CLRM (classical linear regression model), that the regression model is not correctly specified.
  6. Skewness in the distribution of one or more regressors included in the model is another source of heteroscedasticity.
  7. Incorrect data transformation and incorrect functional form (linear or log-linear model) are also the sources of heteroscedasticity
Heteroscedasticity Definition

Consequences of Heteroscedasticity

  1. The OLS estimators and regression predictions based on them remain unbiased and consistent.
  2. The OLS estimators are no longer the BLUE (Best Linear Unbiased Estimators) because they are no longer efficient, so the regression predictions will be inefficient too.
  3. Because of the inconsistency of the covariance matrix of the estimated regression coefficients, the tests of hypotheses, (t-test, F-test) are no longer valid.

Note: Problems of heteroscedasticity are likely to be more common in cross-sectional than in time series data.

Reference
Greene, W.H. (1993). Econometric Analysis, Prentice–Hall, ISBN 0-13-013297-7.
Verbeek, Marno (2004.) A Guide to Modern Econometrics, 2. ed., Chichester: John Wiley & Sons.
Gujarati, D. N. & Porter, D. C. (2008). Basic Econometrics, 5. ed., McGraw Hill/Irwin.

FAQS about Heteroscedasticity

  1. Define heteroscedasticity.
  2. What are the major consequences that may occur if heteroscedasticity occurs?
  3. What does mean by the constant variance of the error term in linear regression models?
  4. What are the possible reasons that make error term variance a variable?
  5. In what kind of data are problems of heteroscedasticity is likely to exist?
https://itfeature.com

Learn R Programming Language

Moments In Statistics (2012)

Introduction to Moments in Statistics

The measure of central tendency (location) and the measure of dispersion (variation) are useful for describing a data set. Both the measure of central tendencies and the measures of dispersion fail to tell anything about the shape of the distribution. We need some other certain measure called the moments. Moments in Statistics are used to identify the shape of the distribution known as skewness and kurtosis.

Moments are fundamental statistical tools for understanding the characteristics of any dataset. They provide quantitative measures that describe the data:

  • Central tendency: The “center” of the data. It is the most common measure of central tendency, but other moments can also be used.
  • Spread: Indicates how scattered the data is around the central tendency. Common measures of spread include variance and standard deviation.
  • Shape: Describes the overall form of the data distribution. For instance, is it symmetrical? Does it have a long tail on one side? Higher-order moments like skewness and kurtosis help analyze the shape.

Moments about Mean

The moments about the mean are the mean of deviations from the mean after raising them to integer powers. The $r$th population moment about the mean is denoted by $\mu_r$ is

\[\mu_r=\frac{\sum\limits^{N}_{i=1}(y_i – \bar{y} )^r}{N}\]

where $r=1,2,\cdots$

The corresponding sample moment denoted by $m_r$ is

\[\mu_r=\frac{\sum\limits^{n}_{i=1}(y_i – \bar{y} )^r}{n}\]

Note that if $r=1$ i.e. the first moment is zero as $\mu_1=\frac{\sum\limits^{n}_{i=1}(y_i – \bar{y} )^1}{n}=0$. So the first moment is always zero.

If $r=2$ then the second moment is variance i.e. \[\mu_2=\frac{\sum\limits^{n}_{i=1}(y_i – \bar{y} )^2}{n}\]

Similarly, the 3rd and 4th moments are

\[\mu_3=\frac{\sum\limits^{n}_{i=1}(y_i – \bar{y} )^3}{n}\]

\[\mu_4=\frac{\sum\limits^{n}_{i=1}(y_i – \bar{y} )^4}{n}\]

For grouped data, the $r$th sample moment  about the sample mean $\bar{y}$ is

\[\mu_r=\frac{\sum\limits^{n}_{i=1}f_i(y_i – \bar{y} )^r}{\sum\limits^{n}_{i=1}f_i}\]

where $\sum\limits^{n}_{i=1}f_i=n$

Moments about Arbitrary Value

The $r$th sample sample moment about any arbitrary origin “a” denoted by $m’_r$ is
\[m’_r = \frac{\sum\limits^{n}_{i=1}(y_i – a)^2}{n} = \frac{\sum\limits^{n}_{i=1}D^r_i}{n}\]
where $D_i=(y_i -a)$ and $r=1,2,\cdots$.

therefore
\begin{eqnarray*}
m’_1&=&\frac{\sum\limits^{n}_{i=1}(y_i – a)}{n}=\frac{\sum\limits^{n}_{i=1}D_i}{n}\\
m’_2&=&\frac{\sum\limits^{n}_{i=1}(y_i – a)^2}{n}=\frac{\sum\limits^{n}_{i=1}D_i ^2}{n}\\
m’_3&=&\frac{\sum\limits^{n}_{i=1}(y_i – a)^3}{n}=\frac{\sum\limits^{n}_{i=1}D_i ^3}{n}\\
m’_4&=&\frac{\sum\limits^{n}_{i=1}(y_i – a)^4}{n}=\frac{\sum\limits^{n}_{i=1}D_i ^4}{n}
\end{eqnarray*}

The $r$th sample moment for grouped data about any arbitrary origin “a” is

$$m’_r=\frac{\sum\limits^{n}_{i=1}f_i(y_i – a)^r}{\sum\limits^{n}_{i=1}f} = \frac{\sum f_i D_i ^r}{\sum f}$$

The moments about the mean are usually called central moments and the moments about any arbitrary origin “a” are called non-central moments or raw moments.

One can calculate the moments about mean from the following relations by calculating the moments about arbitrary value

\begin{eqnarray*}
m_1&=& m’_1 – (m’_1) = 0 \\
m_2 &=& m’_2 – (m’_1)^2\\
m_3 &=& m’_3 – 3m’_2m’_1 +2(m’_1)^3\\
m_4 &=& m’_4 -4 m’_3m’_1 +6m’_2(m’_1)^2 -3(m’_1)^4
\end{eqnarray*}

Moments about Zero

If variable $y$ assumes $n$ values $y_1, y_2, \cdots, y_n$ then $r$th moment about zero can be obtained by taking $a=0$ so the moment about arbitrary value will be
\[m’_r = \frac{\sum y^r}{n}\]

where $r=1,2,3,\cdots$.

therefore
\begin{eqnarray*}
m’_1&=&\frac{\sum y^1}{n}\\
m’_2 &=&\frac{\sum y^2}{n}\\
m’_3 &=&\frac{\sum y^3}{n}\\
m’_4 &=&\frac{\sum y^4}{n}\\
\end{eqnarray*}

The third moment is used to define the skewness of a distribution

\[{\rm Skew ness} = \frac{\sum\limits^{i=1}_n (y_i-\overline{y})^3} {ns^3}\]

If the distribution is symmetric then the skewness will be zero. Skewness will be positive if there is a long tail in the positive direction and skewness will be negative if there is a long tail in the negative direction.

The fourth moment is used to define the kurtosis of a distribution

\[{\rm Kurtosis} = \frac{\sum\limits^{i=1}_{n} (y_i -\overline{y})^4}{ns^4}\]

Moments in Statistics

In summary, moments are quantitative measures that describe the distribution of a dataset around its central tendency. Different types of moments, provide specific information about the shape and characteristics of data. By understanding and utilizing moments, one can get a deeper understanding of the data and make more informed decisions in statistical analysis.

FAQS about Moments in Statistics

  1. Define moments in Statistics.
  2. What is the use of moments?
  3. How moments are used to understand the characteristics of the data?
  4. What is meant by moments about mean?
  5. What are moments about arbitrary value?
  6. What is meant by moments about zero?
  7. Define the different types of moments.
Moments In Statistics (2012)

Online MCQs Test Preparation Website