Properties of Arithmetic Mean with Examples

In this post, we will discuss about properties of Arithmetic mean with Examples.

Arithmetic Mean

The arithmetic mean, often simply referred to as the mean, is a statistical measure that represents the central value of a dataset. The arithmetic mean is calculated by summing all the values in the dataset and then dividing by the total number of observations in the data.

The Sum of Deviations From the Mean is Zero

Property 1: The sum of deviations taken from the mean is always equal to zero. Mathematically $\sum\limits_{i=1}^n (x_i-\overline{x}) = 0$

Consider the ungrouped data case.

Obs. No.$X$$X_i-\overline{X}$
181-19
21000
396-4
41088
590-10
61022
71044
81033
91000
101099
1191-9
1211616
Total$\sum X_i = 1200$$\sum\limits_{i=1}^n (X_i-\overline{X})=0$

For grouped data $\overline{X} = \sum\limits_{i=1}^k f_i(X_i -\overline{X}) =0$, where for grouped data $\overline{X} =\frac{\sum\limits_{i=1}^n M_i f_i}{\sum\limits_{i=1}^n f_i}$. Suppose, we have the following grouped data

Classes$f$$M$$fM$$f_i(M_i – \overline{X})$
65 – 85975675$9\times (75 – 123) = -432$
85 – 1051095950$10\times (95 – 123) = -280$
105 – 125171151955$17\times (115 – 123) = -136$
125 – 145101351350$10\times (135 – 123) = 120$
145 – 1655155775$5\times (155 – 123) = 160$
165 – 1854175700$4\times (175 – 123) = 208$
185 – 2055195975$5\times (195 – 123) = 360$
Total$\Sigma f = n = 60$$\Sigma fM = 7380$$\sum\limits_{i=1}^k f_i(X_i -\overline{X}) =0$

Mean = $\overline{X} = \frac{\Sigma fM}{\Sigma f} = \frac{7380}{60} = 123$ .

The Combined Mean of Different Data Sets

Property 2: If there are different sets of data say $k$ then the combined mean/ average is

\begin{align*}
\overline{X}_c &= \frac{n_1 \overline{x}_1 + n_2\overline{x}_2 +\cdots + n_k \overline{x}_k }{n_1+n_2\cdots + n_k}\\
&=\frac{\Sigma x_1 + \Sigma x_2 + \cdots + \Sigma x_k}{n_1+n_2+\cdots + n_k}
\end{align*}

Suppose, we have data of $k$ groups.

Obs. No.$X_1$$X_2$$X_3$$X_4$$X_5$
1814092107113
2100309511094
396229911493
41085194109119
590101116105
6102103118
7104100115
8103102
9100101
10109
1191
12116
Sum1200143887789524

For \begin{align*}
\overline{X}_1 &= \frac{\sum\limits_{i=1}^n X_1}{n_1} = \frac{1200}{12} = 100\\
\overline{X}_2 &= \frac{\sum\limits_{i=1}^n X_2}{n_2} = \frac{143}{4} = 35.8\\
\overline{X}_3 &= \frac{\sum\limits_{i=1}^n X_3}{n_3} = \frac{887}{9} = 98.6\\
\overline{X}_4 &= \frac{\sum\limits_{i=1}^n X_4}{n_4} = \frac{789}{7} = 112.7\\
\overline{X}_5 &= \frac{\sum\limits_{i=1}^n X_5}{n_5} = \frac{524}{5} = 104.8\\
\overline{X}_c &= \frac{n_1\overline{X}_1 + n_2 \overline{X}_2 + \cdots + n_5 \overline{X}_5}{n_1+n_2+n_3+n_4+n_5}\\
&=\frac{12\times 100 + 4\times 35.8 + 9\times 98.6 + 7\times 112.7 + 5\times 104.8}{12+4+9+7+5} =\frac{3543.5}{37} = 95.7703
\end{align*}

For combined mean, not all the data set needs to be ungrouped or grouped. It may be possible that some data sets are ungrouped and some data sets are grouped.

Sum Squared Deviations from the Mean are Always Minimum

Property 3: The sum of the squared deviations of the observations from the arithmetic mean is minimum, which is less than the sum of the squared deviations of the observations from any other values. In other words, the sum of squared deviations from the mean is less than the sum of squared deviations from any other value. Mathematically,

For Ungrouped Data: $\Sigma (X_i – \overline{X})^2 < \Sigma (X_i – A)^2$

For Grouped Data: $\Sigma f(X_i – \overline{x})^2 < \Sigma f(M_i – A)^2$

where $A$ is any arbitrary value, also known as provisional mean. For this condition, $A$ is not equal to the arithmetic mean.

Note that the difference between the sum of deviations and the sum of squared deviations is that in the sum of deviations we take the difference (subtract) of each observation from the mean and then sum all the differences. In the sum of squared deviations, we take the difference of each observation from the mean, then take the square of all the differences, and then sum all the resultant values at the end.

Properties of Arithmetic Mean with Examples

From the above calculations, it can observed that $\Sigma (X_i – \overline{X})^2 < \Sigma (X_i – 90)^2 < \Sigma (X_i – 99)^2$.

No Resistant to Outliers

Property 4: The arithmetic mean is not resistant to outliers. It means that the arithmetic mean can be misleading if there are extreme values in the data.

Arithmetic Mean is Sensitive to Outliers

Property 5: The arithmetic mean is sensitive to extreme values (outliers) in the data. If there are a few very large or very small values, they can significantly influence the mean.

The Affect of Change in Scale and Origin

Property 6: If a constant value is added or subtracted from each data point, the mean will be changed by the same amount.
Similarly, if a constant value is multiplied or divided by each data point, the mean will be multiplied or divided by the same constant.

Unique Value

Property 7: For any given dataset, there is only one unique arithmetic mean.

In summary, the arithmetic mean is a widely used statistical measure (a measure of central tendency) that provides a central value for a dataset.
However, it is important to be aware of the properties of arithmetic mean and its limitations, especially when dealing with data containing outliers.

FAQs about Arithmetic Mean Properties

  1. Explain how the sum of deviation from the mean is zero.
  2. What is meant by unique arithmetic mean for a data set?
  3. What is arithmetic mean?
  4. How combined mean of different data sets can be computed, explain.
  5. Elaborate Sum of Squared Deviation is minimum?
  6. What is the impact of outliers on arithmetic mean?
  7. How does a change of scale and origin change the arithmetic mean?
Statistics Help https://itfeature.com

https://rfaqs.com

https://gmstat.com

Testing a Claim about a Mean Using a Large Sample: Secrets

In this post, we will learn about “Testing a claim about a Mean” using a Large sample. Before going to the main topic, we need to understand some related basics.

Hypothesis Testing

When a hypothesis test involves a claim about a population parameter (in our case mean/average), we draw a representative sample from the target population and compute the sample mean to test the claim about population. If the sample drawn is large enough ($n\ge 30$), then the Central Limit Theorem (CLT) applies, and the distribution of the sample mean is assumed to be approximately normal, that is we have $\mu_{\overline{x}} = \mu$ and $\sigma_{\overline{x}} = \frac{\sigma}{\sqrt{n}} \approx \frac{s}{\sqrt{c}}$.

Hypothesis Testing: Testing a Claim about a Mean Using a Large Sample

Testing a Claim about a Mean

It is worth noting that $s$ and $n$ are known from the sample data, and we have a good estimate of $\sigma_{\overline{x}}$ but the population mean $\mu$ is not known to us. The $\mu$ is the parameter that we are testing a claim about a mean. To have a value for $\mu$, we will always assume that the null hypothesis is true in any hypothesis test.

It is also worth noting that the null hypothesis must be of one of the following types:

  • $H_0:\mu = \mu_o$
  • $H_0:\mu \ge \mu_0$
  • $H_0:\mu \le \mu_0$

where $\mu_0$ is a constant, and we will always assume that the purpose of our test is that $\mu=mu_0$.

Standardized Test Statistic

To determine whether to reject or not reject the null hypothesis, we have two methods namely (i) a standardized value and (ii) a p-value. In both cases, it will be more convenient to convert the sample mean $\overline{x}$ to a Z-score called the standardized test statistic/score.

Since, we assumed that $\mu=\mu_0$, and we have $\mu_{\overline{x}} =\mu_0$, then the standardized statistic is:

$$Z = \frac{\overline{x} – \mu _{\overline{x}}} {\sigma_{\overline{x}} } = \frac{\overline{x} – \mu _{\overline{x}}} {\frac{s}{\sqrt{n}} }$$

As long as $\mu=\mu_0$ is assumed, the distribution standardized test statistics $Z$ is Standard Normal Distribution.

Example: Testing a Claim about an Average/ Mean

Suppose the average body temperature of a healthy person is less than the commonly accepted temperature of $98.6^{o}F$. Assume that a sample of 60 healthy persons is drawn. The average temperature of these 60 persons is $\overline{x}=98.2^oF$ and the sample standard deviation is $s=1.1^oF$.

The hypothesis of the above statement/claim would be

$H_0:\mu\ge 98.6$
$H_1:\mu < 98.6$

Note that from the alternative hypothesis, we have a left-tailed test with $\mu_0=98.6$.

Based on our sample data, the standardized test statistic is

\begin{align*}
Z &= \frac{\overline{x} – \mu _{\overline{x} } } {\frac{s}{\sqrt{n} } }\\
&=\frac{98.2 – 98.6}{\frac{1.1}{\sqrt{60}}} \approx -2.82
\end{align*}

Learn R Programming Language

Online Quiz Website

Random Variable in Statistics: A Quick Review Notes (2024)

Introduction to a Random Variable in Statistics

A random variable in statistics is a variable whose value depends on the outcome of a probability experiment. As in algebra, random variables are represented by letters such as $X$, $Y$, and $Z$. A random variable in statistics is a variable whose value is determined by chance. A random variable is a function that maps outcomes to numbers. Read more about random variables in Statistics: Random Variable.

Random Variable in Statistics: Some Examples

  • T = the number of tails when a coin is flipped 3 times.
  • s = the sum of the values showing when two dice are rolled.
  • h = the height of a woman chosen at random from a group.
  • V = the liquid volume of soda in a can marked 12 oz.
  • W = The weight of an infant chosen at random in a hospital.

Key Characteristics of a Random Variable

  • Randomness: The value of a random variable is determined by chance.
  • Numerical: It assigns numbers to outcomes.
  • Function: It is technically a function that maps outcomes to numbers.

Types of Random Variables

There are two basic types of random variables.

Discrete Random Variables: A discrete random variable can take on only a countable number of values. It can have a finite or countable number of possible values.

Continuous Random Variables: A continuous random variable Can take on any value within a specified interval. It can take on any value in some interval.

Examples of Discrete and Continuous Random Variables

• The variables $T$ and $s$ from above are discrete random variables
• The variables $h$, $V$, and $W$ from above are continuous random variables.

Random variable in statistics

Importance of Random Variables in Statistics

Random variables are fundamental to statistics. Random variables allow us to:

  • Use mathematical tools to analyze uncertain events.
  • Model the real-world phenomena.
  • Calculate probabilities of events.
  • Compute expected values and variances.
  • Make statistical inferences.

Random variables form the basis for probability distributions and are fundamental to statistical inference. Random variables provide a bridge between the real world of uncertainty and the mathematical world of probability.

Online Quiz Website

R Frequently Asked Questions