You are on page 1of 13

Standard Deviation

68% of measurements are found within one standard deviation σ from the mean μ; 95% are found within
1.96 times the standard deviation from the mean.

Example
A population of liver cells has p.H. with a mean of 5.61 and a standard deviation of 0.33. What are the
upper and lower limits between which 95% of the cells are found? If a cell has a p.H. of 6.10 is it unusual?
Solution
1) 95% of cells will be within 5.61 ± (1.96 x 0.33) = 5.61 ± 0.65 = between 4.96
and 6.26
2) A p.H. of 6.10 is well within these limits so it is not unusual.

The Estimate of the Standard Deviation: We divide by N -1 and not N because we set the mean to be in
the middle of the sample. It might not, in fact be there, and so using N tends to underestimate the
standard deviation. N -1 is called the degrees of freedom

The mean of a sample is likely to vary (but by less than individuals) because you may have picked by
chance samples with more high values of the measurement than average or more low values. The
variability of the estimate of the mean is easily calculated and is called the standard error of the
mean, SE, which is given by the expression: SE = s /√N

The Effect of Sample Size


If you gradually increase your sample size the mean and standard deviation will both gradually level
off, converging on the population mean and standard deviation. Therefore the standard error will
gradually fall.
This is an estimate of the variability of the population you have sampled. When giving results about the
characteristics of a population you are studying always present them as x (s). E.g. the grass had a yield
of 3.43 (1.24) tones -1.
This is an estimate of the likely error of the mean, x, you calculated. It is NOT a characteristic of the
population, but only of the sample you took. Use this to show how much error there will be in the means
you have calculated and present them as x (SE).

Graphing data with error bars. (a) The mean yield of two species of grass with error bars showing their
standard deviation; this emphasises the high degree of variability in each grass, and the fact that the
distributions overlap a good deal. (b) Standard error bars emphasise whether or not the two means are
different; here the error bars do not overlap, suggesting that the means might be significantly different.
However, from the population we have sampled we are not sure of the standard error, SE. All we have is
the estimate, SE, which might be larger or smaller than the real value.

The t Distribution
For this reason the mean has a broader distribution relative to SE called the t distribution. This is
especially wide when the sample size (and hence the number of degrees of freedom, N -1) is low.
However, it approaches the normal distribution when the sample size is high and hence when the
estimated standard error SE approaches that of the population SE.
Normal distribution and t distribution. The distribution of sample means x relative to the estimate of the
standard error SE calculated from samples with 1, 10 and infinite degrees of freedom. With infinite
degrees of freedom the distribution equals the normal distribution. However, it becomes more spread out
as the sample size decreases (fewer degrees of freedom) because the estimate of standard error
becomes less reliable.

Confidence Intervals Of a Known Population


95% of samples of a known population have means within 1.96 x SE of the mean, μ.
From a Sample: 95% of times the sample mean, x, will be within tcrit x SE of the real mean, μ.
Where tcrit is read from the t table for N - 1 degrees of freedom.
Note that as the sample size increases and the estimate for standard error, SE, gets more and more
accurate, tcrit falls and gets closer and closer to 1.96.
Confidence Intervals are written as x ( ±tcrit x SE)
Example 1
1) In the data on sodium ion concentration, what are the 95% confidence intervals for the mean? 2) Is
the real mean likely to be different from 1.01?
Solution
1) 95% Confidence Intervals = x ( ±tcrit x SE)
= 0.86 ( ±t8 x 0.037)
= 0.86 ( ±2.306 x 0.037)
= 0.86 ± 0.085
= 0.76 to 0.94.
2) The 95% confidence intervals for the real mean are below 1.01, suggesting the real mean is likely to
be less than 1.01.

T TESTS
You may want to ask several different sorts of questions about the measurements you have taken.
Is the mean different from an expected value?
Are the means of two measurements made on the same sample different from each other?
Are the means of two separate samples different from each other?
Is the mean of an experimentally treated group different from that of the control?
To answer these questions you need to carry out statistical tests for differences - t tests.
The question is how different must the sample mean be from a particular value, E, to make it unlikely
that the population mean is E?

Figure 1
There's a high probability of getting a mean, x, at least 1 x SE away from the real mean, μ.
There's a very low probability of getting a mean, x, at least 3 x SE away from the real mean, μ.

The Logic of Statistical Testing


Because of this, statistical testing involves rather strange pattern of logic.
You set up a hypothesis that nothing interesting has really occurred; there are no real differences or
associations. (Any that seem to have occurred have done so just because of chance fluctuations). This is
known as a null hypothesis.
You calculate a test statistic from your data that measures the size of the difference or association
relative to the variability in your samples. Using statistical tables, you calculate the probability of getting
such a test statistic if the null hypothesis were true.
You reject the null hypothesis if the probability of getting the test statistic is less than or equal to 0.05 or
5%.
Thus you decide whether something interesting has happened if it's unlikely that nothing interesting
has happened!

The One Sample t Test


Purpose: To test whether the mean value of a measurement is different from an expected value, E.
Rationale: You calculate how many standard errors (SE) your actual mean (x) is from the expected
mean. The further away it is the less likely this is to have happened by chance.
Carrying out the Test: Like all statistical tests there are four stages
Construct a Null Hypothesis: The population mean is NOT different from the expected value, E.
Calculate the test statistic, t.

Compare the absolute value of t, |t| , with the critical value for 5% with N - 1 degrees of freedom. (N =
sample size).
If |t| < tcrit , then there is a greater than 5% probability of this happening by chance. Therefore you have
no evidence to reject the null hypothesis. However if |t| ≥ tcrit there is a less than 5% probability of this
happening by chance. Meaning that you do have evidence to reject the null hypothesis and you can say
the mean is significantly different from the expected mean, E.

Example
In our data for the sodium ion concentration of a sample of 9 cells we found that the mean, x = 0.86 M,
standard deviation, s = 0.11 M, and standard error, SE = 0.037 M.
A textbook says that the concentration in these cells is 0.93 M. Do your cells have a significantly different
mean sodium ion concentration than expected?
Solution
The null hypothesis is that the sodium ion concentration is really the expected value, 0.93.
It is easy to calculate t

Looking in the table for critical values of t, tcrit for 5% at (9 - 1 = 8) degrees of freedom = 2.306
4) 1.89 < 2.306. Therefore there is a greater than 5% probability of this happening by chance. We have
no evidence to reject the null hypothesis. Our cells don't have a significantly different mean sodium
concentration than expected.

The Paired t Test


Purpose: to test whether the means of two sets of paired measurements are different from each other.
Examples: Patients before and after being given drugs, plants during the day and night, peoples'
reaction times with their right and left hand, IQ of married couples.
Rationale: the test has two stages.Calculate the difference, di, between each pair of measurements (eg
before - after).
Use a one sample t test to determine Name Rate Before Rate After Difference
if the mean difference, d, is different
from zero. (There are N -1 degrees of Colin 65 74 9
freedom where n = the number of
pairs of measurements) David 71 78 7

Mark 78 82 4
Example
The Effect of Coffee on Heart Rate: Brian 60 59 -1
1) The Calculation of Mean Difference
Mean Difference, d = 4.2 Paul 45 50 5
Standard Deviation of difference, sd =
3.0 Angus 55 59 4
Standard Error of difference SEd =
1.01 Richard 96 97 1

Peter 51 54 3

Patrick 64 70 6
It looks like people have on average a higher heart rate after drinking coffee, but is this a significant
difference?
Now for the t test itself
2) The t test
The null hypothesis is that the mean difference before and after taking coffee is zero
Calculate t:

3) There are 9 people, so there are 8 degrees of freedom. The critical value of t for 5% at 8 degrees of
freedom t8 = 2.306.
4) 4.16 > 2.306. Therefore there is a less than 5% probability of this happening by chance.
We can say, therefore, that drinking coffee significantly alters heart rate (it increases it, actually).
Testing for Differences between two Groups
In Biology it is much more common to have to see if two groups have measurements with different mean
values.

Examples
Control organisms and experimentally treated ones, males and females of the same species, ntracellular
and extracellular fluid, cells from two different parts of the body. This is difficult because both samples
vary.

To tell what the cut-off is we perform a 2 sample t test. The Two Sample t Test
Purpose: to test whether the mean values of a measurement made on two separate samples are
different from each other.
Rationale: you calculate how different the means of the two samples are.
You compare that with a measure of how variable the difference between the means of the two samples
is.

Examining your Data by Eye


Look at Graphs showing the means ± SE for each sample

a) If the error bars do overlap, the means are not significantly different.
b) If the error bars do not overlap, the means may be significantly different.

Carrying out the Two Sample t Test:


(√[(SE1)2 + (SE2)2] is called the Standard Error of Difference, SEd)
Compare the absolute value of t, |t| , with the critical value for 5% tcrit then there is a greater than 5%
probability of this happening by chance, which means that you have no evidence to reject the null
hypothesis. If |t| > tcrit then there is a less than 5% probability of this happening by chance and you do
have evidence to reject the null hypothesis and you can say the two means are significantly different
from each other.

Example:
Comparing the Longevity of Control and Irradiated Worms
Samples of 16 worms were taken from each group. The results are summarized below.
Control Irradiated

Mean 21.3 days 20.5

Standard deviation 3.4 3.7

Standard Error 0.85 0.93


Does irradiation significantly alter longevity?

Solution
1) The null hypothesis is that there is no difference between the longevity of the two groups.

2)
3) There are 16 worms in each group, so there are (16 + 16 - 2 = 30) degrees of freedom. The critical
value of t for 5% at 30 degrees of freedom t30 = 2.042.
4) 0.63 < 2.042. Therefore there is a greater than 5% probability of this happening by chance.
We have no evidence to reject the null hypothesis. Irradiated and control worms do not have
significantly different longevity.

Types of association
If you take 2 measurements on a single sample you can investigate two different things.
1) Whether the mean of the two measurements is different
eg Is mean heart rate different before and after drinking coffee?
We have seen you can use the paired t test to do this.
2)Whether the two measurements are associated
eg. Do people with high blood pressure also have high heart rates?

Looking for Associations


The most obvious way to investigate whether there is an association between two measurements is to
plot one against the other on a Scatter Plot
Figure 5.2: Ways in which variables can be related. (a) No association. (b) Positive association. (c)
Negative association. (d) A complex curvilinear association.

Limitations to Visual Examination


You might get an apparent association just by chance by accidentally choosing points that seem to
follow a straight line.

In a small sample (a), it's quite likely you can get points that seem to show an association.
In a large sample (b), it's very unlikely you would choose points that all fitted along a straight line.
To determine whether an association is likely to be real, therefore, you need to carry out a statistical
test.
Stage one is to work out the correlation coefficient.

The Correlation Coefficient


The correlation coefficient, r, is a single measure that describes the strength and direction of an
association. It goes from -1 (perfect negative association) to +1 (perfect positive association).
Perfect negative

Association (r = -1)
Perfect Positive

Association (r = 1)
No Association (r = 0)
The further r is from 0, and the larger the sample size, the less likely is the association due to chance.

Calculating the Correlation Coefficient


1) Calculate the means of x and y, x and y.
2) Add up the sum of all (x - x)(y - y) and divide by a scaling factor

Why does this work?


If there is a positive association all points will be either above and to the right of the centre of the
distribution or below and to the left.

When (x - x) is positive, so is (y - y).


When (x - x) is negative, so is (y - y.
Therefore all the values of (x - x)(y - y) will be positive.
If there is a negative association all points will be either above and to the left of the centre of the
distribution or below and to the right.

When (x - x) is positive, (y - y) will be negative.


When (x - x) is negative, (y - y) will be positive.
Therefore all the values of (x - x)(y - y) will be negative.
If there is no association, points will be randomly scattered all around the centre of the distribution.
Positive and negative values of (x - x)(y - y) will cancel each other out.

Carrying out a Test


1) Construct a Null Hypothesis (There is no association between the two sets of measurements)
2) Calculate the correlation coefficient, r
3) Compare the absolute value of r, |r| , with the critical value for 5% significance with N - 2 degrees of
freedom, (N = the number of pairs of measurements.
4) If |r| < rcrit therefore there is a greater than 5% probability of this happening by chance. You have no
evidence to reject the null hypothesis. If |r| ≥ rcrit therefore there is a less than 5% probability of this
happening by chance. You do have evidence to reject the null hypothesis. You can say that there is a
significant correlation between the two measurements.

Example
An investigation seeks to find if brain size of monkeys is related to the size of the gut. The following
results are obtained.
Brain Mass Gut Mass
Species
(% body mass) (% body mass)

1 6 25

2 8 20

3 9 18

4 10 13

5 12 14
Is there any association between the two measurements?

Solution
The first stage is to calculate x and y.
Mean brain mass = 9%
Mean gut mass = 18%
Calculating r
Σ(x - x)(y - y) = (-3 x 7) + (-1 x 2) + (0 x 0) + (1 x -5) + (3 x -4)
= -21 - 2 - 5 - 12
= -40
Σ(x - x)2 = (-3)2 + (-1)2 + (0)2 + (1)2 + (3)2
=9+1+1+9
= 20
Σ(y - y)2 = (7)2 + (2)2 + (0)2 + (-5)2 + (-4)2
= 49 + 4 + 25 + 16
= 94
Therefore

= -40/43.36 = -0.922

Carrying Out the Test


1) The null hypothesis is that there is no association between brain mass and gut mass.
2) Calculate r = -0.922
3) There are 5 pairs of measurements, so there are (5 -2 = 3) degrees of freedom. The critical value of r
for 5% at 3 degrees of freedom r3 = 0.879.
4) 0.922 > 0.879. Therefore there is a less than 5% probability of this happening by chance. We can
reject the null hypothesis
There is a significant association between brain and gut mass. In fact they show negative correlation.
As brain size goes up gut mass goes down, and vice-versa.

Cautions about using Correlation


1) You cannot use correlation analysis if one measurement is independent (ie could not be altered by
the other).

Examples
a) If you are measuring cell mass at different times.
b) If you are controlling one variable yourself in an experiment (eg if you are measuring metabolic rate of
cells at certain temperatures).
2) The fact that there is a correlation does not necessarily imply a causal relationship between the two
sets of measurements.
eg In our example brain mass does not control gut mass or vice-versa. In fact the negative correlation is
related to the diet of different monkeys. Leaf-eating monkeys have larger guts to digest the leaves, but
do not need as big brains as fruit-eating ones because leaves are more plentiful and easier to find than
fruit.
3) The size of the correlation coefficient does not reflect the slope of the relationship, just its closeness.
4) Correlation is only useful for linear relationships, so always look at graphs of your results, first.
CHI Square Tests

Categorical Data: There are many characteristics of organisms that cannot be measured to give numbers
(eg 3.41). Instead the characteristics can only be put into categories.
Eg. Gender (Male/Female), Infection (Present/ Absent), Colour (Red, Yellow, etc), Species (Thrush,
Dunnock, etc)
These can only be quantified by counting the frequency with which each character state turns up. We
might then ask two sorts of questions.
Are the frequencies of particular characteristics different from what we might expect?
Are certain characteristics associated with one another?

The Problem of Sampling


You will only very rarely pick out the same proportion of the different characteristics in your sample as
are present in the population. e.g. If you take a sample of size N from a population with half male and
half female, the numbers of females you pick will vary like a normal distribution curve (bell). This is why
you have to carry out statistical tests to determine if differences or associations are real or could easily
have happened by chance. The test used is the X2 (Chi-squared) test.

The X2 Test for Differences


Purpose: To test whether character frequencies are different from expected values. e.g. Are sex ratios
different from the expected 50:50? Do rats turn more towards or away from a stimulus in a maze?
(expected value towards: away = 50:50) Is a disease more common in a small town than in the nation as
a whole? (expected value = national figure) Are ratios of dominants: recessives different from the
expected 3:1 ratio?

Rationale
To work this out you work out the X2 statistic. This is found by the formula

O = Observed Value
E = Expected Value
The further your values are from the expected ones the bigger is X2 and the less likely the results are to
have happened just by chance.

Carrying out a Test


1) Construct a Null Hypothesis
The frequencies of the character states in the population are equal to the expected frequencies
2) Calculate the value of X2 where

3) Compare X2 with the critical value for 5% significance with N - 1 degrees of freedom, (N = the number
of groups)
4) If X2 < X2crit therefore there is a greater than 5% probability of this happening by chance. You have no
evidence to reject the null hypothesis. If X 2 ≥ X2crit therefore there is a less than 5% probability of this
happening by chance. You do have evidence to reject the null hypothesis. You can say that the
frequencies are significantly different from the expected values.

Example
The incidence of malaria in a region where a new insecticide had been applied was 320 people out of a
population of 15,000. This compares with an incidence in the country as a whole of 3%. Has the
insecticide made a significant difference on malarial infection?
Solution
1) The null hypothesis is that the region has the same incidence of malaria as the country as a whole.
2) The first stage in working out X2 is to calculate the expected numbers of cases. Expected no. cases =
15,000 x 0.03 = 450
(Expected no. of fit people = 15,000 - 450 = 14,550

= (-130)2/450 + 1302/14,450
= 37.56 + 1.17 = 38.73
3) There are two groups (fit and ill) therefore 2 -1 degrees of freedom. X21 = 3.84.
4) 38.73 > 3.84 therefore there is a significant difference in distribution. In fact there are fewer cases
than expected

Warnings about Using X2 Tests


1) Never use percentages. Always use the actual frequencies. This is because bigger samples are less
likely to differ by the same proportion from the expected value. E.g. you are quite likely to throw 2
heads: 0 tails.You are very unlikely to throw 100 heads: 0 tails
2) Bigger sample sizes allow you to spot smaller differences (because of the squared term on top).
3) Cases for which there is an expected outcome are very few and far between. eg If you examine the
numbers of insects on flowers there is no reason to expect any particular ratio. Therefore there is another
sort of test, the X2 test for associations

The X2 Test for Associations


Purpose: To test whether two sets of character states are associated. E.g. Are particular species
associated with particular habitats?
Are particular diseases associated with particular racial groups? Are infections associated with particular
groups of cells?
Rationale: The test determines whether the distribution of character states is different from what it would
be if they were randomly distributed around the population.

Carrying out a Test, an example


This is complex, so we will examine and run through an example. In a study of tail loss by lizards, it was
found that of 60 females found, 18 had lost their tails, while 42 still had them. Of the 30 males, 12 had
lost their tails while 18 still had them.
It looks as if tail loss is more common in males, but is this a significant association?
1) Construct a Null Hypothesis: there is no association between the character states. Here, therefore the
frequencies of tails loss in males and females are the same
2) Calculate the value of X2. To calculate the expected frequencies, first arrange your data in a table.
Minus Tail Plus Tail Total
Males 12 18 30
Females 18 42 60
Total 30 60 90
Now work out how many lizards should be in each category if tail loss was equally distributed between
the sexes. Here 1/3 of all lizards are male and 1/3 of all lizards have lost their tail. Therefore 1/3 x 1/3
should be both male and have lost their tail = 1/9
= 10 lizards.
The general expression to calculate expected values, E, is

And put the results (in brackets) into the Contingency Table:

Minus Tail Plus Tail Total


Males 12 (10) 18 (20) 30
Females 18 (20) 42 (40) 60
Total 30 60 90
c) Finally, calculate X2

= 0.4 + 0.2 + 0.2 + 0.1 = 0.7


3) Compare X2 with the critical value for 5% significance with (C - 1)(R - 1) degrees of freedom, (C and R
are the numbers of columns and rows) Here there are (2 - 1)(2 - 1) = 1 degrees of freedom
X21 = 3.84.
4) If X2 < X2crit therefore there is a greater than 5% probability of this happening by chance. You have no
evidence to reject the null hypothesis. If X 2 ≥ X2crit therefore there is a less than 5% probability of this
happening by chance. Here 0.7 < 3.84, so we have no evidence to reject the null hypothesis. Despite
what it looks like females are not significantly more likely to have intact tails.

Warnings about Using X2 Tests


1)Once again Never use percentages. Always use the actual frequencies.
2) Expected totals should all be greater than 5, or else the test becomes less reliable. Therefore use as
large samples as you can.

ANOVA

Often you might want to carry out rather more complex surveys or experiments. You might want to
compare 2 experimentally treated groups of rats with a control group. You might want to compare
people who have been put on 4 different drug regimes.
You might want to compare 5 strains of bacteria. You might think you could carry out pairwise t tests, eg
1 vs 2, 1 vs 3, 2 vs 3
But there are two problems, you would need to do lots of tests

Number of Groups 3 4 5 6 7 8 9 10

Number of Tests 3 6 10 16 23 31 40 50

Each time you perform a test there is a 5% probability of getting a significant answer by chance alone.
Therefore the more tests you do the more likely you are to get a wrong answer.
For this reason you must use a different test: Analysis of Variance or ANOVA

How ANOVA Works


The t test compares the differences between 2 means with the pooled standard error of the means.
ANOVA compares the amount of variability caused by the means being different with the amount of
variability caused by points within each group being different from the mean.
Take the simplest case of the weights of 2 group of fish (a) below.

The total variability (b) = The


sum of squares of distances of
points to the overall mean.
The between group
variability (c) = The sum of
squares of distances of a
point's group mean to the
overall mean.
The within group variability
(d) = The sum of squares of
distances of points to their
group's mean.

If between group variability is much greater than


within group variability the two groups are
probably different.
Example:
In a) the between group variability is much greater
than the within group variability. There is a
significant difference between the groups.

In b) the between group variability is much less


than the within group variability. There is no
significant difference between the groups.

Working out the Mathematics


Doing the maths is actually rather complex, and involves several stages.
1) Calculating the "Sums of Squares"
a) The Total "Sum of Squares", SSt SSt = Σ so2
Where so is the distance of each point to the overall mean.
b) The Within Sample "Sum of Squares", SSw SSw =Σ sw2
Where sw is the distance of each point to its group mean.
c) The Between Sample "Sum of Squares", SSB SSB = SSt - SSw
We cannot compare these SS directly because they are the result of adding different numbers of points.

2) Calculating the Variance or Mean Squares. This is done by dividing by the correct numbers of
degrees of freedom.
a) Within Sample "Mean Square", MSw
Where n = total number of points. N = number of groups
b) Between Sample "Mean Square", MSB

3) Calculating the ANOVA test statistic, F.


The larger F is, the smaller the probability is of getting the results by chance if there were no
difference between the groups
Fortunately you don't have to work out F yourselves; it can all be done by SPSS.

Carrying Out an ANOVA Test


Construct a Null Hypothesis. Here the null hypothesis is that there is no difference between the groups.
Enter your data in SPSS and use it to calculate the tests statistic, F. If Sig. > 0.05 therefore there is a
greater than 5% probability of this happening by chance. You have no evidence to reject the null
hypothesis. If Sig. ≤ 0.05 therefore there is a less than 5% probability of this happening by chance. You
do have evidence to reject the null hypothesis

Working out Which Groups are Different


That's all fine but it doesn't tell you which groups are different from each other. To work this out,
fortunately, you can use one of several Post Hoc tests - But only if the ANOVA shows a significant
difference between groups. The Tukey and Scheffe tests compare each group with all the others.
Dunnett's test compares a control group with each of the others. In our example there is no clear
control, so we should use the Tukey test.

Regression

The bjectives of this Learning module are to test whether an apparent linear relationship between two
variables is real, or whether it could have happened by chance because of variability. To do this one must
use “Regression Analysis”. In an earlier unit we saw how correlation allows us to test for relationships
between two measurements taken on the same items, both of which might be dependent on the other,
eg Heart rate and blood pressure. However, there are many cases when one measurement is clearly
independent of the other one, Eg Age and Mass of people: age is clearly independent of mass. Time and
pH of cells: you look at the pH of cells at particular times. Reaction rate and temperature: you control the
temperature and this affects the rate of reaction.
You can't investigate these cases using correlation. Instead you must plot the data correctly and use
regression analysis.

Plotting the Data


You must plot the dependent variable on the y axis and the independent variable on the x axis. For
instance the graph below shows how the mass of eggs depends on their age.

This will help you to see how the dependent


What Regression
variableDoes
is affected by the independent variable.
Regression analysis calculates the "line of best fit" through the data points.
It does it by working out the line, y = a + bx, which minimises the sum of
the squares of the distances, si2, of each point to the line.The slope, b, of
the line is given by the equation

a is then calculated by putting the values of x, y and b into the equation


y = a + bx.
The Problem
You could get the same equation just by chance even though the independent variable did not affect the
dependent one, due to scatter.
a) In this case there is significant regression: Σsi2 is
low
b) Here the regression is not significant: Σsi2 is high

Working out if Regressions are Significant


To find if a regression line meaningfully represents your data, and explains much of the variability, you
can use SPSS. It tests
If the slope of the line, b, is significantly different from 0
If the intercept of the line, a, is significantly different from 0

Performing Your Own t Tests


You can also carry out your own t tests from the results of SPSS, to work out if the slope or intercept are
different from any particular values, eg To test if eggs are significantly lighter than 90 g when they are
laid (ie that the intercept < 90).
The null hypothesis is that the eggs do weigh 90 g when laid.
Calculate t using the equation

Compare t with the critical value for N - 2 degrees of freedom, where N = number of points. Here tcrit =
2.069
0.247 < 2.069, so the difference is not significant. Initial weight is not significantly different from 90 g

You might also like