Lecture 2 Chapter 2
Lecture 2 Chapter 2
Chapter 2
Lecture 2
• An engineer is studying the formulation of a Portland cement mortar.
• He has added a polymer latex emulsion during mixing to determine if this
impacts the curing time and tension bond strength of the mortar.
• The experimenter prepared 10 samples of the original formulation and 10
samples of the modified formulation.
• We will refer to the two different formulations as two treatments or as two
levels of the factor formulations.
• When the cure process was completed, the experimenter did find a very
large reduction in the cure time for the modified mortar formulation.
• Then he began to address the tension bond strength of the mortar. If the
new mortar formulation has an adverse effect on bond strength, this could
impact its usefulness.
• The tension bond strength data from
this experiment are shown in Table
2.1 and plotted in Figure 2.1.
• However, it is not obvious that this difference is large enough to imply
that the two formulations really are different.
• A technique of statistical inference called hypothesis testing can be
used to assist the experimenter in comparing these two formulations.
• Hypothesis testing allows the comparison of the two formulations to
be made on objective terms, with knowledge of the risks associated
with reaching the wrong conclusion.
Basic Statistical Concepts
• Each of the observations in the Portland cement experiment described above would be called a run.
• Notice that the individual runs differ, so there is fluctuation, or noise, in the observed bond strengths.
• This noise is usually called experimental error or simply error.
• It is a statistical error, meaning that it arises from variation that is uncontrolled and generally
unavoidable.
• The presence of error or noise implies that the response variable, tension bond strength, is a random
variable.
• A random variable may be either discrete or continuous. If the set of all possible values of the random
variable is either finite or countably infinite, then the random variable is discrete, whereas if the set of all
possible values of the random variable is an interval, then the random variable is continuous.
• For example the number of cars parked in a lot at any given time could be 0, 1, 2, 3, ..., up to the total
capacity of the lot. This is countable and thus a discrete random variable.
• The speed of a car on a highway can take any value within a range, say from 0 km/h to 120 km/h.
Between any two speeds (e.g., 80 km/h and 81 km/h), there are infinitely many possible values. Hence, it
is a continuous random variable.
Graphical Description of Variability
• Where z represents the standard score (also known as a z-score), which measures how many
standard deviations a particular value is from the mean.
• The formula for converting any normally distributed variable y into a standard
normal variable Z is:
• If the null hypothesis is true (the variances are equal), the ratio of the
variances should be close to 1. If the null hypothesis is false (the
variances are different), the ratio will deviate significantly from 1.
Inferences About the Differences in
Means, Randomized Designs
• Hypothesis Testing
• We now reconsider the Portland cement
experiment introduced in Section 2.1. Recall
that we are interested in comparing the
strength of two different formulations: an
unmodified mortar and a modified mortar.
• In general, we can think of these two
formulations as two levels of the factor
“formulations.” Let y11, y12, . . . , represent the
n1 observations from the first factor level and
y21, y22, . . . , represent the n2 observations
from the second factor level.
• We assume that the samples are drawn at
random from two independent normal
populations. Figure 2.9 illustrates the
situation.
Inferences About the Differences in
Means, Randomized Designs
• Hypothesis Testing
• Statistical hypothesis is a statement either about the parameters of a
probability distribution or the parameters of a model. The hypothesis reflects
some conjecture about the problem situation.
• For example, in the Portland cement experiment, we may think that the mean
tension bond strengths of the two mortar formulations are equal.
• where is the mean tension bond strength of the modified mortar and is the
mean tension bond strength of the unmodified mortar.
• The statement : is called the null hypothesis and : is called the alternative
hypothesis. The alternative hypothesis specified here is called a two-sided
alternative hypothesis because it would be true if . Or if
Minitab
Choice of Sample Size
• Selection of an appropriate sample size is one of the most
important parts of any experimental design problem for
ensuring that the study has sufficient power to detect a
statistically significant difference between two groups.
• The sample size needed depends on several factors,
including the expected effect size, the desired level of
power, the significance level (alpha), and the variability in
the data. Steps are:
• Determine Effect Size
• The effect size is the magnitude of the difference you expect to detect
between the two groups. It can be calculated as:
• Set the Desired Power (1 - β)
• Power refers to the probability of correctly rejecting the null hypothesis (i.e., detecting
a true difference when one exists).
• A commonly used value is 0.80 (80%), meaning there is an 80% chance of detecting an
effect if it exists.
• For more rigorous studies, 0.90 or 0.95 (90% or 95%) power might be used.
• Higher power requires larger sample sizes.
• Set the Significance Level (α)
• The significance level (α) is the probability of rejecting the null hypothesis when it is
true (i.e., a Type I error). The most commonly used value is 0.05 (5%), meaning there is
a 5% chance of finding a false positive.
• More conservative significance levels, like 0.01, reduce the chance of false positives
but require larger sample sizes.
• Estimate the Standard Deviation (σ)
• If prior studies or pilot data are available, use them to estimate the pooled
standard deviation. Larger standard deviations imply more variability in the
data, which increases the required sample size.
• Use a Power Analysis Formula or Software
• We can calculate the sample size manually using a formula for a two-sample t-
test:
Inferences About the Differences in Means,
Paired Comparison Designs
• The Paired Comparison Problem
• Consider a hardness testing machine that presses a rod with a pointed tip into a metal
specimen with a known force. By measuring the depth of the depression caused by the
tip, the hardness of the specimen is determined. Two different tips are available for this
machine, and although the precision (variability) of the measurements made by the two
tips seems to be the same, it is suspected that one tip produces different mean hardness
readings than the other.
• An experiment could be performed as follows. A number of metal
specimens (e.g., 20) could be randomly selected. Half of these
specimens could be tested by tip 1 and the other half by tip 2. The
exact assignment of specimens to tips would be randomly
determined. Because this is a completely randomized design, the
average hardness of the two samples could be compared using the t-
test described in Section 2.4.
• A little reflection will reveal a serious disadvantage in the completely
randomized design for this problem. Suppose the metal specimens
were cut from different bar stock that were produced in different heats
or that were not exactly homogeneous in some other way that might
affect the hardness. This lack of homogeneity between specimens will
contribute to the variability of the hardness measurements and will
tend to inflate the experimental error, thus making a true difference
between tips harder to detect.
• To protect against this possibility, consider an alternative experimental
design. Assume that each specimen is large enough so that two
hardness determinations may be made on it. This alternative design
would consist of dividing each specimen into two parts, then randomly
assigning one tip to one-half of each specimen and the other tip to the
remaining half. The order in which the tips are tested for a particular
specimen would also be randomly selected. The experiment, when
performed according to this design with 10 specimens, produced the
Inferences About the Variances of Normal
Distributions
• In many experiments, we are interested in possible differences in the mean
response for two treatments.
• However, in some experiments it is the comparison of variability in the data that is
important.
• In the food and beverage industry, for example, it is important that the variability
of filling equipment be small so that all packages have close to the nominal net
weight or volume of content.
• In chemical laboratories, we may wish to compare the variability of two analytical
methods.