Reliability
Reliability
1. Random error: A source of error in measuring Most appropriate for variables that should be
a targeted variable caused by unpredictable stable over time (e.g., personality) and not
fluctuations and inconsistencies of other appropriate for variables expected to change
variables in the measurement process (i.e., over time (e.g., mood)
noise) As time passes, correlation between the
2. Systematic error: A source of error in scores obtained on each testing decreases
measuring a variable that is typically constant With intervals greater than 6 months, the
or proportionate to what is presumed to be estimate of test-retest reliability is called the
the true value of the variable being measured. coefficient of stability.
Parallel-forms and alternate-forms Average proportional distance (APD): Focuses
on the degree of difference between scores
Coefficient of equivalence: The degree of the on test items; it involves averaging the
relationship between various forms of a test. difference between scores on all of the items,
Parallel forms: For each form of the test, the dividing by the number of response options on
means and the variances of observed test the test, and then subtracting by 1.
scores are equal.
Alternate forms: Different versions of a test MEASURES OF INTER-SCORER RELIABILITY
that have been constructed so as to be
parallel; they do not meet the strict Inter-scorer reliability: The degree of
requirements of parallel forms but item agreement or consistency between two or
content and difficulty are similar between more scorers (or judges or raters) with regard
tests. to a particular measure.
Reliability is checked by administering two It is often used with behavioral measures
forms of a test to the same group; scores may Guards against biases in scoring
be affected by error related to the state of Coefficient of inter-scorer reliability: The
testtakers (e.g.,practice, fatigue, etc.) or item scores from different raters are correlated
sampling. with one another.