3 - Types of Reliability
3 - Types of Reliability
Definition of Reliability
the probability that a system, component, or process will perform
its intended function without failure for a specified period under
stated conditions.
Observer 1
Inter-Rater or Inter-Observer Reliability
Object or
phenomenon
?
=
Observer 1 Observer 2
Inter-Rater or Inter-Observer Reliability
Pilot Study - A pilot study can be defined as a 'small study
to test research protocols, data collection instruments,
sample recruitment strategies, and other research techniques
in preparation for a larger study.
Test = Test
Time 1 Time 2
Test Re-test Reliability
Stability over time
Test = Test
Time 1 Time 2
Test Re-test Reliability
Limitations of this approach:
1. Memory Effect / Carry Over Effect
Particularly holds true when the two administration takes
place within short span of time.
2. Practice Effect
Happens when repeated tests are being taken for the
improvement of test scores, as is typically seen in the case
of classical IQ where there is improvement in the scores.
3. Absence
People remaining absent for re-tests
Parallel Forms Reliability
Known by the various names such as Alternate Forms
Reliability, Equivalent Form Reliability, and Comparable
Form Reliability.
=
Form B
Time 1 Time 2
Parallel Forms Reliability
Form A
Stability across forms
=
Form B
Time 1 Time 2
Parallel Forms Reliability
Limitations of this approach:
1. Difficulty in Creating Truly Equivalent Forms
Designing two or more tests that are truly equivalent in
content, difficulty, and format is challenging. Even small
differences between forms can affect the results, leading to
lower reliability.
2. Practice Effect
If the two forms are administered close in time, test-takers
might remember content from the first form, which could
influence their performance on the second form, potentially
inflating the reliability.
Internal Consistency Reliability
A measure of how consistently different items on a test or
survey assess the same construct or concept.
Key Concepts
1. Homogeneity of Items
2. Correlations Among Items
Internal Consistency Reliability
Measured in:
1. Average Inter-item Correlation
2. Average Item-Total Correlation
3. Split-Half Reliability
4. Cronbach’s Alpha
Internal Consistency Reliability
Average Inter-item Correlation
Average inter-item correlation assesses the degree of
correlation between each pair of items within a test or
survey that is designed to measure the same construct.
Item 1 Average Inter-item Correlation
Item 2
Item 3
Test
Item 4
Item 5
Item 6
Item 1 Average Inter-item Correlation
I1 I2 I3 I4 I5 I6
Item 2
I1 1.00
Item 3 I2 .89 1.00
Test I3 .91 .92 1.00
I4 .88 .93 .95 1.00
Item 4 I5 .84 .86 .92 .85 1.00
I6 .88 .91 .95 .87 .85 1.00
Item 5
Item 6
.90
Internal Consistency Reliability
Average Item-total Correlation
Measures the correlation of each individual item with the
total test score (excluding the item itself). It helps
determine whether each item is contributing effectively to
the overall measurement.
Average Item-total Correlation
Item 1
Item 2
Item 3
Test
Item 4
Item 5
Item 6
Average Item-total Correlation
Item 1
I1 I2 I3 I4 I5 I6
Item 2 I1 1.00
I2 .89 1.00
Item 3 I3 .91 .92 1.00
Test I4 .88 .93 .95 1.00
I5 .84 .86 .92 .85 1.00
Item 4 I6 .88 .91 .95 .87 .85 1.00
Total .84 .88 .86 .87 .83 .82 1.00
Item 5
Item 6
.85
Internal Consistency Reliability
Split-Half Reliability
Tests internal consistency by dividing the test into two
halves and checking how well the scores from each half
correlate with each other. It provides an estimate of
reliability for the full test.
Split-half Reliability
Item 1
Item 2
Item 3
Test
Item 4
Item 5
Item 6
Split-half Reliability
Item 1
Item 3
Test
Item 4
Item 5
Item 2 Item 5 Item 6
Item 6
Split-half Reliability
Item 1
Item 3
Test
Item 4 .87
Item 5
Item 2 Item 5 Item 6
Item 6
Internal Consistency Reliability
Cronbach’s Alpha
Provides an overall measure of internal consistency across
all items in a test. It combines item variance and total
test variance into a single coefficient to assess
reliability.
Item 2