Validity Reliability.2
Validity Reliability.2
VALIDITY AND
RELIABILITY
QUALITIES OF GOOD TEST:
1.VALIDITY
2.RELIABILITY
WHAT IS TEST VALIDITY?
A measure is valid when it
measures what it is supposed
to measure.
• If the quarterly exam is valid, then the
contents should directly measure the
objectives of the curriculum.
Procedure:
The items are compared with the objectives of
the program. The items need to measure
directly the objectives (for achievement) or
definition (for scales). A reviewer conducts the
checking.
FACE VALIDITY
When the test is presented well, free from
errors, and administered well.
Procedure:
The test items and layout are reviewed and
tried out on a small group of respondents. A
manual for administration can be made as a
guide for the test administrator.
PREDICTIVE
VALIDITY
A measure should predict a
future criterion. An example is
an entrance exam predicting the
grades of the students after the
first semester.
CONSTRUCT VALIDITY
The components or factors of the test should
contain items that are strongly correlated.
Procedure:
To determine concurrent validity, you have to
measure the correlation of results from an
existing test and a new test and demonstrate
that the two give similar results.
CONVERGENT VALIDITY
When the components or factors of a test are
hypothesized to have a positive correlation.
• For a test to be considered of high quality, it
should demonstrate convergent validity by
showing strong correlations with other
measures that assess the same or related
constructs. This helps establish the test's
credibility and accuracy.
DIVERGENT VALIDITY
When the components or factors of a test
are hypothesized to have a negative
correlation.
• crucial for a good test as it confirms that
the test is distinct from measures of
unrelated constructs. A high-quality test
should show low correlations with measures
of constructs that are theoretically
unrelated.
Group activity: Analyze the given cases of each type of validity. Then answer the
following question below.
1. Content validity
A coordinator in science is checking the science test paper for grade 4. She asked the
grade 4 science teacher to submit the table of specifications containing the objectives of
the lesson and corresponding items. The coordinator checked whether each item is
aligned with the objectives.
The assistant principal browsed the test paper made by the math
teacher. She checked if the contents of the items are about
mathematics. She examined if instructions are clear. She browsed
through the items if the grammar is correct and if the vocabulary is
within the students’ level of understanding.
1. The number of items in a test- The more items a test has, the
likelihood of reliability is high. The probability of obtaining
consistent scores is high because of the large pool of items.
2. Individual differences of participants- every participant
possesses characteristics that affect their performance in a test,
such as fatigue, concentration, innate ability, perseverance, and
motivation. The individual factors change over time and affect the
consistency of the answer in a test.
3. External environment- the external environment may include
room temperature, noise level depth of instruction, exposure to
materials, and quality of instruction, which could affect changes n
the responses of examinees in a test
What are the different ways to
establish test reliability?
There are different ways in
determining the reliability of a test.
The specific kind of reliability will
depend on the
(1) variable you are measuring,
(2) type of test, and
(3) versions of the test.
The method of Testing Reliability
1. Test-retest
2. Parallel Forms
3. Internal Consistency
4. Inter-rater Reliability
1. Test-retest
The test needs to be administered at one time to a group
of examinees and administered it again at another time to
the “same group” of examinees.
Formula:
n – number of respondents
X – are scores of the 1st test
Y – are the scores of the 2nd test
XY – are the product of the 1st and 2nd test
Y²
RESPONDENTS
(N) X Y X² XY
1 50 51 2500 2601 2550
2 43 42 1849 1764 1806
3 48 48 2304 2304 2304
4 45 44 2025 1936 1980
5 40 41 1600 1681 1640
6 47 47 2209 2209 2209
7 52 51 2704 2601 2652
8 39 38 1521 1444 1482
9 44 43 1936 1849 1892
10 43 41 1849 1681 1763
TOTAL ⅀X =451 ⅀Y= 446 ⅀X²= 20497 ⅀Y²= 20070 ⅀XY= 20278
r
INTERPRETATION:
.70 to 1.00 VERY STRONG
.50 to .70 STRONG
.30 to .50 MODERATE
r 0 to .30 WEAK
0 NONE
Exercise:
Interpretation:
Complete the table an find the reliability of the two test. 0.70 - 1.00 very strong
X Y XY 0.50 - 0.70 strong
0.30 – 0.50 moderate
10 20 0 – 0.30 weak
9 15 0 none
6 12
10 18
12 19 FIND:
4 8 ∑X:
5 7 ∑Y:
7 10
∑:
16 17
∑:
8 13
∑XY:
2. Parallel Forms
• Compares two equivalent forms of a test that
measure the same attribute.
• Kuder-Richardson 20
Parallel Forms-Immediate 2 1
Measure of Equivalence Pearson r
Parallel Forms-Delayed
2 2
Kuder-Richardson
Kuder-Richardson
POSITIVE DISCRIMINATION
-happens when more students in the upper group got the item
correctly than those students in the lower group.
NEGATIVE DISCRIMINATION
-happens when more students in the lower group got the item correctly
than those students in the upper group.
ZERO DISCRIMINATION
Happens when the number of students in the upper and lower group
who answer the test correctly are equal.
Table of Non-plausibility of Distracters
1 14 1 14 1
2 13 0.93 11 0.79
3 14 1 11 0.79
4 14 1 7 0.5
6. Compute the Difficulty Index
Formula:
Item Difficulty =
1 14 1 14 1 1
2 13 0.93 11 0.79 0.86
3 14 1 11 0.79 0.90
DIFFICULTY INDEX REMARK
0.91 above Very Easy
0.76 to 0.90 Easy
0.26 to 0.75 Moderate or average
0.11 to 0.25 Difficult
0.10 below Very Difficult
Item Upper Lower Difficulty REMARKS
No. Group=14 Group=14 Index
F P F P
1 14 1 14 1 1 VERY EASY
2 13 0.93 11 0.79 0.86 EASY
3 14 1 11 0.79 0.90 EASY
7. Compute the Discrimination Index
Formula:
Item Discrimination = pH – pL
DISCRIMINATION REMARK
INDEX
0.40 and above Very good item
0.30 to 0.39 Good Item
0.20 to 0.29 Reasonably good item
0.10 to 0.19 Maginal item
Below 0.10 Poor item
Discrimination Index
Refers to the power of the item to
discriminate the students between
those scored high and low in the overall
test.
Formula:
Item Discrimination = pH – pL
Item Upper Lower Discrimination
Number Group=14 Group=14 Index
F P F P
1 14 1 14 1 0
2 13 0.93 11 0.79 0.14
3 14 1 11 0.79 0.21
4 14 1 7 0.5 0.5
DISCRIMINATION INDEX REMARK
1 14 1 14 1 0 LOW
2 13 0.93 11 0.79 0.14 LOW
3 14 1 11 0.79 0.21 MODERATE
4 14 1 7 0.5 0.5 HIGH
Negative
Discrimination
VERY
1 14 1 14 1 1 0 LOW REJECT
EASY