Chapter 3-Lesson1 Assessment Method
Chapter 3-Lesson1 Assessment Method
ASSESSMENT
METHOD
ASSESSMENT METHOD
There are many characteristics to consider in examining
and designing an appropriate assessment method that
will reflect the established learning goals and activities
which includes: validity, reliability, practically,
comprehensive, relevant, balanced and interpretability.
It is the extent to which the
instrument measures what it
intends to measure.
reliability
with the same subjects. A
measure os considered
reliable if a person's score on
the same test given twice
similar. It is important to
remember that reliability is
not measured, it is estimated.
Practical considerations of the
evaluation instruments cannot
be neglected by teachers.
3
The test must have practical
practicality value from time, economy,
scorability, and administration
point of view.
A good test must cover the
lesson taught. It should assess
knowledge, skills, abilities,
attitudes, values as adequately
4
as possible.
relevant
It measures the desired
learning outcomes.
Validation
Validation
is the process of collecting and analyzing evidence
to support the meaningfulness and usefulness.
VALIDITY
is the extent to which a test measures what it purports to
measure or as referring to the appropriateness,
correctness, meaningfulness and usefulness of the
specific decisions a teacher makes based on the test
results.
i o r
p r
lo w
TYPES OF VALIDITY
Brown, 1996
It usually includes any validity strategies that focus on the correlation of the
test being validated with some-respected outside measure(s) of the same
objectives and specifications.
Refers to the relationship between scores obtained using the instruments and
scores obtained using one or more test (often called criterion).
Criterion-related validity is sometimes called concurrent validity.
Concurrent validity is the degree to which the scores on a test are related to
the scores on another, already established, test administered at the same
time, or to some other valid criterion available at the same time
CRITERION VALIDITY
RELIABILITY
The methods used to determine the reliability are as follows:
1. TEST-RETEST METHOD
This provides an indication of stability over time. A
single test is administered twice, the test and the retest,
to the same group of students. If the test and retest yield
similar results, then the test is said to be reliable. The
stronger the positive correlation of the scores, the more
reliable the test is. It is important to note that in the retest,
the teacher may change the order of the test items.
2. EQUIVALENCE (ALTERNATE-FORM)
METHOD.
In this method, we determine the extent to which different
teachers using the same instrument to measure the same
students at the same time yield consistent results.
Equivalence method may also be done by measuring the
same concept same with different instruments.
3. SPLIT-HALF METHOD.
A single test is divided into two equivalent halves. Each of the
halves measures the same concept but consists of different
test items. The splitting may be done uniquely for each
teacher but the common one is the clustering of even
numbered items and the odd numbered items. If the halves of
the test yield similar results, then the test is said to be reliable.
That is the scores in the first half and the second half must be
strongly (positive) correlated.
Solving for the Cronbach's alpha
(a) (equivalent to Kuder-Richardson formula KR20). Cronbach's alpha is a
measure of internal consistency. It tells about the extent of how a set of items are
closely related as a group.
Technically speaking, the Cronbach's alpha is not a statistical test but simply a
coefficient of reliability. Thus, the higher the alpha (close to or equal to 1), the more
consistent the test is. The alpha (a) can be written as a function of the number of
test items and the average and the average inter-correlation among the items
(IDRE, 2017). The function is defined as:
OTHER QUALITIES OF
GOOD ASSESSMENT
INSTRUMENTS
1. SCORABILITY
It means that a test should be easy to score or check. For
ease of scoring, the test should have clear directions for
scoring, answer sheets should be utilized and answer key
must be made in advance.
2. ADMINISTRABILITY
The test should be directed uniformly to all students so
that the scores obtained will not vary due to factors other
than differences of knowledge and skills of the students.
A test should have a clear provision for directions and test
rules for the students to follow in taking the test.
3. OBJECTIVITY
It represents the agreement of two or more raters (usually
teachers) with regards to the score of a student in the test.
If the raters are influenced by their emotions and personal
biases, objectivity of the test is at stake. Consequently, it
affects the test validity
4. FAIRNESS
It is simply referred to the absence of
discrimination in the test due to skin color,
religion, etc.
5. ADEQUACY