100% found this document useful (3 votes)
3K views

Assessment in Learning

This document discusses key concepts and principles of assessment and evaluation of learning. It defines important terms like tests, measurement, assessment, and evaluation. It also outlines the purposes of classroom assessment including assessment for learning, of learning, and as learning. Several principles of high quality classroom assessment are explained, such as having clear learning targets, using appropriate assessment methods, maintaining balanced assessments, ensuring validity and reliability, and ensuring fairness. Assessment is presented as a continuous process that should occur before, during, and after instruction.
Copyright
© © All Rights Reserved
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
100% found this document useful (3 votes)
3K views

Assessment in Learning

This document discusses key concepts and principles of assessment and evaluation of learning. It defines important terms like tests, measurement, assessment, and evaluation. It also outlines the purposes of classroom assessment including assessment for learning, of learning, and as learning. Several principles of high quality classroom assessment are explained, such as having clear learning targets, using appropriate assessment methods, maintaining balanced assessments, ensuring validity and reliability, and ensuring fairness. Assessment is presented as a continuous process that should occur before, during, and after instruction.
Copyright
© © All Rights Reserved
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 22

PROFESSIONAL EDUCATION

Focus: Assessment and Evaluation of Learning


Competencies:
Apply principles in constructing and interpreting alternative/authentic forms of high quality
assessment.
BASIC CONCEPTS
Test - An instrument designed to measure any characteristic, quality, ability, knowledge or skill.
It comprised of items in the area it is designed to measure.
Measurement - A process of quantifying the degree to which someone/something possesses a
given trait. i.e. quality, characteristics, or feature.
Assessment - A process of gathering and organizing quantitative or qualitative data into an
interpretable form to have a basis for judgment or decision- making.
- It is a prerequisite to evaluation. It provides the information which enables
evaluation to take place.
Evaluation - A process of systematic interpretation, analysis, appraisal or judgment of the worth
of organized data as basis for decision-making. It involves judgment about the desirability of
changes in students.
Traditional Assessment - It refers to the use of pen-and-paper objective test.
Alternative Assessments - It refers to the use of methods other than pen- and-paper objective
test which includes performance test, projects, portfolios, journals, and the likes.
Authentic Assessment - It refers to the use of assessment methods that simulate true-to-life
situations. This could be objective tests that reflect real life situations or alternative methods
that are parallel to what we experience in real life.
PURPOSES OF CLASSROOM ASSESSMENT
Assessment FOR Learning - this includes three types of assessment done before and during
instruction. These are placement, formative and diagnostic.
Placement - done prior to instruction
Its purpose is to assess the needs of the learners to have basis in planning for a relevant
instructions.
Teachers use this assessment to know what their students are bringing into the learning situations
and use this as a starting point for instruction.
The results of this assessment place student in specific learning groups to facilitate teaching and
learning.
Formative – done during instruction
It is this assessment where teachers continuously monitor the students’ level of attainment of the
learning objectives (Stiggins, 2005)
The results of this assessment are communicated clearly and promptly to the students for them to
know their strengths and weaknesses and the progress of their learning.
Diagnostic – done during instruction
This is used to determine students’ recurring or persistent difficulties.
It searches for the underlying causes of student’s learning problems that do not respond to first
aid treatment.
It helps formulate a plan for detailed remedial instruction.
Assessment OF Learning - this is done after instruction. This is usually referred to as the
summative assessment.
It is used to certify what students know and can do and the level of their proficiency or
competency.
It results reveal whether or not instructions have successfully achieved the curriculum outcomes.
The information from assessment of learning is usually expressed as marks or letter grades.
The results of which are communicated to the students, parents and other stakeholders for
decision making.
It is also a powerful factor that could pave the way for educational reforms.
3. Assessment AS Learning – this is done for teachers to understand and perform well their role
of assessing FOR and OF learning. It requires teachers to undergo training on how to assess
learning and be equipped with the following competencies needed in performing their work as
assessors.
PRINCIPLES OF HIGH QUALITY CLASSROOM ASSESSMENT
Principle 1: Clear and Appropriate Learning Targets
Learning targets should be clearly stated, specific, and centers on what is truly important.
Learning Targets
(Mc Millan, 2007; Stiggins, 2007)
Knowledge -student mastery of substantive subject matter
Reasoning -student ability to use knowledge to reason and solve problems
Skills -student ability to demonstrate achievement-related skills
Products -student ability to create achievement-related products
Affective/Disposition -student attainment of affective states such as attitudes, values, interests
and self-efficacy.
Principle 2: Appropriate Methods
Assessment Methods
Objective Supply -Short Answer, Completion Test
Objective Selection -Multiple, Choice, Matching ,True/False
Essay -Restricted, Response Extended, Response
Performance Based -Presentations Papers, Projects, Athletics, Demonstrations, Exhibitions,
Portfolios
Oral Question -Oral, Examinations, Conferences, Interviews
Observation -Informal, Formal
Self-Report -Attitude, Survey, Sociometric, Devices, Questionnaires, Inventories
Modes of Assessment
Traditional -The paper-and-pen-test used in assessing knowledge and thinking skills-
Standardized and teacher-made tests -Scoring is objective
Administration is easy because students can take the test at the same time -Preparation of the
instrument is time consuming.
Prone to guessing and cheating
Performance -A mode of assessment that requires actual demonstration of skills or creation of
products of learning-Practical Test-Oral and Aural Test Projects, etc-Preparation of the
instrument is relatively easy -Measures behavior that cannot be deceived-Scoring tends to be
subjective without rubrics -Administration is time consuming
Portfolio -A process of gathering multiple indicators of student’s progress to support course
goals in dynamic, ongoing and collaborative process.- Working Portfolios, Show Portfolios,
Documentary Portfolios-Measures students growth and development, Intelligence-fair-
Development is time consuming , Rating tends to be subjective without rubrics
Principle 3: Balanced
A balanced assessment sets targets in all domains of learning (cognitive, affective, and
psychomotor) or domains of intelligence (verbal-linguistic, logical-mathematical, bodily-
kinesthetic, visual-spatial, musical-rhythmic, intrapersonal-social, interpersonal-introspection,
physical world-natural-existential-spiritual).
A balanced assessment makes use of both traditional and alternative assessment.
Principle 4: Validity
A. Validity - is the degree to which the assessment instrument measures what it intends to
measure. It is also refers to the usefulness of the instrument for a given purpose. It is the most
important criterion of assessment instrument.
Ways in Establishing Validity
1.Face Validity - is done by examining the physical appearance of the instrument.
2.Content Validity – is done through a careful and critical examination of the objectives of
assessment so that it reflects the curricular objectives.
3.Criterion-related Validity – is established statistically such that a set of scores revealed by
the measuring instrument is correlated with the scores obtained in another external predictor of
measure. It has two purposes.
a. Concurrent validity – describes the present status of the individual by correlating the sets of
scores obtained from two measures given concurrently.
b. Predictive validity – describes the future performance of an individual by correlating the sets
of scores obtained from two measures given at a longer time interval.
4.Construct Validity – is established statistically by comparing psychological traits or factors
that theoretically influence scores in a test.
Convergent Validity - is established if the instrument defines another similar trait other than
what it is intended to measure e.g. Critical Thinking Test may be correlated with Creative
Thinking Test.
Divergent Validity - is established if an instrument can describe only the intended trait and not
the other traits, e.g. Critical Thinking Test may not be correlated with Reading Comprehension
Test.
Principle 5: Reliability
Reliability - it refers to the consistency of scores obtained by the same person when retested
using the same instrument/its parallel or when compared with other students who took the same
test.
Test-Retest-Measure of stability- Give a test twice to the same group with any time interval
between tests from several minutes to several years-Pearson r
Equivalent Forms-Measure of equivalence-Give parallel forms of tests with close time interval
between forms-Pearson r
Test-retest with Equivalent Forms-Measure of stability and equivalence-Give parallel forms of
tests with increased time interval between forms.- Pearson r
Split Half-Measure of Internal Consistency -Give a test once. Score equivalent halves of the
test e.g. odd- and even- numbered items. - Pearson r & Spearman Brown Formula
Kuder-Richardson-Measure of Internal Consistency -Give the test once then correlate the
proportion/percentage of the students passing and not passing a given item-Kuder-Richardson
Formula 20 and 21
Principle 6: Fairness
A fair assessment provides all students with an equal opportunity to demonstrate achievement.
The key to fairness are as follows:
Students have knowledge of learning targets and assessment
Students are given equal opportunity to learn
Students possess the pre-requisite knowledge and skills
Students are free from teacher stereotypes
Students are free from biased assessment tasks and procedures
Principle 7: Assessment should be a continuous process.
Assessment takes place in all phases of instruction. It could be done before, during and after
instruction.
Activities Occurring Prior to Instruction
Understanding students’ cultural backgrounds, interests, skills and abilities as they apply across
a range of learning domains and/or subject areas;
Understanding students’ motivations and their interests in specific class content;
Clarifying and articulating the performance outcomes expected of pupils; and
Planning instruction for individuals or groups of students.
Activities occurring During instruction:
Monitoring pupil progress toward instructional goals;
Identifying gains and difficulties pupils are experiencing in learning and performing;
Adjusting instruction;
Giving contingent, specific, and credible praise and feedback;
Motivating students to learn; and
Judging the extent of pupil attainment of instructional outcomes.
Activities occurring After the appropriate instructional segment (e.g. lesson, class, semester,
grade)
Describing the extent to which each pupil has attained both short- and long-term instructional
goals;
Communicating strengths and weaknesses based on assessment results to students, and parents or
guardians;
Recording and reporting assessment results for school-level analysis, evaluation, and decision-
making;
Analyzing assessment information gathered before and during instruction to understand each
students’ progress to date and to inform future instructional planning;
Evaluating the effectiveness of instruction; and
Evaluating the effectiveness of the curriculum and materials in use
Principle 9: Communication
Assessment targets and standards should be communicated
Assessment results should be communicated to its important users.
Assessment results should be communicated to students through direct interaction or regular
ongoing feedback on their progress.
Principle 10: Positive Consequences
Assessment should have a positive consequence to students, that is, it should motivate them to
learn.
Assessment should have a positive consequence on teachers that is it should help them improve
the effectiveness of their instruction.
Principle 11: Ethics
Teacher should free the student from harmful consequence to students, that is, it should motive
them to learn.
Assessment should have a positive consequence on teachers, that is, it should help them improve
the effectiveness of their instruction
Teacher should be guided by laws and policies that affect their classroom assessment.
Administrators and teachers should understand that it is inappropriate to use standardized
students achievement to measure teaching effectiveness.
PERFORMANCE-BASED ASSESSMENT
Performance-Based Assessment is a process of gathering information about student’s learning
through actual demonstration of essential and observable skills and creation of products that are
grounded in real world contexts and constraints. It is an assessment that is open to many possible
answers and judged using multiple criteria or standards of excellence that are pre-specified and
public.
Reasons for using Performance-Based Assessment
Dissatisfaction of the limited information obtained from selected-response test.
Influence of cognitive psychology, which demands not only for the learning of declarative but
also for procedural knowledge.
Negative impact of conventional test e.g. high-stake assessment, teaching for the test
It is appropriate in experimental, discovery-based, integrated, and problem-based learning
approaches.
Type of Performance-based Task
Demonstration-type - this is a task that requires no product
Examples: constructing a building, cooking demonstrations, entertaining tourist, teamwork,
presentations
Creation-type - this is a task that requires tangible products.
Examples: project plan, research paper, project flyers, discovered.
Methods of Performance-based Assessment
1. Written-open ended - a written prompt is provided
Formats: Essays, open-ended test
2. Behavior-based - utilizes direct observations of behaviors in situation or simulated
contexts.
Formats: structured and unstructured
3. Interview-based - examinees respond in one-to-one conference setting with the examiner to
demonstrate mastery of the skills.
Formats: structured and unstructured
4. Product-based - examines create a work sample or a product utilizing the skills / abilities.
Formats: restricted and extended
5. Portfolio-based - collections of work that are systematically gathered to serve many
purposes.
How to Assess a Performance
1. Identify the competency that has to be demonstrated by the students with or without a product.
2. Describe the task to be performed by the students either individually or as a group, the
resources needed, time allotment and other requirements to be able to assess the focused
competency.
7 Criteria in Selecting a Good Performance Assessment Task
• Generalizability - the likelihood that the students’ performance on the task will generalize to
comparable tasks.
• Authenticity - the task is similar to what the students might encounter in the real world as
opposed to encountering only in the school.
• Multiple Foci - the task measures multiple instructional outcomes.
• Teachability - the task allows one to master the skill that one should be proficient in
• Feasibility - the task is realistically implementable in relation to its cost, space, time, and
equipment requirements
• Scorability - the task can be reliably and accurately evaluated
• Fairness - the task is fair to all the students regardless of their social status or gender
3. Develop a scoring rubric reflecting the criteria, levels of performance and the scores
PORTFOLIO ASSESSMENT
Portfolio Assessment is also an alternative to pen-and-paper objective test. It is a purposeful, on
going, dynamic, and collaborative process of gathering multiple indicators of the learner's
growth and development. Portfolio assessment is also performance-based but more authentic
than any performance-based task.
Reasons for Using Portfolio Assessment
Burke (1999) actually recognizes portfolio as another type of assessment arid considered
authentic because of the following reasons:
It tests what is really happening in the classroom.
It offer; multiple indicators of students' progress.
It gives the students the responsibility of their own learning.
It offers opportunities for students to document reflections of their learning.
It demonstrates what the students know in ways that encompass their personal teaming styles
and multiple intelligences.
It offers teachers new role in the assessment process.
It allows teachers to reflect on the effectiveness of their instruction.
It provides teachers freedom of gaining insights into the student's development or 'achievement
over a period of time.
Types of Portfolios
Portfolios could come in three types: working, show, or documentary.
1. The working portfolio is a collection of a student’s day-to-day works which reflect his/her
learning.
2. The show portfolio is a collection of a student's best works.
3. The documentary portfolio is a combination of a working and a show portfolio.
DEVELOPING RUBRICS
Rubric is a measuring instrument used in rating performance-based tasks. It is the “key to
corrections for assessment tasks designed to measure the attainment of learning competencies
that require demonstration of skills or creation of products of learning, It offers a set of
guidelines or descriptions in scaling different levels of performance or qualities of products of
learning. It can be used in scoring both the process and the products of learning.
Similarity of Rubric with Other Scoring Instruments
Rubric is a modified checklist and rating scale.
1. Checklist
presents the observed characteristics of a desirable performance or product
the rater checks the trait/s that has/have been observed in one's performance or product.
2. Rating Scale
measures the extent or degree to which a trait has been satisfied by one's work or performance
offers an overall description of the different levels of quality of a work or a performance
uses 3 to more levels to describe the work or performance although the most common rating
scales have 4 or 5 performance levels.
Holistic Rubric -It describes the overall quality of e Performance or product. In this rubric, there
is only one rating given to the entire work or performance
Advantages
It allows fast assessment.
It provides one score to describe the overall performance or quality of work.
It can indicate the general strengths and weaknesses of the work or performance
Disadvantages
It does not clearly describe the degree of the criterion satisfied or not by the performance or
product.
It does not permit differential weighting of the qualities of a product or a performance.
Analytic Rubric - it describes the quality of a performance or product in terms of the identified
dimensions and/or criteria for which are rated independently to give a better picture of the
quality of work or performance.
Advantages
It clearly describes the degree of the criterion satisfied or not by the performance or product.
It permits differential weighting of the qualities of a product or a performance.
It helps raters pinpoint specific areas of strengths and weaknesses
Disadvantages
It is more time consuming to use.
It is more difficult to construct
PART II - ANALYZING TEST ITEMS
PROFESSIONAL EDUCATION
Focus: Assessment and Evaluation of Learning 2
Competencies:
1. Apply principles in constructing and interpreting traditional forms of assessment.
2. Utilize processed data and results in reporting and interpreting learners’ performance to
improve teaching and learning.
3. Demonstrate skills in the use of techniques and tools in assessing affective learning.
PART I: CONTENT UPDATE
It is an instrument or systematic procedure which typically consists of a set of questions for
measuring a sample of behavior.
It is a special form of assessment made under contrived circumstances especially so that it may
be administered.
It is a systematic form of assessment that answers the question, “How well does the individual
perform – either in comparison with others or in comparison with a domain of performance task.
An instrument designed to measure any quality, ability, skill or knowledge.
Instructional Uses of Tests
grouping learners for instruction within a class
identifying learners who need corrective and enrichment experiences
measuring class progress for any given period
assigning grades/marks
guiding activities for specific learners (the slow, average, fast)
Guidance Uses of Tests
assisting learners to set educational and vocational goals
improving teacher, counselor and parents’ understanding of children with problems.
preparing information/data to guide conferences with parents about their children.
determining interests in types of occupations not previously considered or known by the students
predicting success in future educational or vocational endeavor.
Administrative Uses of Tests
determining emphasis to be given to the different learning areas in the curriculum
measuring the school progress from year to year
determining how well students are attaining worthwhile educational goals.
determining appropriateness of the school curriculum for students of different levels of ability.
developing adequate basis for pupil promotion or retention.
I. Standardized Tests – tests that have been carefully constructed by experts in the light of
accepted objectives
Ability Tests - combine verbal and numerical ability, reasoning and computations. Ex.: OLSAT
– Otis Lennon Standardized Ability Test
Aptitude Tests - tests which measure potential in a specific field or area; predict the degree to
which an individual will succeed in any given area such art, music, mechanical task or academic
studies. Ex. DAT – Differential Aptitude Test
II. Teacher-Made Tests - tests constructed by classroom teacher which measure and appraise
student progress in terms of specific classroom/instructional objectives.
1. Objective Type – answers are in the form of a single word or phrase or symbol.
a. Limited Response Type – requires the student to select the answer from a given number of
alternatives or choices.
i. Multiple Choice Test - consists of a stem each of which presents three to five alternatives
or options in which only one is correct of definitely better than the other. The correct option
choice or alternative in each item is merely called answer and the rest of the alternatives are
called disasters or decoys or foils.
ii. True – False or Alternative Response - consists of declarative statements that one has to
respond or mark true or false, right or wrong, correct or incorrect, yes or no, fact or opinion,
agree or disagree and the like. It is a test made up of items which allow dichotomous responses.
iii. Matching Type – consists of two parallel columns with each word, number, or symbol in
one column being matched to a word sentence, or phrase in the other column. The items in
Column I or A for which a match is sought are called premises, and the items in Column II or B
from which the selection is made are called responses.
b. Free Response Type or Supply Test – requires the student to supply or give the correct
answer.
i. Short Answer – uses a direct question that can be answered by a word, phrase, number, or
symbol.
ii. Completion Test - consists of an incomplete statement that can also be answered by a
word, phrase, number, or symbol.
2. Essay Type - Essay questions provide freedom of response that is needed to adequately assess
students’ ability to formulate, organize, integrate and evaluate ideas and information or apply
knowledge and skills.
Restricted Essay - limits both the content and the response. Content is usually restricted by the
scope of the topic to be discussed.
Extended Essay - allows the students to select any factual information that they think is
pertinent to organize their answers in accordance with their best judgment and to integrate and
evaluate ideas which they think appropriate.
Use assessment specifications as a guide to item/task writing.
Construct more items/tasks than needed.
Write the items/tasks ahead of the testing date.
Write each test item/task at an appropriate reading level and difficulty.
Write each test item/task in a way that it does not provide help in answering other test items or
tasks.
Write each test item/task so that the task to be performed is clearly defined and it calls forth the
performance described in the intended learning outcome.
Write a test item/task whose answer is one that would be agreed upon by the experts.
Whenever a test is revised, recheck its relevance.
Supply Type of Test
Word the item/s so that the required answer is both brief and specific.
Do not take statements directly from textbooks.
A direct question is generally more desirable than an incomplete statement.
If the item is to be expressed in numerical units, indicate the type of answer wanted.
Blanks for answers should be equal in length and as much as possible in column to the right of
the question.
When completion items are to used, do not include too many blanks.
B. Selective Type of Tests
Alternative – Response
Avoid broad, trivial statements and use of negative words especially double negatives.
Avoid long and complex sentences.
Avoid multiple facts or including two ideas in one statement, unless cause-effect relationship is
being measured.
If opinion is used, attribute it to some source unless the ability to identify opinion is used,
attribute it to some source unless the ability to identify opinion is being specifically measures.
Use proportional number of true statements and false statements.
True statements and false statements should be approximately equal in length.
2. Matching Type
Use only homogeneous material in a single matching exercise.
Include an unequal number of responses and premises and instruct the pupil that responses may
be used once, more that once, or not at all.
Keep the list of items to be matched brief, and place the shorter responses at the right.
Arrange the list of responses in logical order.
Indicate in the directions the basis for matching the responses and premises.
Place all the items for one matching exercise on the same page.
Limit a matching exercise to not more than 10 to 15 items.
3. Multiple Choice
The stem of the item should be meaningful by itself and should present a definite problem.
The item stem should include as much of the item as possible and should be free o irrelevant
material.
Use a negatively stated stem only when significant learning outcomes require it and stress/
highlight the negative words for emphasis.
All the alternatives should be grammatically consistent with the stem of the item.
An item should only contain one correct or clearly best answer.
Items used to measure understanding should contain some novelty, but beware of too much.
g. All distracters should be plausible/attractive.
h. Verbal associations between the stem and the correct answer should be avoided.
i. The relative length of the alternatives/options should not provide a clue to the answer.
j. The alternatives should be arranged logically.
k. The correct answer should appear in each of the alternative positions and approximately
equal number of times but in random order.
l. Use of special alternatives such as “none of the above” or “all of the above” should be done
sparingly.
m. Always have the stem and alternatives on the same page.
n. Do not use multiple choice items when other types are more appropriate.
4. Essay Type of Test
Restrict the use of essay questions to those learning outcomes that cannot be satisfactorily
measured by objective items.
Construct questions that will call forth the skills specified in the learning standards.
Phrase each question so that the student’s task is clearly defined or indicated.
Avoid the use of optional questions.
Indicate the approximate time limit or the number of points for each question
Prepare an outline of the expected answer in advance or scoring rubric.
A. Major Characteristics
Validity – the degree to which a test measures what it is supposed or intends to measure. It is
the usefulness of the test for a given purpose. It is the most important quality/characteristics
desired in an assessment instrument.
Reliability - refers to the consistency of measurement that is how consistent test scores or other
assessment results are from one measurement to another. It the most important characteristics of
an assessment instrument next to validity.
Minor Characteristics
Administrability - the test should be easy to administer such that the directions should clearly
indicate how a student should respond to the test/task items and how much time should he/she
spend for each test item or for the whole test.
Scoreability - the test should be easy to score such that directions for scoring are clear, point/s
for each correct answer(s) is/are specified.
Interpretability - test scores can easily be interpreted and described in terms of the specific
tasks that a student can perform of his/her relative position in a clearly defined group.
Economy - the test should be given in the cheapest way in terms of time and effort spent for
administration of the test and answer sheets must be provided so the test can be given from time
to time.
Factors Influencing the Validity of an Assessment Instrument
1. Unclear directions. Directions that do not clearly indicate to the students how to respond to
the tasks and how to record the responses tend to reduce validity.
2. Reading vocabulary and sentence structure too difficult. Vocabulary and sentence
structure that are too complicated for the students result in the assessment of reading
comprehension thus altering the meaning of assessment result.
3. Ambiguity. Ambiguous statements in assessment tasks contribute to misinterpretations and
confusion. Ambiguity sometimes confuses the better students more that it does the poor students.
4. Inadequate time limits. Time limits that do not provide students with enough time to
consider the tasks and provide thoughtful responses can reduce the validity of interpretations of
results. Rather than measuring what a student knows about a topic or is able to do given adequate
time, the assessment may become a measure of the speed with which the student can respond.
For some content (e.g. a typing test), speed may be important. However, most assessments of
achievement should minimize the effects of speed on student performance.
5. Overemphasis of easy - to assess aspects et domain at the expense of important, but hard —
to assess aspects (construct under representation). it is easy to develop test questions that assess
factual recall and generally harder to develop ones that tap conceptual understanding or higher
order thinking processes such as the evaluation of competing positions or arguments. Hence, it is
important to guard against under representation of tasks getting at the important, but .more
difficult to assess aspects of achievement.
6. Test items inappropriate for the outcomes being measured. Attempting to measure
understanding, thinking skills, and other complex types of achievement with test forms that are
appropriate only for measuring factual knowledge will invalidate the results.
7. Poorly constructed test items. Test items that unintentionally provide clues to the answer
tend to measure the students' alertness in detecting clues as well as mastery of skins or
knowledge the test is intended to measure.
8. Test too short If a test is too short to provide a representative sample of the performance we
are interested in, its validity will suffer accordingly.
9. Improper arrangement of items. Test items are typically arranged in order of difficulty, with
the easiest items first. Placing difficult items first in the test may cause students to spend too
much time on these and prevent them from reaching items they could easily answer. Improper
arrangement may also influence validity by having a detrimental effect on student motivation.
10. Identifiable pattern of answer. Placing correct answers in some systematic pattern (e.g.., T,
1, F, F, or B, B, B, C, C, C, D, D, D) enables students to guess the answers to some items more
easily, and this lowers validity.
Improving Test Reliability
Several test characteristics affect reliablity. They include the following:
1. Test length. In general, a longer test is more reliable than a shorter one because longer tests
sample the instructional objectives more adequately.
2. Spread of scores. The type of students taking the test can influence Reliability. A group of
students with heterogeneous ability will produce a larger spread of test scores than a group with
homogeneous ability.
3. Item difficulty. In general, tests composed of items of moderate or average difficulty (.30 to
JO) will have more influence on reliability than those composed primarily of easy or very
difficult items.
4. Item discrimination. In general, tests composed of more discriminating items will have
greater reliability than those composed of less discriminating items.
5. Time limits. Adding a time factor may improve reliability for lower — level cognitive test
items. Since all students do not function at the same pace, a time factor adds another criterion to
the test that causes discrimination, thus improving reliability. 'Teachers should not, however,
arbitrarily impose a time limit. For higher - level cognitive test items, the imposition of a time
limit may defeat the intended purpose of the items.
Symmetrically Shaped Test Score Distributions
Normal Distribution or Bell Shaped Curve
Rectangular Distribution
U-Shaped Curve
B. Skewed Distribution of Test Score
Positively Skewed Distribution (mean>median> mode)
Negatively Skewed Distribution (move>median>mean)
Skewness is the degree of asymmetry, or departure from symmetry of a distribution.
Skewed to the right: (positive skewness): if the frequency curve of a distribution has a longer
“tail” to the left of the central maximum than to the right. Most scores are below the mean.
Skewed to the left: (negative skewness): if the frequency curve of a distribution has a longer
“tail” to the right of the central maximum than to the left. Most scores are above the mean and
there are extremely low scores.
Kurtosis -is the degree of peakedness of a distribution, usually taken relative to a normal
distribution.
Leptokurtic: distribution having a realtively high peak.
Platykurtic: a distribution having flat-topped
Mesokurtic: a distribution which is moderately peaked.
Descriptive Statistics – the first step in data analysis is to describe, or summarize the data using
descriptive statistics
Measure of Central Tendency
numerical values which describe the average or typical performance of given group in terms of
certain attributes
Basis in determining whether the group is performing better or poorer than the other groups
Mean-Arithmetic average, used when the distribution is normal/symmetrical or bell – shaped.
Most reliable/ affected by extreme scores
Median-Point in a distribution above and below which are 50% of the scores/cases;
Midpoint of a distribution;
Used when the distribution is skewed/ most stable
Mode-Most frequent/common score in a distribution;
Opposite of the mean, unreliable/unstable;
Used as quick description in term of average/typical performance of the group.
Measures of Variability
indicate or describe how spread the scores are. The larger the measure of variability the more
spread the score are and the group is said to be heterogeneous; the smaller the less spread the
scores are and the group is said to be homogeneous.
Range-The difference between the highest and lowest score plus one;
Counterpart of the mode it is also unreliable/unstable;
Used as a quick, rough estimate of measure of variability.
Standard Deviation-The counterpart of the mean, used also when the distribution is normal or
symmetrical;
Reliable/stable and so widely used
Z-score AS applied to test results
Study this group of test which was administered with the following results, and then answer the
question that follows.
Z score= score – mean / sd
Assessment and Evaluation of Learning Part 1
1. It is the engagement to make sure learners' success in working from guided to independent
display of knowledge, comprehension, and skills. Among the following, which is the theoretical
basis of the said system?
A. Multiple Intelligence Theory
B. Information Processing Theory
C. Zone of Proximal Development
D. Dale's Cone of Experiences
Answer: C

Rationalization: DepEd Order No. 8, s. 2015 explicitly states the theoretical basis of classroom
assessment as the Zone of Proximal Development by Lev Vygotsky which is focused on
ensuring learners' success in moving from guided to independent display of knowledge,
understanding, and skills.

2. James was given a module by his teacher for him to catch up with the lessons he missed. The
module includes a series of objective type test items after each lessons for James to check and
monitor his progress as he move to the next lesson. What form of assessment is being applied by
the teacher?
A. Assessment of learning
B. Assessment for learning
C. Assessment of and for learning
D. Assessment as learning
Answer: D
Rationalization: In the situation, James is using the objective test as a form of self-assessment to
check and monitor his progress as he move to the next lesson. Self-assessment is associated with
the concept of assessment as learning.

3. A learning outcome in Grade 12 research subject reads 'At the end of the lesson, the students
will be exposed to various research designs'. Is this a good target for assessment?
A. Yes, the action is observable.
B. Yes, the action can be measured.
C. No, the action is not done by the students.
D. No, the action is not attainable.
Answer: C
Rationalization: The behavior contained in the learning outcome 'will be exposed' is primarily an
action that reflects an input and not an outcome. Is it an action that will be done by the teacher.

4. Teacher Anne asks her students to submit a portfolio that showcases their best literary works
from all their finished compositions. What type of portfolio is teacher Anne asking?
A. Display Portfolio
B. Working Portfolio
C. Documentation Portfolio
D. Growth Portfolio
Answer: A
Rationalization: A portfolio that shows students' best literary composition is characteristic of a
display portfolio, which is known as showcase or best work portfolio. Such portfolio aims to
show the highest level of achievement attained by the students.

5. In the present marking system of K-12, how are learners' behavior reported?
A. Numerical rating
B. Non-numerical rating
C. Pass-fail system
D. Anecdotal record
Answer: B
Rationalization: DepEd Order No. 8, s. 2015 explicitly states that learners' behavior will be
reported through non-numerical rating based on observations by the teachers.

6. Mr. Adamson administered an achievement test in Mathematics. He is in the process of


counting the scores obtained by his students. This process if called _____.
A. Assessment
B. Measurement
C. Evaluation
D. Testing
Answer: B
Rationalization: Measurement is the most specific answer among the choices, as the situation is
focused on the process of obtaining a numerical description of the degree to which an individual
possess a particularly characteristics, which in this case is the achievement in Mathematics.

7. Which of the following is the advantage of portfolio assessment?


A. Learners are assessed based on a wide range of evidences of learning.
B. Learers are assessed objective.
C. Learners are assessd in a manner that is from the teachers' biases.
D. Learners can demonstrate what they can do without any evidences of learning.
Answer: A
Rationalization: A primary advantage of portfolio assessment is that it allows learners to show
varied evidences of learning. A certain level of objectivity may also be achieved by portfolio
assessment thru the use of scoring guides, but this is not an advantage of portfolio over other
forms of assessment, which are already objective in nature.

8. What is the lowest mark that can be reflectes in the report card?
A. 60
B. 65
C. 70
D. 75
Answer: A
Rationalization: DepEd Order No. 8, s. 2015 states that the lowest mark that can appear on the
report card is 60 for Quarterly Grade and Final Grades.

9. Which of the following test in considered to be relatively less reliable for scoring and grading?
A. Completion
B. Alternative Response
C. Essay
D. Matching Type
Answer: C
Rationalization: Essay tests are considered as a non-objective test, which means that relative to
other type of test, it is considered to be less reliable in terms of scoring.

10. Teacher Vivian is asking her students to identify the equipment in the Chemistry Laboratory.
Under the cognitive domain, in which assessment level is this?
A. Remembering
B. Understanding
C. Analyzing
D. Creating
Answer: A
Rationalization: Identifying laboratory equipments is an activity involving recalling of facts.
Thus, it is considered to be in the knowledge level or the remembering level in the revised
Bloom's taxonomy.

You might also like