Psychological Testing Assessment
Psychological Testing Assessment
(TESTING) ASSESSMENT
A review…
Why are you here?
•House Rules…
Before exam!
During exam!
After exam!
Topic Outline
• Basic Definitions…
• Basic Concepts/Assumptions in Psychological Testing and Assessment
• History of Psychological Testing (History of Psychometrics )
• Basic Psychological Statistics
• Basic Principles and Tools of (Psychological) Measurement
• Norms
• Reliability & Validity
• Test Development
• Psychological Tests and Applications of Tests
• Ethics in Psychological Testing
Psychological Assessment
• Psychological assessment is a process of testing that uses a combination of
techniques to help arrive at some hypotheses about a person and their behavior,
personality and capabilities.
• According to Philippine law, RA 10029 (Philippine Psychology Act of 2009),
• Psychological assessment is the: "gathering and integration of
psychology-related data for the purpose of making a psychological
evaluation, accomplished through a variety of tools, including
individual tests, projective tests, clinical interview and other
psychological assessment tools, for the purpose of assessing diverse
psychological functions including cognitive abilities, aptitudes,
personality characteristics, attitudes, values, interests, emotions
and motivations, among others, in support of psychological
counseling, psychotherapy and other psychological interventions."
• In the book authored by Munarriz and Cervera, Psychological
Testing in the Philippines: Practice, Directions and Perspectives
(2013), they made a clear distinction with words such as
psychometrics, psychological testing, and psychological
assessment:
• Psychometrics then refers to the theory, technique
and development of psychological measurements;
• Psychological testing, to the application of tests,
and;
• Psychological assessment, to the use of tests,
among others to do a psychological evaluation.
Who/What is a Psychometrician?
Psychological Testing
and Assessment
Basic Concepts and its Differences?
Psychological Testing
✓ an objective and standardized measure sample of
behavior;
✓ process of measuring psychology-variables by
means of devices or procedures designed to
obtained a sample of behavior;
✓ process of administering, scoring and interpreting
Ψ test.
• There are a number of core principles that form the foundation for
psychological assessment:
• Tests are samples of behavior.
• Tests do not directly reveal traits or capacities, but may allow inferences to be
made about the person being examined.
• Tests should have adequate reliability and validity.
• Test scores and other test performances may be adversely affected by temporary
states of fatigue, anxiety, or stress; by disturbances in temperament or
personality; or by brain damage.
• Test results should be interpreted in light of the person’s cultural background,
primary language, and any handicaps.
• Test results are dependent on the person’s cooperation and motivation.
• Tests purporting to measure the same ability may produce different scores for
that ability.
• Test results should be interpreted in relation to other behavioral data and to case
history information, never in isolation.
Psychological Assessment
✓ The gathering and integration of psychology-related data for the purpose
of making a Ψ evaluation, accomplished through the use of tools such as
tests, interviews, case studies, behavioral observation & other designed
apparatus and measurement procedure;
✓ Extends beyond obtaining a number;
✓ Focuses on how the individual processes than the results of that
processing;
✓ Less emphasis on the measurement of strength of traits; and
✓ More emphasis on the understanding of problems in their social context
• Psychological assessment is a powerful tool, but its
effectiveness depends upon the skill and knowledge of the
person administering and interpreting the test. When used
wisely and in a cautious manner, psychological assessment
can help a person learn more about themselves and gain
valuable insights. When used inappropriately, psychological
testing can mislead a person who is making an important
life decision or decision about treatment, possibly causing
harm.
BASIC
ASSUMPTIONS
In Psychological Testing and
Assessment
Assumption 1: Psychological Traits and States Exist
47
History of Psychometrics: Chinese influence
• 2000 B.C.E.
• Scattered evidence of civil service testing in China
• Every 3rd year in China, oral examinations were given to
help determine work evaluations and promotion decisions.
• 206 B.C.E. to 220 C.E.
• Han Dynasty in China develops test batteries
• two or more tests used in conjunction.
• Test topics include civil law, military affairs, agriculture,
revenue, geography
48
History of Psychometrics: Chinese influence
• 1368 C.E. to 1644 C.E.
• Ming Dynasty in China develops multistage testing
• Local tests lead to provincial capital tests; capital tests lead
to national capital tests
• Only those that passed the national tests were eligible for
public office
• 1832
• EnglishEast India Company copies Chinese system to
select employees for overseas duty.
49
History of Psychometrics: Chinese influence
• 1855
• British Government adopts English East India
Company selection examinations.
• French & German governments follow shortly.
• 1883
• United States establishes the American Civil Service
Commission
• Developed & administered competitive examinations
for government service jobs.
50
History of Psychometrics: Individual
Differences, Darwin and Galton
• Individual differences - despite our similarities,
no two humans are exactly alike.
• Charles Darwin
• some of these individual differences are more
“adaptive” than others
• these individual differences, over time, lead to
more complex, intelligent organisms
51
History of Psychometrics: Individual
Differences, Darwin and Galton
• Sir Francis Galton
• “Applied Darwinist”: some people possessed
characteristics that made them “more fit” than others.
• Wrote Hereditary Genius (1869)
• Sets up an anthropometric (measurement of the human
individual) laboratory at the International Exposition of
1884
• For 3 pence, visitors could be measured with:
• The Galton Bar - visual discrimination of length
• The Galton Whistle (aka “dog whistle” - determining
highest audible pitch
52
History of Psychometrics: Individual
Differences, Darwin and Galton
• Galton’s Anthropometric Lab
53
Individual Differences:
Darwin and Galton
54
History of Psychometrics:
Galton’s Famous Students
• Karl Pearson
• Does the name Pearson sound familiar?
• student of Galton
• extended Galton’s early work with statistical
regression
• James McKeen Cattell
• first to use the term “mental test”
• U.S.dissertation on individual differences in
reaction time based upon Galton’s work
55
History of Psychometrics:
Early Experimental Psychologists
• Early 19th century scientists, generally interested in
identifying common aspects, rather than individual
differences.
• Differences between individuals was considered a
source of error which rendered human measurement
inexact.
• Sounds a lot like things from your past (e.g. ANOVA)
and your coming future
56
History of Psychometrics:
Early Experimental Psychologists
• JohanFriedrich Herbart - mathematical models of the
mind; founder of pedagogy as an academic discipline;
went against Kant;
• Ernst Heinrich Weber - sensory thresholds; just
noticeable difference (JND)
• Gustav Theodor Fechner - mathematics of sensory
thresholds of experience; founder of psychophysics;
considered of one founders of experimental psychology;
Weber-Fechner Law first to relate sensation and
stimulus; considered by some the founder of
psychometrics
57
History of Psychometrics:
Early Experimental Psychologists
• Fechner influenced many prominent
psychologists (e.g. Wundt, Freud)
• Wilhelm Wundt – considered one of the
founders of psychology; first to set up a psych
laboratory
• Edward Titchener – succeeded Wundt;
brought Structuralism to America; His brain is
still on display in the psychology department
at Cornell
58
History of Psychometrics:
Early Experimental Psychologists
• Guy Montrose Whipple – Student of Titchener; pioneer of human
ability testing; conducted seminars that changed the field of psych
testing; APA issued its first set of standards for professional
psychological testing because of his criticisms
62
History of Psychometrics:
Intelligence Testing
• Alfred Binet
• First Intelligence Test: Binet-Simon Scale of 1905
• 30 items of increasing difficulty
• Standardized administration
• Same instructions & format for ALL children
• Standardization sample
• Created norms by which performance one child can be
compared with other children.
63
History of Psychometrics:
Intelligence Testing
• Alfred Binet
• Standardization Sample
• 50 Normal children aged 3-11yrs
• “Some” mentally retarded children and adults
64
History of Psychometrics:
Intelligence Testing
• Alfred Binet’s legacy
• 1911 Binet-Simon, minor revision
• Binet dies
• 1912 Kuhlmann-Binet revision
• Extends testing downward to 3 months of age
• 1916 Lewis Madison Terman & Stanford Colleagues
revise Binet’s test for use in the United States
• More psychometrically sound
• Introduction of the term IQ
• Mental Age / Chronological Age x 100 = IQ
65
History of Psychometrics:
Intelligence Testing
• World War I - Robert Yerkes
• Need for large-scale group administered ability tests by the
army
• Army commissions Yerkes, then head of the American
Psychological Association, to develop two structured tests
of human abilities
• Army Alpha - required reading ability
• Army Beta - did not require reading ability
66
History of Psychometrics:
Intelligence Testing
• Testing Frenzy of the 1930’s
• 1937 Revision of the Stanford-Binet includes over 3000
individuals in standardization
• 1939 Wechsler-Bellevue Intelligence Scale
• David Wechsler
• Subscales were “adopted” from the Army Scales
• Produces several scores of intellectual ability rather than
Binet’s single scores (e.g. Verbal, Performance, Full-Scale)
• Evolves into the Wechsler Series of intelligence tests (e.g.
WAIS, WISC, etc.)
67
History of Psychometrics:
Personality Testing
• Rise – 1920s, Fall – 1930s, Slow Rise – 1940s
• Intended to measure personality traits
• Trait: relatively enduring dispositions (tendencies to act,
think or feel in a certain manner in any given circumstance)
that distinguish one individual from another
• NOT temporary states
68
History of Psychometrics:
Personality Testing
• First Rise and Fall: Structured Tests
• Woodworth Personal Data Sheet
• First objective personality test meant to aid in psychiatric
interviews
• Developed during World War I
• Designed to screen out soldiers unfit for duty
• Mistakenly assumed that a subjects response could be
taken at face value
69
History of Psychometrics:
Personality Testing
• Woodworth
Test Item Yes No
1. I wet the bed.
2. I drink a quart of whiskey each day.
3. I am afraid of closed spaces.
4. I believe I am being followed.
5. People are out to get me.
6. Sometimes I see or hear things that other
people do not hear or see.
70
History of Psychometrics:
Personality Testing
• Slow Rise: Projective Tests
• Herman Rorschach inkblot test (1921)
• Started with great suspicion; first serious
study in 1932.
• Symmetric colored & b/w inkblots.
71
History of Psychometrics:
Personality Testing
72
History of Psychometrics:
Personality Testing
• Thematic Apperception Test
• Henry Murray and Christina Morgan (1935)
• “Ambiguous” pictures though considerably more
structured than the Rorschach
• Subjects are shown the pictures and asked to write a
story including:
• what has led up to the event shown
• what is happening at the moment
• what the characters are feeling and thinking, and
• what the outcome of the story was.
73
History of Psychometrics:
Personality Testing
Thematic Apperception Test Example
74
History of Psychometrics:
Personality Testing
• Second coming of the Structured Test
• Early 1940s – Structured Tests were being developed based
on better psychometric properties.
• Minnesota Multiphasic Personality Inventory (MMPI; 1943)
• Tests like the Woodworth made too many assumptions
• The meaning of the test response could only be determined
by empirical research
• Most widely used (MMPI-2, MMPI-A)
75
History of Psychometrics:
Personality Testing
76
BASIC PRINCIPLES
AND TOOLS OF
MEASURMENT
SCALES OF MEASUREMENT
Measurement - the act of assigning numbers or
symbols to characteristics of things, people or
events according to rules.
Scale - a set of numbers or other symbols whose
properties model empirical properties of the
objects to which the numbers are assigned.
Measurement
• The process of assigning numbers to objects in such a way that
specific properties of the objects are faithfully represented by
specific properties of the numbers.
• Psychological tests do not attempt to measure the total person,
but only a specific set of attributes.
• Measurement is used to capture some “construct”
- For example, if research is needed on the construct of
“depression”, it is likely that some systematic measurement tool
will be needed to assess depression.
Types of Measurement Scales
1. Nominal
2. Ordinal
3. Interval
4. Ratio
Types of Measurement Scales
Nominal Scales –
✓there must be distinct classes but these classes have no
quantitative properties. Therefore, no comparison can be
made in terms of one being category, being higher than
the other;
✓involves CLASSIFICATION or CATEGORIZATION based on
one or more distinguishing characteristics.
✓Mutually exclusive and exhaustive.
✓Cannot be meaningfully added, subtracted, ranked or
averaged.
For example - there are two classes for the variable gender --
males and females. There are no quantitative properties for
this variable or these classes and, therefore, gender is a
nominal variable.
Other Examples:
country of origin Disorders listed in DSM V
Yes/No responses
animal or non-animal married vs. single
Nominal Scale
• Sometimes numbers are used to designate category
membership
Example:
Country of Origin
1 = United States 3 = Canada
2 = Mexico 4 = Other
• However, in this case, it is important to keep in mind
that the numbers do not have intrinsic meaning
Ordinal Scales
• There are distinct classes but these classes have a
natural ordering or ranking. The differences can be
ordered on the basis of magnitude (property of
moreness); have no absolute zero point.
Example:
Final position of horses in a thoroughbred race.
The horses finish first, second, third, fourth, and so
on. The difference between first and second is not
necessarily equivalent to the difference between
second and third, or between third and fourth.
Ordinal Scales
• Doesnot assume that the intervals between
numbers are equal.
Example:
finishing place in a race (first place, second place)
• For example:
• Heart beats per minute has a very natural zero point.
Zero means no heart beats. Weight (in grams) is also a
ratio variable. Again, the zero value is meaningful, zero
grams means the absence of weight. The number of
intimate relationships a person has had 0 quite literally
means none. A person who has had 4 relationships has
had twice as many as someone who has had 2.
• Each of these scales have different properties (i.e.,
difference, magnitude, equal intervals, or a true zero
point) and allows for different interpretations
Reference Group - any group of people against which test scores are
compared.
Types of Norms
Norms can be Developed:
• Locally
• Regionally
• Nationally
Normative Data Can be Expressed By:
• Percentile Ranks
• Age Norms
• Grade Norms
Local Norms
• Test users may wish to evaluate scores on the basis of
reference groups drawn from specific geographic or
institutional setting.
For Example
Norms can be created employees of a particular
company or the students of a certain university.
- Age
- Sex
- Occupation
- Education Level
Or any other variable that may have a significant impact on test scores
or yield comparisons of interest.
Percentile Ranks
• The most common form of norms and is the simplest method of
presenting test data for comparative purposes.
• The percentile rank represents the percentage of the norm group that
earned a raw score less than or equal to the score of that particular
individual.
For example, a score at the 50th percentile indicates that the individual
did as well or better on the test than 50% of the norm group.
For Example - if most people in class get an F on a test and you get a D,
your performance in comparison to the normative group is good.
However, overall your score is not good.
• Theoretically, the normal curve is bell shaped with the highest point at its
center. The curve is perfectly symmetrical, with no skewness (i.e., where
symmetry is absent). If you fold it in half at the mean, both sides are exactly
the same.
•In theory, the distribution of the normal curve ranges from negative infinity
to positive infinity.
Skewness
• Skewness is the nature and extent to which symmetry is
absent.
Positive Skewness - when relatively few of the scores fall at
the high end of the distribution.
For Example - positively skewed examination results may indicate that a test was too difficult.
A RELIABLE TEST
Produces similar scores across various
conditions and situations, including different
evaluators and testing environments.
Reliability
• Reliability refers to a measure’s ability to capture an
individual’s true score, i.e. to distinguish accurately
one person from another
• While a reliable measure will be consistent,
consistency can actually be seen as a by-product of
reliability, and in a case where we had perfect
consistency (everyone scores the same and gets the
same score repeatedly), reliability coefficients could
not be calculated
How do we account for an individual who
does not get exactly the same test score
every time he or she takes the test?
1. Test-taker’s temporary psychological or
physical state
2. Environmental factors
3. Test form
4. Multiple raters
TYPES OF RELIABILITY
1. Test-retest
2. Split half (internal consistency)
3. Inter-rater
4. Alternate forms
TEST-RETEST RELIABILITY
• Suggests that subjects tend to obtain the same score when tested at
different times.
• Correlation between time 1 & time 2 (Product-moment
correlation (r))
• Test-retest reliability refers to the consistency of participant’s
responses over time (usually a few weeks, why?)
• Assumes the characteristic being measured is stable over time—
not expected to change between test and retest.
Split-Half Reliability
•Sometimes referred to as internal
consistency
•Indicates that subjects’ scores on
some trials consistently match their
scores on other trials.
Internal Consistency Reliability
• Relevant for measures that consist of more than 1 item
(e.g., total scores on scales, or when several behavioral
observations are used to obtain a single score).
• Internal consistency refers to inter-item reliability, and
assesses the degree of consistency among the items in a
scale, or the different observations used to derive a
score.
• Want to be sure that all the items (or observations) are
measuring the same construct.
Internal Consistency
• We can get a sort of average correlation among items to
assess the reliability of some measure (Cronbach’s alpha)
• Correlations amongst multiple items in a factor
• As one would most likely intuitively assume, having more
measures of something is better than few
• It is the case that having more items which correlate with
one another will increase the test’s reliability
• Is a multi-item scale measuring a single concept?
• Are items in scale consistent with one another?
Estimates of Internal Consistency
• Item-total score consistency
• Split-half reliability: randomly divide items into 2
subsets and examine the consistency in total scores
across the 2 subsets (any drawbacks?)
• Cronbach’s Alpha: conceptually, it is the average
consistency across all possible split-half reliabilities
• Cronbach’s Alpha can be directly computed from data
INTER-RATER RELIABILITY
Involves having two raters independently
observe and record specified behaviors,
such as hitting, crying, yelling, and getting
out of the seat, during the same time period.
TARGET BEHAVIOR
A specific behavior the observer is looking to
record
• If a measurement involves behavioral
ratings by an observer/rater, we would
expect consistency among raters for a
reliable measure
• Best to use at least 2 independent raters,
‘blind’ to the ratings of other observers
• Precise operational definitions and well-
trained observers improve inter-rater
reliability
ALTERNATE FORMS RELIABILITY
• Also known as equivalent forms reliability or
parallel forms reliability
• Obtained by administering two equivalent
tests to the same group of examinees
• Items are matched for difficulty on each
test
• It is necessary that the time frame between
giving the two forms be as short as possible
RELIABILITY COEFFICIENTS
•The statistic for expressing reliability.
•Expresses the degree of consistency in
the measurement of test scores.
•Denoted by the letter r with two
identical subscripts (rxx).
FACTORS AFFECTING RELIABILITY
1. Test length
2. Test-retest interval
3. Variability of scores
4. Guessing
5. Variation within the test situation
Reliability Interpretation
<.6 = not reliable
.6 = OK
.7 = reasonably reliable
.8 = good, strong reliability
.9 = excellent, very reliable
>.9 = potentially overly reliable or redundant
measurement – this is subjective and whether a scale is
overly reliable depends also on the nature what is being
measured
Estimating Reliability
✓Reliability can range from 0 to 1.0.
✓When a reliability coefficient equals 0, the
scores reflect nothing but measurement error.
✓Rule of Thumb: measures with reliability
coefficients of 70% or greater have acceptable
reliability
How Many Items per Factor?
• More items -> greater reliability
(The more items, the more ‘rounded’ the
measure)
• Min. = 3
• Max. = unlimited
RELIABILITY
Reproducibility of a measurement
VALIDITY
VALIDITY
The extent to which an instrument
actually measures what it purports to
measure.
Validity
• Validity refers to the question of whether
our measurements are actually hitting on
the construct we think they are.
• While we can obtain specific statistics for
reliability (even different types), validity is
more of a global assessment based on the
evidence available.
• Denotes the extent to which an instrument is
measuring what it is supposed to measure.
• We can have reliable measurements that are
invalid
• Classic example: The scale which is
consistent and able to distinguish from one
person to the next but actually off by 5
pounds.
Face Validity
• Prima facie extent to which an item is judged to reflect target
construct
• Refers to the extent to which a measure ‘appears’ to measure
what it is supposed to measure
• Not statistical—involves the judgment of the researcher (and the
participants)
• A measure has face validity—’if people think it does’
• Just because a measure has face validity does not ensure that it is
a valid measure (and measures lacking face validity can be valid).
Validity Criteria in Psychological Testing
1. Content validity
2. Criterion validity
• Concurrent
• Predictive
3. Construct-related validity
• Convergent
• Discriminant
Content Validity
1. Whether the individual items of a test represent what you actually want
to assess.
2. Items represent the kinds of material (or content areas) they are supposed
to represent
• Are the questions worth a flip in the sense they cover all domains of a
given construct?
• E.g. job satisfaction = salary, relationship w/ boss, relationship w/
coworkers etc.
3. Systematic examination of the extent to which test content covers a
representative sample of the domain to be measured – e.g. sources,
• existing literature
• expert panels
• qualitative interviews / focus groups with target sample
CRITERION VALIDITY
• A method for assessing the validity of an instrument
by comparing its scores with another criterion
known already to be a measure of the same trait or
skill.
• Test Construction
• Test Tryout
• Item Analysis
• Test Revision
TEST CONCEPTUALIZATION
SOME PRELIMINARY
QUESTIONS
• What is the test designed to measure?
• Who will use this test?
• Who benefits from an administration of
this test?
• What is the ideal form of the test?
• Is there a need for this test?
• What is the objective of this test?
Norm Reference vs. Criterion-Reference
• Good item = High Score • High Scores does not count
and is irrelevant
• Mastery
• Accuracy
PLEASE RATE THE EMPLOYEE ON ABILITY TO COOPERATE AND GET ALONG WITH FELLOW
EMPLOYEES:
EXCELLENT____/____/____/____/____/____/____/____/UNSATISFACTORY
SCALING METHODS
LIKERT SCALE
• Consists of
5 alternative responses, usually on an
agree/disagree or approve/disapprove basis (sometimes 7)
Driving without seatbelt is:
1 2 3 4 5 6 7 8 9 10
Never Always
Justified Justified
• Test takers are presented with pairs of stimuli which they are
asked to compare
SELECTED-RESPONSE CONSTRUCTED-RESPONSE
SELECTED-RESPONSE ITEM
MULTIPLE CHOICE
Has 3 elements :
• a stem,
• a correct alternative or option, and
• several incorrect alternatives or options variously
referred to as distractors or foils.
A good multiple-choice item in an
achievement test :
a. Has one correct alternative
b. Has grammatically parallel alternatives
c. Has alternatives of similar strength
d. Has alternatives that fit grammatically with the stem
e. Includes as much of the item as possible in the stem to avoid unnecessary
repetition
f. Avoids ridiculous distractors
g. Is not excessively long
h. All of the above
i. None of the above
Matching Item
• The test taker is presented with two columns, premises on
the left and responses on the right.
• Must be homogeneous
For example…
1. Britney Spears a. Hero
2. Backstreet Boys b. Drop it Like Its Hot
3. N’Sync c. Beautiful
4. Jennifer Lopez d. If You Wanna Be My Lover
5. Christina Aguilera e. One Time
6. Boyzone f. Get Down
7. Mariah Carey g. Pretty Boy
8. M2M h. Pyramid
9. Snoop Dogg i. Oops! I Did It Again
10. Destiny’s Child j. Lets Get Loud
11. Spice Girls k. Its Raining Men
l. This I Promise You
TRUE / FALSE ITEM
True False
True False
True False
CONSTRUCTED-
RESPONSE ITEM
Completion item
• Requires the examinee to provide a word or phrase
that completes a sentence.
• Should be worded so that the correct answer is
specific.
• For example:
• The Culinary Arts focuses mainly on ________.
• A “mariposa” is a type of ________.
SHORT-answer item
216
Types of Tests
• Ability Tests
• Achievement Tests
• evaluates what an individual has learned
• measures prior activity
• Aptitude Tests
• evaluates what an individual is capable of learning
• measures capacity or future potential
• Intelligence Tests
• Measures a person’s general potential to solve problems,
adapt to novel situations and profit from experience
217
Types of Tests
• Personality Tests: Objective & Projective
• Objective Personality Tests
• present specific stimuli and ask for specific responses
(e.g. true/false questions) .
• Projective Personality Tests
• present more ambiguous stimuli and ask for less specific
responses (e.g. inkblots, drawings, photographs,
Rorschach, TAT)
218
APPLICATIONS OF TESTS
Application of Psychological Measurement
➢ Educational Testing
➢Personnel Testing
➢ Clinical Testing
Educational Testing
• Intelligence tests and achievement tests are used from an
early age in the U.S and Canada. From kindergarten on, tests
are used for placement and advancement.
•Educational institutions have to make admissions and
advancement decisions regarding students. e.g., SAT, GRE,
subject placement tests
• Used to assess students for special education programs.
Also, used in diagnosing learning difficulties.
•Guidance counselors use instruments for advising students.
•Investigates school curriculum.
Personnel Testing
➢The investigator has the responsibility to make a careful evaluation of its ethical
acceptability.
➢The researcher must evaluate whether participants are considered “Subject at risk”
or “Subject at minimal risk” - No appreciable risk (physical risk, mental harm).
➢The principal investigator always retains the responsibility for ensuring ethical
practice in research. That is, the principal researcher is responsible for the ethical
practices of collaborators, assistants, employees, etc. (all of whom are also responsible
for their own ethical behavior).
➢After the data are collected the investigator provides participants with
information about the nature of the study and attempts to remove any
misconceptions that may have arisen.
➢The investigator has the responsibility to detect and remove any undesirable
consequences to the participant that may occur due to the research.
Define what each test measures and what the test should be used for.
Provide test takers the information they need to be familiar with the
coverage of the test, the types of question formats, the directions, and
appropriate test-taking strategies. Strive to make such information equally
available to all test takers.
Tell test takers or their parents/guardians how long scores will be kept on
file and indicate to whom and under what circumstances test scores will or will
not be released.
Responsibility of The Tester
• 1. Individual or
Institutional
• 2. Comparative or Absolute
THE END!
THANK YOU!!!