0% found this document useful (0 votes)
20 views4 pages

Program Evaluation

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
20 views4 pages

Program Evaluation

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 4

The nature of program evaluation

1. 1. The Nature of Program Evaluation Carlo Magno, PhD Counseling and


Educational Psychology Department
2. 2. Answer the following questions: • Why is program evaluation needed?
• What are the roles of a professional program evaluator?
3. 3. Program evaluation is needed because… • Policy makers need good
information about the relative effectiveness of the program. – Which
programs are working well? – Which poorly? – What are the program’s
relative cost and benefits? – Which parts of the program are working? –
What can be done with those parts that are not working well? – Have al
parts of the program been thought through carefully at the planning
stage? – What is the theory or logic model for the program effectiveness?
– What adaptations would make the program more effective?
4. 4. Program Evaluation • Systematic investigation of the merit, worth or
significance of an object (Scriven, 1999), hence assigning “value” to a
program’s efforts means addressing those three inter-related domains: –
Merit (or quality) – Worth (or value, i.e., cost-effectiveness) –
Significance (or importance) • The identification, clarification, and
application of defensible criteria to determine an object’s value in relation
to those criteria (Fitzpatrick, WEorthen, & Sanders, 2004).
5. 5. Prerequisite to evaluation • Need a program: - an organized action –
Direct service interventions – Community mobilization efforts – Research
initiatives – Surveillance systems – Policy development activities –
Outbreak investigations – Laboratory diagnostics – Communication
campaigns – Infrastructure building projects – Training and education
services – Administrative systems
6. 6. Inquiry and Judgment in Evaluation • (1) Determining standards for
judging quality and deciding whether those standards should be relative
or absolute. • (2) Collecting relevant information • (3) Applying the
standards to determine value, quality, utility, effectiveness, or
significance.
7. 7. Evidence of value and judgement: • What will be evaluated? (i.e., what
is "the program" and in what context does it exist?) • What aspects of the
program will be considered when judging program performance? • What
standards (i.e., type or level of performance) must be reached for the
program to be considered successful? • What evidence will be used to
indicate how the program has performed? • What conclusions regarding
program performance are justified by comparing the available evidence to
the selected standards? • How will the lessons learned from the inquiry be
used to improve public health effectiveness?
8. 8. Difference between Research and Evaluation • Purpose • Approaches •
Who sets the agenda? • Generalizability of results • Criteria and standards
• Preparation
9. 9. Difference in Purpose • Research – Add knowledge in a field,
contribute to theory – Seeks conclusion • Evaluation – Help those who
hold a stake in whatever is being evaluated – Leads to judgments
10. 10. Difference in Approaches • Research – Quest for laws – Explore and
establish causal relationships • Evaluation – Describing a phenomenon
may use causal relationships – Causal relationships will depend on the
needs of the stakeholders
11. 11. Difference on who sets the agenda • Research – The hypothesis
investigated is chosen by the researcher and the appropriate steps in
developing the theory. • Evaluation – Questions to be answered comes
form many sources (stakeholders). – Consults with stakeholders to
determine the focus of the study.
12. 12. Difference in generalizability of results • Research – Methods are
designed to maximize generalizability to many different settings •
Evaluation – Specific to the context which evaluation object rests.
13. 13. Difference in Criteria and standards • Research – Internal validity
(causality), – external validity (generalizability) • Evaluation – Accuracy
(corresponding to reality) – Utility (results serve practical information) –
Feasibility (realistic, prudent, diplomatic, frugal) – Propriety (done
legally and ethiocally)
14. 14. Difference in Preparation • Research – In depth training on a single
discipline in their field of inquiry. • Evaluation – Responds to the needs
of clients and stakeholders with many information needs and operating in
many different settings. – Interdisciplinary: Sensitive to a wide range of
phenomenon that they must attend to. – Familiar with a wide variety of
methods – Establish personal working relationships with clients
(interpersonal and communication skills)
15. 15. Competencies needed by professional Evaluators (Sanders, 1999) •
Ability to describe the object and context of an evaluation •
Conceptualize appropriate purposes and framework for evaluation •
Identify and select appropriate evaluation questions, information needs,
and sources of information • Select mans for collecting and analyzing
information • Determine the value of the object of an evaluation •
Communicate plans and results effectively to audiences • Manage the
evaluation • Maintain ethical standards • Adjust to external factors
influencing the evaluation • Evaluate the evaluation
16. 16. Purposes of Evaluation • Talmage (1982) – Render judgment in the
worth of the program – Assist decision makers responsible for deciding
policy – Serve a political function • Rallis and Rossman (2000) –
Learning, helping practitioners and others better understand and interpret
their observations •
17. 17. Purposes of Evaluation • Weiss (1988) and Henry (2000) – Bring
about social betterment • Mark, Henry, and Julnes (1999) – Betterment –
alleviation of social problems, meeting of human needs • Chelimsky
(1997) – takes a global perspective: new technologies, demographic
imbalance, environmental protection, sustainable development, terrorism,
human rights
18. 18. Purposes of Evaluation • House and Howe (1999) – Foster deliberate
democracy-work to help less powerful stakeholders gain a voice and to
stimulate dialogue among stakeholders in a democratic fashion. • Mark,
Henry, and Julnes (1999) – Assessment of merit and worth – Oversight
and compliance – Program and organizational improvement – Knowledge
development
19. 19. Roles of the Professional Evaluator • Rallis and Rossman (2000) –
Critical friend: “someone the emperor knows and can listen to. She is
more friend than judge, although she is not afraid to offer judgment” (p.
83) • Schwant (2001) – Helping practitioners develop critical judgment
20. 20. Roles of the Professional Evaluator • Patton (1996) – Facilitator –
Collaborator – Teacher management consultant – OD specialist – Social-
change agent • Preskilll and Torres (1999) – Bring about organizational
learning and instilling a learning environment
21. 21. Roles of the Professional Evaluator • Mertens (1999), Chelimsky
(1998), and Greene (1997) – Including the stakeholders as part of the
evaluation process • House and Howe (1999) – Stimulating dialogue
among various groups
22. 22. Roles of the Professional Evaluator • Bickman (2001) and Chen
(1990) – Take part in program planning – Help articulate program
theories or logic model • Wholey (1996) – Help policy makers and
managers select the performance dimension to be measured as well as the
tools to use in measuring those dimensions
23. 23. Roles of the Professional Evaluator • Lipsey (2000) – Provides
expertise to track things down, systematically observe and measure them,
and compare, analyze, and interpret with a good faith attempt at
aobjectivity.
24. 24. Roles of the Professional Evaluator • Fitzpatrick, Worthen, and
Sanders (2004) – Negotiating with stakeholders group to define the
purpose of evaluation – Developing contracts – Hiring and overseeing
staff – Managing budgets – Identifying disenfranchised or
underrepresented groups – Working with advisory panels – Collecting
and analyzing and interpreting qualitative and quantitative information –
Communicating frequently with various stakeholders to seek input into
the evaluation and to report results – Writing reports – Considering
effective ways to disseminate information – Meeting with the press and
other representatives to report on progress and results – Recruiting others
to evaluate the evaluation
25. 25. Examples of evaluation use in Education • To empower teachers to
have more say about how school budget are allocated • To judge the
quality of the school curricula in specific content areas • To accredit
schools that meet minimum accreditation standards • To determine the
value of a middle school’s block scheduling • To satisfy an external
funding agency’s demands for reports on effectiveness of school
programs it supports • To assist parents and students in selecting schools
in a district with school choice • To help teachers improve their reading
program to encourage more voluntary reading
26. 26. Examples of evaluation use in other public and Nonprofit sectors • To
decide whether to implement an urban development program • To
establish the value of a job-training program • To decide whether to
modify a low-cost housing project’s rental policies • To improve a
recruitment program for blood donors • To determine the impact of a
prison’s early release program in recidivism • To gauge community
reaction to proposed fire-burning restrictions to improve air quality • To
determine the cost-benefit contribution of a new sports stadium for a
metropolitan area
27. 27. Examples of evaluation use in Business and industry • To improve a
commercial product • To judge the effectiveness of a corporate training
program on teamwork • To determine the effect of a new flextime policy
on productivity, recruitment, and retention • To identify the contributions
of specific programs to corporate profits • To determine the public’s
perception of a corporation’s environmental image • To recommend ways
to improve retention among younger employees • To study the quality of
performance-appraisal dfeedback
28. 28. Formative and Summative Evaluation • Formative – provide
information for program improvement. Judgment of a part of a program. •
Summative – concerned with providing information to serve decisions or
assist in making judgments about program adoption, continuation or
expansion.

You might also like