Академический Документы
Профессиональный Документы
Культура Документы
]
Chapter 5: Program Evaluation: Concepts
and Perspectives
Concepts and perspectives
James W.AltschuldJames T.Austin
At first it might seem strange or even somewhat misplaced that a handbook dealing with
research in psychology would contain a chapter on evaluation. After all, evaluation is
an activity that is carried out for a purpose quite different from that of research, that is,
to guide practical decision making in applied settings. Add to this the fact that although
there is scientific study of evaluation, many scholars and empirical researchers still tend
to look with disdain at the topic. It is perceived to be poor research, that is, research
often conducted with much less control than the norm in articles in prestigious, refereed
academic journals. Beyond these issues is the further concern that evaluation is usually
not based on theoretical understandings. For these reasons collectively, some social
scientists may not ascribe much value to evaluation.
So we must begin by addressing the subtle, implicit questions in the first paragraph.
Why should evaluation be in this book, and why is it of importance and value for
practicing psychologists, psychology students, educational psychologists, and for that
matter, others in social science? We will provide several answers to this question.
A first answer is that psychologists in a variety of environments (businesses,
government organizations, educational institutions, etc.), in the course of their work,
might be called on to evaluate an activity or a program whether or not they have been
trained as evaluators or have in-depth comprehension of it. Davidson (2002), for
instance, shows how this could occur for industrial-organizational psychologists. They
are there in the situation, and they have had methodological training, so they must
be able to do it. Essentially, this is a reason by default. Secondly, evaluation has a
number of unique methodological and substantive content areas that would be useful
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 4 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
for psychologists to understand and be able to use in their work. A third answer is that
some psychologists will undoubtedly lead or participate in evaluations as their main
professional focus.
Therefore, the purposes of this chapter are to provide an overview of the nature of
evaluation, to define it conceptually, to show its relationship to research, and to give
a sense of what might have to be done to conduct an evaluation. Specifically, the
chapter is organized as follows: Some background related to defining evaluation is
described; defining evaluation with a comparison of evaluation and research, based
on the definition, comes next; and implications of the definition for practice and generic
types of [p. 76
] evaluation are the latter parts of the text. Last, a hands-on exercise,
which helps in pulling together ideas, will form the chapter conclusion.
Toward a Definition of Evaluation
In 2002, the lead author (Altschuld, 2002) wrote about a presentation that he had
attended at an annual meeting of the American Evaluation Association (http://
www.aea.org), a presentation that is relevant for defining evaluation. In a panel
conducted by individuals who were nearing completion of graduate programs or
who had just received their degrees, perspectives and views of the quality of the
educational experience were being shared and examined. The panelists explained
their backgrounds and then briefly discussed what they thought about their graduate
education in terms of learning and preparation for the job market.
One participant, a recent graduate, elaborated about his extensive coursework in
statistics and measurement and the skills he had gained in these areas. Later when
questions from the audience were appropriate, the coauthor asked him what in this set
of courses would constitute training for the practice of evaluation. Stated alternatively,
what really is evaluation, what does it mean to be an evaluator, what does an evaluator
do, how should one train for work as an evaluator, what should an evaluator know, and
what specific skills should an evaluator have? After all, would it not be better to think of
what he had taken as simply methodology and just call it that? Further, he was asked
if he had any courses that could be somehow thought of or construed as specifically
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 5 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
emphasizing evaluation concepts, whatever those concepts might be, even if they were
distantly related to evaluation.
He pondered for a moment and to his credit replied that he had never considered this
line of inquiry, and in reality, his background could not be viewed as actually dealing
with evaluation. There was nothing specific in it related to, or called, evaluation.
The point here is not to discount being well grounded in methodology or that evaluators
need sound methodological understandings in order to be successful in the field. That
is neither an issue nor debatable. Indeed, our advisees are required to take many
quantitative and qualitative methods courses. Rather, methodology, although necessary
and critically important for evaluation, constitutes, by itself, an insufficient condition for
being adequately trained in the field and understanding its substance.
Evidence to support the above assertion can be found in international studies of
evaluation training conducted by Altschuld, Engle, Cullen, Kim, and Macce (1994) and
by Engle and Altschuld (2003/2004). Also, the writing of individuals concerned with
the training of evaluators is pertinent here. For example, see Scriven (1996), Altschuld
(1995, 2002), Mertens (1994), and the recent work of King, Stevahn, Ghere, and
Minnema (2001) in regard to the critical skill domains required to be an evaluator.
In 1994, Altschuld and his colleagues identified 47 university-based evaluation training
programs, and in a repeat of the same study in 2001, Engle and Altschuld found 29
programs. Although the number of programs is smaller, it is also apparent from studying
trends across the two time periods that there has been a slow, steady, and noticeable
emergence of a unique set of courses with specialized evaluation content. That content
is quite distinct from research and/or measurement methodology. Examples include
evaluation theory and models, applied evaluation design for field settings, evaluation
and program planning, evaluation and its relationship to the formation of policy, cost-
benefit analysis, readings and research in evaluation, and others (Engle, Altschuld,
& Kim, 2005). Nearly all of these topics are not covered in traditional methodology
offerings or other courses in psychology and other social science disciplines, or they are
covered only tangentially.
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 6 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
Several writers have discussed the skill sets needed for evaluation work. Mertens
(1994) presented a taxonomy derived rationally from literature reviews and discussions
with colleagues. It consisted of grouping the knowledge and skills into the following
areas: research methodology, borrowed from other areas, and unique to specific
disciplines. The work of King and her associates (King et al., 2001) is particularly
relevant because of its combination of rational and empirical methods. King etal.
(2001) engaged 31 individuals representing a diverse set [p. 77
] of backgrounds
and contexts of practice in rating the relative importance for 69 domains or items of
evaluation capability. Using what they describe as a multiattribute consensus-building
process, they reported that they achieved significant agreement on 54 of the 69 (78%)
competencies. King etal. (2001) proposed four skill domains in which evaluators must
be successful, not only doing evaluations but obtaining the results of evaluations
accepted and used in making decisions. Those domains are (a) systematic inquiry,
(b) competent evaluation practice, (c) general skills for evaluation practice, and (d)
evaluation professionalism. Stevahn, King, Minnema, and Ghere (2005) followed up
with additional research and refinement. Finally, a project by the Canadian Evaluation
Society (CES) that focused on advocacy and professional development also addressed
capacity building. The literature review by McGuire (2002) focused on the benefits,
outputs, processes, and knowledge elements of evaluation.
Given the nature of emerging curricula, the evolving dimensions of evaluative thinking,
and the skill domains, what is evaluation? In what ways does it differ from research, and
in what ways is it similar?
Defining Evaluation and Some Useful
Distinctions
From many sources (Scriven, 1991; Stuffle-beam, 1971, 1980, 2002; Worthen,
Sanders, & Fitzpatrick, 1997), there are common features to the definition of evaluation.
Synthesizing across these authors, we could define or think of evaluation as a process
or set of procedures for providing information for judging decision alternatives or for
determining the merit and worth of a process, product, or program.
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 7 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
Imbedded in this collated, conceptual definition are subtleties that help us to distinguish
between research and evaluation, as well as to explain the practice of evaluation. In
Table 5.1, evaluation and research are briefly compared in terms of several dimensions
emanating from the definition. When examining the table, keep in mind that the
comparisons are somewhat exaggerated to aid the discussion and your understanding.
Probably a better way to think about the linkage between evaluation and research is to
use Venn diagrams with areas of overlap as well as areas of uniqueness.
In Table 5.1, evaluation is presented as being conducted in field situations to guide and
facilitate the making of decisions. This immediately implies that evaluators must identify
who the decision makers and key stakeholding groups are, the degree of their influence
on the entity being evaluated, and how their value positions and perspectives will
affect the evaluation. This point is so important that audience identification for results
and involvement in the process of evaluation is highlighted as the first standard in the
Program Evaluation Standards (American Evaluation Association, 1995). Evaluators
essentially work at the behest of decision makers, and the needs of multiple decision-
making audiences have to be factored into all evaluations. Evaluators have to carefully
maintain their objectivity and avoid being co-opted (handling their results to favor a
political or vested position), given the social and political environments in which they
exist.
Researchers, on the other hand, in the best of all possible worlds, are not working on
someone else's priorities but more in terms of their own curiosity about phenomena.
Their wonderment, their questioning, their desire to know and understand, and even
their awe are what drives the enterprise, not what someone else wants. This is an ideal
point of view, but at the same time, we recognize that politics and values can influence
research. Political forces frequently push certain substantive areas or stress what
methodologies might be acceptable at a point in time. Despite that, the ideal is one that
we should strive for and cherish with regard to research.
Looking at the comparisons in the second row of the table, the relationship of evaluation
to theory is not as strong or prominent as it is in research. Although there are theory-
driven evaluations and the idea of theory in a variety of forms and connotations
does affect evaluation (see Chen, 1990), for the most part evaluations are directed
toward practical concerns of decision makers and stakeholders, not toward theory.
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 8 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
Proving (demonstrating) the validity of a theory or proposition versus depicting
how well something works in a field setting (its [p. 78
] [p. 81
]
Generic Types of Evaluation
Reviewing the prior sections of this chapter, we find that different types of evaluation
have been woven into or alluded to in the discussion. Other classifications of evaluation
are possible with the one given below having been used in the teaching of evaluation
classes at The Ohio State University. For another example of a classification scheme
see the Worthen, Sanders, and Fitzpatrick (1997) text cited earlier.
In Table 5.3, an informal listing of types of evaluation is offered. It is intended to convey
the range of activities often conducted under the rubric or title of evaluation. With the
entries in the table, a more comprehensive picture (in conjunction with the definition of
evaluation, its comparison to research, and key aspects of what evaluators do) should
now be coming into view.
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 15 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
The entries in the table are, for the most part, self explanatory with perhaps the
exception of needs assessment (NA). Need is defined as the measurable discrepancy
between the what-is status and the what-should-be status of an entity. Generally,
the what-should-be status comes from collective value judgment. Examples [p. 83
]
would be the following: What should be the outcomes of a high school education? What
should be the wellness level of the U.S. population? What constitutes reasonable and
cost-feasible access to drug therapies? And so forth. Current or what-is status can
be ascertained from a variety of methods including records, surveys, observations,
databases, and so forth. Numerous methodological approaches are available for
assessing needs, and over the last 20 years many texts have been written describing
the nature of needs and procedures for assessing them (Altschuld & Witkin, 2000;
McKillip, 1998; Witkin & Altschuld, 1995).
After the states have been determined, the process continues with prioritizing the
resulting needs or discrepancies and transforming the high priorities needs into
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 16 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
organizational action plans to resolve or reduce the problem underlying or causing the
discrepancy. In other words, decisions to design and implement programs should be
based on clearly delineated needs. From some perspectives, needs assessment might
appear to be more of a planning topic type of activity, so why was it placed in Table 5.3
as a type of evaluation work?
Several reasons support such placement. A well-implemented NA illuminates major
problems, what key variables relate to them or are causing them, and what likely
solutions might be. It is the essential condition for designing and implementing good
programs; and by virtue of its analytic properties, it should enhance the overall
evaluation of solution strategies that are developed and used. Therefore, one of the
reasons is that evaluation and NA are inextricably intertwined; they are two ends of one
continuum.
A second reason is basically an offshoot of the first. A small (almost informal) national
group of needs assessors felt that they could not survive as a small group; therefore,
in the late 1980s they began to cast about for a larger society with which to affiliate.
They chose the American Evaluation Association (AEA), based on the recognition that
evaluation and NA were highly related. Now, as a partial outgrowth of that event, most
evaluators see NA as important to the practice of evaluation. Other entries in Table 5.3
could have been discussed in a similar fashion. Needs assessment was singled out as
an exemplar to demonstrate the thought process.
Toward the Future
As we near the end of this exposition, consider for a moment evaluation as a
profession. As a discipline, evaluation emerged in psychology and in education during
the 1930s. Ralph Tyler's Eight-Year Study and notions of curriculum evaluation were
pioneer landmarks in education, and Rossi, Lipsey, and Freeman (2004) point to
field studies of evaluation conducted by Kurt Lewin and other social psychologists.
Nonetheless, these were isolated instances and were not formalized in terms of
department or area status, curriculum, journals, and professional organizations. Rossi
et al. (2004) went on to define a boom period in evaluation following World War II and
pointed out its emergence as a specialized area during the 1970s. Their indicators of
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 17 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
emergence included books, journals, and associations. The books that have shaped
the discipline are numerous. They include pioneering classics (Suchman, 1967),
methodological contributions (Campbell & Stanley, 1966; Cook & Campbell, 1979), and
textbooks across multiple editions (Fitzpatrick, Sanders, & Worthen, 2003; Rossi etal.,
2004). Prominent journals devoted to evaluation range from the first, Evaluation Review
(since 1976), to the American Journal of Evaluation (since 1978), and Evaluation
and Program Planning (since 1980). Two key professional associations in the United
States are the American Evaluation Association (http://www.eval.org), created in 1986
through a merger between the Evaluation Network (ENet) and the Evaluation Research
Society (ERS), and the American Educational Research Association Evaluation
Division (http://aera.net). Many other professional organizations exist across nations,
including the Canadian Evaluation Association (http://www.evaluationcanada.ca/), the
European Evaluation Society (http://www.europeanevaluation.org), the U.K. Evaluation
Society (http://www.evaluation.uk), and the Australasian Evaluation Society (http://
www.parklane.com.au/aes). Instances of these features are depicted in Table 5.4.
[p. 84
]
Why focus on the future of evaluation? We might reply, For one major reason.
The demand for evaluation is increasing rapidly, even as university-based training
programs are tending to disappear (Engle, Altschuld, & Kim, in process). Thus, a
priority need is to resuscitate graduate or other types of training in this important
discipline. Perhaps the multidisciplinary nature of program evaluation contributes to
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 18 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
training issues due to the fact that evaluation training is housed in varied colleges and
departments (education, psychology, social work, public policy, and management).
Clearly, affiliation between psychologists and evaluators could be valuable for both
groups. Davidson (2002), in the aforementioned article, indicated that there is a
great interest in the evaluation community about industrial-organizational psychology
issues and that many opportunities exist for forms of involvement. Edwards, Scott,
and Raju (2003) edited a wide-ranging application of evaluation to human resource
management. Consider further that the major professional society for evaluators, the
American Evaluation Association, runs a major international discussion list, EVALTALK
(http://bama.ua.edu/archives/evaltalk.html). An informal group that coalesced at a
recent Society for Industrial Organizational Psychology conference, the Strategic
Evaluation Network, also runs an electronic mailing list specifically for people who do
evaluation in organizational settings (see http://evaluation.wmich.edu/archives). The
American Evaluation Association has a Business and Industry (B & I) Topical Interest
Group on evaluation (http://www.evaluationsolutions.com/aea-bi-tig), which welcomes
presentation proposals (and audience members!) for the annual conference each
November (details at http://eval.org).
Other features of the field include knowledge dissemination and globalization. The
Evaluation Center at Western Michigan University is a professional clearinghouse for
information, including bibliographies and checklists. The checklist project provides a
number of useful checklists for various evaluation functions (http://www.wmich.edu/
evalctr/checklists/). The checklists are organized under categories: evaluation
management, evaluation models, evaluation values and criteria, metaevaluation, and
other. There is even a checklist for developing checklists by Stufflebeam [p. 85
]
(http://www.wmich.edu/evalctr/checklists/cdc.htm). Globalization is clearly evident at the
Canadian Evaluation Association's Web site (http://consultation.evaluationcanada.ca/
resources.htm). There are links to evaluation standards, for example from associations
(Joint Committee on Standards for Educational Evaluation, 1994) and from nations
(Germany, Canada, United States).
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 19 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
Conclusion
In this postscript, we attempt to tie together the strands of evaluation presented in
this chapter. First, the placement of evaluation in a psychology research handbook
makes good sense for several reasons. Readers may be called on to consume or
produce evaluations. Respectively, these two terms mean reading and reacting to
evaluation reports or designing, conducting, and reporting a program evaluation. An
individual trained in social psychology, for instance, might be asked to conduct a needs
assessment in order to design and evaluate a community-based collaborative conflict
resolution program. On the other hand, an industrial-organizational psychologist might
be charged with doing a training needs assessment in order to design a leadership
development program for middle and top managers.
Second, the nature of evaluation can be viewed as dynamic with exciting emergent
dimensions and twists. Third, using the definition of evaluation as a process or set of
procedures for providing information for judging decision alternatives or for determining
the merit and worth of a process, product, or program, we presented useful distinctions
as to several types of evaluation (Table 5.3). Fourth, we discussed the structure of
evaluation as a profession in a limited way to provide some context. Two key sets of
professional standards are the five AEA Guiding Principles (1995) and the Program
Evaluation Standards (Joint Committee on Standards for Educational Evaluation,
1994). Fifth and finally, we have provided three brain-storming exercises to tie your new
understandings together into a scaffolded mental model. As you practice evaluation,
read evaluation research, and talk to evaluation practitioners, this scaffolding will
eventually disappear, and you will be able to use the mental model for practice and
research.
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 20 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
Exercises
Pulling Your Understandings Together
In evaluation workshops conducted for the National Association of School
Psychologists, Altschuld and Ronning (2000) developed a brainstorming activity that
fits well with the content of this chapter. To do this exercise, pretend that you are in
the specialized area of school psychology or modify the activity in accord with the
psychological work context that has most meaning for you. The activity consists of three
brainstorming or idea-generating parts and then examples of responses to them. Do not
peek ahead, but rather try your hand at the activity before glancing at our ideas about
possible responses. We believe that the activity constitutes an interesting, different, and
good way to summarize the concepts contained in this chapter.
Spend a minute or two thinking about the possibilities for school psychologists to
become involved in program evaluation (or substitute for school psychologists your
particular interest in psychology). Additional material on evaluation exercises and
activities may be found in Mertens (1989).
For example, what decisions are various stakeholders in your district facing? Are there
specific paperwork demands, time limits on referrals, or accountability reports? Do you
work under local guidelines and restrictions, the values of which are questionable? Are
your methods, results, and roles being scrutinized or changed erratically?
First, fill in the blanks from your experiences. These responses may vary widely as each
person works in, or will work in, different settings under different rules and regulations
and with different administrations and colleagues.
[p. 86
] Step 1: List up to 10 reasons why you might be called upon or wish to conduct
an evaluation. Reflect back on what was presented in the chapter, what evaluators do,
and the various types of evaluations. There are no wrong or right answers. Rather the
purpose here is to encourage you to think about all of the possibilities for conducting an
evaluation. Later, when you examine the reasons we supplied, please note that they
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 21 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
are not definitive answers but come from a brainstorming activity very similar to what we
asked you to do.
[p. 87
]
Step 2: In the spaces provided below, identify as many programs as you can in which
evaluations could be conducted. What programs are you or your colleagues currently
involved with that may be appropriate for using evaluation techniques to determine their
worth or merit? For example, school psychologists in private practice may link closely
with community agencies and schools to provide mental health services. Is this more
effective than having the school psychologist assigned to specific buildings provide this
service? Or, you may be responsible for grief counseling of children in cooperation with
the local hospitals. List program titles in the appropriate space, or use a short phrase to
describe the program. List as many as you can.
Program Titles or Descriptive Phrases
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 22 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
Step 3. Speculate about the types of activities evaluators might do in the course of an
evaluation and briefly describe them below.
Sampling of Responses to Brainstorming Activity
Step 1. List reasons why you might be called upon or wish to conduct an evaluation.
[p. 88
] Step 2: What programs are you or your colleagues currently involved with that
may be appropriate for using evaluation techniques to determine their worth or merit?
1. Community involvement programs
2. Districtwide proficiency interventions
3. Reading evaluation (primary)
4. Provision of school psychological services
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 23 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
5. Therapy outcomes in conjunction with
6. Career planning or experience program
7. Reporting test results to parents, etc.
8. Parenting programs
9. Special education programs
10. Multihandicapped programs
11. Community outreach programs
12. Efficacy of multiage classrooms
13. School-based provision of health benefits
14. Head Start
15. All day, everyday kindergarten other providers (e.g., AA)
16. Preschool programs
17. Districtwide testing program
18. Statewide testing program
19. Effectiveness of research efforts
20. Effectiveness of local efforts to serve
Step 3. What types of activities do program evaluators do in the course of an
evaluation?
[p. 89
]
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 24 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
Recommended Readings
The field of evaluation includes educational researchers and psychological researchers.
Among the resources for further understanding are the following: Witkin and Altschuld
(1995); the Encyclopedia of Evaluation (Mathison, 2004); and Rossi et al. (2004).
Bickman and Rog (1998) provide an edited collection of applied research methods
relevant to evaluation; Chelimsky and Shadish (1997) provide an edited collection
of forward-looking chapters on evaluation topics. Caron (1993) presents a Canadian
perspective on the body of knowledge for evaluation practice; Mark, Henry, and Julnes
(1999) articulated a framework for evaluation practice. The 1987 program evaluation
tool kit from Sage is another valuable resource. The evaluation tool kit at the Web site of
the W. K. Kellogg Foundation is exceptionally valuable as a distance resource (it can be
accessed at http://www.wkkf.org/).
References
Altschuld, J. W. Developing an evaluation program: Challenges in the teaching
of evaluation . Evaluation and Program Planning 18(3) (1995). 259265. http://
dx.doi.org/10.1016/S0149-7189%2895%2900014-3
Altschuld, J. W. The preparation of professional evaluators: Past tense and future
perfect . Japanese Journal of Evaluation Studies 2(1) (2002). 19.
Altschuld, J. W.Engle, M.Cullen, CKim, I.Macce, B. R. The 1994 directory of evaluation
training programs . New Directions in Program Evaluation 62 (1994). 7194. http://
dx.doi.org/10.1002/ev.1678
Altschuld, J. W.Kumar, D. D. Program evaluation in science education: The model
perspective . New Directions for Program Evaluation 65 (1995). 517. http://
dx.doi.org/10.1002/ev.1699
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 25 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
Altschuld, J. W., & Kumar, D. D. (2002). Evaluation of science and technology
education at the dawn of a new millennium . New York: Kluwer/Plenum. http://
dx.doi.org/10.1007/0-306-47560-X
Altschuld, J. W., & Ronning, M. (2000). Program evaluation: Planning a program
evaluation . Invited Workshop presented at 32nd annual convention of the National
Association of School Psychologists, New Orleans, LA.
Altschuld, J. W., & Witkin, B. R. (2000). From needs assessment to action:
Transforming needs into solution strategies . Thousand Oaks, CA: Sage.
Altschuld, J. WYoon, J. S.Cullen, C. The utilization of needs assessment
results . Evaluation and Program Planning 16 (1993). 279285. http://
dx.doi.org/10.1016/0149-7189%2893%2990040-F
American Evaluation Association . Guiding principles for evaluators . New Directions for
Program Evaluation 66 (1995). 1926.
Bickman, L., ed. , & Rog, D. (Eds.). (1998). Handbook of applied social research .
Thousand Oaks, CA: Sage.
Campbell, D. T, & Stanley, J. C. (1966). Experimental and quasi-experimental designs
for research . Chicago: Rand McNally.
Caracelli, V. J. Structured conceptualization: A framework for interpreting
evaluation results . Evaluation and Program Planning 12 (1989). 4552. http://
dx.doi.org/10.1016/0149-7189%2889%2990021-9
Caron, D. J. Knowledge required to perform the duties of an evaluator . Canadian
Journal of Program Evaluation 8 (1993). 5979.
Chelimsky, E., ed. , & Shadish, W. R. (Eds.). (1997). Evaluation for the 21st century: A
handbook . Thousand Oaks, CA: Sage.
Chen, H-T. (1990). Theory-driven evaluations . Newbury Park, CA: Sage.
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 26 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
Cook, T. D., & Campbell, D. T. (1979). Quasi-exper-imentation: Design and analysis
issues for field settings . Chicago: Rand McNally.
Davidson, E. J. The discipline of evaluation: A helicopter tour for I/O psychologists .
Industrial/Organizational Psychologist 40(2)(2002, October). 3135.
Edwards, J. E., ed. , Scott, J. C, ed. , & Raju, N. S. (Eds.). (2003). The human
resources program evaluation handbook . Thousand Oaks, CA: Sage. http://
dx.doi.org/10.4135/9781412986199
Engle, M.Altschuld, J. W An update on university-based training . Evaluation Exchange
9(4)(2003/2004). 13.
Engle, M., Altschuld, J. W., & Kim, Y. C. (2005). Emerging dimensions of university-
based evaluator preparation programs . Manuscript in progress. Corvallis: Oregon State
University.
Fitzpatrick, J. L. The politics of evaluation: Who is the audience? Evaluation Review
13(6) (1989). 563578. http://dx.doi.org/10.1177/0193841X8901300601
Fitzpatrick, J. L., Sanders, J. R., & Worthen, B. R. (2003). Program evaluation:
Alternative approaches and practical guidelines (3rd ed.) . Boston: Allyn & Bacon.
Greene, J. C, ed. , & Caracelli, V. J. (Eds.). (1997). Advances in mixed-method
evaluation: The challenges and benefits of integrating diverse paradigms . San
Francisco: Jossey-Bass.
Holton, E. F., IIIBates, R. A.Naquin, S. S. Large-scale performance-driven training
needs assessment: A case study . Public Personnel Management 29 (2000). 249268.
Joint Committee on Standards for Educational Evaluation . (1994). The program
evaluation standards (2nd ed.) . Thousand Oaks, CA: Sage.
King, J. A.Stevahn, L.Ghere, G.Minnema, J. Toward a taxonomy of essential evaluator
competencies . American Journal of Evaluation 22 (2001). 229247.
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 27 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
Kirkpatrick, D. L. Techniques for evaluating training programs . Journal of the American
Society for Training and Development 13 (1959). 332.
Kirkpatrick, D. L. (1998). Evaluating training programs: The four levels (2nd ed.) . San
Francisco: Berrett-Koehler.
Leviton, L. CHughes, E. F. X. Research on the utilization of evaluations:
A review and synthesis . Evaluation Review 5 (1981). 525548. http://
dx.doi.org/10.1177/0193841X8100500405
Mark, M. M.Henry, G. TJulnes, G. Toward an integrative framework for evaluation
practice . American Journal of Evaluation 20 (1999). 193212.
Mark, M. M., ed. , & Shotland, R. L. (Eds.). (1987). Multiple methods in program
evaluation (New Directions in Program Evaluation, Vol. 35) . San Francisco: Jossey-
Bass.
Mathison, S. (Ed.). (2004). Encyclopedia of evaluation . Thousand Oaks, CA: Sage.
http://dx.doi.org/10.4135/9781412950558
McGuire, M. (2002, October). Canadian Evaluation Society project in support of
advocacy and professional development: Literature review . Toronto: Zorzi.
McKillip, J. (1998). Needs analysis: Processes and techniques . In L. Bickman, ed. & D.
Rog (Eds.), Handbook of applied social research (pp. 261284) . Thousand Oaks, CA:
Sage.
Mertens, D. Training evaluators: Unique skills and knowledge . New Directions for
Program Evaluation 62 (1994). 1727. http://dx.doi.org/10.1002/ev.1673
Mertens, D. M. (1989). Creative ideas for teaching evaluation: Activities, assignments,
and resources (Evaluation in Education and Human Services Series) . New York:
Kluwer.
Rossi, P. H., Lipsey, M. W., & Freeman, H. E. (2004). Evaluation: A systematic
approach (7th ed.) . Thousand Oaks, CA: Sage.
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 28 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
Scriven, M. (1991). Evaluation thesaurus (4th ed.) . Newbury Park, CA: Sage.
Scriven, M. The final synthesis . Evaluation Practice 15 (1994). 367382. http://
dx.doi.org/10.1016/0886-1633%2894%2990031-0
Scriven, M. Types of evaluation and evaluators . Evaluation Practice 17 (1996). 151
161. http://dx.doi.org/10.1016/S0886-1633%2896%2990020-3
Stevahn, L.King, J. A.Ghere, G.Minnema, J. Establishing essential competencies
for program evaluators . American Journal of Evaluation 26(1) (2005). 4359. http://
dx.doi.org/10.1177/1098214004273180
Stufflebeam, D. L. The relevance of the CIPP evaluation model for educational
accountability . Journal of Research and Development in Education 5(1) (1971). 1925.
Stufflebeam, D. L. An EEPA interview with Daniel L. Stufflebeam . Educational
Evaluation and Policy Analysis 4 (1980). 8590.
Stufflebeam, D. L. (2002). The CIPP model for evaluation . In T. Kellaghan, ed. & D.
L. Stufflebeam (Eds.), International handbook of educational evaluation . Dordrecht:
Kluwer.
Suchman, E. A. (1967). Evaluative research: Principles and practice in public service of
social action programs . New York: Russell Sage Foundation.
Tashakkori, A., ed. , & Teddlie, C. (Eds.). (2002). Handbook of mixed methods for the
social and behavioral sciences . Thousand Oaks, CA: Sage.
Wholey, J. S., ed. , Hatry, H. P., ed. , & Newcomer, K. E. (Eds.). (1994). Handbook of
accreditation standards for practical program evaluation . San Francisco: Jossey-Bass.
Witkin, B. R., & Altschuld, J. W. (1995). Planning and conducting needs assessments: A
practical guide . Thousand Oaks, CA: Sage.
Worthen, B. R., & Sanders, J. R. (Compilers). (1973). Educational evaluation: Theory
and practice . Worthington, OH: C. A. Jones.
SK Marketing
2006 SAGE Publications, Inc. All Rights Reserved. SAGE knowledge
Page 29 of 29 The Psychology Research Handbook: A Guide
for Graduate Students and Research Assistants:
Program Evaluation: Concepts and Perspectives
Worthen, B. R., & Sanders, J. R. (1987). Educational evaluation . New York: Longman.
Worthen, B. R., Sanders, J. R., & Fitzpatrick, J. L. (1997). Program evaluation:
Alternative approaches and guidelines (2nd ed.) . New York: Longman.
http://dx.doi.org/10.4135/9781412976626.n5