I’ll admit it. When
John Camey, Interim Dean of the School of Business and Justice Studies, told me
he published a paper titled “The Wicked Problem of Assessment,” I reacted
defensively. Imagine, I thought, if I
referred to someone’s discipline or the work they had been doing for over two
decades as a “wicked problem.”
But let’s face it. We
learn more from diverse views than from those that reflect our own. As Walter B. Greenwood, my undergraduate
professor in contemporary lit, said, “Having old certainties destroyed by new
considerations is one of the hazards of reading.”
I read the paper.
Far from disagreeing with what John Camey and his co-author,
Carolyn E. Johnson, said, I thought they made a lot of practical sense.
A wicked problem, I learned, was not one that is inherently
evil, but is one for which there is no straightforward solution. Camey and Johnson offer ten reasons why
assessment meets the criteria of a wicked problem. A few of these help explain why assessment
frustrates faculty, and, more importantly, how assessment professionals and
accrediting bodies may have unwittingly been the cause of certain
vexations.
The authors note that when accreditors first required
institutions to assess student learning, or provide “assurance of learning,” they
offered little assistance with how to achieve this. Faculty were supposed to
“figure it out.” However, the methods they had been using for years—i.e.
grades—were not considered valid measures.
So how were faculty supposed to navigate this new terrain? As Camey and Johnson observe, “An entire
industry of workshops, seminars, conferences and travelling consultants [grew] up
to help.” Needless to say, each of these had a price tag, and some were quite
hefty.
Wicked problems, Camey and Johnson explain, do not lead to
right or wrong answers, but only good or bad solutions. This might be where assessment is most
frustrating, because “good” or “bad” are arbitrary judgements. We see this in the accreditation
process. One visiting team may consider
the assessment efforts at an institution to be acceptable, while a second
visiting team rules them not good enough.
Even within an institution, what constitutes “good enough” for one group
of members on an assessment committee may not be sufficient for the team that
replaces them. No small wonder that in 2018, 36% of faculty responding to a survey on assessment culture agreed that assessment is "based on the whims of the people in charge," and in
2017, UC faculty described assessment as a moving target where the expectations
continuously changed.
At most institutions, the assessment cycle is the academic
year. On an annual basis, faculty assess
student learning and document the results in some type of report. Besides reporting results, faculty are
expected to reflect upon the findings and articulate how they will be used to
improve teaching and learning. In
subsequent assessment cycles, they should provide evidence as to whether the
changes they made resulted in better learning.
This process, which many institutions like to graphically
represent as a continuing sequence in a circular flow, may not facilitate
effective, meaningful assessment. Camey
and Johnson assert, it might take “multiple semesters before sufficient data
can be gathered to determine whether our solution is good, bad, or not quite
good enough.” This claim echoes Trudy
Banta’s and Charles Blaich’s conclusion that
“Collecting and reviewing reliable evidence from multiple sources can take
several years” but “state mandates or impatient campus leaders may exert
pressure for immediate action.”
Camey and Johnson offer a solution to the wicked problem of
assessment: an assessment committee that
is chaired by a designated leader, an “assessment czar,” and comprised of
faculty who rotate off periodically. At
UC, we already have such a solution in place, but is it sufficient?
It strikes me that it is also important to have clearly
defined criteria that communicate what “good” assessment is, and it is critical
to view assessment as a process of inquiry and analysis, not a
fill-in-the-blank, paint-by-number activity.
But let’s hear from faculty on this topic. How might we collaborate not to solve but to
address this wicked problem of assessment?
Works Cited
Banta, T.W., Blaich, C.
(2011). Closing the Assessment Loop. Change: The Magazine of Higher Learning, 22- 27.
Camey, J. P.,
Johnson, C. E. (2011). The Wicked Problem of Assessment. Journal of Business and Behavioral Sciences, 23(2), 68-83.
Sam Houston University. (2018). 2018 National Faculty Survey of Assessment Culture. https://www.shsu.edu/research/survey-of-assessment-ulture/documents/Nationwide%20Faculty.pdf
Sam Houston University. (2018). 2018 National Faculty Survey of Assessment Culture. https://www.shsu.edu/research/survey-of-assessment-ulture/documents/Nationwide%20Faculty.pdf
No comments:
Post a Comment