Holistic Assessment and Problem Based Learning Author & presenter: Glen O’Grady Type of Presentation: Oral Theme: Assessment Student assessment (assignments, examinations, presentations, portfolios etc) are widely accepted as an important part of the learning process. This paper explores how assessment underpins the quality of learning in problem based learning. It also reports the effects certain types of assessment have upon the quality of problem based learning. The paper concludes that assessment in PBL needs to be holistic in nature, meaning the raft of assessment tasks should both foster and measure a wide array of skills such that the rich nature of learning promised by PBL is achieved. Glen O'Grady Deputy Director Centre for Educational Development Republic Polytechnic Tanglin Campus 1 Kay Siang Road, Singapore 248922 DID : (65) 6376-8151 Fax : (65) 6415-1310 www.rp.edu.sg
Holistic Assessment and Problem Based Learning Glen O’Grady 5th Asia -Pacific Conference on PBL : 16 - 17 March 2004 Introduction In this paper I examine the role of student assessment in learning, in particular problem based learning (PBL). I will concentrate on the summative and formative functions of assessment in respect to learning. However to talk about assessment is to discuss “several things at once” (Ramdsen 1992). Assessment encompasses measuring, defining and summarizing what students can do and/or inferring what students could do. Assessment can entail gate keeping where students are selected or rejected on the basis of norm or criterion referenced standards. Assessment is about diagnosing the quality of learning so that learners strengths can be identified and weaknesses addressed. It can also involve fostering change both in the learner and the teacher as they both seek to negotiate their symbiotic relationship (Rowntree 1987). All of this points to the fact that when considering assessment one cannot reduce it to the common duality of formative and summative functions. However, a detailed exploration of the many factors embodied in student assessment is beyond the scope of this paper, instead this paper will focus on the connection between assessment and learning while not forgetting that the practice of assessment may encompass more than just its association with learning. This paper will show how a selection of assessment methods, in a PBL setting can be used to both foster and measure learning that can be characterised as “understanding”. In reporting the effects different types of assessment methods have upon learning in PBL, I will argue that assessment needs to be holistic in nature, meaning a raft of assessment tasks should be used collectively when fostering and measuring understanding. Learning is Understanding Everyone agrees that learning is the object of education; however many differ in their conceptions of learning. Dewey (1972) criticizes conceptions of learning that define learning as “fragments” and “dualisms” which are simply synthesised. Rather, Dewey describes learning as a habit formation that arises from the continual reconstruction of existing dualisms into entirely new unitary functions and understandings. Recent findings about the brain architecture and cognition support Dewey’s view of learning. Researchers (Lakomski 2002, Minsky 1986) have suggested that the brain does not appear to operate in a linear processing fashion of matching individual stimuli to pre- existing pieces of information. Those advocating connectionism contend that learning occurs when neurons, that make up neural nets, fire in certain unique patterns. These connections form a pattern that can be strengthen through regular activation and recognition of these patterns. In other words learning encompasses more than absorbing or identifying discrete bits of information and then mapping them against other similar bits of information, rather it is the interaction of a new idea with the sum of all that we know, to the effect of transforming existing cognitive or neural structures.
Situated cognition (Brown, et. al. 1989) further suggests that an activity in which knowledge is developed and deployed, is not separable from or ancillary to learning and cognition. Rather, it is an integral part of what is learned since every situation triggers a unique sets of connections that have not already been experienced by the mind. Every act of speaking, every motion of the pen, each gesture, turn of head, or any idea at all is produced by the cognitive architecture as a matter of course, as a new neurological coordination (Clancey, 1993, pp. 111-112). The familiarity of cognitive patterns is often characterised as prior knowledge. While prior knowledge influences the way we make new meaning, there is also the potential for new understandings to transform existing thinking structures leading to new conceptions of the world. In other words prior knowledge that is brought to bear is itself constructed, rather than retrieved intact from memory when it’s brought to bear on new situations. Social constructivists add that learning is not only a process that occurs in the mind of the learner, but it is negotiated through the interaction with others (von Glasserfield 1988). How people make sense or understand the world and themselves is a broader view of learning encompassing various cognitive perspectives (Bruner 1990). Fosnot (1996) argues that if learning is about sense making then the goal of instruction is not to shape behaviours or skills but rather to influence conceptual development and deep understanding. A conception of learning that focuses upon “understanding” has serious implications for assessment. Learning for Understanding: Implications for Assessment Assessment has for a long time been recognised as a powerful driver of learning, and therefore something that teachers could leverage on to ensure students achieve desired objectives. Unfortunately however, this belief about assessment, has contributed to, inadvertently perhaps, assessment grades being used as a carrot or stick resulting in learning for the sake of its external value and not the intrinsic value of understanding. In one study an examiner lamented: In an ideal world quantitative marks would be abolished…this is because students tend to see their mark as a valuation of their worth (Warren Piper et.al. 1996:83) This belief that assessment drives learning focused on the external value of learning has helped to sustain simplistic conceptions of learning and assessment. In an outcome based system focused on the external value of learning what has become important is the performance on the test rather than the learning that facilitates doing well on a test. In these instances learning has become subject to the instrumental ideals of efficiency, accountability, that which is easily tested and measured, and the ranking of individuals and/or institutions. In this context assessment has developed into a quality assurance mechanism that validates the transfer of measurable knowledge and skills from the expert to the novice. Furthermore, assessment as an objective judgement has helped to drive the
idea that assessment, learning and teaching are all separate activities (Huba and Freed 2000). However, as outlined above there are alternative views of learning and along with these views a commitment to re-defining of the role of assessment in learning: Assessment is the process of gathering and discussing information from multiple and diverse sources in order to develop a deep understanding of what students know, understand and can do with their knowledge as a result of their educational experiences; the process culminates when assessment results are used to improve subsequent learning (Huba and Freed 2000:8) Herrick (1996) in his summary of Dewey’s views on assessment suggests: Assessments are valuable when they are "low stakes," that is, when students and teachers use them to clarify achievement, interest, and aptitude for the purpose of stimulating new learning. Assessments are valuable when they are authentic to life experiences as much as possible, thus encouraging the integration of vocational and general education. Assessments are valuable when they promote positive consequences for students, thus helping them form new and more meaningful habits. Assessments are valuable when they do not separate students into performance categories, but rather help students better understand their own growth. Assessments are valuable when they match instructional aims, which, in turn, match societal aims for education. In short, assessments are valuable when they expand opportunities for students rather than limit opportunities. Dewey (1980) Perry (1970) and Biggs and Collins (1982) all point out in their studies of student intellectual development that assessment focused on feedback is essential for the development of a deep understanding. Assessment is so critical in learning for understanding that the processes of teaching, learning and assessment cannot be easily separated. Rowntree (1987:24) perhaps best locates assessment when he describes it as the “life-blood of learning”. Assessment rather than being a driver of learning outcomes, external to understanding, assessment can be an enabler for understanding. The difference is significant. As an enabler for understanding assessment becomes part of the teaching and learning process by virtue of the fact that understanding is a process and not simply a product that can be objectified and easily measured. Furthermore, assessment as an integral part of the learning process requires a shift away from teaching models that define learning in terms of discrete bits of information that are dispensed and passively absorbed, to models that provide opportunities for the student to want to make sense and understand the world around them (von Glasersfeld, 1996). This different conception of assessment requires a re-think about the purpose of different assessment methods. Methods that focus on measuring discrete or observable behaviours, the hallmark of traditional examinations (Warren Piper et.al. 1996: 10-13), may not be appropriate for assessing understanding. Numerous studies have shown that many current approaches to assessment are not encouraging students to learn in a deep fashion
(Ramsden 1992: 181-213) or are failing to assess deep learning or understanding. One example is a study by Nurrenbern (1987) who reported how 88% of students tested using the algorithmic questions were able to correctly answer but when tested conceptually only 31% were able to respond appropriately. The researchers also concluded that success on traditional exam type questions did not necessarily indicate conceptual understanding. PBL offers an alternative approach to learning where assessment is an integral element in both the facilitation and measurement of understanding. Understanding: the Promise of Problem-based Learning The underlying principle in PBL is "all learning begins with a problem". The problem provides the direction of learning, the motivation for learning and the application of learning. (Barrows & Tamblyn, 1980; Boud & Feletti, 1991; Schmidt, 1983). The appeal of PBL is its enormous potential for developing understanding since encapsulated in the PBL are explicit expectations that students: • explore knowledge concepts within different contexts; • articulate what they already know about a problem (prior knowledge) • identify and then find information in respect to what “they don’t know”; • specify how new information connects with prior knowledge; • share and test the viability of new conceptions; and • reflect how they personally constructed knowledge and became meaning makers (meta-cognitive). (O’Grady 2002) The potential for developing understanding through PBL, is further enhanced by a focus on not only developing cognitive awareness (meta-cognition) but also in engaging the students in “learning for the whole person” (Heron 1993). Two examples of this holistic approach to learning are an emphasis on developing general dispositions of the learner which Kay and Bawden (1996) suggests includes, “itch to know” (propositional), “itch to do” (practical), and “itch to be” (experimental) and, the importance of working in a collaborative setting (Rhem 1998). But how does assessment work in PBL? Nowak and Plucker (1999) point out the need for alignment between the intentions of PBL and assessment. Studies have shown that when standardized tests are used, students doing PBL generally do not score as well as those being taught using traditional teaching methods. However, when non-standardized forms of assessment are employed, students doing PBL are at least as good as other students (Vernon & Blake, 1993) and when you account for long-term retention, PBL students do better (Farnsworth, 1994). Learning defined as understanding requires assessments that focus on the process of developing understanding and not just the products of thinking (Nightingale et. al. 1996). Assessment that focuses simply on what content knowledge students posses may fail to measure the way in which students have reasoned out their answers and hence whether the answer indeed represents understanding. Or worse still it sends a message to students that the inquiry process does not really matter as long the “correct” answer is reproduced.
Leaving aside the epistemological problems of right or wrong answers, assessment of answers without checking how ideas are understood and synthesised together in order to derive answers, also makes it difficult for the assessor to determine what are the difficulties or impediments in understanding. While the assessment of content is necessary the intention in PBL is to evaluate how well understanding has been developed such that meaningful responses to problems can be generated. In PBL with the emphasis on students learning how to learn (how to understand) students use the context and experience of working on a problem to demonstrate how they develop an understanding of ideas so they can meaningfully apply these ideas to a problem. Assessment therefore is just not focused on the application of knowledge but also the process of meaning making so that meaningful application is possible. PBL at the Republic Polytechnic PBL at the Republic Polytechnic is used as the exclusive educative approach. PBL at the Republic Polytechnic has a number of distinct features including a one-day, one problem cycle, one facilitator for 5 teams of 5 students per class, an integrative curriculum and a holistic approach to grading and assessment (O’Grady 2002). One of the distinctive features is the requirement that facilitators assess students’ performance daily and give them both formative and summative feedback. In order for staff to assess students in ways meaningful so that it both fosters and measures understanding, staff are expected to collect evidence about students understanding. Facilitators, with the help of technology, collect evidence which allows them to make an overall assessment about students understanding. The methods used to collect evidence about understanding at the Republic Polytechnic include: • Team Presentations • Peer & Self Assessment • Learning Journal • Tests and Quizzes • Observations of student discussions Samuelowicz (1994) points out that it is not the assessment task which is critical in developing understanding but the context in which these methods are put into practice. Each of these methods in the context of PBL have an important instructional and assessment dimension in the development of understanding. For the remainder of the paper I shall explore how this approach to PBL, premised on students developing understanding, employs these methods of assessment (individually and collectively) to both foster and measure learning for understanding. Team Presentations In the Republic Polytechnic students are expected each day to develop a presentation that is in response to the problem they have been presented. The formal presentations are a culmination of a team of five students’ research and discussions. The presentation is shared with other teams and are compared and critiqued by their peers (who also offer presentations) and the facilitator. Student presentations are extensively used in many PBL courses (see table 1)
While presentations have long been advocated as a valuable instructional and assessment process, the emphasis is often upon the development of communication skills (Hay 1994). However, the value of the presentation is more than the practice of “communication skills”. The function of presentations at the Republic Polytechnic is to give a focus for the students collaborative efforts in expecting the team to organise and explain their ideas. A pupil does not really know what he has learned till he has organised it and explained it to someone else. The mere recognition of what is right in someone else’s wording is only the beginning of the awareness of truth. (Rowntree 1987:65). Table 1: PBL courses using class presentations • Northern Arizona University College of Education http://jan.ucc.nau.edu/~lam37/eci350/syllsprg04.htm! • Dept of health management , University of Missouri ,Columbia http://hmi.missouri.edu/course_materials/Residential_Informatics/semesters/W2000_Materials/401_ha les/MI_401_sec2.htm!: • University of Hong Kong sociology department http://www.hku.hk/sociodep/courses/soci8004_03outline.doc!: • University of Maryland School of Pharmacy https://rxsecure.umaryland.edu/courses/PHAR540/assign.htm!; • Indiana state university http://www.indstate.edu/thcme/PSP/pbl.html!: • :Department of Mechanical Engineering, Hong Kong Polytechnic University http://tlrc.edc.polyu.edu.hk/html/theme/pbl_casestudy/ChuenChunWah.html#8! • Cornell medical college http://class2004.med.cornell.edu/applicants/pbl.html! There are numerous studies on the positive effects of collaborative learning, one example reports that students who collaborated in the preparation of a shared assignment tended to engage in higher level cognitive strategies than those who worked by themselves (Tang 1998). There are two major challenges in using classroom presentations in the instruction and assessment of student understanding in PBL. The first is the establishing of criteria that specifies the expectations of students’ presentations (Hay 1994). Without criteria students are unsure of what is expected of them. At RP we have developed criteria (see table 2) that specifies the expectation as to what would demonstrate understanding. Table 2:Criteria for Student Presentations • Statement of strategy for proposed solution to problem • Information collation and analysis • Rationale for problem solving strategy (statement of how the solution was reached) • Ability to address questioning (for justification of proposed solution) • Ability to compare and analyse multiple solutions/perspectives • Communication or presentation skills (spans the other five dimensions) The second challenge relates to the more fundamental problem that any presentation or articulation of ideas is limited in its capacity to represent ones understanding, since what we know and understand is more than the words or symbols we use to represent it (Lakomski 2002 ). Given the limitations of representations of knowledge, a prepared
presentation should be regarded as the tip of an ice-berg in respect to what students actually know. In PBL the facilitator (and peers) will probe further on the insights, in particular how these insights relate to the problem trigger. While this interactive questioning and probing helps to avoid presentations being unmindful collections of information (information that is “copied and pasted” without a real sense of how the information relates to the problem), it also affords the opportunity for students to build on their insights and further extend their understanding by constructing responses or questions. Peer Assessment & Self Assessment Peer assessment is the act of students assessing one another (Boud et. al. 2001). Self- assessment is judging the quality of ones own work. While there is some debate about whether self and peer assessment is workable in PBL (Woods 1994), there is little doubt that there is immense value in helping students to develop an understanding and judgement of the quality of their own and the work of others. At the Republic Polytechnic students are constantly asked to make informal judgements about one anothers’ comments and team presentations. However, in a more formal sense (formal meaning recorded) students are asked to complete a peer and self assessment questionnaire. These questionnaires are constructed using research on the behaviours that facilitate student centred discussions. Research (Shoop 2000) has shown that these behaviours underpin meaningful collaboration which is a major premise for understanding in PBL. By asking students to regularly reflect and hold themselves and each other accountable for these behaviours, it serves to both instruct what behaviours are desirable and also provides information that maybe useful for the facilitator in making an assessment of students’ understanding. There is a good deal of scepticism about the validity of peer and self assessments (Nightingale et. al. 1996: 123-124) and there are staff at the Republic Polytechnic who share this scepticism. Some staff perceive students as not taking the self assessment seriously and randomly selecting ratings in the self & peer assessment. A random selection of 50 students’ peer evaluations show that students were making distinctions from individual to individual. The ratings were analysed to see if the variations were due to random selection by comparing the rating of peers to overall daily grade. What we found was that while the correlations were relatively low, they were statistically significant suggesting that there was more than just a random selection of ratings. Another concern is that students, knowing this assessment is taken into account in the summative judgement of their understanding, use this tool to rate themselves and their peers in the best light possible (or in the worst possible light if they do not “get on”). This relates to criticism of peer and self-assessment centres as being less accurate than expert judgements. A comparison between peer and self assessment and the facilitators judgement about students performance showed a fairly large difference between the two judgements. However, when comparing self and peer assessments, peers tend to rate
others higher than their own evaluation suggesting students were not simply rating on the basis of self interest. One feasible conclusion from the analysis is that students are still coming to terms with peer and self assessment and while there may be occasions where they overstate their abilities, when compared with expert judgements, there is a demonstrated willingness to make distinctions between various performance between peers and between self and peers. The value of peer and self assessment extend beyond the validity of peer and self- assessments to include the organic experience of trying to develop an understanding of one self and others work. Facilitators signal the importance and help students in developing judgement about their own and others performance by taking peer and self assessment into account when assessing students. Learning Journal A learning journal is for students to record written reflections about learning. The act of writing helps to organise thoughts and facilitate understanding. Donald Schon (1983) coined the term reflective practitioner to describe how professionals think in action. The idea of reflection has since been used to describe an important process in learning. Jim Butler’s (1994) model of human action and learning places reflection at the centre as “the open, active communication channel between the social context and the inner self”. Reflection is the crucible or catalyst which transforms “public knowledge” – theories, book-knowledge, received as rules and codes – into “PPK – Private practical knowledge – “what I know and use”. PPK is constantly enriched and renewed by vivid experiences and can be transformed by paradigm shifts and value re-alignment. The most recent layers can be examined, but much of it is probably so deeply embedded as to be unavailable for conscious recall. Using PPK in the public domain of professional practice is possible again only through reflection. Reflections journals are recommended as an important assessment method in PBL (Woods 1994). At the Republic Polytechnic students are expected to keep a reflection journal for each of their modules and record an entry each day. Upon the completion of each day facilitators at the Republic Polytechnic provide a question that students can choose to respond to in their reflection journal. Students then record their thoughts about the day and the problem they have worked on. The expectation for students as to what constitutes a meaningful reflection is expressed in the criteria facilitators use to assess the journal. Table 3 Criteria for Reflection Journal • Clarity in reasoning • Strength to take and express a stance • Evidence of milestones • Awareness of learning preferences What is expected of students extends beyond merely describing the events of the day or listing the contents encountered while they were working on the problem. Students are expected to articulate how they have personalised public knowledge or made sense of what they have done during the day. A content analysis of 50 random selected journal
entries revealed that the majority of the entries (38 out 50) contained personal insights about how they are learning and the personal significance of what it is they have learnt rather than just descriptions of the day’s events. Tests and Quizzes In many PBL courses various types of invigilated tests are used. Despite the criticism levelled against tests, particularly objective tests where answers are predetermined, multiple choice questions (MCQs) under semi-formal test conditions can be used to assess reasoning processes (Gibbs 1992b). However for MCQs to be effective in fostering learning for understanding, it is important that the questions focus on requiring more than recall. While recall is an important learning process, by itself it may not be a very good indicator of reasoning and understanding. The challenge of using MCQs is that in selecting an answer the facilitator may not be aware of the students’ reasoning process and is forced to infer the reasoning through the selection of the various responses. While all assessment requires some inference, the lack of gradation in the measurement of understanding or misunderstanding makes this inference more difficult than other assessments which allow students to express the varying quality of their understanding. At the end of each day, students at the Republic Polytechnic take a short quiz that focuses on understanding of specific aspects of content that was addressed throughout the day; the quiz is an opportunity to compare their understanding with the understanding of a subject expert. Students are generally given detailed verbal feedback about their responses in class. Given the quiz is completed at the end of the day it is perhaps not surprising that in a review of a 50 randomly selected students that majority scored fairly well on their daily quizzes. While there has been some concern about collusion, given the tests are not invigilated like more formal exams, the quiz has become regarded as a practice for the more formal regular understanding tests which are conducted under invigilated conditions. Facilitators generally agree that this is one method of assessment that needs to be considered in light of the other forms of assessment when evaluating the validity of the answers both from the point of view of understanding and reliability. Observations of student discussions The role of the facilitator in PBL is to observe and guide students learning and discussions (Barrows and Tamblyn 1980, Barrows 1988). The facilitator assists student learning through guidance at the meta-cognitive level. It is the facilitator’s expertise in this process, not in the content areas in which the students are studying, that is important. The students are expected to acquire the knowledge they need from content experts who serve as consultants, as well as books, journals, and automated information sources. The facilitator guides the students through repeated practices in reasoning and gives encouragement and advice on how to enhance their ability to do self-directed study. In the course of this guidance the facilitators are able to observe students working through these processes and are able to formulate judgements about students understanding based on their engagement with the processes of learning e.g. problem definition, identification of learning issues, reading, discussing, testing theories or possible solutions etc.
Facilitators at the Republic Polytechnic observe students in the classroom. Huba and Freed (2000) suggest facilitators see the learning process further allowing facilitators to learn how students learn. However, facilitators’ perspectives of the learning process can influence what they observe and fail to observe (Pratt 1998). A group of facilitators were asked what they observe when facilitating. In the table 4 below are three different response that represent a range of dimensions facilitators observe when in the classroom. Table 4 What facilitators observe in the classroom Facilitator 1 Student’s level of engagement with the problem • Motivation level • Whether he/she is genuinely trying or just putting up a show • Questions that they ask would indicate level of understanding • Their response to questions asked • Amount of progress from last week • Can students break down the problem • Whether they are able to explain the problem in a coherent way • Whether they are able to explain their solution (even if it is wrong), in a coherent way • Whether they actively try to solve the problem that is stopping them or do they just give up • Whether once the problem is solved do they try to pursue a better solution • What they have found out and how they present it • Whether it is individual work or copied • If they got help, I would look out for coherence in their explanation, at the same time • Though they got help !I would check if they implemented what they have learnt from their • Friends or is it just a relatively blind/vague copy to get pass the day • Ability to coach others Facilitator 2 • Peer Evaluations • How they work in their teams (when I go around or peek around or sudden checks) • Punctuality • Bonus for additional work/initiatives Facilitator 3 • The quality of questions asked by students • Ability to explain or defend their points when challenged by facilitator or other students • Showing understanding of the day's topic by being able to apply it to different situations • RJ has to!demonstrate understanding of the day's!lesson. ! • Good behaviour in class (ie not being late or sleeping in class, being respectful when other teams are presenting) • Showing progress from 1st meeting to 3rd meeting • Quality of class participation and team participation • Interesting/creative solutions (eg making a movie instead of powerpoint) are a bonus. These dimensions may reflect on the one hand the personal bias different facilitators have in what they choose to observe but they also each reflect the different contexts with which facilitators work in. However, one could make an argument that the emphasis in most cases is upon observing understanding in a broader sense than just what information they are able to recall.
So far I have specified the different assessment methods used in the daily assessment at the Republic Polytechnic. Each of these assessment methods despite their weaknesses and strengths is a means of fostering learning for understanding as well as collecting information or evidence about the quality of students understanding. So far I have emphasised the formative nature of assessment. However, facilitators are also expected to use this information to derive a summative judgement – a daily grade. The challenge in assessment for learning (PBL) is how to reconcile formative and summative judgements. I have already alluded to the potential for learning to be undermined by the assignment of a grade that can be interpreted as an outcome that appears to be external to the process of learning. So how can a summative assessment be made without their being an adverse affect (often unintended) on learning for understanding? One way of reconciling the formative and summative elements of assessment is to employ a holistic approach. Holistic Assessment Holistic assessment entails using a number of assessments and then combining these assessments using a rationale that is defensible to the objective of assessing understanding. Or in other words employing a “professional judgement” in the assessment of students. Professional judgement implies a professional body to which the judgement can be ascribed. Warren Piper (1994) suggest professional bodies have a mastery and shared view of a body of knowledge, have a commitment to the needs of a cliental, and sense of responsibility for decisions made by individuals counted as members of the profession. Warren Piper goes on to ask the question whether teachers are professional? As earlier stated there are competing perspectives suggesting a wide array of teaching practices (and malpractice). Perhaps the best that can be hoped for is that within certain perspectives or theoretical positions there are shared views about learning, the learner and the responsibilities of a teacher, however this seems somewhat narrow and would undermine the idea of academic standards that are agreed upon, at least notionally across higher education systems. So what is needed is a organising belief or philosophy that teachers can judge good teaching and learning and hence inform the exercise of “professional judgement” particularly in the realm of assessment where the “professionalism” of a teachers is open to greatest scrutiny. In this section of the paper I describe how learning for understanding as a principle for an holistic approach to formative and summative judgements in PBL, is an attempt to be professional. For formative assessment facilitators at the Republic Polytechnic are expected to use the information as it is collected to guide the ongoing facilitation process. Some information, collected informally during class: observations and team presentations, may be used immediately to prompt a question, comment or deliberate silence by the facilitator. Other information, available only at the end of the day after students have submitted them ( reflection journal, quiz, peer and self assessments, along with evidence from observations and team presentations) may be used to inform facilitation in subsequent problems. The information is also used to formulate detailed written feedback for students which is attached to students submitted assessment. The comments are relatively brief but are meant to focus on specifying how students can further develop their capacity to understand (make sense of the world around them). The critical decision the facilitator
needs to make is given all the information they have collected what do they comment on? It is simply not possible for the facilitator to respond to all the information they have about a student. So they must choose what comment would best help the student develop their capacities for understanding. This requires the judgement of the facilitator after considering what she/he knows about the student. For summative assessment the facilitator is expected to derive a daily grade that will be used along with other daily grades and performance on a series of understanding tests to compute an overall module grade. Facilitators are once again faced with the problem of having to try and summarise a significant amount of information into broad grading bands. To encourage a professional judgement based on the knowledge known about the student no weightage of evidence or algorithm for calculation of the grade is given to the staff. What is expected of facilitators, is that like doctors, who are required to make a diagnosis based on pieces of information they collect using different tests (blood pressure tests, x-ray, throat examination etc), they should judge each person on the merit of the evidence at hand. How they determine which pieces of evidences are regarded as being critical for each student is left up to the professional judgement of the facilitator. This holistic approach to judging allows facilitators to judge students understanding according to however the student is best able to demonstrate, and facilitator interpret, understanding (or misunderstanding). This is not an easy task because the evidence collected from the different assessment methods may paint a complex picture of varying degrees of understanding on issues related to the problem. In fact sometimes the evidence is perceived to be contradictory. However, all this has simply reinforces the idea that understanding is a complex process that is not easily represented and interpreted. Given the complex nature of understanding as learning, it seems reasonable to argue that facilitators need to collect as much evidence as possible. One assessment method (a 3 hour invigilated exam) may produce a more manageable picture of students performance but it may fail to recognise the complexity of thought and seduces examiners into thinking understanding is easily tested and measured. Worse still it may give students the impression that learning is simply the sum of the expectations and requirements expected from that single test. A holistic approach is sometimes criticised as being subjective and hence inherently unfair because the facilitator is judging each student differently. However, I would argue that this differentiation in judgement is precisely what makes the formative and summative judgement fair. Students in holistic assessment are still judged against general criteria and the objectives of the problem but in the measurement and fostering of understanding the uniqueness of each student in how they understand is contextualised into the judgement. Formative assessment is not undermined by holistic assessment, since students are afforded multiple opportunities to demonstrate their understanding and get feedback on how to improve. Also by considering each piece of assessment in relation to other pieces of assessment facilitators can construct a more complex picture of students understanding
and respond in more meaningful ways rather than relying on inferences that maybe more tenuous because they are based on assessment that has less dimensionality. Summative assessment or the measurement of learning and its summation as a grade is not weakened by holistic assessment since multiple assessments help to ensure greater defensibility in the determination of the grade. The triangulation and verification of the various pieces of evidence about understanding can help the facilitator determine a grade that suits the criteria associated with each grade band. Students in receiving the grade will have attached to it substantive formative feedback that not only indicates how they can improve but the rationale as to why they have received their specific daily grade. Conclusion Learning defined as deep understanding has serious implications for how we both teach and assess. I have suggested that the processes of PBL lends itself well to the definition of learning as understanding because assessment is regarded as an integral element in the facilitation of learning. In reviewing each of the assessment methods used for the daily assessment of students at the Republic polytechnic I have tried to demonstrate how they have both an instructional and measurement element. Furthermore I have argued that by taking a holistic approach to summative and formative functions of assessment it can reconcile the potential conflict between fostering learning and the measurement of learning.
References Barrows, H. S., Tamblyn R. M. 1980. Problem-Based Learning: An Approach to Medical Education. Springer, New York Barrows, H. S. 1988. The Tutorial Process. Southern Illinois School of Medicine, Springfield, Illinois Biggs, J. B. & Collins, K.F. (1982) Evaluating the Quality of Learning: The SOLO Taxonomy. New York, Academic Press. Boud, D. Cohen, R, and Sampson, J. (2001) Peer learning and assessment In Peer Learning in Higher Education. London, Kogan Page. p67-81. Boud, D. & Feletti, G. (1991). The Challenge of Problem Based Learning. London. Kogan Page. Brown, John Seely, Collins, Allan and Duguid, (1989) Paul Situated Cognition and the Culture of Learning Educational Researcher; v18 n1, pp. 32-42, Jan-Feb. Bruner, J.S. (1990). Acts of meaning. Cambridge, MA: Harvard University Press. Butler, Jim (1994) From action to thought: The fulfilment of human potential. In J. Edwards (ed.) Thinking International Interdisciplinary Perspectives. Melbourne, Hawker Brownlow, pp. 16-22. Carter, Annie and Palermo, Josephine (2004) Reflections on Learning and Self Assessment: A Case Study of Problem Based Learning http://www.tamil.net/people/indy/Carter.htm Clancey, W. J. (1993). Situated action: a neuropsychological interpretation. Response to Vera and Simon. Cognitive Science, 17, 87-116. Dewey, J. (1980). Thinking in education. In J. Boydston (Ed.), John Dewey: The middle works, 1899-1924: (Vol. 9, pp. 159-170). Carbondale and Edwardsville, IL: Southern Illinois University Press. (Original work published in 1916). Farnsworth, C. C. (1994). Using computer simulations in problem-based learning. In M. Orey (Ed.), Proceedings of the Thirty-fifth ADCIS Conference (pp. 137-140). Nashville, TN: Omni Press. Fosnot, C. (1996). Constructivism: A Psychological theory of learning. In C. Fosnot (Ed.) Constructivism: Theory, perspectives, and practice, (pp.8-33). New York: Teachers College Press. Gibbs, G. 1992a. Improving the Quality of Student Learning. Technical and Educational Services, Bristol. Gibbs, G. 1992b. Assessing More Student. Oxford, Oxford Centre for Staff Development. Hay, I. (1994) Justifying and applying oral presentations ing geographical education. Journal of Geography in Higher Education. 18 (1) 43-55. Heron, J. (1993) Group Facilitation London, Kogan Page.
Herrick, Michael J. (1996) Assessment of Student Achievement and Learning, What Would Dewey Say? A 'Recent' Interview with John Dewey. Journal of Vocational and Technical Education. Volume 13, Number 1. Hoffman, Bob., Jones, Diana., Lewis, Dave., Lucas, Gail. and Ritchie, Donn. (1996) Assessment of problem based learning; students and classes: CSU Instructional Technology Initiatives http://edweb.sdsu.edu/clrit/webassess/studentNclasses.html The California State University Huba, M.E. and Freed, J.E. (2000) Learner-Centered Assessment on College Campuses Boston: Allyn and Bacon. Maudsley, G. (1999) Do we all mean the same thing by Problem-based Learning? A review of the concepts and a formulation of the ground rules. In Academic Medicine, 74 (2). 178-184. Kay, R. and R. Bawden (1996). “Learning to be systematic: some reflections from a learning organization.” The Learning Organization 3(5). http://www.emerald-library.com/brev/11903ec1.htm Lakomski, G. (2002) How we think – Modern cognitive science and what it tells us about teaching and learning. Proceedings of the Second Symposium on Teaching and Learning in Higher Education, CDTL, National University of Singapore. 4-6 September 2002. Minsky, Marvin (1986) The Society of Mind New York, Simon and Schuster. Mohanan K.P. (200 4) PBL in Perspective, CDTL, The National University of Singapore http://www.cdtl.nus.edu/publications/pbl/default.htm Downloaded: Feb 15 2004 Nightingale, P., Te Wiata, I., Toohey, S., Ryan, G., Hughes, C., and Magin, D. (1996) Assessing Learning in Universities (1st ed.). Sydney: University of New South Wales Press. Nowak, Jeffrey A. and Plucker, Jonathan A. (1999) Do as I Say, Not as I Do? Student Assessment in Problem Based Learning Indiana University http://www.indiana.edu/~legobots/q515/pbl.html Nurrenbern, S.C., & Pickering, M. (1987). Concept learning versus problem-solving: Is there a difference? Journal of Chemical Education, 64(6), 508-510 O’Grady, Glen & Alwis, W.A.M (2002) One Day, One Problem: PBL at the Republic Polytechnic 4 th Asia Pacific Conference in PBL. Hatyai, Thailand December 2002 Perry, W.G. (1970) Forms of Intellectual and Ethical Development in the College Years: A Scheme New York: Holt, Reinehart and Winston. Pratt, D. and associates (1998) Five Perspective on teaching and Learning in adult Higher Education. Florida, Krieger Publishing. Ramsden, P. (1992) Learning to teach in Higher Education. London, Routledge. Rhem James (1998) Problem-Based Learning: An Introduction The National Teaching & Learning Forum on-line edition December Vol. 8 No. 1 Rowntree, D. (1987) Assessing Students: How Shall We Know Them? (2 nd revised ed.). London: Kogan Page.
Samuelowicz, K. (1994) Teaching conceptions and teaching practice: a case of assessment. In Phenomenography – Philosophy and Practice. Brisbane: Queensland Uiversity of Technology. (343-354). Schon, Donald (1983) The Reflective Practitioner: How professionals think in action New York, Basic Books. Shoop, Linda (2000) Student-Centred Discussion Workshop Notes Ngee Ann Polytechnic, singapore May 25 th and 26 th 2000. Schmidt, H.G. (1983). Problem-based learning. rationale and description. Medical Education, 17, 11-16 Tang, C. (1998) Collaborative Learning and Quality of Assignments in Teaching and Learning in Higher Education. B. Dart & G. Boulton-Lewis (eds.) ACER. Camberwell, Victoria. Vernon, D.T. & Blake, R.L. (1993). Does problem-based learning work? A meta-analysis of evaluative research. Academic Medicine, 68(7) 550-563. von Glasersfeld, E. (1996).Introduction: Aspects of constructivism. In C. Fosnot (Ed.), Constructivism: Theory, perspectives, and practice, (pp.3-7). New York: Teachers College Press. von Glasersfeld, E. (1998).Cognition, construction of knowledge and teaching. (ERIC Document reproduction Service No. ED 294 754. Warren Piper, D., Nulty, D. D., and O'Grady G. (1996) Examination Practices and Procedures in Australian University. Higher Education Division, Department of Employment Education and Training. Evaluations and Investigations Program. AGPS, Canberra. Warren Piper, D. (1994) Are Professors Professional: the organisation of University examinations. London, Jessica Kingsley Publishers. Woods, D.R. (1994). Problem-based Learning: how to gain the most from PBL. Waterdown, Canada |