INTRODUCING ASSESSMENT-BASED LEARNING TO A COMMERCE TOPIC David Forsaith School of Commerce Flinders University, Adelaide SCHOOL OF COMMERCE RESEARCH PAPER SERIES: 01-1 ISSN: 1441-3906 Abstract The response of a class of undergraduate students studying a commerce topic to the introduction of assessment-based learning methods was investigated through short open-ended questionnaires administered in lecture classes. The assessment dispensed with the previous heavy reliance on a final examination, replacing it with a series of continuous non-exam assessments. The surveys sought feedback on student perceptions of how assessment methods influenced their study methods, the understanding they gained, and their attitude to the topic. Overall, the comments were very positive which, combined with a higher average mark, suggest that the change in assessment methods was successful in achieving its objectives of facilitating enhanced student learning. Introduction Traditional assessment methods in tertiary education, especially in business related courses, rely heavily on examinations conducted at the end of the course. Increasingly, the value of this form of assessment as an aid to student learning is being questioned. This paper reports on the effects on student learning, as perceived by students themselves, of non-examination based through-the-semester methods of assessment introduced into an undergraduate course on investments. Traditional tertiary education assessment Tertiary teachers have generally not been trained as educators. Until recently, most of them developed their assessment procedures “on the job” (Murray-Harvey et al, 1996, p25). Ramsden (1992, p183) argues that “teachers in higher education frequently assess as amateurs when the task demands grave professionalism”, and refers to academics’ “ignorance about how to do the job properly”. This picture has been changing as programs to prepare university teachers for their tasks are increasingly being implemented, particularly in Australasia and Europe, although they affect new teachers more than existing teachers (Gibbs, 1998). Academics often see assessment as the last task in a linear sequence of teaching duties to perform (Ramsden, 1992). They perceive assessment as “very separate from the teaching and learning process, something to think about once the curriculum has been devised and plans for delivery finalised” (Brown, 1999, p3). Consequently, they focus mainly on setting exam questions that reflect what they have taught and are of an appropriate degree of difficulty (Ramsden, 1992). Assessment is generally regarded as an institutional requirement. Universities require academics to grade students at the end of topics and courses. More broadly, there is a social expectation that graduates receive a statement of qualification, and assessment is a necessary part of that process. “Employers may want accounts of what the student can do and the student …will also want such a summative verdict” on their university studies (Brown and Knight, 1994, p13). The end-of-semester examination is generally seen by academics as the most expedient way to discharge their assessor obligations to their universities, while the universities have generally accepted it as a sufficiently rigorous process for the awarding of qualifications to their students. Assessment based learning Education theory and research suggests a much broader and more central role for assessment. In this role, assessment is seen as driving the student learning process. For students “assessment always defined the actual curriculum” (Ramsden, 1992, p187). Gibbs (1999, p41) sees assessment as “the most powerful lever teachers have to influence the way students respond to courses and behave as learners”. Students use assessment as a basis for deciding where and when to direct their learning effort and how much effort to make (Brown and Knight, 1994). This suggests students are extrinsicly motivated in the way they study. They seek qualifications and progression more than being motivated by the subject matter (Entwistle, 1998). Students are heavily influenced by the hidden curriculum. They look for assessment clues and use these to drive their study effort. Very little out-of-class student learning activity is unrelated to assessment (Gibbs, 1999). Empirical research over the last quarter of a century has distinguished between surface and deep approaches to learning (Marton and Saljo, 1976; Entwistle and Ramsden, 1983; Biggs, 1989; Entwistle, 1998). The surface approach is an “intention to cope with the task in a minimalist fashion” (Entwistle, 1998, p84) and is generally characterised by rote learning, not integrating the constituent parts of a course and not reflecting on the subject material. In contrast, the deep learning approach seeks understanding of the subject material and being able to relate it to previous knowledge and experience. In the words of Marton and Saljo (1984, p46), the deep approach “is the best, indeed the only, way to understand learning materials” (their emphasis). While students may have a general disposition to either the surface or deep learning approach, they have the capacity to use either approach and will be influenced in their choice in a particular situation by environmental factors, including degree of difficulty of the task, interest in the material, assessment structure and time constraints (Wilson, 1981; Entwistle and Ramsden, 1983; Ramsden, 1992; Entwistle, 1998). This offers tertiary teachers the opportunity to influence students' learning approaches through, inter alia, the assessment structure. Assessment dominated by an end-of-semester examination leads students to distribute their learning efforts unequally over the semester (Gibbs, 1999). Their study is heavily oriented towards intense effort near the end of the semester as the examination draws near. Moreover, that study effort tends to be substantially directed at short-term memorisation rather than deeper understanding (Tan, 1992). This surface approach to study tends to be necessitated by the extreme pressures on students immediately prior to examinations as well as questions often asked in examinations, which require recall as the main skill necessary to pass. (This latter issue is a criticism of the questions that are set rather than examination as a form of assessment per se.) Distributing the assessment throughout the semester is likely to encourage more continuous learning effort by students. This has many advantages. Learning “is an evolutionary process… not a revolutionary process” (Pine and Horne, 1969, p109) and so distributing it over the whole semester is more likely to enhance the learning process. However, frequent testing per se does not necessarily improve student learning if the nature of the assessment reinforces the rote learning approach of students, or if students do not use feedback from assessments to remedy weaknesses in their understanding because it is no longer seen as assessable (Tan, 1992). Therefore, assessment should be designed to further the learning objectives of the curriculum. “Assessment needs to be incremental” so that students have the opportunity to use the assessments as a guide for improving their learning and understanding (Brown, 1999, p11). The use of several assessments over the semester provides the opportunity to introduce variety into the assessment rather than relying solely on formal, summative examinations. This enables a wider range of desired skills to be assessed as examinations are not the most efficient means of assessing all skills (Crooks, 1988a). Critical thinking, for example, may be better assessed in a research essay. A research essay also is likely to encourage deeper learning than an examination requires. Using a variety of assessment formats also caters for differences in the relative strengths and weaknesses in the student population (Brown and Knight, 1994). Deeper learning is more likely where assessments facilitate research effort by the student. The satisfaction the student gains from discovery provides intrinsic motivation that can lead to greater understanding. Even where this assessment is summative, the process should make it less threatening to students and so encourage higher order learning (Fransson, 1977; Tan, 1992). Assessment driven learning creates the opportunity for greater formative assessment to be introduced to the curriculum. An end-of-semester examination is entirely summative in nature. Students only receive a grade. All assessment that contributes directly to a final grade is also summative by definition, but where it is not examination assessment and/or is undertaken during the semester, it provides the opportunity to introduce formative elements. These formative elements include advice or assistance given to students prior to submission, comments on early drafts and comments on the final submission, provided they are timely and of substance. Ideally, the formative element enables students to gain a clear view of their learning goals and for them to more closely approach achievement of those goals (Sadler, 1983). The literature is divided on the virtues of combining formative and summative elements into the same assessment (Murray-Harvey et al, 1996). Tan (1992) and Newble and Cannon (1989) argue for separating these elements while Crooks (1988b) and Sadler (1983,1989) support combining these elements in assessment. A persuasive argument for the latter view is the importance of summative assessment in driving the student learning effort. According to Brown (1999, p6), they are not opposites but “ends of the same continuum”. Changes introduced to the Investments course assessment structure The Investments course at Flinders University had been presented in the traditional format of tertiary education since its inception. The end-of-semester examination was the principal means of assessment contributing around 70-80% to students’ final grades. While the teaching had always been enthusiastic and well intentioned, assessment had generally been seen as following the learning process rather than being part of, or even driving, the learning process. In 1999 a decision was made to move to a through-the-semester assessment regime with the main intention of achieving a more consistent learning effort from students over the course of the semester. It was also intended to encourage and facilitate deeper student learning. This assessment structure was very different from previous years in the Investments subject. It was also very different from the assessment, which the students (who were generally in their final semester of a three-year degree) were familiar with from other subjects in their degree studies. The increase in the number of assessments that this entailed also provided the opportunity for additional variety in assessments that were set. While this was a major change to the assessment structure, it is regarded as one step in an on-going process which will see all aspects of the assessments closely examined to improve the contribution they can make to the learning process of students. The revised assessment structure comprised a research essay (10% of total assessment), an investments game played over approximately two-thirds of the semester (20%), three problem assignments (20%, 15% and 15% respectively) and student participation in tutorial discussion throughout the semester (20%). The due dates of the assessments were distributed as evenly as possible over the semester with no examination as part of the assessment structure. There was no flexibility in the assessment to allow students to opt for certain forms of assessment and avoid others. If this were introduced an end-of-semester examination could be necessary as a quality control (Freeman, 1997). The tutorial participation assessment was essentially formative in nature to encourage preparation, attendance and discussion in the classes. Incorrect answers were not penalised in the assessment as learning from mistakes was considered to be an important part of the learning process. The investments game was also largely formative. Students were encouraged to discuss their weekly trading activities in class and there were ample opportunities for students to seek assistance and obtain feedback before they had to submit the assignment for final assessment. Student effort was a major contributor to their mark in this assignment. The other assessments were essentially summative assessments, although feedback obtained from the earlier assessments was designed to assist students in preparing later assignments. Most assessments, therefore, included both summative and formative elements. The summative element was present in all assessments, with the formative element more important in some assessments than in others. The survey instrument In order to assess the impact of the new assessment structure on student learning, the class was surveyed by written questionnaires at the beginning and end of the semester. The surveys were anonymous in order to encourage a high response rate. However this did not allow the responses of individual students at the beginning and end of the semester to be matched. The surveys were administered in the first and last lectures of the semester and the response rates of those present was close to 100%. The first survey was administered before students were informed of the assessment structure for the subject. Students were asked their preferred form of assessment, choosing between exams only, a mix of exams and other assessment and no exams. They were also asked their reasons for their preference and how the structure of the assessment mix would affect their learning in the topic. Students were then asked about their study habits. This questionnaire is reproduced in appendix A (some additional questions at the end of the survey that were not related to assessment are not reproduced). The survey conducted at the end of the semester was open-ended. Students were asked how the set of assessment methods used influenced or affected the way they studied, the understanding they gained and their attitude to the topic. This questionnaire is reproduced in appendix B. From a class with just over 100 students enrolled, 88 surveys were completed in the first lecture and 57 in the final lecture. (Attendance is generally high for the first lecture where students find out about the course and its requirements and because there are few competing pressures on students’ time. By the last lecture, students are under greater pressure to complete final assignments in most of their subjects and this adversely affects attendance.) The student population was overwhelmingly fulltime students under the age of 25. The class was approximately evenly comprised of males and females in both surveys. Australian students comprised 59% of respondents to the first survey but only 46% in the second survey indicating that foreign students were more consistent in their attendance at lectures. Results Beginning-of-semester survey Almost 60% of respondents to the initial survey preferred a combination of examination and assignments, while one-third preferred no examination. Less than 5% preferred the examination only option. In their previous studies, students had generally encountered assessment structures comprising examinations and one or two assignments, with the examination typically comprising at least two-thirds of the final grade in most topics. Many respondents in the first survey argued for greater weighting to be given to assignments. Students were asked to explain their preference and many did. Those wanting assignments included in the assessment gave a range of reasons including excessive pressure of examinations, incentive to study during the semester, and deeper learning throughout the semester. Two of the respondents arguing for no examinations gave subject specific reasons along the lines that investments is “a practical subject” closely related to current issues and should be based on research assignments. The main reason advanced by students for including examinations in the assessment was to ensure assessment covered the whole topic. The concern of some students that assignments do not adequately cover the subject content was consistent with their previous academic experience where typically only one or two assignments were set covering limited areas of content and counting for a minor part of the overall assessment grade for the subject. Interestingly, many comments from students supported the arguments that assessment drives student learning effort and that examinations are low level assessors that do not adequately test students’ understanding. “I find that I would learn more with just assignments as it would make me review work during the semester instead of leaving it until the end when I tend to learn what I think will make me pass rather than all topics.” “Exam is about memorising, not understanding.” One informed foreign student, in arguing for assessment based on assignments only, noted that this was: “a modern approach to study … lots of universities tend to follow this style … a better understanding of the subject will be achieved.” There were 89 responses to the question on study habits with 29 indicating that most study is done just before exams, 21 suggesting that most work during the semester is on assignments and 39 pointing to a consistent learning effort applied to all content throughout the semester. Three respondents chose both the first two options and there were two non-respondents to this question. End-of-semester survey Overall, the comments of students in the end-of-semester survey were very positive towards their experience in the topic. Approaches to learning The most frequent comments about the effect of the assessment on the way students studied were that they were more consistent in their study efforts during the semester (21 included this comment) and that the amount of study they did increased (16 comments). “It assisted me a great deal in my studying. By doing an assignment on nearly every topic, it enabled me to understand the topic presented a lot better than just the lecture by itself, since for the assignments further research was required.” Only one student reported a reduced study effort. Impact on student understanding About half of the students commented that the assessment methods improved the knowledge and/or understanding that they gained from the topic, while only seven responses suggested that it made no difference. “I feel that I have a good understanding of this topic due to all the work I have put in…having a good understanding of the topic content has resulted in me having a positive attitude towards this topic.” Those respondents who used the word “knowledge” and those who used the word “understanding” were approximately evenly divided. Whether they had the same perceptions of the difference in meaning as used by authors such as Brown et al (1997) or Entwistle (1998) is arguable. These authors regard the former as surface learning and only the latter as deep learning. Whatever the case, students perceived that the assessment methods had increased their learning. When asked what they liked about the assessment methods, half of the respondents commented on the absence of examinations. Other frequent responses were greater understanding gained (7 responses) and more consistent workload (6 responses). Most of the suggestions on improving the assessment procedures were comments on matters of detail. Six students suggested that the structure of assessment be fundamentally changed by reintroducing examinations into the assessment. Academic performance of students Student grades in the topic in 1999 were, on average, a little higher than in previous years. This result is consistent with the positive feedback from the end-of-semester survey, and is suggestive of the success of assessment methods on student learning. However, other explanatory factors cannot be ruled out. The absence of examinations, with their associated pressure and inevitable surprises, may have been a factor. Moreover, the scope for weaker students to “free-ride” on the efforts of stronger students increases with certain non-examination forms of assessment. This may also have contributed to the higher average result, despite the efforts of teaching staff to be vigilant for such behaviour. Impact on teacher workload The more intensive assessment introduced certainly increased the marking load of the teaching staff, although this was not accurately measured. With additional resources unlikely to be available, this presents challenges for the staff in coming years to reduce their time input without adversely affecting the achievement of the objectives of the change in assessment structure. The use of alternative assessment methods will be explored. Conclusion The assessment changes introduced to the Investments topic represented a major break from the previous assessment structure in two ways: assessment was spread evenly throughout the semester and there was no examination. These changes appear to have had strongly positive effects on students’ learning approaches, their perceptions of their understanding and on their overall attitude to the topic. They have also resulted in an overall improvement in student grades. The changes were strongly endorsed by the student group with criticism largely being confined to matters of detail. All of this suggests that the changes have succeeded in improving student learning in the topic. Many improvements can still be made to the assessment structure. Curriculum development is an on-going process and the assessments in this topic will be progressively refined and developed in the years ahead. The results to date endorse the importance of assessment in driving student learning and the challenge ahead is to enhance that process with further improvements in assessment design. In the immediate future, the design of assessments will be modified to achieve a closer nexus between the assessment content and learning objectives and to enable greater and more timely feedback to students. A wider range of assessment methods will be utilised, with more emphasis on group presentations in class with some peer assessment being introduced. This added variety is desirable for the intrinsic learning merits of these assessment methods and to relieve some of the additional assessment workload that this continuous assessment has brought upon the teaching staff. Appendix A SURVEY OF LEARNING ACTIVITY AND ASSESSMENT ISSUES Please answer the following questions. Tick or circle the appropriate responses where applicable. All responses are anonymous. SECTION A: CLASSIFICATION DATA 1 Respondent’s enrolment status full-time / part-time 2 Respondent’s age less than 25 / 25 or over 3 Respondent’s gender female / male 4 Respondent’s nationality Australian / foreign SECTION B: ASSESSMENT AND LEARNING ISSUES 5 What form of assessment would you prefer? 100% weighting on exam weighting for assignments and exam no exam, 100% weighting on assignments other (please specify) 6 Please give reasons for your answer to the previous question. 7 How does the structure of the assessment affect your learning in a topic? SECTION C: LEARNING ACTIVITIES 8 Which responses best describe your study habits? Most of my study is done just before the exams During the semester, I work mainly on assignments I try to learn all of the content in the topic throughout the semester Appendix B SURVEY OF LEARNING ACTIVITY AND ASSESSMENT ISSUES Please answer the following questions. All answers are anonymous. SECTION A: CLASSIFICATION DATA 1 Respondent’s enrolment status full-time / part-time 2 Respondent’s age less than 25 / 25 or over 3 Respondent’s gender female / male 4 Respondent’s nationality Australian / foreign SECTION B: ASSESSMENT AND LEARNING ISSUES The assessment methods in this topic centred on multiple assignments rather than being based primarily on an end-of-semester exam. 5 Compared with your previous experiences of topics assessed mainly through examination, how did the set of assessment methods used in this topic influence and affect; (i) the way in which you studied this topic? (ii) the understanding you gained from this topic? (iii) your attitude to this topic? 6 What did you like about this set of assessment methods? 7 What changes should be made to the assessment procedures, and why? 8 Are there any other comments you would like to make about this topic? Bibliography Biggs, J. B. (1989), “Approaches to the enhancement of tertiary education”, Higher Education Research and Development 8, 7-25 Brown, G., Bull, J. and Pendlebury, M. (1997), Assessing Student Learning in Higher Education, Routledge, London. Brown, S. (1999), “Institutional Strategies for Assessment”, chapter 1 in S. Brown and A. Glasner (eds) (1999), Assessment Matters in Higher Education, Society for Research into Higher Education and Open University Press, Buckingham, UK Brown, S. and Knight, P. (1994), Assessing Learners in Higher Education, Kogan Page, London. Crooks, T. (1988a) Assessing Student Performance. HERDSA, Sydney. Crooks, T.J. (1988b) “The Impact of Classroom Practices on Students”, Review of Educational Research 58, 438-481. Entwistle, N. (1998), “Improving Teaching through Research on Student Learning”, chapter 4 in J. J. F. Forest (ed), University Teaching: International Perspectives, Garland Publishing, New York Entwistle, N. J. and Ramsden, P. (1983), Understanding Student Learning, Croom Helm, London. Fransson, A. (1977), “On qualitative differences in learning: IV – Effects of intrinsic motivation and extrinsic test anxiety on process and outcome”, British Journal of Educational Psychology, 42, 244-257. Freeman, M. (1997), “Flexibility and Assessment: A Combination Sure to Motivate Student Learning”, in R. Ballantyne, J.Bain and J. Packer (eds) (1997), Reflecting on University Teaching: Academics’ Stories, AGPS, Canberra. Gibbs, G. (1998), “The Preparation of University Teachers: A Cross-National Approach”, chapter 9 in J. J. F. Forest (ed), University Teaching: International Perspectives, Garland Publishing, New York Gibbs, G. (1999), “Using Assessment Strategically to Change the Way Students Learn”, chapter 4 in S. Brown and A. Glasner (eds) (1999), Assessment Matters in Higher Education, Society for Research into Higher Education and Open University Press, Buckingham, UK Marton, F. and Saljo, R. (1976), “On qualitative differences in learning, outcome and process”, British Journal of Educational Psychology, 46, 4-11. Marton, F. and Saljo, R. (1984), “Approaches to Learning” in F. Marton, D. Hounsell and N. Entwistle (eds) The Experience of Learning, Scottish Academic Press, Edinburgh. Murray-Harvey, R., Silins, H and Orrell, J. (1996), Assessment for Learning: A Guide for Academics, School of Education, Flinders University, Adelaide. Newble, D. and Cannon. R. (1989), A Handbook for Teachers in Universities and Colleges, Kogan Page, London. Pine G J and Horne P J (1969) “Principles and Conditions for Learning in Adult Education”, Adult Leadership, Oct 1969, 108-133. Ramsden, P. (1992), Learning to Teach in Higher Education, Routledge, New York. Sadler, D R. (1983) “Evaluation and the Improvement of Academic Learning”, Journal of Higher Education, 54, 60-79. Sadler, DR (1989), “Formative Assessment and the Design of Instructional Systems”, Instructional Science, 18, 119-144. Tan, C M. (1992), “An Evaluation of Continuous Assessment in the Teaching of Physiology”, Higher Education, 23, 255-272. Wilson, J.D. (1981), Student Learning in Higher Education, Croom Helm, London.