This document discusses the redesign of rubrics for assessment at Deakin University to focus on learning standards rather than marks. It describes issues with the previous system where students focused on marks rather than feedback. The researchers redesigned rubrics with clear standards and descriptors to provide better feedback for students. They trained markers and had students self-assess drafts using the rubrics. The new system led to improved feedback and fewer remark requests. The document advocates for standards-based assessment to better support student learning.
1. Deakin University CRICOS Provider Code: 00113B
Susie Macfarlane, Dr Dominique
Condo & Dr Claire Margerison
Assessment for learning not
marks:
Designing standards-based
rubrics for students’ learning
and feedback literacy
3. Deakin University CRICOS Provider Code: 00113B
Issues: Unit / course context
1. Students weren’t looking at rubric
2. Students complaining about marks, but were not
focussed on learning or feedback, as focussed on marks
3. There is no clear statement of the standard we expect
students to achieve
4. Students didn’t use feedback to improve their work
4. Deakin University CRICOS Provider Code: 00113B
Literature:
Impact of
rubrics, grades
and feedback on
learning
5. Deakin University CRICOS Provider Code: 00113B
Q1. How does the assessment support students’
understanding of the standard?
Do grades tell us what a student can or can’t do?
Does a pass mark represent the expected standard?
Boud, D. (2017). Standards-based assessment for an era of increasing transparency. In Carless, D., Bridges, S., Chan, C.
and Glofcheski, R. (Eds). Scaling up assessment for learning in higher education. Dordrecht: Springer, 19-31.
6. Deakin University CRICOS Provider Code: 00113B
Marks (out of 100) provide false precision about
standard of work
“If only three or four categories of difference can
be ascertained, then this is all that can be
legitimately reported. Any finer grain of reporting
is spurious and cannot be justified.”
(Boud, 2017)
Q1. How does the assessment support students’
understanding of the standard?
7. Deakin University CRICOS Provider Code: 00113B
Q2. How does the assessment support learning?
Marks provide very little
information the student can
use to improve
Students need rich information
Boud, D. (2017). Standards-based assessment for an era of increasing transparency. In Carless, D., Bridges, S., Chan, C.
and Glofcheski, R. (Eds). Scaling up assessment for learning in higher education. Dordrecht: Springer, 19-31.
8. Deakin University CRICOS Provider Code: 00113B
Who sets the standard?
Do students engage with and trying to understand the
standard?
Who evaluates the work?
Who provides feedback?
Is feedback a one-way flow of information, or a dialogue?
Do students use the feedback?
Do we know if students’ work improved?
Q3. How does the assessment support students’ active participation?
9. Deakin University CRICOS Provider Code: 00113B
Definition:
A rubric must have:
• evaluative criteria
• quality definitions for those criteria at particular levels, and
• a scoring strategy.
Dawson, 2015
Popham, 1997
10. Deakin University CRICOS Provider Code: 00113B
Rubrics:
• Make expectations and criteria explicit
• Facilitate feedback and self-assessment
• Support students’ self-regulated learning
• Students like rubrics and find them useful
(Jonsson and Svingby, 2007)
(Jonsson and Svingby, 2007)
Jonsson and Pandero, 2017
Reddy and Adrade, 2017
12. Aims
Students engage with
feedback
Focus students and
markers on assessment
criteria (standards and
capabilities)
(rather than marks)
Students evaluate work
against the expected
standard
13. Feedback is information
about the gap
between the actual level and the reference level
which is used to alter the gap in some way.
(Ramaprasad 1983 p.4, cited in Boud and Molloy, 2013)
Feedback definition:
14. Aims
Students use feedback
to improve their work
Students and markers
evaluate work against
the expected standard
Focus students and
markers on assessment
criteria (standards and
capabilities)
(rather than marks)
16. Aim
Focus on standards and
capabilities
(rather than marks)
Rewrote assessment criteria: authentic cognitive skills or task that can be evidenced
Rewrote the rubric with 3 levels, one at the expected standard, clear descriptors
Framed language around development and learning
Strategies
21. Aim
Focus on standards and
capabilities
(rather than marks)
Students and markers
evaluate work against
the expected standard
Students submit a draft and evaluate it using the rubric.
Markers also provide feedback.
Removed marks from the rubric students used to self evaluate
Strategies
Rewrote assessment criteria: authentic cognitive skills or task that can be evidenced
Rewrote the rubric with 3 levels, one at the expected standard, clear descriptors
Framed language around development and learning
22. Engaging students in the rubric.
Deakin University CRICOS Provider Code: 00113B
Students create the rubric
Students critique or contribute to the rubric
Students use the rubric to evaluate their own or others’ work
Students use the rubric to evaluate an example
Teacher uses rubric to evaluate an example
Teacher shows high and low quality work
Teacher explains rubric
Nicol (2010) Assess. & Eval. in HE
Student engagement with rubric
Fromlowesttohighestlevelofparticipation
24. Rewrote assessment criteria: authentic cognitive skills or task that can be evidenced
Rewrote the rubric with 3 levels, one at the expected standard, clear descriptors
Framed language around development and learning
Aim
Students use feedback
to improve their work
Focus on standards and
capabilities
(rather than marks)
Students self review
and receive feedback
on draft
Strategies
Students submit a draft and evaluate it using the rubric.
Markers also provide feedback on draft.
Removed marks from the rubric students used to self evaluate
26. Students submit a draft and evaluate it using the rubric.
Markers also provide feedback on draft.
Rewrote assessment criteria: authentic cognitive skills or task that can be evidenced
Rewrote the rubric with 3 levels, one at the expected standard, clear descriptors
Framed language around development and learning
Aim
Students use feedback
to improve their work
Students resubmit assignment explaining their use of feedback
PD of Unit Chairs
Trained markers to provide feedback on the standard student has achieved on Assessment Criteria
Focus on standards and
capabilities
(rather than marks)
Students self review
and receive feedback
on draft
Strategies
Removed marks from the rubric students used to self evaluate
29. Unit Chair perspective
No remark requests since new rubric
Feedback students can use: 50% on draft
Students’ self assessment
Alignment - Markers provide feedback on students’ self evaluation
Confidence as assessor
30. Deakin University CRICOS Provider Code: 00113B
Response from markers
“As a marker, I felt that self assessed rubrics were very helpful for providing meaningful
and individualised feedback.”
“Knowing the level the student felt they were at definitely helped direct feedback and
helped to justify any discrepancies between marks given and the students perceived
level.”
“Being able to provide feedback based not only on the work the student produced, but
also on their planned direction (self identified areas to improve) I believe meant
constructive feedback was easier to provide.”
31. Deakin University CRICOS Provider Code: 00113B
Outcomes: self evaluation
Asking students to self-mark helps fine tune their abilities to review and improve
their work and helps me provide more individualised feedback. Self-marking uniquely
enables me to provide feedback on how accurately students perceive their own work, for
example a student may think that they should focus on citing more references when
really they need to provide more details about the references they identify.
By helping students review their own work more accurately they are better able to
improve their work independently.
I can also better provide feedback on the skills individual students find the most difficult
and want help with e.g. if a student identified that they found it difficult to source and
discuss peer reviewed literature I can provide specific feedback about how they did or
didn't achieve this and further recommend how to improve. At the end of this
process if a student improved on a weakness it is easier for an assessor like me to identify
this and commend them on their hard work and show that the effort paid off.
32. Deakin University CRICOS Provider Code: 00113B
Response from markers
“I think the rubric is easy to use and provide feedback with.
It ensures more consistency in the assessment/feedback process and I think
makes the expectations clear to the marker and student alike.
However it can sometimes be a little too simple and there is room for
interpretation within each of the criteria within the rubric.”
33. Deakin University CRICOS Provider Code: 00113B
1. Replaced exams with other assessment
2. Made a common structure for rubric
3. Implemented students’ self assessment using the rubric
Course wide changes in M.Diet
35. Deakin University CRICOS Provider Code: 00113B
References
Ajjawi & Boud, D (2018) Examining the nature and effects of feedback dialogue, Assessment & Evaluation in Higher
Education, 43:7, 1106-1119, DOI: 10.1080/02602938.2018.1434128
Barton, K, Schofield, S, McAleer, S, Ajjawi, R. (2016) Translating evidence-based guidelines to improve feedback practices: the
interact case study, BMC Medical Education, 16: 53
Bearman M and Ajjawi R (2018) From “Seeing Through” to “Seeing With”: Assessment Criteria and the Myths of
Transparency. Frontiers in Education, 3: 96, doi: 10.3389/feduc.2018.00096
Boud, D. (2017). Standards-based assessment for an era of increasing transparency. In Carless, D., Bridges, S., Chan, C. and
Glofcheski, R. (Eds). Scaling up assessment for learning in higher education. Dordrecht: Springer, 19-31.
Boud, D & Molloy, E, (2013) Feedback in Higher and Professional Education: Understanding it and doing it well, Routledge,
UK.
Carless, D (2006) Differing perceptions in the feedback process, Studies in Higher Education, 31(2) 219 – 233
Dawson, P. (2015). Assessment rubrics: towards clearer and more replicable design, research and practice. Assess. Eval. High.
Educ. 42, 347–360. doi: 10.1080/02602938.2015.1111294
Jönsson A., Panadero E. (2017) The Use and Design of Rubrics to Support Assessment for Learning. In: Carless D., Bridges S.,
Chan C., Glofcheski R. (eds) Scaling up Assessment for Learning in Higher Education. The Enabling Power of
Assessment, vol 5. Springer, Singapore
Hattie, J and Timperley, H (2007) The Power of Feedback, Review of Educational Research, 77: 81-112
36. Deakin University CRICOS Provider Code: 00113B
References
Hounsell, D (2015) Flipping feedback, WISE Assessment Briefing #12, CETL, Available at https://t.co/cq1cek57ui
Nicol, D. (2010). From monologue to dialogue: Improving written feedback processes in mass higher education. Assessment
and Evaluation in Higher Education, 35(5), 501-517. doi:10.1080/02602931003786559
Reddy, Y. M., and Andrade, H. (2010). A review of rubric use in higher education. Assessment and Evaluation in Higher
Education, 35, 435–448. doi: 10.1080/02602930902862859
Sadler, D. Royce(2009) 'Grade integrity and the representation of academic achievement', Studies in Higher Education, 34: 7,
807 — 826
37. Key elements of feedback
s
Hattie & Timperley (2007). The power of feedback
Where am
I going?
How am I
going?
Where to
next?
ULOs
Task
Rubrics
Rubrics
Self evaluation
Drafts
Peer feedback
Feedback response
Feedback is used
38. Deakin University CRICOS Provider Code: 00113B
Programmatic assessment
“Marks and grades may for convenience be
aggregated across a programme, but such a process
says little about how a student meets the outcomes
for a programme as a whole.”
Boud, D. (2017). Standards-based assessment for an era of increasing transparency. In Carless, D., Bridges, S., Chan, C.
and Glofcheski, R. (Eds). Scaling up assessment for learning in higher education. Dordrecht: Springer, 19-31.
39. Deakin University CRICOS Provider Code: 00113B
False rigour
“…an explicit model produces outputs (grades) that appear to have
been substantially ‘validated’ through careful attention to all the
steps.
However, the model itself is characterized by indeterminacy, that is
inherently weak. … its implementation creates a veil of rigour that
makes it difficult for learners to question either the process or the
outcome”
Sadler (2009, p.5)
40. 2015 SENS Rubric enhancement project
RUBRIC PRINCIPLES
1. Assessment Criteria that clearly express and allow judgement of the key learning outcomes of the
unit, and are have a clear and obvious to connection to ULOs, CLOs and GLOs
2. Clearly describe the observable, measurable learning outcomes the students demonstrate, not
content topics
3. Focus the assessment criteria on authentic graduate capabilities and knowledge, not the specifics of
a particular assessment task
4. Specifically describe the different levels of performance for each criteria, to give students a clearer
understanding of the characteristics of a good performance and what they need to do to improve
next time
5. Write the Quality Descriptors to clearly communicate the authentic characteristics of a good
performance in the world (e.g. using the Paul and Elder (2010) Universal intellectual standards of
critical thinking: clarity, accuracy, precision, relevance, depth, breadth, significance, logic, fairness)
6. Performance level labels refer to the level of achievement, and clearly indicate the level students are
expected to achieve.
41. Deakin University CRICOS Provider Code: 00113B
Issues: M.Diet
Students: High achieving, competitive & marks
focussed
Staff: Small course team, willingness to make
changes and interested in continuous
improvement cycle
Everyone in team, willingness to come to a
shared understanding
42. Deakin University CRICOS Provider Code: 00113B
Course wide standards based assessment
• Introduced competency based assessment across whole course
• No marks are given
• Students collect portfolio of evidence against the DAA
competencies
• use portfolio in their mock interviews
Added in transferable skills / GLOs
Notes de l'éditeur
Dominique – 2 mins
Dominique
Susie – 2 mins on literature
Susie
“Does the result of an assessment act relate directly to an outcome and standard? Does a grade signify what a student can or can’t do?
Setting a pass mark is not setting a standard. Unless a pass means something real in terms of what is a particular learning outcome, it is irrelevant whether it is 40%, 50% or 60%.
Indeed marks that do not address an outcome are rendered meaningless in this conception of assessment.”
Susie
“Does the result of an assessment act relate directly to an outcome and standard? Does a grade signify what a student can or can’t do?
Setting a pass mark is not setting a standard. Unless a pass means something real in terms of what is a particular learning outcome, it is irrelevant whether it is 40%, 50% or 60%.
Indeed marks that do not address an outcome are rendered meaningless in this conception of assessment.”
Susie
“…an explicit model produces outputs (grades) that appear to have been substantially ‘validated’ through careful attention to all the steps.
However, the model itself is characterized by indeterminacy, that is inherently weak. … its implementation creates a veil of rigour that makes it difficult for learners to question either the process or the outcome”
Susie
We have a policy at Deakin that assessment panels must create a rubric
Rubrics are built into our LMS
Susie
Susie – 3 mins on aims and rubric design
Susie
Susie
Susie
Susie
Susie
Employers tell us graduates cant articulate what they have learnt and what they can do
Susie
Replace vague quality terms with precise terms to help students understand the standard.
Susie
Susie
Susie
Susie
Susie
Susie
The cognitive, social-affective and structural dimensions
Feedback Mark 2 focuses on developing capacity in the learner, to enable them to provide feedback to themselves.
Feedback Mark 1 relies on the teacher providing feedback to the learner, but feedback Mark 2 focuses on developing capacity in the learner, to enable them to provide feedback to themselves.
Initially feedback was conceived of from a behaviouralist perspective –external provision of information based on observable action –feedback was conceptualised in terms of ‘telling’. However this ignored the need for volition on the student’s behalf in order to interact with the feedback in order to make a change. Junior students see feedback as a much more passive process, whereas senior students are more proactive in seeking out feedback.
So we move to a situation where the student must engage with the feedback in order to make the change.
Affective
Relational (Educational alliance)–the relationship between the assessor and the learner is really important if you want students to engage in feedback –the learner must trust and value the opinion of the one giving feedback.
Many of the practical/lab based classes that are in your existing curriculum lend themselves to this feedback ecosystem –where dialogic feedback can occur in informal settings from a range of sources.
Do students think about feedback as a dialogue, or a one way flow of information from the teacher to the student.
Lecturer, tutor, other students, peers (in group assessments, discussions) think of all the examples of when you give informal feedback. Do students recognise all of these as feedback. What strategy could you use to make sure students know that this is how we think about feedback?
Feedback shouldn’t be just linked to assessment tasks, but encourage students to see feedback as coming from many different areas…
So the question is how do we get to this point, how do we move along this continuum to make feedback more sustainable and powerful?
Susie
Susie
Dominique – 3 mins on impact and outcomes
Dominique
Since using the new rubric there have been no requests for re-marking or student complaints
Half of the allocated marking time is given to the marking of the draft. Markers are instructed to provide feedback on the draft and give three main points for improvement before submission of the final assignment.
Markers are instructed to mark the rubric of the draft submission and focus feedback on areas where there is discrepancy between the student and marker.
This method of marking and rubric design has reduced the subjectivity of marking, increasing my confidence as an assessor. It has also allowed students to assess their understanding of the content and assessment task.
Dominique
Dominique
Easy to use
Feedback
Consistency between markers
Makes expectations clearer
Little too simple and frustrated she couldn’t give a wider range of marks
Dominique
How am I going? Can you get students themselves and/or peers to assist in this part, do you ask for drafts, do you ask students to self-mark according to the rubric? Are students given opportunities to develop their evaluative judgement on their own abilities. Are there opportunities for students to engage in any form of dialogue about their feedback? One student on the panel said that clinical skills practice with peer feedback would be effective, as it would be seen as less daunting for students.