No Cover Image

Conference Paper/Proceeding/Abstract 806 views

Using Elo Rating as a Metric for Comparative Judgement in Educational Assessment

Andy Gray, Alma A.M Rahat, Tom Crick, Stephen Lindsay, Darren Wallace, ANDREW GRAY, Alma Rahat Orcid Logo, Tom Crick Orcid Logo

Proceedings of the 6th International Conference on Education and Multimedia Technology, Volume: ICEMT'22, Issue: July 2022, Pages: 272 - 278

Swansea University Authors: ANDREW GRAY, Alma Rahat Orcid Logo, Tom Crick Orcid Logo

Full text not available from this repository: check for access using links below.

DOI (Published version): 10.1145/3551708.3556204

Abstract

Marking and feedback are essential features of teaching and learning, across the overwhelming majority of educational settings and contexts. However, it can take a great deal of time and effort for teachers to mark assessments, and to provide useful feedback to the students. Furthermore, it also cre...

Full description

Published in: Proceedings of the 6th International Conference on Education and Multimedia Technology
ISBN: 978-1-4503-9645-5/22/07
Published: New York, NY, USA ACM 2022
Online Access: http://dx.doi.org/10.1145/3551708.3556204
URI: https://cronfa.swan.ac.uk/Record/cronfa60788
Tags: Add Tag
No Tags, Be the first to tag this record!
first_indexed 2022-08-23T11:51:54Z
last_indexed 2023-01-13T19:21:10Z
id cronfa60788
recordtype SURis
fullrecord <?xml version="1.0" encoding="utf-8"?><rfc1807 xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:xsd="http://www.w3.org/2001/XMLSchema"><bib-version>v2</bib-version><id>60788</id><entry>2022-08-10</entry><title>Using Elo Rating as a Metric for Comparative Judgement in Educational Assessment</title><swanseaauthors><author><sid>70200df2c37f5d14360e8660d93f57a8</sid><firstname>ANDREW</firstname><surname>GRAY</surname><name>ANDREW GRAY</name><active>true</active><ethesisStudent>false</ethesisStudent></author><author><sid>6206f027aca1e3a5ff6b8cd224248bc2</sid><ORCID>0000-0002-5023-1371</ORCID><firstname>Alma</firstname><surname>Rahat</surname><name>Alma Rahat</name><active>true</active><ethesisStudent>false</ethesisStudent></author><author><sid>200c66ef0fc55391f736f6e926fb4b99</sid><ORCID>0000-0001-5196-9389</ORCID><firstname>Tom</firstname><surname>Crick</surname><name>Tom Crick</name><active>true</active><ethesisStudent>false</ethesisStudent></author></swanseaauthors><date>2022-08-10</date><abstract>Marking and feedback are essential features of teaching and learning, across the overwhelming majority of educational settings and contexts. However, it can take a great deal of time and effort for teachers to mark assessments, and to provide useful feedback to the students. Furthermore, it also creates a significant cognitive load on the assessors, especially in ensuring fairness and equity. Therefore, an alternative approach to marking called comparative judgement (CJ) has been proposed in the educational space. Inspired by the law of comparative judgment (LCJ). The key idea here is that the better submission between a pair will be identified by a suitably qualified or experienced assessor. This pairwise comparison for as many pairs as possible can then be used to rank all submissions. Studies suggest that CJ is highly reliable and accurate while making it quick for the teachers. Alternative studies have questioned this claim suggesting that the process can increase bias in the results as the same submission is shown many times to an assessor for increasing reliability. Additionally, studies have also found that CJ can result in the overall marking process taking longer than a more traditional method of marking as information about many pairs must be collected.There is a clear necessity to investigate the efficacy of alternative rating and ranking systems that do not require extensive data on every pair of submissions, to reduce the temporal and cognitive burden on assessors, and bias from observing the same submission repeatedly. In this paper, we investigate Elo, which has been extensively used in rating players in zero-sum games such as chess – for devising a ranking between submissions in a comparative judgement context. We experimented on a large-scale Twitter dataset on the topic of a recent major UK political event ("Brexit", the UK’s political exit from the European Union) to ask users which tweet they found funnier between a pair selected from ten tweets. Our analysis of the data reveals that the Elo rating is statistically significantly similar to the CJ ranking with a Kendall’s tau score of 0.96 and a p-value of 1.5 × 10^−5. We finish with an informed discussion regarding the potential wider application of this approach to a range of educational contexts.</abstract><type>Conference Paper/Proceeding/Abstract</type><journal>Proceedings of the 6th International Conference on Education and Multimedia Technology</journal><volume>ICEMT'22</volume><journalNumber>July 2022</journalNumber><paginationStart>272</paginationStart><paginationEnd>278</paginationEnd><publisher>ACM</publisher><placeOfPublication>New York, NY, USA</placeOfPublication><isbnPrint/><isbnElectronic>978-1-4503-9645-5/22/07</isbnElectronic><issnPrint/><issnElectronic/><keywords/><publishedDay>13</publishedDay><publishedMonth>7</publishedMonth><publishedYear>2022</publishedYear><publishedDate>2022-07-13</publishedDate><doi>10.1145/3551708.3556204</doi><url>http://dx.doi.org/10.1145/3551708.3556204</url><notes/><college>COLLEGE NANME</college><CollegeCode>COLLEGE CODE</CollegeCode><institution>Swansea University</institution><apcterm>Not Required</apcterm><funders/><projectreference/><lastEdited>2023-11-20T15:11:42.6982289</lastEdited><Created>2022-08-10T17:58:41.4310406</Created><path><level id="1">Faculty of Science and Engineering</level><level id="2">School of Mathematics and Computer Science - Computer Science</level></path><authors><author><firstname>Andy</firstname><surname>Gray</surname><order>1</order></author><author><firstname>Alma A.M</firstname><surname>Rahat</surname><order>2</order></author><author><firstname>Tom</firstname><surname>Crick</surname><order>3</order></author><author><firstname>Stephen</firstname><surname>Lindsay</surname><order>4</order></author><author><firstname>Darren</firstname><surname>Wallace</surname><order>5</order></author><author><firstname>ANDREW</firstname><surname>GRAY</surname><order>6</order></author><author><firstname>Alma</firstname><surname>Rahat</surname><orcid>0000-0002-5023-1371</orcid><order>7</order></author><author><firstname>Tom</firstname><surname>Crick</surname><orcid>0000-0001-5196-9389</orcid><order>8</order></author></authors><documents/><OutputDurs/></rfc1807>
spelling v2 60788 2022-08-10 Using Elo Rating as a Metric for Comparative Judgement in Educational Assessment 70200df2c37f5d14360e8660d93f57a8 ANDREW GRAY ANDREW GRAY true false 6206f027aca1e3a5ff6b8cd224248bc2 0000-0002-5023-1371 Alma Rahat Alma Rahat true false 200c66ef0fc55391f736f6e926fb4b99 0000-0001-5196-9389 Tom Crick Tom Crick true false 2022-08-10 Marking and feedback are essential features of teaching and learning, across the overwhelming majority of educational settings and contexts. However, it can take a great deal of time and effort for teachers to mark assessments, and to provide useful feedback to the students. Furthermore, it also creates a significant cognitive load on the assessors, especially in ensuring fairness and equity. Therefore, an alternative approach to marking called comparative judgement (CJ) has been proposed in the educational space. Inspired by the law of comparative judgment (LCJ). The key idea here is that the better submission between a pair will be identified by a suitably qualified or experienced assessor. This pairwise comparison for as many pairs as possible can then be used to rank all submissions. Studies suggest that CJ is highly reliable and accurate while making it quick for the teachers. Alternative studies have questioned this claim suggesting that the process can increase bias in the results as the same submission is shown many times to an assessor for increasing reliability. Additionally, studies have also found that CJ can result in the overall marking process taking longer than a more traditional method of marking as information about many pairs must be collected.There is a clear necessity to investigate the efficacy of alternative rating and ranking systems that do not require extensive data on every pair of submissions, to reduce the temporal and cognitive burden on assessors, and bias from observing the same submission repeatedly. In this paper, we investigate Elo, which has been extensively used in rating players in zero-sum games such as chess – for devising a ranking between submissions in a comparative judgement context. We experimented on a large-scale Twitter dataset on the topic of a recent major UK political event ("Brexit", the UK’s political exit from the European Union) to ask users which tweet they found funnier between a pair selected from ten tweets. Our analysis of the data reveals that the Elo rating is statistically significantly similar to the CJ ranking with a Kendall’s tau score of 0.96 and a p-value of 1.5 × 10^−5. We finish with an informed discussion regarding the potential wider application of this approach to a range of educational contexts. Conference Paper/Proceeding/Abstract Proceedings of the 6th International Conference on Education and Multimedia Technology ICEMT'22 July 2022 272 278 ACM New York, NY, USA 978-1-4503-9645-5/22/07 13 7 2022 2022-07-13 10.1145/3551708.3556204 http://dx.doi.org/10.1145/3551708.3556204 COLLEGE NANME COLLEGE CODE Swansea University Not Required 2023-11-20T15:11:42.6982289 2022-08-10T17:58:41.4310406 Faculty of Science and Engineering School of Mathematics and Computer Science - Computer Science Andy Gray 1 Alma A.M Rahat 2 Tom Crick 3 Stephen Lindsay 4 Darren Wallace 5 ANDREW GRAY 6 Alma Rahat 0000-0002-5023-1371 7 Tom Crick 0000-0001-5196-9389 8
title Using Elo Rating as a Metric for Comparative Judgement in Educational Assessment
spellingShingle Using Elo Rating as a Metric for Comparative Judgement in Educational Assessment
ANDREW GRAY
Alma Rahat
Tom Crick
title_short Using Elo Rating as a Metric for Comparative Judgement in Educational Assessment
title_full Using Elo Rating as a Metric for Comparative Judgement in Educational Assessment
title_fullStr Using Elo Rating as a Metric for Comparative Judgement in Educational Assessment
title_full_unstemmed Using Elo Rating as a Metric for Comparative Judgement in Educational Assessment
title_sort Using Elo Rating as a Metric for Comparative Judgement in Educational Assessment
author_id_str_mv 70200df2c37f5d14360e8660d93f57a8
6206f027aca1e3a5ff6b8cd224248bc2
200c66ef0fc55391f736f6e926fb4b99
author_id_fullname_str_mv 70200df2c37f5d14360e8660d93f57a8_***_ANDREW GRAY
6206f027aca1e3a5ff6b8cd224248bc2_***_Alma Rahat
200c66ef0fc55391f736f6e926fb4b99_***_Tom Crick
author ANDREW GRAY
Alma Rahat
Tom Crick
author2 Andy Gray
Alma A.M Rahat
Tom Crick
Stephen Lindsay
Darren Wallace
ANDREW GRAY
Alma Rahat
Tom Crick
format Conference Paper/Proceeding/Abstract
container_title Proceedings of the 6th International Conference on Education and Multimedia Technology
container_volume ICEMT'22
container_issue July 2022
container_start_page 272
publishDate 2022
institution Swansea University
isbn 978-1-4503-9645-5/22/07
doi_str_mv 10.1145/3551708.3556204
publisher ACM
college_str Faculty of Science and Engineering
hierarchytype
hierarchy_top_id facultyofscienceandengineering
hierarchy_top_title Faculty of Science and Engineering
hierarchy_parent_id facultyofscienceandengineering
hierarchy_parent_title Faculty of Science and Engineering
department_str School of Mathematics and Computer Science - Computer Science{{{_:::_}}}Faculty of Science and Engineering{{{_:::_}}}School of Mathematics and Computer Science - Computer Science
url http://dx.doi.org/10.1145/3551708.3556204
document_store_str 0
active_str 0
description Marking and feedback are essential features of teaching and learning, across the overwhelming majority of educational settings and contexts. However, it can take a great deal of time and effort for teachers to mark assessments, and to provide useful feedback to the students. Furthermore, it also creates a significant cognitive load on the assessors, especially in ensuring fairness and equity. Therefore, an alternative approach to marking called comparative judgement (CJ) has been proposed in the educational space. Inspired by the law of comparative judgment (LCJ). The key idea here is that the better submission between a pair will be identified by a suitably qualified or experienced assessor. This pairwise comparison for as many pairs as possible can then be used to rank all submissions. Studies suggest that CJ is highly reliable and accurate while making it quick for the teachers. Alternative studies have questioned this claim suggesting that the process can increase bias in the results as the same submission is shown many times to an assessor for increasing reliability. Additionally, studies have also found that CJ can result in the overall marking process taking longer than a more traditional method of marking as information about many pairs must be collected.There is a clear necessity to investigate the efficacy of alternative rating and ranking systems that do not require extensive data on every pair of submissions, to reduce the temporal and cognitive burden on assessors, and bias from observing the same submission repeatedly. In this paper, we investigate Elo, which has been extensively used in rating players in zero-sum games such as chess – for devising a ranking between submissions in a comparative judgement context. We experimented on a large-scale Twitter dataset on the topic of a recent major UK political event ("Brexit", the UK’s political exit from the European Union) to ask users which tweet they found funnier between a pair selected from ten tweets. Our analysis of the data reveals that the Elo rating is statistically significantly similar to the CJ ranking with a Kendall’s tau score of 0.96 and a p-value of 1.5 × 10^−5. We finish with an informed discussion regarding the potential wider application of this approach to a range of educational contexts.
published_date 2022-07-13T15:11:43Z
_version_ 1783096256674922496
score 11.037144