No Cover Image

Conference Paper/Proceeding/Abstract 143 views 34 downloads

Optimising Human Trust in Robots: A Reinforcement Learning Approach

Abdullah Alzahrani, Muneeb Ahmad Orcid Logo

Proceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction, Pages: 1202 - 1206

Swansea University Authors: Abdullah Alzahrani, Muneeb Ahmad Orcid Logo

  • 68696.pdf

    PDF | Accepted Manuscript

    Author accepted manuscript document released under the terms of a Creative Commons CC-BY licence using the Swansea University Research Publications Policy (rights retention).

    Download (2.42MB)

Abstract

This study explores optimising human-robot trust using reinforcement learning (RL) in simulated environments. Establishing trust in human-robot interaction (HRI) is crucial for effective collaboration, but misaligned trust levels can restrict successful task completion. Current RL approaches mainlyp...

Full description

Published in: Proceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction
ISBN: 979-8-3503-7893-1
Published: IEEE 2025
Online Access: https://dl.acm.org/doi/10.5555/3721488.3721647
URI: https://cronfa.swan.ac.uk/Record/cronfa68696
first_indexed 2025-01-15T12:26:10Z
last_indexed 2025-04-01T04:42:40Z
id cronfa68696
recordtype SURis
fullrecord <?xml version="1.0"?><rfc1807><datestamp>2025-03-31T15:14:00.4099834</datestamp><bib-version>v2</bib-version><id>68696</id><entry>2025-01-15</entry><title>Optimising Human Trust in Robots: A Reinforcement Learning Approach</title><swanseaauthors><author><sid>d2f9f67e9bfd515f861a917fe1d00321</sid><firstname>Abdullah</firstname><surname>Alzahrani</surname><name>Abdullah Alzahrani</name><active>true</active><ethesisStudent>false</ethesisStudent></author><author><sid>9c42fd947397b1ad2bfa9107457974d5</sid><ORCID>0000-0001-8111-9967</ORCID><firstname>Muneeb</firstname><surname>Ahmad</surname><name>Muneeb Ahmad</name><active>true</active><ethesisStudent>false</ethesisStudent></author></swanseaauthors><date>2025-01-15</date><abstract>This study explores optimising human-robot trust using reinforcement learning (RL) in simulated environments. Establishing trust in human-robot interaction (HRI) is crucial for effective collaboration, but misaligned trust levels can restrict successful task completion. Current RL approaches mainlyprioritise performance metrics without directly addressing trust management. To bridge this gap, we integrated a validated mathematical trust model into an RL framework and conducted experiments in two simulated environments: Frozen Lake and Battleship. The results showed that the RL model facilitated trust by dynamically adjusting it based on task outcomes, enhancing task performance and reducing the risks of insufficient or extreme trust. Our findings highlight the potential of RL to enhance human-robot collaboration (HRC) and trust calibration in different experimental HRI settings.</abstract><type>Conference Paper/Proceeding/Abstract</type><journal>Proceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction</journal><volume/><journalNumber/><paginationStart>1202</paginationStart><paginationEnd>1206</paginationEnd><publisher>IEEE</publisher><placeOfPublication/><isbnPrint/><isbnElectronic>979-8-3503-7893-1</isbnElectronic><issnPrint/><issnElectronic/><keywords/><publishedDay>4</publishedDay><publishedMonth>3</publishedMonth><publishedYear>2025</publishedYear><publishedDate>2025-03-04</publishedDate><doi/><url>https://dl.acm.org/doi/10.5555/3721488.3721647</url><notes/><college>COLLEGE NANME</college><CollegeCode>COLLEGE CODE</CollegeCode><institution>Swansea University</institution><apcterm/><funders/><projectreference/><lastEdited>2025-03-31T15:14:00.4099834</lastEdited><Created>2025-01-15T12:23:30.0099957</Created><path><level id="1">Faculty of Science and Engineering</level><level id="2">School of Mathematics and Computer Science - Computer Science</level></path><authors><author><firstname>Abdullah</firstname><surname>Alzahrani</surname><order>1</order></author><author><firstname>Muneeb</firstname><surname>Ahmad</surname><orcid>0000-0001-8111-9967</orcid><order>2</order></author></authors><documents><document><filename>68696__33335__eb6ce8181fbb4cc6a9549762138f9f61.pdf</filename><originalFilename>68696.pdf</originalFilename><uploaded>2025-01-15T12:26:01.1156360</uploaded><type>Output</type><contentLength>2539729</contentLength><contentType>application/pdf</contentType><version>Accepted Manuscript</version><cronfaStatus>true</cronfaStatus><documentNotes>Author accepted manuscript document released under the terms of a Creative Commons CC-BY licence using the Swansea University Research Publications Policy (rights retention).</documentNotes><copyrightCorrect>true</copyrightCorrect><language>eng</language><licence>https://creativecommons.org/licenses/by/4.0/deed.en</licence></document></documents><OutputDurs/></rfc1807>
spelling 2025-03-31T15:14:00.4099834 v2 68696 2025-01-15 Optimising Human Trust in Robots: A Reinforcement Learning Approach d2f9f67e9bfd515f861a917fe1d00321 Abdullah Alzahrani Abdullah Alzahrani true false 9c42fd947397b1ad2bfa9107457974d5 0000-0001-8111-9967 Muneeb Ahmad Muneeb Ahmad true false 2025-01-15 This study explores optimising human-robot trust using reinforcement learning (RL) in simulated environments. Establishing trust in human-robot interaction (HRI) is crucial for effective collaboration, but misaligned trust levels can restrict successful task completion. Current RL approaches mainlyprioritise performance metrics without directly addressing trust management. To bridge this gap, we integrated a validated mathematical trust model into an RL framework and conducted experiments in two simulated environments: Frozen Lake and Battleship. The results showed that the RL model facilitated trust by dynamically adjusting it based on task outcomes, enhancing task performance and reducing the risks of insufficient or extreme trust. Our findings highlight the potential of RL to enhance human-robot collaboration (HRC) and trust calibration in different experimental HRI settings. Conference Paper/Proceeding/Abstract Proceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction 1202 1206 IEEE 979-8-3503-7893-1 4 3 2025 2025-03-04 https://dl.acm.org/doi/10.5555/3721488.3721647 COLLEGE NANME COLLEGE CODE Swansea University 2025-03-31T15:14:00.4099834 2025-01-15T12:23:30.0099957 Faculty of Science and Engineering School of Mathematics and Computer Science - Computer Science Abdullah Alzahrani 1 Muneeb Ahmad 0000-0001-8111-9967 2 68696__33335__eb6ce8181fbb4cc6a9549762138f9f61.pdf 68696.pdf 2025-01-15T12:26:01.1156360 Output 2539729 application/pdf Accepted Manuscript true Author accepted manuscript document released under the terms of a Creative Commons CC-BY licence using the Swansea University Research Publications Policy (rights retention). true eng https://creativecommons.org/licenses/by/4.0/deed.en
title Optimising Human Trust in Robots: A Reinforcement Learning Approach
spellingShingle Optimising Human Trust in Robots: A Reinforcement Learning Approach
Abdullah Alzahrani
Muneeb Ahmad
title_short Optimising Human Trust in Robots: A Reinforcement Learning Approach
title_full Optimising Human Trust in Robots: A Reinforcement Learning Approach
title_fullStr Optimising Human Trust in Robots: A Reinforcement Learning Approach
title_full_unstemmed Optimising Human Trust in Robots: A Reinforcement Learning Approach
title_sort Optimising Human Trust in Robots: A Reinforcement Learning Approach
author_id_str_mv d2f9f67e9bfd515f861a917fe1d00321
9c42fd947397b1ad2bfa9107457974d5
author_id_fullname_str_mv d2f9f67e9bfd515f861a917fe1d00321_***_Abdullah Alzahrani
9c42fd947397b1ad2bfa9107457974d5_***_Muneeb Ahmad
author Abdullah Alzahrani
Muneeb Ahmad
author2 Abdullah Alzahrani
Muneeb Ahmad
format Conference Paper/Proceeding/Abstract
container_title Proceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction
container_start_page 1202
publishDate 2025
institution Swansea University
isbn 979-8-3503-7893-1
publisher IEEE
college_str Faculty of Science and Engineering
hierarchytype
hierarchy_top_id facultyofscienceandengineering
hierarchy_top_title Faculty of Science and Engineering
hierarchy_parent_id facultyofscienceandengineering
hierarchy_parent_title Faculty of Science and Engineering
department_str School of Mathematics and Computer Science - Computer Science{{{_:::_}}}Faculty of Science and Engineering{{{_:::_}}}School of Mathematics and Computer Science - Computer Science
url https://dl.acm.org/doi/10.5555/3721488.3721647
document_store_str 1
active_str 0
description This study explores optimising human-robot trust using reinforcement learning (RL) in simulated environments. Establishing trust in human-robot interaction (HRI) is crucial for effective collaboration, but misaligned trust levels can restrict successful task completion. Current RL approaches mainlyprioritise performance metrics without directly addressing trust management. To bridge this gap, we integrated a validated mathematical trust model into an RL framework and conducted experiments in two simulated environments: Frozen Lake and Battleship. The results showed that the RL model facilitated trust by dynamically adjusting it based on task outcomes, enhancing task performance and reducing the risks of insufficient or extreme trust. Our findings highlight the potential of RL to enhance human-robot collaboration (HRC) and trust calibration in different experimental HRI settings.
published_date 2025-03-04T08:21:03Z
_version_ 1829542887980793856
score 11.058845