Conference Paper/Proceeding/Abstract 143 views 34 downloads
Optimising Human Trust in Robots: A Reinforcement Learning Approach
Proceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction, Pages: 1202 - 1206
Swansea University Authors:
Abdullah Alzahrani, Muneeb Ahmad
-
PDF | Accepted Manuscript
Author accepted manuscript document released under the terms of a Creative Commons CC-BY licence using the Swansea University Research Publications Policy (rights retention).
Download (2.42MB)
Abstract
This study explores optimising human-robot trust using reinforcement learning (RL) in simulated environments. Establishing trust in human-robot interaction (HRI) is crucial for effective collaboration, but misaligned trust levels can restrict successful task completion. Current RL approaches mainlyp...
Published in: | Proceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction |
---|---|
ISBN: | 979-8-3503-7893-1 |
Published: |
IEEE
2025
|
Online Access: |
https://dl.acm.org/doi/10.5555/3721488.3721647 |
URI: | https://cronfa.swan.ac.uk/Record/cronfa68696 |
first_indexed |
2025-01-15T12:26:10Z |
---|---|
last_indexed |
2025-04-01T04:42:40Z |
id |
cronfa68696 |
recordtype |
SURis |
fullrecord |
<?xml version="1.0"?><rfc1807><datestamp>2025-03-31T15:14:00.4099834</datestamp><bib-version>v2</bib-version><id>68696</id><entry>2025-01-15</entry><title>Optimising Human Trust in Robots: A Reinforcement Learning Approach</title><swanseaauthors><author><sid>d2f9f67e9bfd515f861a917fe1d00321</sid><firstname>Abdullah</firstname><surname>Alzahrani</surname><name>Abdullah Alzahrani</name><active>true</active><ethesisStudent>false</ethesisStudent></author><author><sid>9c42fd947397b1ad2bfa9107457974d5</sid><ORCID>0000-0001-8111-9967</ORCID><firstname>Muneeb</firstname><surname>Ahmad</surname><name>Muneeb Ahmad</name><active>true</active><ethesisStudent>false</ethesisStudent></author></swanseaauthors><date>2025-01-15</date><abstract>This study explores optimising human-robot trust using reinforcement learning (RL) in simulated environments. Establishing trust in human-robot interaction (HRI) is crucial for effective collaboration, but misaligned trust levels can restrict successful task completion. Current RL approaches mainlyprioritise performance metrics without directly addressing trust management. To bridge this gap, we integrated a validated mathematical trust model into an RL framework and conducted experiments in two simulated environments: Frozen Lake and Battleship. The results showed that the RL model facilitated trust by dynamically adjusting it based on task outcomes, enhancing task performance and reducing the risks of insufficient or extreme trust. Our findings highlight the potential of RL to enhance human-robot collaboration (HRC) and trust calibration in different experimental HRI settings.</abstract><type>Conference Paper/Proceeding/Abstract</type><journal>Proceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction</journal><volume/><journalNumber/><paginationStart>1202</paginationStart><paginationEnd>1206</paginationEnd><publisher>IEEE</publisher><placeOfPublication/><isbnPrint/><isbnElectronic>979-8-3503-7893-1</isbnElectronic><issnPrint/><issnElectronic/><keywords/><publishedDay>4</publishedDay><publishedMonth>3</publishedMonth><publishedYear>2025</publishedYear><publishedDate>2025-03-04</publishedDate><doi/><url>https://dl.acm.org/doi/10.5555/3721488.3721647</url><notes/><college>COLLEGE NANME</college><CollegeCode>COLLEGE CODE</CollegeCode><institution>Swansea University</institution><apcterm/><funders/><projectreference/><lastEdited>2025-03-31T15:14:00.4099834</lastEdited><Created>2025-01-15T12:23:30.0099957</Created><path><level id="1">Faculty of Science and Engineering</level><level id="2">School of Mathematics and Computer Science - Computer Science</level></path><authors><author><firstname>Abdullah</firstname><surname>Alzahrani</surname><order>1</order></author><author><firstname>Muneeb</firstname><surname>Ahmad</surname><orcid>0000-0001-8111-9967</orcid><order>2</order></author></authors><documents><document><filename>68696__33335__eb6ce8181fbb4cc6a9549762138f9f61.pdf</filename><originalFilename>68696.pdf</originalFilename><uploaded>2025-01-15T12:26:01.1156360</uploaded><type>Output</type><contentLength>2539729</contentLength><contentType>application/pdf</contentType><version>Accepted Manuscript</version><cronfaStatus>true</cronfaStatus><documentNotes>Author accepted manuscript document released under the terms of a Creative Commons CC-BY licence using the Swansea University Research Publications Policy (rights retention).</documentNotes><copyrightCorrect>true</copyrightCorrect><language>eng</language><licence>https://creativecommons.org/licenses/by/4.0/deed.en</licence></document></documents><OutputDurs/></rfc1807> |
spelling |
2025-03-31T15:14:00.4099834 v2 68696 2025-01-15 Optimising Human Trust in Robots: A Reinforcement Learning Approach d2f9f67e9bfd515f861a917fe1d00321 Abdullah Alzahrani Abdullah Alzahrani true false 9c42fd947397b1ad2bfa9107457974d5 0000-0001-8111-9967 Muneeb Ahmad Muneeb Ahmad true false 2025-01-15 This study explores optimising human-robot trust using reinforcement learning (RL) in simulated environments. Establishing trust in human-robot interaction (HRI) is crucial for effective collaboration, but misaligned trust levels can restrict successful task completion. Current RL approaches mainlyprioritise performance metrics without directly addressing trust management. To bridge this gap, we integrated a validated mathematical trust model into an RL framework and conducted experiments in two simulated environments: Frozen Lake and Battleship. The results showed that the RL model facilitated trust by dynamically adjusting it based on task outcomes, enhancing task performance and reducing the risks of insufficient or extreme trust. Our findings highlight the potential of RL to enhance human-robot collaboration (HRC) and trust calibration in different experimental HRI settings. Conference Paper/Proceeding/Abstract Proceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction 1202 1206 IEEE 979-8-3503-7893-1 4 3 2025 2025-03-04 https://dl.acm.org/doi/10.5555/3721488.3721647 COLLEGE NANME COLLEGE CODE Swansea University 2025-03-31T15:14:00.4099834 2025-01-15T12:23:30.0099957 Faculty of Science and Engineering School of Mathematics and Computer Science - Computer Science Abdullah Alzahrani 1 Muneeb Ahmad 0000-0001-8111-9967 2 68696__33335__eb6ce8181fbb4cc6a9549762138f9f61.pdf 68696.pdf 2025-01-15T12:26:01.1156360 Output 2539729 application/pdf Accepted Manuscript true Author accepted manuscript document released under the terms of a Creative Commons CC-BY licence using the Swansea University Research Publications Policy (rights retention). true eng https://creativecommons.org/licenses/by/4.0/deed.en |
title |
Optimising Human Trust in Robots: A Reinforcement Learning Approach |
spellingShingle |
Optimising Human Trust in Robots: A Reinforcement Learning Approach Abdullah Alzahrani Muneeb Ahmad |
title_short |
Optimising Human Trust in Robots: A Reinforcement Learning Approach |
title_full |
Optimising Human Trust in Robots: A Reinforcement Learning Approach |
title_fullStr |
Optimising Human Trust in Robots: A Reinforcement Learning Approach |
title_full_unstemmed |
Optimising Human Trust in Robots: A Reinforcement Learning Approach |
title_sort |
Optimising Human Trust in Robots: A Reinforcement Learning Approach |
author_id_str_mv |
d2f9f67e9bfd515f861a917fe1d00321 9c42fd947397b1ad2bfa9107457974d5 |
author_id_fullname_str_mv |
d2f9f67e9bfd515f861a917fe1d00321_***_Abdullah Alzahrani 9c42fd947397b1ad2bfa9107457974d5_***_Muneeb Ahmad |
author |
Abdullah Alzahrani Muneeb Ahmad |
author2 |
Abdullah Alzahrani Muneeb Ahmad |
format |
Conference Paper/Proceeding/Abstract |
container_title |
Proceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction |
container_start_page |
1202 |
publishDate |
2025 |
institution |
Swansea University |
isbn |
979-8-3503-7893-1 |
publisher |
IEEE |
college_str |
Faculty of Science and Engineering |
hierarchytype |
|
hierarchy_top_id |
facultyofscienceandengineering |
hierarchy_top_title |
Faculty of Science and Engineering |
hierarchy_parent_id |
facultyofscienceandengineering |
hierarchy_parent_title |
Faculty of Science and Engineering |
department_str |
School of Mathematics and Computer Science - Computer Science{{{_:::_}}}Faculty of Science and Engineering{{{_:::_}}}School of Mathematics and Computer Science - Computer Science |
url |
https://dl.acm.org/doi/10.5555/3721488.3721647 |
document_store_str |
1 |
active_str |
0 |
description |
This study explores optimising human-robot trust using reinforcement learning (RL) in simulated environments. Establishing trust in human-robot interaction (HRI) is crucial for effective collaboration, but misaligned trust levels can restrict successful task completion. Current RL approaches mainlyprioritise performance metrics without directly addressing trust management. To bridge this gap, we integrated a validated mathematical trust model into an RL framework and conducted experiments in two simulated environments: Frozen Lake and Battleship. The results showed that the RL model facilitated trust by dynamically adjusting it based on task outcomes, enhancing task performance and reducing the risks of insufficient or extreme trust. Our findings highlight the potential of RL to enhance human-robot collaboration (HRC) and trust calibration in different experimental HRI settings. |
published_date |
2025-03-04T08:21:03Z |
_version_ |
1829542887980793856 |
score |
11.058845 |