No Cover Image

E-Thesis 87 views 43 downloads

Explainable Artificial Intelligence for Medical Science / JAMIE DUELL

Swansea University Author: JAMIE DUELL

DOI (Published version): 10.23889/SUthesis.66607

Abstract

Explainable Artificial Intelligence (XAI) is at the forefront of Artificial Intelligence (AI) research. As the development of AI has become increasingly complex with modern day computational capabilities, the transparency of the AI models decreases. This promotes the necessity of XAI, as it is illic...

Full description

Published: Swansea University, Wales, UK 2024
Institution: Swansea University
Degree level: Doctoral
Degree name: Ph.D
Supervisor: Seisenberger, M.; & Aarts, G
URI: https://cronfa.swan.ac.uk/Record/cronfa66607
Tags: Add Tag
No Tags, Be the first to tag this record!
Abstract: Explainable Artificial Intelligence (XAI) is at the forefront of Artificial Intelligence (AI) research. As the development of AI has become increasingly complex with modern day computational capabilities, the transparency of the AI models decreases. This promotes the necessity of XAI, as it is illicit as per the General Data Protection Regulations (GDPR) “right to an explanation” to not provide a person with an explanation given a decision reached after algorithmic judgement. The latter is crucial in critical fields such as Healthcare, Finance and Law. For this thesis, the Healthcare field and morespecifically Electronic Health Records are the main focus for the development and application of XAI methods.This thesis offers prospective approaches to enhance the explainability of Electronic Health Records (EHRs). It presents three different perspectives that encompass the Model, Data, and the User, aimed at elevating explainability. The model perspective draws upon improvements to the local explainability of black-box AI methods. The data perspective enables an improvement to the quality of the data provided for AI methods, such that the XAI methods applied to the AI models account for a key property of missingness. Finally, the user perspective provides an accessible form of explainability by allowing less experienced users to have an interface to use both AI and XAI methods.Thereby, this thesis provides new innovative approaches to improve the explanations that are given for EHRs. This is verified through empirical and theoretical analysis of a collection of introduced and existing methods. We propose a selection of XAI methods that collectively build upon current leading literature in the field. Here we propose the methods Polynomial Adaptive Local Explanations (PALE) for patient specific explanations, both Counterfactual-Integrated Gradients (CF-IG) and QuantifiedUncertainty Counterfactual Explanations (QUCE) that utilise counterfactual thinking, Batch-Integrated Gradients (Batch-IG) to address the temporal nature of EHR data and Surrogate Set Imputation (SSI) that addresses missing value imputation. Finally, we propose a tool called ExMed that utilises XAI methods and allows for the ease of access for AI and XAI methods.
Item Description: A selection of content is redacted or is partially redacted from this thesis to protect sensitive and personal information
Keywords: Explainability, Interpretability, Artificial Intelligence, Machine Learning, Deep Learning
College: Faculty of Science and Engineering
Funders: EPSRC Doctoral Training Grant