There are many image modalities such as conventional imaging, Ultrasound (US), Magnetic Resonance Imaging (MRI), and Computed Tomography (CT) that are routinely utilised by the clinicians to determine the presence and nature of a disease. Artificial Intelligence (AI) application in medical imaging is based on complex models that provide state of the art results in image classification to positively identify the onset of a disease. Recent AI techniques have transformed the classification process and can provide results comparable to or better than human expert. However, compared to earlier machine learning models that had explainability, it is no longer possible to observe and trace the steps followed by a model.
The complexity of the models and their ‘black box’ nature makes it difficult to understand why the model made a particular decision which also affects the accompanying trust. This information is very important due to the criticality of the correct diagnosis of a disease where an incorrect diagnosis can be fatal, or result in unnecessary or inappropriate treatment affecting patient outcomes. Explainable AI aims to make it easier for the human in the loop to understand not only ‘what’ decision was made but ‘why’ it was made. This can not only help in making the process understandable but also increases the degree of trust and confidence level in the output of an AI based system.
This project will develop algorithms and models for making medical image classification process understandable for medical professionals. The major objectives are as follows:
(i) Critically analysing the state of the art and the gap of knowledge in medical image classification and explainable AI in medical applications;
(ii) Designing a medical image analysis framework with justifiable explainability levels for medical experts;
(iii) Experimentally evaluating the effectiveness and performance of explainable medical image classification system
There is also a possibility of working with National Health Service (NHS) Scotland for this project.
The information you submit to Glasgow Caledonian University will only be used by them or their data partners to deal with your enquiry, according to their privacy notice. For more information on how we use and store your data, please read our privacy statement.
Based on your current searches we recommend the following search filters.
Check out our other PhDs in Glasgow, United Kingdom
Check out our other PhDs in Data Analysis
Start a new search with our database of over 4,000 PhDs
Based on your current search criteria we thought you might be interested in these.
Doctor of Engineering (EngD): More efficient deep learning for medical image analysis (Canon Medical)
Heriot-Watt University
Medical Image Analysis using Deep Learning
Kingston University
Understanding multi-component remedies from historical medical texts
University of Warwick