10.00 – 10.45: Trial lecture. Title: "Evaluation and validation of ML models in the medical context"
12.00 – 16.00: Public defence
Ordinary opponents
- First opponent: Professor Kerstin Bach, Norwegian University of Science and Technology (NTNU), Norway
- Second opponent: Professor Jean-Baptiste Woillard, University of Limoges, France
- Leader of the evaluation committee: Professor Pedro Lind, OsloMet, Norway
Leader of the public defence
Associate Professor Siri Fagernes, OsloMet, Norway
Supervisors
- Main supervisor: Professor Pål Halvorsen, Simula Metropolitan Center for Digital Engineering (SimulaMet)
- Co-supervisor: Professor Michael Riegler, Simula Metropolitan Center for Digital Engineering (SimulaMet)
- Co-supervisor: Professor Tor Paaske Utheim, Oslo University Hospital
- Co-supervisor: Associate Professor Inga Strümke, Norwegian University of Science and Technology (NTNU)
Summary
Artificial intelligence and machine learning (ML) have become a part of our everyday lives. The advancement of such technology affects the society by assisting us in several tasks.
Healthcare and medicine are some domains where ML is believed to play an important role in the future. Studies present impressive results when training ML models to solve medical tasks that are typically performed by humans today.
Implementing ML systems in the clinic can among other things lead to faster and more accurate diagnosis, earlier detection of diseases and personalized treatment planning and patient follow-up.
Taking the shortage of healthcare personnel into account, ML systems could improve the efficiency of the healthcare system and let the medical experts focus on the patients and challenging medical tasks rather than administrative and repetitive simpler tasks.
New challenges with new technology
However, new technology also brings new challenges. Lack of transparency in ML systems for medical applications is likely to limit their use in the clinic. The ML models are typically ‘black boxes’ that healthcare workers do not understand how work.
Moreover, the models might be developed on private datasets where details regarding demographics, age, gender and ethnicity are unavailable. If the development dataset does not represent the population that the model will be used in, the model might fail drastically or discriminate against minority groups.
Further on, the model evaluation process might not be thoroughly described, making it difficult to evaluate the applicability of the ML system.
This thesis contributes to solve the above-mentioned challenges by exploring methods that improve the transparency of medical ML systems.
In addition to applying explanation methods for a variety of medical use cases, we publish open medical datasets, investigate techniques for synthetic healthcare data generation and collaborate with medical experts for evaluating the ML models and explanations.
Expert feedback is important to detect improvements that are necessary for successfully explain the ML systems to healthcare personnel. Moreover, new medical insights might be gained by exploring how the ML models work and analyse medical data.
Important to explain machine learning to medical professionals
The results indicate that ML systems will most likely serve as a useful tool in the clinic in the future.
However, existing explanation methods should be tailored to the specific medical use case to meet the expectations from healthcare personnel.
Explaining the ML systems to experienced experts in the field can give rise to new discoveries about diseases and how to treat them.
Finally, combining knowledge from computer science and medicine enables us to find solutions to tasks that would be impossible to solve without interdisciplinary collaboration.