XIMED: A Dual-Loop Evaluation Framework Integrating Predictive Model and Human-Centered Approaches for Explainable AI in Medical Imaging

Guardado en:
Detalles Bibliográficos
Publicado en:Machine Learning and Knowledge Extraction vol. 7, no. 4 (2025), p. 168-205
Autor principal: Karagoz Gizem
Otros Autores: Tanir, Ozcelebi, Meratnia Nirvana
Publicado:
MDPI AG
Materias:
Acceso en línea:Citation/Abstract
Full Text + Graphics
Full Text - PDF
Etiquetas: Agregar Etiqueta
Sin Etiquetas, Sea el primero en etiquetar este registro!

MARC

LEADER 00000nab a2200000uu 4500
001 3286316714
003 UK-CbPIL
022 |a 2504-4990 
024 7 |a 10.3390/make7040168  |2 doi 
035 |a 3286316714 
045 2 |b d20251001  |b d20251231 
100 1 |a Karagoz Gizem 
245 1 |a XIMED: A Dual-Loop Evaluation Framework Integrating Predictive Model and Human-Centered Approaches for Explainable AI in Medical Imaging 
260 |b MDPI AG  |c 2025 
513 |a Journal Article 
520 3 |a In this study, a structured and methodological evaluation approach for eXplainable Artificial Intelligence (XAI) methods in medical image classification is proposed and implemented using LIME and SHAP explanations for chest X-ray interpretations. The evaluation framework integrates two critical perspectives: predictive model-centered and human-centered evaluations. Predictive model-centered evaluations examine the explanations’ ability to reflect changes in input and output data and the internal model structure. Human-centered evaluations, conducted with 97 medical experts, assess trust, confidence, and agreements with AI’s indicative and contra-indicative reasoning as well as their changes before and after provision of explainability. Key findings of our study include explanation of sensitivity of LIME and SHAP to model changes, their effectiveness in identifying critical features, and SHAP’s significant impact on diagnosis changes. Our results show that both LIME and SHAP negatively affected contra-indicative agreement. Case-based analysis revealed AI explanations reinforce trust and agreement when participant’s initial diagnoses are correct. In these cases, SHAP effectively facilitated correct diagnostic changes. This study establishes a benchmark for future research in XAI for medical image analysis, providing a robust foundation for evaluating and comparing different XAI methods. 
653 |a Image classification 
653 |a Clinical decision making 
653 |a Image analysis 
653 |a Artificial intelligence 
653 |a Localization 
653 |a Medical imaging 
653 |a Prediction models 
653 |a Explainable artificial intelligence 
653 |a Evaluation 
653 |a Medical personnel 
700 1 |a Tanir, Ozcelebi 
700 1 |a Meratnia Nirvana 
773 0 |t Machine Learning and Knowledge Extraction  |g vol. 7, no. 4 (2025), p. 168-205 
786 0 |d ProQuest  |t Advanced Technologies & Aerospace Database 
856 4 1 |3 Citation/Abstract  |u https://www.proquest.com/docview/3286316714/abstract/embedded/7BTGNMKEMPT1V9Z2?source=fedsrch 
856 4 0 |3 Full Text + Graphics  |u https://www.proquest.com/docview/3286316714/fulltextwithgraphics/embedded/7BTGNMKEMPT1V9Z2?source=fedsrch 
856 4 0 |3 Full Text - PDF  |u https://www.proquest.com/docview/3286316714/fulltextPDF/embedded/7BTGNMKEMPT1V9Z2?source=fedsrch