Skip to content

Investigating Superpixels in LIME for Explaining Predictions of Facial Images

Investigating explainability algorithms for granular healthcare images

From left to right connected by arrows: 1. Loading a pretrained/fine-tuned DCNN model and an example image. 2. Super pixels. 3. Random Perturbations. 4. Predictions. 5. Weights. 6. Surrogate Model. 7. Top Features and Explanations Figure 1: An illustration of LIME workflow

A work experience project investigating the application of a Local Interpretable Model-agnostic Explanations (LIME) technique to an image classification task around identifying Rosacea. A binary classification model was trained on the normal and Rosacea faces to generate the LIME explanation for Rosacea faces. Secondly, the fine-tuned model was integrated into the LIME pipeline to generate explanations based on the crucial features on which predictions were made in the classification model. Hence the experimentations helped in understanding the features the classification model took.

Results

Image pre-processing (through contrast enhancement) improved the LIME by increasing the number of superpixels for the given input image so that some broad features such as discolouration and rashes were identified. However, the coarseness of the superpixels generated through the pipeline was not sufficient to pick up the features which would discern between different types of Rosacea. It may require domain/ imaging modality-specific pre-processing tasks to enhance the quality of explanation by improving the distinctiveness of the features that may help pick the right number of superpixels.

Output Link
Open Source Code & Documentation Github
Case Study n/a
Technical report Here