LIME
- @ribeiroWhyShouldTrust2016
- novel model-agnostic modular and extensible explanation technique that explains the predictions of any classifier in an interpretable and faithful manner
- learning an interpretable model locally around the prediction
- SP-LIME
- method to explain models by selecting representative individual predictions and their explanations in a non-redundant way, framing the task as a submodular optimization problem and providing a global view of the model to users
- flexibility of these methods by explaining different models for text (e.g random forests) and image classification (e.g neural networks)
- usefulness of explanations is shown via novel experiments, both simulated and with human subjects
