Papers
arxiv:2207.01420

Comparing Feature Importance and Rule Extraction for Interpretability on Text Data

Published on Jul 4, 2022
Authors:

Abstract

Different interpretability methods for text data can produce varied explanations, even for simple models, and a new approach is proposed to compare these explanations.

AI-generated summary

Complex machine learning algorithms are used more and more often in critical tasks involving text data, leading to the development of interpretability methods. Among local methods, two families have emerged: those computing importance scores for each feature and those extracting simple logical rules. In this paper we show that using different methods can lead to unexpectedly different explanations, even when applied to simple models for which we would expect qualitative coincidence. To quantify this effect, we propose a new approach to compare explanations produced by different methods.

Community

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2207.01420 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2207.01420 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2207.01420 in a Space README.md to link it from this page.

Collections including this paper 1