Class Explanations: the Role of Domain-Specific Content and Stop Words
Paper in proceeding, 2023

We address two understudied areas related to explainability for neural text models. First, class explanations. What features are descriptive across a class, rather than explaining single input instances? Second, the type of features that are used for providing explanations. Does the explanation involve the statistical pattern of word usage or the presence of domain-specific content words? Here, we present a method to extract both class explanations and strategies to differentiate between two types of explanations – domain-specific signals or statistical variations in frequencies of common words. We demonstrate our method using a case study in which we analyse transcripts of political debates in the Swedish Riksdag.

Author

Denitsa Saynova

Chalmers, Computer Science and Engineering (Chalmers), Data Science and AI

Sebastianus Cornelis Jacobus Bruinsma

Chalmers, Computer Science and Engineering (Chalmers), Data Science and AI

Moa Johansson

AoA Information and Communications technology

Richard Johansson

University of Gothenburg

Chalmers, Computer Science and Engineering (Chalmers), Data Science

Proceedings of the 24th Nordic Conference on Computational Linguistics, NoDaLiDa 2023

103-112
9789916219997 (ISBN)

24th Nordic Conference on Computational Linguistics, NoDaLiDa 2023
Torshavn, Faroe Islands,

Subject Categories (SSIF 2025)

Natural Language Processing

More information

Latest update

2/14/2025