Fact and Ideology in the Machine: Modelling Knowledge and Belief in Neural Models from Text
Doctoral thesis, 2025

This thesis explores questions of knowledge, language, and neural network models. Motivated by an increasing need for insight into complex political and social science phenomena, we study how methods within natural language processing (NLP) can help us gain such insight. With a particular focus on a model's knowledge, how it is structured, and how we can access and assess it, we study two important aspects of NLP models.
First, we investigate their capabilities and limitations, focusing on how they can capture political and social signals. We use embedding models to capture and reveal distinctions in policy and ideology in Swedish political parties, discussing the strengths and drawbacks of the approach. We also investigate the presence of more complex social knowledge in large pre-trained language models. We prompt models to produce synthetic samples of responses to social science experiments and access if effects calculated from the synthetic data can be used to predict a study's replicability. A central limitation we find in these studies is the lack of robustness, which we explore in depth by studying what influences model consistency in a more simplified setting, namely, recalling facts.
Second, we aim to bridge the gap between the model and the domain expert by developing and improving interpretability insights of model behaviour. We develop a method for aggregating class-level explanations for a text classifier and demonstrate its utility in the context of Swedish political texts. We also develop the understanding of how models store and access factual information. We propose a taxonomy of possible language model behaviours for fact completion and, based on our novel testing data set, examine internal knowledge structures using established mechanistic interpretability methods.

Representation

Explainability

Evaluation

Natural Language Processing

Political Science

SB-H3, Sven Hultins Gata 6
Opponent: Prof. Dirk Hovy, Bocconi University

Author

Denitsa Saynova

Data Science and AI 2

Word embeddings on ideology and issues from Swedish parliamentarians’ motions: a comparative approach

Journal of Elections, Public Opinion and Parties,;Vol. In Press(2024)

Journal article

Class Explanations: the Role of Domain-Specific Content and Stop Words

Proceedings of the 24th Nordic Conference on Computational Linguistics, NoDaLiDa 2023,;(2023)p. 103-112

Paper in proceeding

The Effect of Scaling, Retrieval Augmentation and Form on the Factual Consistency of Language Models

EMNLP 2023 - 2023 Conference on Empirical Methods in Natural Language Processing, Proceedings,;(2023)p. 5457-5476

Paper in proceeding

Fact Recall, Heuristics or Pure Guesswork? Precise Interpretations of Language Models for Fact Completion

Findings of the Association for Computational Linguistics: ACL 2025,;(2025)p. 18322-18349

Paper in proceeding

D. Saynova, K. Hansson, B. Bruinsma, A. Fredén, M. Johansson Identifying Non-Replicable Social Science Studies with Language Models

Bias and methods of AI technology studying political behaviour

Marianne och Marcus Wallenberg Foundation (M&MWallenbergsStiftelse), 2020-01-01 -- 2023-12-31.

Subject Categories (SSIF 2025)

Natural Language Processing

ISBN

978-91-8103-272-7

Doktorsavhandlingar vid Chalmers tekniska högskola. Ny serie: 5730

Publisher

Chalmers

SB-H3, Sven Hultins Gata 6

Online

Opponent: Prof. Dirk Hovy, Bocconi University

More information

Latest update

8/26/2025