Fact and Ideology in the Machine: Modelling Knowledge and Belief in Neural Models from Text
Doctoral thesis, 2025
First, we investigate their capabilities and limitations, focusing on how they can capture political and social signals. We use embedding models to capture and reveal distinctions in policy and ideology in Swedish political parties, discussing the strengths and drawbacks of the approach. We also investigate the presence of more complex social knowledge in large pre-trained language models. We prompt models to produce synthetic samples of responses to social science experiments and access if effects calculated from the synthetic data can be used to predict a study's replicability. A central limitation we find in these studies is the lack of robustness, which we explore in depth by studying what influences model consistency in a more simplified setting, namely, recalling facts.
Second, we aim to bridge the gap between the model and the domain expert by developing and improving interpretability insights of model behaviour. We develop a method for aggregating class-level explanations for a text classifier and demonstrate its utility in the context of Swedish political texts. We also develop the understanding of how models store and access factual information. We propose a taxonomy of possible language model behaviours for fact completion and, based on our novel testing data set, examine internal knowledge structures using established mechanistic interpretability methods.
Representation
Explainability
Evaluation
Natural Language Processing
Political Science
Author
Denitsa Saynova
Data Science and AI 2
Word embeddings on ideology and issues from Swedish parliamentarians’ motions: a comparative approach
Journal of Elections, Public Opinion and Parties,;Vol. In Press(2024)
Journal article
Class Explanations: the Role of Domain-Specific Content and Stop Words
Proceedings of the 24th Nordic Conference on Computational Linguistics, NoDaLiDa 2023,;(2023)p. 103-112
Paper in proceeding
The Effect of Scaling, Retrieval Augmentation and Form on the Factual Consistency of Language Models
EMNLP 2023 - 2023 Conference on Empirical Methods in Natural Language Processing, Proceedings,;(2023)p. 5457-5476
Paper in proceeding
Fact Recall, Heuristics or Pure Guesswork? Precise Interpretations of Language Models for Fact Completion
Findings of the Association for Computational Linguistics: ACL 2025,;(2025)p. 18322-18349
Paper in proceeding
D. Saynova, K. Hansson, B. Bruinsma, A. Fredén, M. Johansson Identifying Non-Replicable Social Science Studies with Language Models
Bias and methods of AI technology studying political behaviour
Marianne och Marcus Wallenberg Foundation (M&MWallenbergsStiftelse), 2020-01-01 -- 2023-12-31.
Subject Categories (SSIF 2025)
Natural Language Processing
ISBN
978-91-8103-272-7
Doktorsavhandlingar vid Chalmers tekniska högskola. Ny serie: 5730
Publisher
Chalmers