Dual-layer prompt ensembles: Leveraging system- and user-level instructions for robust LLM-based query expansion and rank fusion
Journal article, 2026

Large Language Models (LLMs) show strong potential for query expansion (QE), but their effectiveness is highly sensitive to prompt design. This paper investigates whether exploiting the system-user prompt distinction in chat-based LLMs improves QE, and how multiple expansions should be combined. We propose Dual-Layer Prompt Ensembles, which pair a behavioural system prompt with varied user prompts to generate diverse expansions, and aggregate their BM25-ranked lists using lightweight SU-RankFusion schemes. Experiments on six heterogeneous datasets show that dual-layer prompting consistently outperforms strong single-prompt baselines. For example, on Touche-2020 a dual-layer configuration improves nDCG@10 from 0.4177 (QE-CoT) to 0.4696, and SU-RankFusion further raises it to 0.4797. On Robust04 and DBPedia, SU-RankFusion improves nDCG@10 over BM25 by 24.7% and 25.5%, respectively, with similar gains on NFCorpus, FiQA, and TREC-COVID. These results demonstrate that system-user prompt ensembles are effective for QE, and that simple fusion transforms prompt-level diversity into stable retrieval improvements.

Information retrieval

Prompt engineering

Query expansion

BM25

Large language models

Rank fusion

Author

Minghan Li

Soochow University

Ercong Nie

Ludwig Maximilian University of Munich (LMU)

Munich Center for Machine Learning

Huiping Huang

Chalmers, Electrical Engineering, Communication, Antennas and Optical Networks

Xinxuan Lv

Soochow University

Guodong Zhou

Soochow University

Information Fusion

1566-2535 (ISSN) 18726305 (eISSN)

Vol. 131 104160

Subject Categories (SSIF 2025)

Other Electrical Engineering, Electronic Engineering, Information Engineering

Embedded Systems

Telecommunications

Computer Systems

Control Engineering

DOI

10.1016/j.inffus.2026.104160

More information

Latest update

2/6/2026 9