FedGT: Identification of Malicious Clients in Federated Learning with Secure Aggregation
Journal article, 2025

Federated learning (FL) has emerged as a promising approach for collaboratively training machine learning models while preserving data privacy. Due to its decentralized nature, FL is vulnerable to poisoning attacks, where malicious clients compromise the global model through altered data or updates. Identifying such malicious clients is crucial for ensuring the integrity of FL systems. This task becomes particularly challenging under privacy-enhancing protocols such as secure aggregation, creating a fundamental trade-off between privacy and security. In this work, we propose FedGT, a novel framework designed to identify malicious clients in FL with secure aggregation while preserving privacy. Drawing inspiration from group testing, FedGT leverages overlapping groups of clients to identify the presence of malicious clients via a decoding operation. The clients identified as malicious are then removed from the model training, which is performed over the remaining clients. By choosing the size, number, and overlap between groups, FedGT strikes a balance between privacy and security. Specifically, the server learns the aggregated model of the clients in each group - vanilla federated learning and secure aggregation correspond to the extreme cases of FedGT with group size equal to one and the total number of clients, respectively. The effectiveness of FedGT is demonstrated through extensive experiments on three datasets in a cross-silo setting under different data-poisoning attacks. These experiments showcase FedGT's ability to identify malicious clients, resulting in high model utility. We further show that FedGT significantly outperforms the private robust aggregation approach based on the geometric median recently proposed by Pillutla et al. and the robust aggregation technique Multi-Krum in multiple settings.

secure aggregation

malicious clients

group testing

poisoning attacks

federated learning

privacy

AI security

security

Author

Marvin Xhemrishi

Technical University of Munich

Johan Östman

AI Sweden

Antonia Wachter-Zeh

Technical University of Munich

Alexandre Graell I Amat

Chalmers, Electrical Engineering, Communication, Antennas and Optical Networks

IEEE Transactions on Information Forensics and Security

1556-6013 (ISSN) 15566021 (eISSN)

Vol. In Press

Subject Categories (SSIF 2025)

Communication Systems

Computer Sciences

Computer Systems

DOI

10.1109/TIFS.2025.3539964

More information

Latest update

2/26/2025