You are required to read and agree to the below before accessing a full-text version of an article in the IDE article repository.

The full-text document you are about to access is subject to national and international copyright laws. In most cases (but not necessarily all) the consequence is that personal use is allowed given that the copyright owner is duly acknowledged and respected. All other use (typically) require an explicit permission (often in writing) by the copyright owner.

For the reports in this repository we specifically note that

  • the use of articles under IEEE copyright is governed by the IEEE copyright policy (available at http://www.ieee.org/web/publications/rights/copyrightpolicy.html)
  • the use of articles under ACM copyright is governed by the ACM copyright policy (available at http://www.acm.org/pubs/copyright_policy/)
  • technical reports and other articles issued by M‰lardalen University is free for personal use. For other use, the explicit consent of the authors is required
  • in other cases, please contact the copyright owner for detailed information

By accepting I agree to acknowledge and respect the rights of the copyright owner of the document I am about to access.

If you are in doubt, feel free to contact webmaster@ide.mdh.se

EncCluster: Bringing Functional Encryption in Federated Foundational Models

Fulltext:


Authors:

Vasileios Tsouvalas , Samaneh Mohammadi, Ali Balador, Tanir Ozcelebi , Francesco Flammini, Nirvana Meratnia

Publication Type:

Conference/Workshop Paper

Venue:

2024 Conference on Neural Information Processing Systems


Abstract

Federated Learning (FL) decentralizes model training by transmitting local model updates to a central server, yet it remains vulnerable to inference attacks during these transmissions. Existing solutions, such as Differential Privacy (DP) and Functional Encryption (FE), often degrade performance or impose significant operational burdens on clients. Meanwhile, the advent of Foundation Models (FMs) has trans- formed FL with their adaptability and high performance across diverse tasks. How- ever, delivering strong privacy guarantees with these highly parameterized FMs in FL using existing privacy-preserving frameworks amplifies existing challenges and further complicates the efficiency-privacy trade-off. We present EncCluster†, a novel method that integrates model compression through weight clustering with decentralized FE and privacy-enhancing data encoding using probabilistic filters to deliver strong privacy guarantees in FL without affecting model performance or adding unnecessary burdens to clients. We perform a comprehensive evaluation, spanning 4 datasets and 5 architectures, to demonstrate EncCluster scalability across encryption levels. Our findings reveal that EncCluster significantly re- duces communication costs — below even conventional FedAvg — and accelerates encryption up to 1000× over baselines; at the same time, it maintains high model accuracy and enhanced privacy assurances.

Bibtex

@inproceedings{Tsouvalas7122,
author = {Vasileios Tsouvalas and Samaneh Mohammadi and Ali Balador and Tanir Ozcelebi and Francesco Flammini and Nirvana Meratnia},
title = {EncCluster: Bringing Functional Encryption in Federated Foundational Models},
month = {December},
year = {2024},
booktitle = {2024 Conference on Neural Information Processing Systems},
url = {http://www.es.mdu.se/publications/7122-}
}