1 code implementation • 25 Jan 2024 • Vasileios Tsouvalas, Aaqib Saeed, Tanir Ozcelebi, Nirvana Meratnia
Federated Learning (FL) is a promising technique for the collaborative training of deep neural networks across multiple devices while preserving data privacy.
no code implementations • 29 Nov 2023 • Vasileios Tsouvalas, Yuki Asano, Aaqib Saeed
Foundation Models (FMs) have revolutionized machine learning with their adaptability and high performance across tasks; yet, their integration into Federated Learning (FL) is challenging due to substantial communication overhead from their extensive parameterization.
no code implementations • 15 Nov 2023 • Saeed Khalilian, Vasileios Tsouvalas, Tanir Ozcelebi, Nirvana Meratnia
To ensure a smooth learning curve and proper calibration of clusters between the server and the clients, FedCode periodically transfers model weights after multiple rounds of solely communicating codebooks.
1 code implementation • 3 May 2023 • Aaqib Saeed, Vasileios Tsouvalas
As technology advances and digital devices become prevalent, seamless human-machine communication is increasingly gaining significance.
1 code implementation • 19 Aug 2022 • Vasileios Tsouvalas, Aaqib Saeed, Tanir Ozcelebi, Nirvana Meratnia
Federated Learning (FL) is a distributed machine learning paradigm that enables learning models from decentralized private datasets, where the labeling effort is entrusted to the clients.
1 code implementation • 5 Feb 2022 • Vasileios Tsouvalas, Tanir Ozcelebi, Nirvana Meratnia
To the best of our knowledge, this is the first federated SER approach, which utilizes self-training learning in conjunction with federated learning to exploit both labeled and unlabeled on-device data.
1 code implementation • 14 Jul 2021 • Vasileios Tsouvalas, Aaqib Saeed, Tanir Ozcelebi
Notably, we show that with as little as 3% labeled data available, FedSTAR on average can improve the recognition rate by 13. 28% compared to the fully supervised federated model.