no code implementations • 9 Dec 2023 • Ege Erdogan, Simon Geisler, Stephan Günnemann
It is well-known that deep learning models are vulnerable to small input perturbations.
no code implementations • 14 Sep 2023 • Mahdi Dhaini, Wessel Poelman, Ege Erdogan
While recent advancements in the capabilities and widespread accessibility of generative language models, such as ChatGPT (OpenAI, 2022), have brought about various benefits by generating fluent human-like text, the task of distinguishing between human- and large language model (LLM) generated text has emerged as a crucial problem.
2 code implementations • 16 Feb 2023 • Ege Erdogan, Unat Teksen, Mehmet Salih Celiktenyildiz, Alptekin Kupcu, A. Ercument Cicek
Split learning enables efficient and privacy-aware training of a deep neural network by splitting a neural network so that the clients (data holders) compute the first layers and only share the intermediate output with the central compute-heavy server.
1 code implementation • 20 Aug 2021 • Ege Erdogan, Alptekin Kupcu, A. Ercument Cicek
(1) We show that an honest-but-curious split learning server, equipped only with the knowledge of the client neural network architecture, can recover the input samples and obtain a functionally similar model to the client model, without being detected.
1 code implementation • 20 Aug 2021 • Ege Erdogan, Alptekin Kupcu, A. Ercument Cicek
Distributed deep learning frameworks such as split learning provide great benefits with regards to the computational cost of training deep neural networks and the privacy-aware utilization of the collective data of a group of data-holders.