Search Results for author: Or Zamir

Found 4 papers, 1 papers with code

Excuse me, sir? Your language model is leaking (information)

1 code implementation18 Jan 2024 Or Zamir

We introduce a cryptographic method to hide an arbitrary secret payload in the response of a Large Language Model (LLM).

Language Modelling Large Language Model

Undetectable Watermarks for Language Models

no code implementations25 May 2023 Miranda Christ, Sam Gunn, Or Zamir

We ask: Is it possible to introduce a watermark without incurring any detectable change to the output distribution?

Planting Undetectable Backdoors in Machine Learning Models

no code implementations14 Apr 2022 Shafi Goldwasser, Michael P. Kim, Vinod Vaikuntanathan, Or Zamir

Second, we demonstrate how to insert undetectable backdoors in models trained using the Random Fourier Features (RFF) learning paradigm or in Random ReLU networks.

Adversarial Robustness BIG-bench Machine Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.