no code implementations • 26 Mar 2024 • Uri Hacohen, Adi Haviv, Shahar Sarfaty, Bruria Friedman, Niva Elkin-Koren, Roi Livni, Amit H Bermano
The advent of Generative Artificial Intelligence (GenAI) models, including GitHub Copilot, OpenAI GPT, and Stable Diffusion, has revolutionized content creation, enabling non-professionals to produce high-quality content across various domains.
no code implementations • 5 Oct 2023 • Ofir Bar Tal, Adi Haviv, Amit H. Bermano
Evasion Attacks (EA) are used to test the robustness of trained neural networks by distorting input data to misguide the model into incorrect classifications.
1 code implementation • 7 Oct 2022 • Adi Haviv, Ido Cohen, Jacob Gidron, Roei Schuster, Yoav Goldberg, Mor Geva
In this work, we offer the first methodological framework for probing and characterizing recall of memorized sequences in transformer LMs.
1 code implementation • 30 Mar 2022 • Adi Haviv, Ori Ram, Ofir Press, Peter Izsak, Omer Levy
Causal transformer language models (LMs), such as GPT-3, typically require some form of positional encoding, such as positional embeddings.
2 code implementations • 10 Jan 2022 • Uri Shaham, Elad Segal, Maor Ivgi, Avia Efrat, Ori Yoran, Adi Haviv, Ankit Gupta, Wenhan Xiong, Mor Geva, Jonathan Berant, Omer Levy
NLP benchmarks have largely focused on short texts, such as sentences and paragraphs, even though long texts comprise a considerable amount of natural language in the wild.
Ranked #8 on Long-range modeling on SCROLLS
no code implementations • NAACL 2021 • Adi Haviv, Lior Vassertail, Omer Levy
Latent alignment objectives such as CTC and AXE significantly improve non-autoregressive machine translation models.
no code implementations • EACL 2021 • Adi Haviv, Jonathan Berant, Amir Globerson
In this work, we propose a method for automatically rewriting queries into "BERTese", a paraphrase query that is directly optimized towards better knowledge extraction.