no code implementations • 31 Mar 2024 • Itai Kreisler, Maor Ivgi, Oliver Hinder, Yair Carmon
We propose a method that achieves near-optimal rates for smooth stochastic convex optimization and requires essentially no prior knowledge of problem parameters.
1 code implementation • 23 May 2023 • Uri Shaham, Maor Ivgi, Avia Efrat, Jonathan Berant, Omer Levy
We introduce ZeroSCROLLS, a zero-shot benchmark for natural language understanding over long texts, which contains only test and small validation sets, without training data.
1 code implementation • 8 Feb 2023 • Maor Ivgi, Oliver Hinder, Yair Carmon
Empirically, we consider a broad range of vision and language transfer learning tasks, and show that DoG's performance is close to that of SGD with tuned learning rate.
1 code implementation • 1 Aug 2022 • Maor Ivgi, Uri Shaham, Jonathan Berant
Transformer-based pretrained language models (LMs) are ubiquitous across natural language understanding, but cannot be applied to long sequences such as stories, scientific articles and long documents, due to their quadratic complexity.
Ranked #6 on Long-range modeling on SCROLLS
no code implementations • 13 Feb 2022 • Maor Ivgi, Yair Carmon, Jonathan Berant
Neural scaling laws define a predictable relationship between a model's parameter count and its performance after training in the form of a power law.
2 code implementations • 10 Jan 2022 • Uri Shaham, Elad Segal, Maor Ivgi, Avia Efrat, Ori Yoran, Adi Haviv, Ankit Gupta, Wenhan Xiong, Mor Geva, Jonathan Berant, Omer Levy
NLP benchmarks have largely focused on short texts, such as sentences and paragraphs, even though long texts comprise a considerable amount of natural language in the wild.
Ranked #8 on Long-range modeling on SCROLLS
no code implementations • 10 Nov 2021 • Amirata Ghorbani, Dina Berenbaum, Maor Ivgi, Yuval Dafna, James Zou
We address this limitation by introducing Feature Vectors, a new global interpretability method designed for tabular datasets.
1 code implementation • EMNLP 2021 • Maor Ivgi, Jonathan Berant
In this work, we address this gap and leverage discrete attacks for online augmentation, where adversarial examples are generated at every training step, adapting to the changing nature of the model.
no code implementations • 23 Sep 2020 • Maor Ivgi, Yaniv Benny, Avichai Ben-David, Jonathan Berant, Lior Wolf
We empirically show on the COCO-STUFF dataset that our approach improves the quality of both the intermediate layout and the final image.