no code implementations • 27 Sep 2024 • Daniel Jiwoong Im, Kevin Zhang, Nakul Verma, Kyunghyun Cho
We propose an autoregressive (AR) CI framework capable of handling complex confounders and sequential actions common in modern applications.
1 code implementation • 14 Jun 2023 • Narutatsu Ri, Fei-Tzin Lee, Nakul Verma
While static word embedding models are known to represent linguistic analogies as parallel lines in high-dimensional space, the underlying mechanism as to why they result in such geometric structures remains obscure.
1 code implementation • 9 Sep 2022 • Xiao Yu, Nakul Verma
In this work, we demonstrate that more sophisticated label representations are better for classification than the usual one-hot encoding.
1 code implementation • 31 Dec 2021 • Iddo Drori, Sarah Zhang, Reece Shuttleworth, Leonard Tang, Albert Lu, Elizabeth Ke, Kevin Liu, Linda Chen, Sunny Tran, Newman Cheng, Roman Wang, Nikhil Singh, Taylor L. Patti, Jayson Lynch, Avi Shporer, Nakul Verma, Eugene Wu, Gilbert Strang
We automatically synthesize programs using few-shot learning and OpenAI's Codex transformer and execute them to solve course problems at 81% automatic accuracy.
no code implementations • 27 Nov 2021 • Fei-Tzin Lee, Chris Kedzie, Nakul Verma, Kathleen McKeown
Prior work in AMR-based summarization has automatically merged the individual sentence graphs into a document graph, but the method of merging and its effects on summary content selection have not been independently evaluated.
no code implementations • 16 Nov 2021 • Leonard Tang, Elizabeth Ke, Nikhil Singh, Nakul Verma, Iddo Drori
Our work is the first to introduce a new dataset of university-level probability and statistics problems and solve these problems in a scalable fashion using the program synthesis capabilities of large language models.
no code implementations • 16 Nov 2021 • Iddo Drori, Nakul Verma
We solve MIT's Linear Algebra 18. 06 course and Columbia University's Computational Linear Algebra COMS3251 courses with perfect accuracy by interactive program synthesis.
no code implementations • AAAI 2021 • Hengrui Xing, Ansaf Salleb-Aouissi, Nakul Verma
This is not a trivial problem as it involves searching for a complex mathematical relationship over a large set of explanatory variables and operators that can be combined in an infinite number of ways.
no code implementations • 4 Dec 2020 • Bo Cowgill, Fabrizio Dell'Acqua, Samuel Deng, Daniel Hsu, Nakul Verma, Augustin Chaintreau
We find that biased predictions are mostly caused by biased training data.
no code implementations • 30 Oct 2019 • Yibo Jiang, Nakul Verma
By providing multiple types of training datasets as inputs, our model has the ability to generalize well on unseen datasets (new clustering tasks).
no code implementations • 16 Oct 2019 • Daniel Jiwoong Im, Yibo Jiang, Nakul Verma
By leveraging this refined control, we demonstrate that there are multiple principled ways to update MAML and show that the classic MAML optimization is simply a special case of second-order Runge-Kutta method that mainly focuses on fast-adaptation.
no code implementations • 5 Feb 2019 • Max Aalto, Nakul Verma
Recent literature has shown that symbolic data, such as text and graphs, is often better represented by points on a curved manifold, rather than in Euclidean space.
1 code implementation • NeurIPS 2019 • Alexandre Louis Lamy, Ziyuan Zhong, Aditya Krishna Menon, Nakul Verma
We finally show that our procedure is empirically effective on two case-studies involving sensitive feature censoring.
no code implementations • 3 Nov 2018 • Daniel Jiwoong Im, Nakul Verma, Kristin Branson
A common concern with $t$-SNE criterion is that it is optimized using gradient descent, and can become stuck in poor local minima.
no code implementations • NeurIPS 2015 • Nakul Verma, Kristin Branson
Metric learning seeks a transformation of the feature space that enhances prediction quality for the given task at hand.
no code implementations • NeurIPS 2007 • Yoav Freund, Sanjoy Dasgupta, Mayank Kabra, Nakul Verma
We present a simple variant of the k-d tree which automatically adapts to intrinsic low dimensional structure in data.